[Date Prev][Date Next] [Thread Prev][Thread Next] [Date Index] [Thread Index]

Re: Unidentified subject!



> Well the 2T memory everybody was curious about 3 weeks ago got here early.
>
> From dmesg after plugging one in:
> [629240.916163] usb 1-2: new high-speed USB device number 39 using xhci_hcd
> [629241.066221] usb 1-2: New USB device found, idVendor=048d,
> idProduct=1234, bcdDevice= 2.00
> [629241.066234] usb 1-2: New USB device strings: Mfr=1, Product=2,
> SerialNumber=3
> [629241.066239] usb 1-2: Product: Disk 3.0
> [629241.066242] usb 1-2: Manufacturer: USB
> [629241.066246] usb 1-2: SerialNumber: 2697241127107725123
> [629241.069485] usb-storage 1-2:1.0: USB Mass Storage device detected
> [629241.074187] scsi host37: usb-storage 1-2:1.0
> [629242.100738] scsi 37:0:0:0: Direct-Access              SSD 3.0    2.00
> PQ: 0 ANSI: 4
> [629242.100959] sd 37:0:0:0: Attached scsi generic sg13 type 0
> [629242.101190] sd 37:0:0:0: [sdm] 4096000000 512-byte logical blocks: (2.10
> TB/1.91 TiB)
> [629242.101289] sd 37:0:0:0: [sdm] Write Protect is off
> [629242.101290] sd 37:0:0:0: [sdm] Mode Sense: 03 00 00 00
> [629242.101409] sd 37:0:0:0: [sdm] No Caching mode page found
> [629242.101410] sd 37:0:0:0: [sdm] Assuming drive cache: write through
> [629242.103927]  sdm: sdm1
> [629242.104047] sd 37:0:0:0: [sdm] Attached SCSI disk
> gene@coyote:
>
> Looks like a reasonable facsimile of a 2T disk to me.

AFAIK the bogus 128TB drives do properly report such ridiculous sizes:
the reality only hits when you try to actually store that amount of
information on them.

[ I'm not sure how it works under the hood, but since SSDs store their
  data "anywhere" in the flash, they can easily pretend to have any size
  they want, and allocate the physical flash blocks only on-the-fly as
  logical blocks are being written.
  Also, some Flash controllers use compression, so if you store data
  that compresses well, they can let you store a lot more than if you
  store already compressed data.  ]

IOW, to really check, try to save 2TB of videos (or other already
compressed data), and then try and read it back.


        Stefan


Reply to: