Alex de Oliveira Silva <firstname.lastname@example.org> writes: > Ola a todos da lista. Olá Alex > Acabei de fazer uma burrada, deu um kernel panic e deu uns erros e > estava querendo lembrar o nome do fsck so q acabei botando mkfs > (realmente foi um vacilo muito feio) bem isso foi na minha particao do > win por lado foi bom ter apagado o win por outro não já que tinha muitas > mp3, clippes q documentos que não podria ter perdido. Se possivel alguem > q saiba recuperar esses arquivos ou recuperar tudo me da uma força. Não > sei nem por onde começar. Desde já obrigado e `s =) Anexo envio uma página (perdi o endereço, por isto a envio, e não o link) que pode lhe ajudar. Talvez seja interesse de mais gente na lista, por isto mando-a publicamente. Siga as instruções da página. Eu usei o programa citado, mas não sei se foi o programa que resolveu o problema (até agora ainda não consegui descobrir). É uma opção a se testar. Siga-a por sua conta e risco. Reporte-nos resultados e/ou dúvidas.Title: Recovering lost ext2 Linux filesystems
Recently we had the unpleasant experience of having to recover data from a crashed hard drive, and we're documenting our complete success here to hopefully help others in the same boat. It was a very, very long process that was made worse by the irony of the failure's occurrance just as we were bringing the machine down to install a tape drive. Grrrr.
We ran some DOS-based low-level diagnostics (TuffTEST-Pro) that showed only about two dozen blocks were physically bad, so this suggested that we'd be able to recover at least some of the data. But the partition table was hit, so this made it quite a bit more interesting.
These notes were written to reflect work done on Red Hat Linux 6.0 on the Intel platform (the 2.2.5 kernel). This is quite old, but this particular version was required for an embedded development project that this machine was used for. All drives were IDE.
These notes are also not terribly complete. Most of them were written while drives were being copied or scanned or otherwise crunching: once we actually recovered all our data, our interest in writing about this saga dropped rapidly in favor of getting back to actual work. Sorry if it's a bit unpolished.
WARNING - The procedures described here can be very dangerous if applied to the wrong drive, and it's remarkably easy to do this. Don't do this on a live machine with un-backed-up data unless you are really sure of yourself. Please be careful.
The first step was to make a physical copy of the drive so we'd have something to work with in case the failed drive took a complete dump. We were quite sure this would happen eventually, and we'd like to have as much data onto good media as long as the Gods were smiling. We had a larger drive lying around so installed it into the same system.
IDE drives in Linux are addressed this way:
Copying a drive is usually done with the dd command, but a few non-obvious options are required in order for this:
# dd if=/dev/hdb of=/dev/hdd bs=512 conv=noerror,syncHere we have:
Using a small blocksize makes the copy go much more slowly, but we think that this minimizes the data that gets filled in with zeros on an error: if we used 64k blocksizes, we're afraid of one bad spot zeroing out most of the large block. We really need to research this.
If the bad spots are all known to be early in the drive, it's possible to do a small blocksized copy for just that area and run a second dd command with larger blocksizes later, but we weren't in enough of a rush to warrant that much trouble.
But a custom C program holds some promise too. Not only could it manage buffering properly to optimize speed and minimize data loss, but it could also save the list of bad blocks to a file and report on progress (percentage completion would be really helpful). Writing this kind of utility is on our list...
Ultimately we never were able to recover anything with this tool, and we're not sure why. The drive had previously been repartitioned several times, so there were a lot of phantom partition blocks left over from prior runs, and the overwhelming amount of data made it hard for us to know what we were looking at, and we think our own inexperience with the ext2 filesytesm contributed to this. Looking back we see that the partition we ultimately found was pointed out to us by gpart.
After looking for several other tools, we decided to do it ourselves. We were pretty sure that the data was still on the drive, and this was as good an opportunity to really dig into filesystem recovery as any. We've done this before with the older (and simpler) System V filesystems, but never with BSD or ext2. Now we know.
But one point we'll note: since the geometry of the spare drive and the failed drive are probably not the same, it's important to not rely on anything that refers to a "cylinder boundary". Only if the drives are identical can you take these optimizations.
For very large drives, the simple act of scanning them can take hours. So we decided to split up the process into "scanning" and "analysis": one simple program optimized for speed would run through the drive one time and create a list of "interesting" sectors, and a later program would take this as input for analysis. When it takes more than two hours to scan a large drive, it's simply too much work to keep fine-tuning this process over and over.
Instead, our "scandrive" program makes one very fast run through the drive and notes everything with a broad definition of "interesting". The idea is that we cast a wide net with a one-time scan, and use later tools to process all the blocks located: this allows us to refine our tools without having to re-scan many gigabytes of data. It also lends itself to using perl for later parts of the process.
Scandrive knows about three kinds of "interesting" items: possible partition tables, possible ext2 superblocks, and possible root directory entries. Each of these is reported with type and sector number, and though we expect to get some mistaken entries in the list, later software can process them in more detail without having to scan the whole drive.
We ran scandrive on the whole drive - it took more than two hours on the 20GB part that was a mirror of the failed drive - and the result was a logfile of several thousand lines. In this process we have found that the partition table entries weren't that helpful, so we've ignored them here.
Winnowing the wheat from the chaff can be a bit trying: not only do we have the routine false positives where random data are seen as something interesting, but any previous formatting of the drive can leave remnants behind in the otherwise empty space. This leads to a lot of "clutter" in the output.
Much of it could be reduced by a bit of postprocessing, but we have found that just by "eyeballing" the data we can locate the data that forms the most likely suspects.
All our ext2 filesystems use 1k "blocks" (two "sectors"), and the first super block is the second "block" on the partition. This means that sectors #0 & 1 are reserved for the OS, and sectors #2 & 3 are the first super block.
In each super block are lots of key parameters, but two of them are of particular interest: the filesytem size in blocks, and the block group number. Each super block inside a filesystem has the same "size" value, measured in (usually) 1 kbyte sizes. Seeing a string of candidate super blocks with the same filesystem size suggests that they all go together.
The second field of interest is the block group number: this is the sequence of super block within the filesystem, and they progress from #0, #1, ... up to the last super block in the filesystem.
If we see a super block #0 with a "sane" size, followed shortly by a root directory, then by a block #1 with the same size, this is a good clue that we have found the start of an ext2 partition. Not a complete clue, but a good one.
We can do a bit of automatec processing on the logfile that will help dump some of the entries that are clearly bogus.
We've created a small perl program that follows most of these rules, and it produces an output file that we summarize here. We've highlighted the #0 lines which draw most of our attention.
# Started scan of /dev/hdd at Thu Jan 31 09:41:34 2002 # buffer size = 1024 sectors # capacity = 40088160 sectors many obviously bogus lines deleted ... ext2 5189252 # size=32098 #0 root 5189762 ext2 5205636 # size=32098 #1 ext2 5222020 # size=32098 #2 ext2 5238404 # size=32098 #3 ext2 5526617 # size=530113 #0 root 5527139 ext2 5543001 # size=530113 #1 ext2 5559385 # size=530113 #2 ext2 5575769 # size=530113 #3 ext2 5592153 # size=530113 #4 ext2 5608537 # size=530113 #5 ext2 5624921 # size=530113 #6 ext2 5641305 # size=530113 #7 ext2 5657689 # size=530113 #8 ext2 5674073 # size=530113 #9 ext2 5690457 # size=530113 #10 ext2 5706841 # size=530113 #11 ext2 5723225 # size=530113 #12 ext2 5739609 # size=530113 #13 ext2 5755993 # size=530113 #14 ext2 5772377 # size=530113 #15 ext2 5788761 # size=530113 #16 ext2 5805145 # size=530113 #17 ext2 5821529 # size=530113 #18 ext2 5837913 # size=530113 #19 ext2 5854297 # size=530113 #20 ext2 5870681 # size=530113 #21 ext2 5887065 # size=530113 #22 ext2 5903449 # size=530113 #23 ext2 5919833 # size=530113 #24 ext2 5936217 # size=530113 #25 ext2 5952601 # size=530113 #26 ext2 5968985 # size=530113 #27 ext2 5985369 # size=530113 #28 ext2 6001753 # size=530113 #29 ext2 6018137 # size=530113 #30 ext2 6034521 # size=530113 #31 ext2 6050905 # size=530113 #32 ext2 6067289 # size=530113 #33 ext2 6083673 # size=530113 #34 ext2 6100057 # size=530113 #35 ext2 6116441 # size=530113 #36 ext2 6132825 # size=530113 #37 ext2 6149209 # size=530113 #38 ext2 6165593 # size=530113 #39 ext2 6181977 # size=530113 #40 ext2 6198361 # size=530113 #41 ext2 6214745 # size=530113 #42 ext2 6231129 # size=530113 #43 ext2 6247513 # size=530113 #44 ext2 6263897 # size=530113 #45 ext2 6280281 # size=530113 #46 ext2 6296665 # size=530113 #47 ext2 6313049 # size=530113 #48 ext2 6313802 # size=16884283 #0 root 6314450 ext2 6329433 # size=530113 #49 ext2 6330186 # size=16884283 #1 ext2 6345817 # size=530113 #50 ext2 6346570 # size=16884283 #2 ext2 6362954 # size=16884283 #3 ext2 6379338 # size=16884283 #4 ext2 6394969 # size=530113 #53 ext2 6395722 # size=16884283 #5 ext2 6411353 # size=530113 #54 ext2 6412106 # size=16884283 #6 ext2 6428490 # size=16884283 #7 ext2 6444874 # size=16884283 #8 ext2 6461258 # size=16884283 #9 ext2 6477642 # size=16884283 #10 ... (many lines deleted)
The #0 entries with root tags right after are highly likely candidates for partitions. The "size=" parameter is measured in "blocks", and for our purposes we simply assume 1kbyte blocks (though we really ought to compute it from the fields in the super block). Since each super blocks starts at sector #2, we subtract two from each starting offset to get the real partition starting point. From this we can also compute the ending block of the partition, which allows us to find (improper) overlap.
|#||Line from scandrive||Partition start||Size (sectors)||Last sector|
|#1||ext2 5189252 # size=32098 #0||5189250||64196 (31.3MB)||5253445|
|#2||ext2 5526617 # size=530113 #0||5526615||6586843 (517.7M)||12113457|
|#3||ext2 6313802 # size=16884283 #0||6313800||33768566 (16.4G)||40082364|
We see that partition #2 overlaps with partition #3 - this couldn't be valid - and since we remember having a very large root partition, we suspect that #2 is a phantom leftover from previous partitioning and that #3 is the real one. We are also pretty sure that #1 is the /boot partition, which is usually small and always contained within the first 4G of the drive.
Once we have a handle on what partitions are what, we need to put the partition table back onto the drive. We're assuming here that the current partition table is bogus and must be recreated from scratch. There are several tools that can be used for partitioning under Linux, but we only use sfdisk. This tool allows much more control over exactly where the partition boundaries lie, and we have used this tool for other projects well enough to be comfortable with it.
Always start by taking a snapshot of the current partitioning of the drive just for good measure: better to have it and not need it than to need it and not have it:
# sfdisk -d /dev/hdd > /tmp/partition.saveThe -d parameter reads the table and dumps it to the standard output in a format that can be fed back to sfdisk to put the table back.
Each partition has three parts: a starting sector, a size, and a type. It's important that they don't overlap (fortunately, sfdisk will do a lot of checking for you). Using our constructed table from above, we are able to create the sfdisk input.
HOWEVER - The partition table at the start of the drive only holds four entries, and if more than four partitions are required, they are done with "extended partitions". This substantially complicates the calculations due to the linked list of extra partition tables, and though we understand this process pretty well, it's much easier to just deal with up to four partitions at a time.
Remember that all sfdisk parameters are measured in sectors, and if a super block is at sector n, then the partition starts at sector n-2. We also must provide a "type", which gives a hint as to the contents of the partition. 83 means "Linux native", and 0 means "unused". sfdisk --list-types shows the whole list of types known to this program.
The constructed sfdisk input file now looks like:
# partition table of /dev/hdd unit: sectors /dev/hdd1 : start= 5189250, size= 64196, Id=83 /dev/hdd2 : start= 6313800, size= 33768566, Id=83 /dev/hdd3 : start= 0, size= 0, Id= 0 /dev/hdd4 : start= 0, size= 0, Id= 0
The partition file created, we must now apply it to the drive. Fortunately, we're able to "test" the configuration before we actually "make it so".
The -n parameter says to go through the motions but to not actually write to the disk, and it will allow you check your work before writing to the drive. sfdisk usually complains about things not being aligned on a cylinder boundary, but these can be ignored. Modern drives no longer really have fixed geometries (outer tracks have higher densities than inner tracks), so "cylinder/head/sector" is more an addressing scheme than a real physical geometry.# sfdisk -n /dev/hdd < /tmp/new.partition
But partition overlap and those that extend beyond the size of the drive are to be taken seriously. Fix them in the partition file and try again. Once done, we can actually apply the changes. This is done with the --force command-line parameter, which insists that sfdisk do its work even if it doesn't care for what it sees. This will write to the drive, so making the mistake of mentioning the wrong drive can be very painful.
This produces a fair amount of information, causes the kernel to reread the drive's new partition table, and when finished the /dev/hddX devices are available to the user.# sfdisk --force /dev/hdd < /tmp/new.partition
Our first step is to try to inspect each partition to see if it might be any good: this is what the fsck is for:
In the first go-round, we don't normally allow writes to the drive (we say "no" to all questions) while we get a feel for what kind of shape the filesystem is in. If the damage looks minimal, we may actually run it again and allow full repair, but serious damage is often a sign that the partitioning is not right.# fsck /dev/hdd1 # fsck /dev/hdd2
To actually get to the data, we mount the filesystem readonly onto a common /mnt/hdd mount point. In our case we had only two partitions, so we could be reasonably sure which was which:
but in more advanced schemes, you might have to mount the partitions one at a time in order to look around. Once mounted, we suggest taking all the data off, either by backing up to tape or by copying the data to a "good" drive (not the spare used for recovery).# mkdir /mnt/hdd # mount -r /dev/hdd2 /mnt/hdd <-- root partition # mount -r /dev/hdd1 /mnt/hdd/boot <-- boot partition
We do not trust the partitioning enough to actually go with this live. Instead, we copy the data elsewhere, repartition the spare drive via the "normal" routes, reformat the filesystems, and copy the data back.
Building the program is very easy:
and it takes these command-line parameters:# cc -O scandrive.cpp -o scandrive
|(REQUIRED) DEV is the device to scan. Most of this time this is the full drive ( /dev/hdX ), but it can be an individual partition if this is appropriate for the current situtation.|
|(OPTIONAL) size is the number of 512-byte sectors to use for I/O buffering. The default is probably fine.|
|(OPTIONAL) capacity is the size of the drive in 512-byte sectors. This is used to limit the range of the scan - we'll stop when we hit the end - as well as to report percent-done status. When scanning a really large drive it's nice to know how far along in the process it is. If omitted, the whole drive is scanned.|
|(OPTIONAL) Append summarized results into logfile FILE. This file has a format intended to be easy to parse and process, and it is a bit more terse than the standard output. The intent is that the standard output need not be captured if this option is used.|
The output file created with the -L parameter is in straight ASCII, and it always appends to the file if found: scanning is too slow of an operation to risk blowing an important previous file away.
Blank lines and those starting with a # should be treated as comments and ignored, and the rest are lines with a "type" and the starting sector number, and possibly a comment after. A sample output file:
The types known are:# Started scan of /dev/hda at Thu Jan 31 15:50:51 2002 # buffer size = 256 sectors pt 0 ext2 65 # size=24066 #0 root 575 pt 4033 pt 5869 ext2 16449 # size=24066 #1 pt 48195 ext2 48260 # size=1496045 #0 root 52362 pt 69234 pt 69283 pt 72122 pt 72171 pt 75010
|pt||This entry might be a Partition Table, but the bar is very low for this test: any sector with 55AA as the last two bytes will be reported this way.|
|ext2||This sector might be an ext2 filesystem super block as recognized by the magic number being in the right place. Each filesytem can have more than one super block (they are spread out for redundancy), so we also report the size of the partition in "blocks" and the number of the super block group (#0, #1, ...).|
|root||This sector might contain an ext2 root directory as recognized by the "." and ".." directories at inode number 2. This should be found shortly after the first super block inside a partition, and it may help isolate where the real filesystem starts.|
# uname -a
# sfdisk -d /dev/hda
# cat /etc/fstab
Navigate: More Tech Tips