blob: d966ea98bff766d87ef60ddc2b80b68b686398d3 [file] [log] [blame]
User request:
BTW: Could you please add some sort of deleted and possibly corrupted file
and inode list to e2fsck report. There should be filenames deleted
from directory inodes, files with duplicate blocks e.t.c.
It's pretty annoying to filter this information from e2fsck output
by hand :-
------------------------------------------
Add a "answer Yes always to this class of question" response.
----------------------------------
ext2fs_flush() should return a different error message for primary
versus backup superblock flushing, so that mke2fs can print an
appropriate error message.
---------------------------------
Date: Mon, 08 Mar 1999 21:46:14 +0100
From: Sergio Polini <s.polini@mclink.it>
I'm reading the sorce code of e2fsck 1.14.
In pass2.c, lines 352-357, I read:
if ((dirent->name_len & 0xFF) > EXT2_NAME_LEN) {
if (fix_problem(ctx, PR_2_FILENAME_LONG, &cd->pctx)) {
dirent->name_len = EXT2_NAME_LEN;
dir_modified++;
}
}
I think that I'll never see any messages about too long filenames,
because "whatever & 0xFF" can never be "> 0xFF".
Am I wrong?
--------------------------------------
Add chmod command to debugfs.
------------------------------------------
Maybe a bug in debugfs v.1.14:
if a file has more than one hardlink, only the first filename is shown when
using command
ncheck <inode>
------------------------------------
Add a filesystem creation date to the superblock
-----------------------------------
Date: Tue, 18 Jan 2000 17:54:53 -0800 (PST)
From: Alan Blanchard <alan@abraxas.to>
To: tytso@MIT.EDU
Subject: DEBUGFS - thanks and a feature idea
Content-Type: TEXT/PLAIN; charset=US-ASCII
Theodore:
First, let me thank you for writing debugfs. Recently, my Linux box
(RH 6.0, 400 MHz PIII, on a DSL line) was hacked into. The intruder did
an "rm -Rf" on a 34 GB drive with about 5GB of data on it. I was able to
restore essentially the entire thing with debugfs and a bit of C code and Perl.
Actually, I could have done the entire thing with debugfs and Perl, but I
thought it would be too slow.
During this exercise, I noticed that one small feature was lacking that would
have made my job a bit easier. The length of a deleted directory is
reported as 0, hence debugfs won't dump the contents of the directory to a
file using the "dump" command. The only thing that saved me was that the
list of disk blocks is not zeroed out. I was able to dump the contents of the
directories by using debugfs to get the relevant block numbers, then
using dd to get the actual data.
If debugfs had a feature where it ignored the size of a directory reported by
the inode and instead just dumped all the blocks, it would have facilited
things a bit. This seems like a very easy feature to add.
Again, thanks for writing debugfs (and all the other Linux stuff you've written!).
Cheers,
Alan Blanchard
alan@abraxas.to
-------------------------------------------------------------------
Date: Fri, 21 Jan 2000 14:07:12 -0800
From: "H. Peter Anvin" <hpa@www.transmeta.com>
Subject: mkfs -cc and fsck -c
b) An option to mkfs to zero the partition. Yes, it can be done with
dd, but it would be a nicer way of doing it.
------------------------------------------------------------------
Add support for in ext2fs_block_iterate() for a returning the
compressed flag blocks to block_iterate. Change default to not return
EXT2_COMPRESSED_BLKADDR. Change e2fsck to pass this flag in.
(The old compression patches did this by default all the time, which
is bad, since it meant e2fsck never saw the EXT2_COMPRESSED_BLKADDR
flagword.
------------------------------------------------------------
E2fsck should offer to clear all the blocks in an indirect block, not
the entire inode, so there's better recovery for when an indirect
block gets trashed.
-------------------------------------------------------------
From: Yann Dirson - LOGATIQUE <Yann.Dirson@France.Sun.COM>
Date: Thu, 2 Mar 2000 13:52:13 +0100 (MET)
During my experiments on the broken system, I noticed the following in
the badblocks program (which I'm aware is not designed for IDE drives)
- I'd probably have already fixed them if my home system was up :(
* the syntax summary documents 2nd arg as blocks_count, which should
probably read something like end_count.
* testing past end of device is not detected, and lists those blocks
as bad, whereas they simply do not exist.
I think I'll probably add a "max count" option to findsuper(8), so
that I do not have to wait for the whole disk to be scanned when the
system had to be launched with "init=/bin/sh", in which case Ctrl-[CZ]
and friends appear to be absolutely ignored.
Somewhat unrelated, I just noticed the
http://web.mit.edu/tytso/www/linux/ext2.html could be updated:
- could mention SGI xfs (http://oss.sgi.com/projects/xfs/ - they just
release 0.03 snapshot)
----------------------------------------------------------------
Return-Path: <tytso@MIT.EDU>
Date: Thu, 10 Feb 2000 13:20:14 -0500
From: "Theodore Y. Ts'o" <tytso@MIT.EDU>
To: R.E.Wolff@BitWizard.nl
In-Reply-To: Rogier Wolff's message of Thu, 10 Feb 2000 08:46:30 +0100 (MET),
<200002100746.IAA24573@cave.bitwizard.nl>
Subject: Re: e2fsck request for enhancement.
Phone: (781) 391-3464
Date: Thu, 10 Feb 2000 08:46:30 +0100 (MET)
From: R.E.Wolff@BitWizard.nl (Rogier Wolff)
Lately, while trying to recover a broken disk, my system froze (twice,
until I tried something else) while copying the disk.
So I had a file of about 50Mb that was growing frantically at the
moment of the crash.
e2fsck, then finds an indirect block that is completely bogus. It
starts by asking me if it's ok to clear a few of the referenced
blocks. I say yes. Then it comes to the conclusion:
too many invalid blocks. Clear inode?
and then I get the option to delete the whole file. Not to truncate
the file to a "working" size.
I'd MUCH rather have e2fsck say something like:
inode 1234 references an invalid block 134345454. Hmm.
inode 1234 references 567 out of 50176 invalid blocks,
all near the end. Truncate file to 49152 blocks?
Here you can see that of the 1024 blocks near the end of the file,
only 567 were detected as invalid. However now 48Mb of the file will
be recovered, instead of thrown away.
That's a good point. Actually, the right thing is for e2fsck to offer
to clear all of the bad blocks in a particular indirect block. I don't
know how hard it would be to do that, but I'll put it on my e2fsprogs
TODO list.
- Ted
-----------------------------------------------------------------
Debugfs's link command should set the file type information
---------------------------------------------------------------