How to report bugs (Re: VERY frustrated with FreeBSD/UFS stability - please help or comment...)

Eric Anderson anderson at freebsd.org
Tue May 22 12:30:19 UTC 2007


On 05/21/07 15:37, Gore Jarold wrote:
> --- Kris Kennaway <kris at obsecurity.org> wrote:
> 
>> On Mon, May 21, 2007 at 12:16:33PM -0700, Gore
>> Jarold wrote:
>>
>>>>> a) am I really the only person in the world
>> that
>>>> moves
>>>>> around millions of inodes throughout the day ?
>>  Am
>>>> I
>>>>> the only person in the world that has ever
>> filled
>>>> up a
>>>>> snapshotted FS (or a quota'd FS, for that
>> matter)
>>>> ? 
> 
> 
> (snip)
> 
> 
>> You are certainly not the only persion who operates
>> on millions of
>> inodes, but it is disingenuous to suggest that this
>> is either a
>> "mainstream" or "simple" workload.  Also, I
>> personally know of several
>> people who do this without apparent problem, so that
>> is further
>> evidence that whatever problems you are seeing are
>> something specific
>> to your workload or configuration, or you are just
>> unlucky.
> 
> 
> Ok.  In my defense, I have to say that as a
> non-developer, end user, it's hard to watch people
> installing ZFS on FreeBSD and running with journaling
> and newfs'ing raw disk with 7.0-current, etc., and not
> feel like I am an extremely pedestrian use case.
> 
> I had no idea I was so cutting edge :)
> 
> 
> 
>> The larger issue here is that apparently you have
>> been suffering in
>> silence for many years with your various
>> frustrations and they have
>> finally exploded into this email.  This is really a
>> poor way to
>> approach the goal of getting your problems solved:
>> it is fundamentally
>> a failure of your expectations to think that without
>> adequately
>> reporting your bugs that they will somehow get
>> fixed.
> 
> 
> I need to clarify and respond to this ... my point was
> that every release since 5.0 has had some new and
> interesting instability in this regard.  Every time a
> new release comes out, it seems to be "fixed", only to
> reveal some other new and interesting instability.
> 
> So, no, I have not silently suffered with _any one_
> particular problem - they never seem to last more than
> one release or two.  It is only now, however, that I
> have come to realize that I am in the same spot
> (overall) today as I was in early 2004.  The details
> are slightly different, but the end result is that my
> rsyncs and cps and rms are too much for FreeBSD, and
> have been for 3 years now.
> 
> So what I am saying is, individual causes of
> instability (seem to) come and go, but I am not any
> better of today than I was with 5.0.  I have just
> realized this, and that is why I make my frustration
> known today.
> 
> 
>> Without these two things there is really very little
>> that a developer
>> can do to try and guess what might possibly be
>> happening on your
>> system.  However, it appears that we might now be
>> making some
>> progress:
>>
>>> ssh user at host rm -rf backup.2
>>> ssh user at host mv backup.1 backup.2
>>> ssh user at host cp -al backup.0 backup.1
>>> rsync /files user at host:/backup.0
>>>
>>> The /files in question range from .2 to 2.2
>> million
>>> files, all told. This means that when this script
>>> runs, it first either deletes OR unlinks up to 2
>>> million items.  Then it does a (presumably) zero
>> cost
>>> move operation.  Then it does a hard-link-creating
>> cp
>>> of the same (up to 2 million) items.
>> Please provide additional details of how the
>> filesystems in question
>> are configured, your kernel configuration, hardware
>> configuration, and
>> the debugging data referred to in 2) above.
> 
> 
> I will collect all of this and submit it the next time
> the system crashes...

For whatever it might be worth, I'm doing a very similar task (using 
rsnapshot), for backing up a decent amount of data, with a nightly 
difference of a million or so files, touching ~200million files nightly. 
  I currently have 5 10TB filesystems running, with 20TB more coming 
online today or tomorrow.

Here's a df output:

# df -ilk
Filesystem     1024-blocks       Used      Avail Capacity   iused 
ifree %iused  Mounted on
/dev/amrd0s3a     20308398    1114658   17569070     6%     23080 
2614742    1%   /
devfs                    1          1          0   100%         0 
    0  100%   /dev
/dev/amrd0s3e     18441132    3794030   13171812    22%        96 
2402206    0%   /tmp
/dev/amrd0s3d     20308398    3164982   15518746    17%    250111 
2387711    9%   /usr
/dev/ufs/vol1   9926678106 5030793092 4101750766    55%  38875256 
1244237702    3%   /vol1
/dev/ufs/vol2   9926678106 3668501950 5464041908    40%  67622249 
1215490709    5%   /vol2
/dev/ufs/vol3   9926678106 5937797134 3194746724    65%     97153 
1283015805    0%   /vol3
/dev/ufs/vol10  9925732858 8054663156 1077011074    88%  97873355 
1185121843    8%   /vol10
/dev/ufs/vol11  9925732858 7288510876 1843163354    80% 126038333 
1156956865   10%   /vol11


I have roughly 50-60 hardlinks per file (for about 80% of the files).

Obviously fsck is not an option (due to memory/time constraints) so I've 
been using gjournaling (thanks to PJD).

I *do* have one issue that has cropped up on one of these file systems, 
which I just recently found. I'll send a separate email with details.

I don't use snapshots, or background fsck on these at all, nor do I use 
quotas.

So, it can be done, and it can be done with pretty good reliability 
(whatever that might mean to any particular person).


Eric






More information about the freebsd-fs mailing list