zfs on nvme: gnop breaks pool, zfs gets stuck
Karli Sjöberg
karli.sjoberg at slu.se
Thu Apr 28 05:52:53 UTC 2016
On Thu, 2016-04-28 at 07:48 +0200, Gerrit Kühn wrote:
> On Wed, 27 Apr 2016 15:14:36 +0100 Gary Palmer <gpalmer at freebsd.org>
> wrote
> about Re: zfs on nvme: gnop breaks pool, zfs gets stuck:
>
> GP> vfs.zfs.min_auto_ashift
> GP>
> GP> which lets you manage the ashift on a new pool without having to
> try
> GP> the gnop trick
>
> I just tried this, and it appears to work fine:
>
> ---
> root at storage:~ # sysctl vfs.zfs.min_auto_ashift
> vfs.zfs.min_auto_ashift: 12
>
> root at storage:~ # zpool status
> pool: data
> state: ONLINE
> scan: none requested
> config:
>
> NAME STATE READ WRITE CKSUM
> data ONLINE 0 0 0
> raidz2-0 ONLINE 0 0 0
> gpt/disk0 ONLINE 0 0 0
> gpt/disk1 ONLINE 0 0 0
> gpt/disk2 ONLINE 0 0 0
> gpt/disk3 ONLINE 0 0 0
> gpt/disk4 ONLINE 0 0 0
> gpt/disk5 ONLINE 0 0 0
> gpt/disk6 ONLINE 0 0 0
> gpt/disk7 ONLINE 0 0 0
> gpt/disk8 ONLINE 0 0 0
> gpt/disk9 ONLINE 0 0 0
> gpt/disk10 ONLINE 0 0 0
> gpt/disk11 ONLINE 0 0 0
>
> errors: No known data errors
>
> pool: flash
> state: ONLINE
> scan: none requested
> config:
>
> NAME STATE READ WRITE CKSUM
> flash ONLINE 0 0 0
> raidz1-0 ONLINE 0 0 0
> gpt/flash0 ONLINE 0 0 0
> gpt/flash1 ONLINE 0 0 0
> gpt/flash2 ONLINE 0 0 0
>
> errors: No known data errors
>
> root at storage:~ # zdb | grep ashift
> ashift: 12
> ashift: 12
>
> root at storage:~ # zpool list
> NAME SIZE ALLOC FREE EXPANDSZ FRAG CAP DEDUP HEALTH AL
> TROOT
> data 65T 1.88M 65.0T - 0% 0% 1.00x ONLINE -
> flash 1.39T 800K 1.39T - 0% 0% 1.00x ONLINE -
>
> ---
>
>
> I still wonder why the gnop workaround went so terribly wrong.
Again, because you need to tell zfs where the providers are:
# zpool import -d /dev/gpt flash
/K
> Anyway,
> thanks again for pointing out this new sysctl to me!
>
> And for the record: this is what I get with a simple linear write
> test:
>
> ---
> root at storage:~ # dd if=/dev/zero of=/flash/Z bs=1024k count=10000
> 10000+0 records in
> 10000+0 records out
> 10485760000 bytes transferred in 3.912829 secs (2679840997 bytes/sec)
> ---
>
>
> I guess I won't complain...
>
>
> cu
> Gerrit
> _______________________________________________
> freebsd-fs at freebsd.org mailing list
> https://lists.freebsd.org/mailman/listinfo/freebsd-fs
> To unsubscribe, send any mail to "freebsd-fs-unsubscribe at freebsd.org"
More information about the freebsd-fs
mailing list