FreeBSD10 Stable + ZFS + PostgreSQL + SSD performance drop < 24 hours

Caza, Aaron Aaron.Caza at ca.weatherford.com
Sat Jun 17 05:16:28 UTC 2017


Regarding this issue, I've now conducted testing using merely a FreeBSD 10.3 Stable amd64 GENERIC kernel and using dd to read a large file.  The following is a log, taken hourly, of the degradation which occurred at just over 9 hours of uptime.  As the original is quite large, I've removed some sections; however, these can be supplied if desired.
Supplied are the initial dmesg and zpool status, logged only on startup, followed by uptime, uname -a, and zfs-stats -a output, each of which are logged hourly.


Copyright (c) 1992-2017 The FreeBSD Project.
Copyright (c) 1979, 1980, 1983, 1986, 1988, 1989, 1991, 1992, 1993, 1994
The Regents of the University of California. All rights reserved.
FreeBSD is a registered trademark of The FreeBSD Foundation.
FreeBSD 10.3-STABLE #0 r319701: Mon Jun 12 19:23:44 UTC 2017
    root at releng1.nyi.freebsd.org:/usr/obj/usr/src/sys/GENERIC amd64
FreeBSD clang version 3.4.1 (tags/RELEASE_34/dot1-final 208032) 20140512
CPU: Intel(R) Xeon(R) CPU E31240 @ 3.30GHz (3292.60-MHz K8-class CPU)
  Origin="GenuineIntel"  Id=0x206a7  Family=0x6  Model=0x2a  Stepping=7
  Features=0xbfebfbff<FPU,VME,DE,PSE,TSC,MSR,PAE,MCE,CX8,APIC,SEP,MTRR,PGE,MCA,CMOV,PAT,PSE36,CLFLUSH,DTS,ACPI,MMX,FXSR,SSE,SSE2,SS,HTT,TM,PBE>
  Features2=0x1dbae3ff<SSE3,PCLMULQDQ,DTES64,MON,DS_CPL,VMX,SMX,EST,TM2,SSSE3,CX16,xTPR,PDCM,PCID,SSE4.1,SSE4.2,x2APIC,POPCNT,TSCDLT,XSAVE,OSXSAVE,AVX>
  AMD Features=0x28100800<SYSCALL,NX,RDTSCP,LM>
  AMD Features2=0x1<LAHF>
  XSAVE Features=0x1<XSAVEOPT>
  VT-x: PAT,HLT,MTF,PAUSE,EPT,UG,VPID
  TSC: P-state invariant, performance statistics
real memory  = 8589934592 (8192 MB)
avail memory = 8219299840 (7838 MB)
Event timer "LAPIC" quality 600
ACPI APIC Table: <SUPERM SMCI--MB>
FreeBSD/SMP: Multiprocessor System Detected: 8 CPUs
FreeBSD/SMP: 1 package(s) x 4 core(s) x 2 SMT threads
 cpu0 (BSP): APIC ID:  0
 cpu1 (AP): APIC ID:  1
 cpu2 (AP): APIC ID:  2
 cpu3 (AP): APIC ID:  3
 cpu4 (AP): APIC ID:  4
 cpu5 (AP): APIC ID:  5
 cpu6 (AP): APIC ID:  6
 cpu7 (AP): APIC ID:  7
random: <Software, Yarrow> initialized
ioapic0 <Version 2.0> irqs 0-23 on motherboard
kbd1 at kbdmux0
cryptosoft0: <software crypto> on motherboard
acpi0: <SUPERM SMCI--MB> on motherboard
acpi0: Power Button (fixed)
cpu0: <ACPI CPU> on acpi0
cpu1: <ACPI CPU> on acpi0
cpu2: <ACPI CPU> on acpi0
cpu3: <ACPI CPU> on acpi0
cpu4: <ACPI CPU> on acpi0
cpu5: <ACPI CPU> on acpi0
cpu6: <ACPI CPU> on acpi0
cpu7: <ACPI CPU> on acpi0
attimer0: <AT timer> port 0x40-0x43 irq 0 on acpi0
Timecounter "i8254" frequency 1193182 Hz quality 0
Event timer "i8254" frequency 1193182 Hz quality 100
atrtc0: <AT realtime clock> port 0x70-0x71 irq 8 on acpi0
Event timer "RTC" frequency 32768 Hz quality 0
hpet0: <High Precision Event Timer> iomem 0xfed00000-0xfed003ff on acpi0
Timecounter "HPET" frequency 14318180 Hz quality 950
Event timer "HPET" frequency 14318180 Hz quality 550
Timecounter "ACPI-fast" frequency 3579545 Hz quality 900
acpi_timer0: <24-bit timer at 3.579545MHz> port 0x408-0x40b on acpi0
pcib0: <ACPI Host-PCI bridge> port 0xcf8-0xcff on acpi0
pci0: <ACPI PCI bus> on pcib0
em0: <Intel(R) PRO/1000 Network Connection 7.6.1-k> port 0xf020-0xf03f mem 0xfba00000-0xfba1ffff,0xfba24000-0xfba24fff irq 20 at device 25.0 on pci0
em0: Using an MSI interrupt
em0: Ethernet address: 00:25:90:76:6b:41
ehci0: <Intel Cougar Point USB 2.0 controller> mem 0xfba23000-0xfba233ff irq 16 at device 26.0 on pci0
usbus0: EHCI version 1.0
usbus0 on ehci0
pcib1: <ACPI PCI-PCI bridge> irq 17 at device 28.0 on pci0
pci1: <ACPI PCI bus> on pcib1
pcib2: <ACPI PCI-PCI bridge> irq 17 at device 28.4 on pci0
pci2: <ACPI PCI bus> on pcib2
em1: <Intel(R) PRO/1000 Network Connection 7.6.1-k> port 0xe000-0xe01f mem 0xfb900000-0xfb91ffff,0xfb920000-0xfb923fff irq 16 at device 0.0 on pci2
em1: Using MSIX interrupts with 3 vectors
em1: Ethernet address: 00:25:90:76:6b:40
ehci1: <Intel Cougar Point USB 2.0 controller> mem 0xfba22000-0xfba223ff irq 23 at device 29.0 on pci0
usbus1: EHCI version 1.0
usbus1 on ehci1
pcib3: <ACPI PCI-PCI bridge> at device 30.0 on pci0
pci3: <ACPI PCI bus> on pcib3
vgapci0: <VGA-compatible display> mem 0xfe000000-0xfe7fffff,0xfb800000-0xfb803fff,0xfb000000-0xfb7fffff irq 23 at device 3.0 on pci3
vgapci0: Boot video device
isab0: <PCI-ISA bridge> at device 31.0 on pci0
isa0: <ISA bus> on isab0
ahci0: <Intel Cougar Point AHCI SATA controller> port 0xf070-0xf077,0xf060-0xf063,0xf050-0xf057,0xf040-0xf043,0xf000-0xf01f mem 0xfba21000-0xfba217ff irq 19 at device 31.2 on pci0
ahci0: AHCI v1.30 with 6 6Gbps ports, Port Multiplier not supported
ahcich0: <AHCI channel> at channel 0 on ahci0
ahcich1: <AHCI channel> at channel 1 on ahci0
ahciem0: <AHCI enclosure management bridge> on ahci0
acpi_button0: <Power Button> on acpi0
atkbdc0: <Keyboard controller (i8042)> port 0x60,0x64 irq 1 on acpi0
atkbd0: <AT Keyboard> irq 1 on atkbdc0
kbd0 at atkbd0
atkbd0: [GIANT-LOCKED]
psm0: <PS/2 Mouse> irq 12 on atkbdc0
psm0: [GIANT-LOCKED]
psm0: model IntelliMouse Explorer, device ID 4
uart0: <16550 or compatible> port 0x3f8-0x3ff irq 4 flags 0x10 on acpi0
orm0: <ISA Option ROMs> at iomem 0xc0000-0xc7fff,0xc8000-0xc8fff on isa0
sc0: <System console> at flags 0x100 on isa0
sc0: VGA <16 virtual consoles, flags=0x300>
vga0: <Generic ISA VGA> at port 0x3c0-0x3df iomem 0xa0000-0xbffff on isa0
ppc0: cannot reserve I/O port range
est0: <Enhanced SpeedStep Frequency Control> on cpu0
est1: <Enhanced SpeedStep Frequency Control> on cpu1
est2: <Enhanced SpeedStep Frequency Control> on cpu2
est3: <Enhanced SpeedStep Frequency Control> on cpu3
est4: <Enhanced SpeedStep Frequency Control> on cpu4
est5: <Enhanced SpeedStep Frequency Control> on cpu5
est6: <Enhanced SpeedStep Frequency Control> on cpu6
est7: <Enhanced SpeedStep Frequency Control> on cpu7
ZFS filesystem version: 5
ZFS storage pool version: features support (5000)
Timecounters tick every 1.000 msec
md0: Preloaded image </boot/mfsroot> 17686528 bytes at 0xffffffff81daa1b8
random: unblocking device.
usbus0: 480Mbps High Speed USB v2.0
usbus1: 480Mbps High Speed USB v2.0
ugen0.1: <Intel EHCI root HUB> at usbus0
uhub0: <Intel EHCI root HUB, class 9/0, rev 2.00/1.00, addr 1> on usbus0
ugen1.1: <Intel EHCI root HUB> at usbus1
uhub1: <Intel EHCI root HUB, class 9/0, rev 2.00/1.00, addr 1> on usbus1
ada0 at ahcich0 bus 0 scbus0 target 0 lun 0
ada0: <Samsung SSD 850 PRO 256GB EXM03B6Q> ACS-2 ATA SATA 3.x device
ada0: Serial Number S39KNB0HB00482Y
ada0: 600.000MB/s transfers (SATA 3.x, UDMA6, PIO 512bytes)
ada0: Command Queueing enabled
ada0: 244198MB (500118192 512 byte sectors)
ada0: quirks=0x1<4K>
ada1 at ahcich1 bus 0 scbus1 target 0 lun 0
ada1: <Samsung SSD 850 PRO 256GB EXM03B6Q> ACS-2 ATA SATA 3.x device
ada1: Serial Number S39KNB0HB00473Z
ada1: 600.000MB/s transfers (SATA 3.x, UDMA6, PIO 512bytes)
ada1: Command Queueing enabled
ada1: 244198MB (500118192 512 byte sectors)
ada1: quirks=0x1<4K>
ses0 at ahciem0 bus 0 scbus2 target 0 lun 0
ses0: <AHCI SGPIO Enclosure 1.00 0001> SEMB S-E-S 2.00 device
ses0: SEMB SES Device
SMP: AP CPU #1 Launched!
SMP: AP CPU #6 Launched!
SMP: AP CPU #3 Launched!
SMP: AP CPU #5 Launched!
SMP: AP CPU #2 Launched!
SMP: AP CPU #4 Launched!
SMP: AP CPU #7 Launched!
Timecounter "TSC-low" frequency 1646298306 Hz quality 1000
Root mount waiting for: usbus1 usbus0
uhub1: 2 ports with 2 removable, self powered
uhub0: 2 ports with 2 removable, self powered
Root mount waiting for: usbus1 usbus0
ugen1.2: <vendor 0x8087 product 0x0024> at usbus1
uhub2: <vendor 0x8087 product 0x0024, class 9/0, rev 2.00/0.00, addr 2> on usbus1
ugen0.2: <vendor 0x8087 product 0x0024> at usbus0
uhub3: <vendor 0x8087 product 0x0024, class 9/0, rev 2.00/0.00, addr 2> on usbus0
Root mount waiting for: usbus1 usbus0
uhub2: 6 ports with 6 removable, self powered
uhub3: 6 ports with 6 removable, self powered
ugen1.3: <Weatherford SPD> at usbus1
Trying to mount root from ufs:/dev/md0 []...
bridge0: Ethernet address: 02:5d:9c:c3:f4:00
bridge0: link state changed to UP
em0: promiscuous mode enabled
em1: promiscuous mode enabled
em0: link state changed to UP

Zpool Status:
  pool: wwbase
 state: ONLINE
  scan: scrub repaired 0 in 0h1m with 0 errors on Sat Jun 10 18:01:26 2017
config:

NAME                                          STATE     READ WRITE CKSUM
wwbase                                        ONLINE       0     0     0
  mirror-0                                    ONLINE       0     0     0
    gpt/83b2ce34-4b9f-11e7-8d12-002590766b41  ONLINE       0     0     0
    gpt/8ec0d395-4b9f-11e7-8d12-002590766b41  ONLINE       0     0     0

errors: No known data errors

-------------------------------------------------------------------------------

Testing SSD performance @ Fri Jun 16 19:00:00 UTC 2017

FreeBSD xyz.test 10.3-STABLE FreeBSD 10.3-STABLE #0 r319701: Mon Jun 12 19:23:44 UTC 2017     root at releng1.nyi.freebsd.org:/usr/obj/usr/src/sys/GENERIC  amd64

 7:00PM  up 12 mins, 1 user, load averages: 0.00, 0.04, 0.07

Starting 'dd' test of large file...please wait
16000+0 records in
16000+0 records out
16777216000 bytes transferred in 18.844918 secs (890278004 bytes/sec)


------------------------------------------------------------------------
ZFS Subsystem ReportFri Jun 16 19:00:18 2017
------------------------------------------------------------------------

System Information:

Kernel Version:1003514 (osreldate)
Hardware Platform:amd64
Processor Architecture:amd64

ZFS Storage pool Version:5000
ZFS Filesystem Version:5

FreeBSD 10.3-STABLE #0 r319701: Mon Jun 12 19:23:44 UTC 2017 root
 7:00PM  up 13 mins, 1 user, load averages: 2.27, 0.57, 0.25

------------------------------------------------------------------------

System Memory:

0.19%14.90MiB Active,0.38%29.62MiB Inact
4.31%339.96MiB Wired,0.00%0 Cache
95.13%7.33GiB Free,0.00%4.00KiB Gap

Real Installed:8.00GiB
Real Available:99.18%7.93GiB
Real Managed:97.11%7.71GiB

Logical Total:8.00GiB
Logical Used:8.02%656.83MiB
Logical Free:91.98%7.36GiB

Kernel Memory:102.71MiB
Data:73.04%75.02MiB
Text:26.96%27.69MiB

Kernel Memory Map:7.71GiB
Size:2.48%196.06MiB
Free:97.52%7.51GiB

------------------------------------------------------------------------

ARC Summary: (HEALTHY)
Memory Throttle Count:0

ARC Misc:
Deleted:2.06m
Recycle Misses:0
Mutex Misses:10.06k
Evict Skips:1.49m

ARC Size:103.44%52.76MiB
Target Size: (Adaptive)100.00%51.00MiB
Min Size (Hard Limit):98.04%50.00MiB
Max Size (High Water):1:151.00MiB

ARC Size Breakdown:
Recently Used Cache Size:89.95%47.45MiB
Frequently Used Cache Size:10.05%5.30MiB

ARC Hash Breakdown:
Elements Max:11.92k
Elements Current:73.83%8.80k
Collisions:19.65k
Chain Max:2
Chains:25

------------------------------------------------------------------------

ARC Efficiency:4.06m
Cache Hit Ratio:49.05%1.99m
Cache Miss Ratio:50.95%2.07m
Actual Hit Ratio:49.05%1.99m

Data Demand Efficiency:99.91%1.94m
Data Prefetch Efficiency:0.01%2.05m

CACHE HITS BY CACHE LIST:
  Anonymously Used:0.01%159
  Most Recently Used:99.82%1.99m
  Most Frequently Used:0.16%3.28k
  Most Recently Used Ghost:0.00%9
  Most Frequently Used Ghost:0.00%83

CACHE HITS BY DATA TYPE:
  Demand Data:97.40%1.94m
  Prefetch Data:0.01%151
  Demand Metadata:2.59%51.44k
  Prefetch Metadata:0.01%100

CACHE MISSES BY DATA TYPE:
  Demand Data:0.09%1.76k
  Prefetch Data:99.07%2.05m
  Demand Metadata:0.07%1.50k
  Prefetch Metadata:0.77%16.01k

------------------------------------------------------------------------

L2ARC is disabled

------------------------------------------------------------------------

File-Level Prefetch: (HEALTHY)

DMU Efficiency:46.70k
Hit Ratio:34.40%16.06k
Miss Ratio:65.60%30.64k

Colinear:0
  Hit Ratio:100.00%0
  Miss Ratio:100.00%0

Stride:0
  Hit Ratio:100.00%0
  Miss Ratio:100.00%0

DMU Misc:
Reclaim:0
  Successes:100.00%0
  Failures:100.00%0

Streams:0
  +Resets:100.00%0
  -Resets:100.00%0
  Bogus:0

------------------------------------------------------------------------

VDEV Cache Summary:16.52k
Hit Ratio:1.65%272
Miss Ratio:93.47%15.44k
Delegations:4.88%806

------------------------------------------------------------------------

ZFS Tunables (sysctl):
kern.maxusers                           843
vm.kmem_size                            8273297408
vm.kmem_size_scale                      1
vm.kmem_size_min                        0
vm.kmem_size_max                        1319413950874
vfs.zfs.trim.max_interval               1
vfs.zfs.trim.timeout                    30
vfs.zfs.trim.txg_delay                  32
vfs.zfs.trim.enabled                    0
vfs.zfs.vol.unmap_enabled               1
vfs.zfs.vol.mode                        1
vfs.zfs.version.zpl                     5
vfs.zfs.version.spa                     5000
vfs.zfs.version.acl                     1
vfs.zfs.version.ioctl                   7
vfs.zfs.debug                           0
vfs.zfs.super_owner                     0
vfs.zfs.sync_pass_rewrite               2
vfs.zfs.sync_pass_dont_compress         5
vfs.zfs.sync_pass_deferred_free         2
vfs.zfs.zio.dva_throttle_enabled        1
vfs.zfs.zio.exclude_metadata            0
vfs.zfs.zio.use_uma                     1
vfs.zfs.cache_flush_disable             0
vfs.zfs.zil_replay_disable              0
vfs.zfs.min_auto_ashift                 9
vfs.zfs.max_auto_ashift                 13
vfs.zfs.vdev.trim_max_pending           10000
vfs.zfs.vdev.bio_delete_disable         0
vfs.zfs.vdev.bio_flush_disable          0
vfs.zfs.vdev.queue_depth_pct            1000
vfs.zfs.vdev.write_gap_limit            4096
vfs.zfs.vdev.read_gap_limit             32768
vfs.zfs.vdev.aggregation_limit          131072
vfs.zfs.vdev.trim_max_active            64
vfs.zfs.vdev.trim_min_active            1
vfs.zfs.vdev.scrub_max_active           2
vfs.zfs.vdev.scrub_min_active           1
vfs.zfs.vdev.async_write_max_active     10
vfs.zfs.vdev.async_write_min_active     1
vfs.zfs.vdev.async_read_max_active      3
vfs.zfs.vdev.async_read_min_active      1
vfs.zfs.vdev.sync_write_max_active      10
vfs.zfs.vdev.sync_write_min_active      10
vfs.zfs.vdev.sync_read_max_active       10
vfs.zfs.vdev.sync_read_min_active       10
vfs.zfs.vdev.max_active                 1000
vfs.zfs.vdev.async_write_active_max_dirty_percent60
vfs.zfs.vdev.async_write_active_min_dirty_percent30
vfs.zfs.vdev.mirror.non_rotating_seek_inc1
vfs.zfs.vdev.mirror.non_rotating_inc    0
vfs.zfs.vdev.mirror.rotating_seek_offset1048576
vfs.zfs.vdev.mirror.rotating_seek_inc   5
vfs.zfs.vdev.mirror.rotating_inc        0
vfs.zfs.vdev.trim_on_init               1
vfs.zfs.vdev.cache.bshift               16
vfs.zfs.vdev.cache.size                 10485760
vfs.zfs.vdev.cache.max                  16384
vfs.zfs.vdev.metaslabs_per_vdev         200
vfs.zfs.txg.timeout                     5
vfs.zfs.space_map_blksz                 4096
vfs.zfs.spa_min_slop                    134217728
vfs.zfs.spa_slop_shift                  5
vfs.zfs.spa_asize_inflation             24
vfs.zfs.deadman_enabled                 1
vfs.zfs.deadman_checktime_ms            5000
vfs.zfs.deadman_synctime_ms             1000000
vfs.zfs.debug_flags                     0
vfs.zfs.debugflags                      0
vfs.zfs.recover                         0
vfs.zfs.spa_load_verify_data            1
vfs.zfs.spa_load_verify_metadata        1
vfs.zfs.spa_load_verify_maxinflight     10000
vfs.zfs.ccw_retry_interval              300
vfs.zfs.check_hostid                    1
vfs.zfs.mg_fragmentation_threshold      85
vfs.zfs.mg_noalloc_threshold            0
vfs.zfs.condense_pct                    200
vfs.zfs.metaslab.bias_enabled           1
vfs.zfs.metaslab.lba_weighting_enabled  1
vfs.zfs.metaslab.fragmentation_factor_enabled1
vfs.zfs.metaslab.preload_enabled        1
vfs.zfs.metaslab.preload_limit          3
vfs.zfs.metaslab.unload_delay           8
vfs.zfs.metaslab.load_pct               50
vfs.zfs.metaslab.min_alloc_size         33554432
vfs.zfs.metaslab.df_free_pct            4
vfs.zfs.metaslab.df_alloc_threshold     131072
vfs.zfs.metaslab.debug_unload           0
vfs.zfs.metaslab.debug_load             0
vfs.zfs.metaslab.fragmentation_threshold70
vfs.zfs.metaslab.gang_bang              16777217
vfs.zfs.free_bpobj_enabled              1
vfs.zfs.free_max_blocks                 -1
vfs.zfs.no_scrub_prefetch               0
vfs.zfs.no_scrub_io                     0
vfs.zfs.resilver_min_time_ms            3000
vfs.zfs.free_min_time_ms                1000
vfs.zfs.scan_min_time_ms                1000
vfs.zfs.scan_idle                       50
vfs.zfs.scrub_delay                     4
vfs.zfs.resilver_delay                  2
vfs.zfs.top_maxinflight                 32
vfs.zfs.zfetch.array_rd_sz              1048576
vfs.zfs.zfetch.max_distance             8388608
vfs.zfs.zfetch.min_sec_reap             2
vfs.zfs.zfetch.max_streams              8
vfs.zfs.prefetch_disable                0
vfs.zfs.delay_scale                     500000
vfs.zfs.delay_min_dirty_percent         60
vfs.zfs.dirty_data_sync                 67108864
vfs.zfs.dirty_data_max_percent          10
vfs.zfs.dirty_data_max_max              4294967296
vfs.zfs.dirty_data_max                  851961036
vfs.zfs.max_recordsize                  1048576
vfs.zfs.send_holes_without_birth_time   1
vfs.zfs.mdcomp_disable                  0
vfs.zfs.nopwrite_enabled                1
vfs.zfs.dedup.prefetch                  1
vfs.zfs.l2c_only_size                   0
vfs.zfs.mfu_ghost_data_esize            0
vfs.zfs.mfu_ghost_metadata_esize        39650304
vfs.zfs.mfu_ghost_size                  39650304
vfs.zfs.mfu_data_esize                  0
vfs.zfs.mfu_metadata_esize              2785280
vfs.zfs.mfu_size                        2839040
vfs.zfs.mru_ghost_data_esize            253952
vfs.zfs.mru_ghost_metadata_esize        7712768
vfs.zfs.mru_ghost_size                  7966720
vfs.zfs.mru_data_esize                  40481280
vfs.zfs.mru_metadata_esize              3874816
vfs.zfs.mru_size                        49604608
vfs.zfs.anon_data_esize                 0
vfs.zfs.anon_metadata_esize             0
vfs.zfs.anon_size                       28672
vfs.zfs.l2arc_norw                      1
vfs.zfs.l2arc_feed_again                1
vfs.zfs.l2arc_noprefetch                1
vfs.zfs.l2arc_feed_min_ms               200
vfs.zfs.l2arc_feed_secs                 1
vfs.zfs.l2arc_headroom                  2
vfs.zfs.l2arc_write_boost               8388608
vfs.zfs.l2arc_write_max                 8388608
vfs.zfs.arc_meta_limit                  13369344
vfs.zfs.arc_free_target                 14047
vfs.zfs.compressed_arc_enabled          1
vfs.zfs.arc_shrink_shift                7
vfs.zfs.arc_average_blocksize           8192
vfs.zfs.arc_min                         52428800
vfs.zfs.arc_max                         53477376

------------------------------------------------------------------------


SSD performance testing completed @ Fri Jun 16 19:00:19 UTC 2017

-------------------------------------------------------------------------------




      This section was removed in the interests of brevity but can be
      supplied is required.




-------------------------------------------------------------------------------

Testing SSD performance @ Sat Jun 17 03:00:00 UTC 2017

FreeBSD xyz.test 10.3-STABLE FreeBSD 10.3-STABLE #0 r319701: Mon Jun 12 19:23:44 UTC 2017     root at releng1.nyi.freebsd.org:/usr/obj/usr/src/sys/GENERIC  amd64

 3:00AM  up  8:12, 0 users, load averages: 0.00, 0.00, 0.00

Starting 'dd' test of large file...please wait
16000+0 records in
16000+0 records out
16777216000 bytes transferred in 18.995611 secs (883215382 bytes/sec)


------------------------------------------------------------------------
ZFS Subsystem ReportSat Jun 17 03:00:19 2017
------------------------------------------------------------------------

System Information:

Kernel Version:1003514 (osreldate)
Hardware Platform:amd64
Processor Architecture:amd64

ZFS Storage pool Version:5000
ZFS Filesystem Version:5

FreeBSD 10.3-STABLE #0 r319701: Mon Jun 12 19:23:44 UTC 2017 root
 3:00AM  up  8:13, 0 users, load averages: 3.41, 0.80, 0.29

------------------------------------------------------------------------

System Memory:

0.04%2.78MiB Active,0.55%43.71MiB Inact
4.59%361.96MiB Wired,0.01%632.00KiB Cache
94.82%7.31GiB Free,0.00%4.00KiB Gap

Real Installed:8.00GiB
Real Available:99.18%7.93GiB
Real Managed:97.11%7.71GiB

Logical Total:8.00GiB
Logical Used:8.14%666.72MiB
Logical Free:91.86%7.35GiB

Kernel Memory:103.60MiB
Data:73.27%75.91MiB
Text:26.73%27.69MiB

Kernel Memory Map:7.71GiB
Size:2.66%210.04MiB
Free:97.34%7.50GiB

------------------------------------------------------------------------

ARC Summary: (HEALTHY)
Memory Throttle Count:0

ARC Misc:
Deleted:18.57m
Recycle Misses:0
Mutex Misses:91.37k
Evict Skips:13.33m

ARC Size:110.62%56.41MiB
Target Size: (Adaptive)100.00%51.00MiB
Min Size (Hard Limit):98.04%50.00MiB
Max Size (High Water):1:151.00MiB

ARC Size Breakdown:
Recently Used Cache Size:84.43%47.63MiB
Frequently Used Cache Size:15.57%8.78MiB

ARC Hash Breakdown:
Elements Max:11.93k
Elements Current:80.43%9.59k
Collisions:193.31k
Chain Max:3
Chains:30

------------------------------------------------------------------------

ARC Efficiency:36.57m
Cache Hit Ratio:49.18%17.98m
Cache Miss Ratio:50.82%18.58m
Actual Hit Ratio:49.18%17.98m

Data Demand Efficiency:99.96%17.53m
Data Prefetch Efficiency:0.00%18.43m

CACHE HITS BY CACHE LIST:
  Anonymously Used:0.00%588
  Most Recently Used:99.83%17.95m
  Most Frequently Used:0.16%29.16k
  Most Recently Used Ghost:0.00%56
  Most Frequently Used Ghost:0.00%263

CACHE HITS BY DATA TYPE:
  Demand Data:97.46%17.53m
  Prefetch Data:0.00%633
  Demand Metadata:2.54%456.26k
  Prefetch Metadata:0.00%275

CACHE MISSES BY DATA TYPE:
  Demand Data:0.04%7.74k
  Prefetch Data:99.15%18.43m
  Demand Metadata:0.03%5.77k
  Prefetch Metadata:0.77%143.89k

------------------------------------------------------------------------

L2ARC is disabled

------------------------------------------------------------------------

File-Level Prefetch: (HEALTHY)

DMU Efficiency:597.55k
Hit Ratio:24.22%144.70k
Miss Ratio:75.78%452.85k

Colinear:0
  Hit Ratio:100.00%0
  Miss Ratio:100.00%0

Stride:0
  Hit Ratio:100.00%0
  Miss Ratio:100.00%0

DMU Misc:
Reclaim:0
  Successes:100.00%0
  Failures:100.00%0

Streams:0
  +Resets:100.00%0
  -Resets:100.00%0
  Bogus:0

------------------------------------------------------------------------

VDEV Cache Summary:138.31k
Hit Ratio:0.57%794
Miss Ratio:98.82%136.67k
Delegations:0.61%840

------------------------------------------------------------------------

ZFS Tunables (sysctl):
kern.maxusers                           843
vm.kmem_size                            8273297408
vm.kmem_size_scale                      1
vm.kmem_size_min                        0
vm.kmem_size_max                        1319413950874
vfs.zfs.trim.max_interval               1
vfs.zfs.trim.timeout                    30
vfs.zfs.trim.txg_delay                  32
vfs.zfs.trim.enabled                    0
vfs.zfs.vol.unmap_enabled               1
vfs.zfs.vol.mode                        1
vfs.zfs.version.zpl                     5
vfs.zfs.version.spa                     5000
vfs.zfs.version.acl                     1
vfs.zfs.version.ioctl                   7
vfs.zfs.debug                           0
vfs.zfs.super_owner                     0
vfs.zfs.sync_pass_rewrite               2
vfs.zfs.sync_pass_dont_compress         5
vfs.zfs.sync_pass_deferred_free         2
vfs.zfs.zio.dva_throttle_enabled        1
vfs.zfs.zio.exclude_metadata            0
vfs.zfs.zio.use_uma                     1
vfs.zfs.cache_flush_disable             0
vfs.zfs.zil_replay_disable              0
vfs.zfs.min_auto_ashift                 9
vfs.zfs.max_auto_ashift                 13
vfs.zfs.vdev.trim_max_pending           10000
vfs.zfs.vdev.bio_delete_disable         0
vfs.zfs.vdev.bio_flush_disable          0
vfs.zfs.vdev.queue_depth_pct            1000
vfs.zfs.vdev.write_gap_limit            4096
vfs.zfs.vdev.read_gap_limit             32768
vfs.zfs.vdev.aggregation_limit          131072
vfs.zfs.vdev.trim_max_active            64
vfs.zfs.vdev.trim_min_active            1
vfs.zfs.vdev.scrub_max_active           2
vfs.zfs.vdev.scrub_min_active           1
vfs.zfs.vdev.async_write_max_active     10
vfs.zfs.vdev.async_write_min_active     1
vfs.zfs.vdev.async_read_max_active      3
vfs.zfs.vdev.async_read_min_active      1
vfs.zfs.vdev.sync_write_max_active      10
vfs.zfs.vdev.sync_write_min_active      10
vfs.zfs.vdev.sync_read_max_active       10
vfs.zfs.vdev.sync_read_min_active       10
vfs.zfs.vdev.max_active                 1000
vfs.zfs.vdev.async_write_active_max_dirty_percent60
vfs.zfs.vdev.async_write_active_min_dirty_percent30
vfs.zfs.vdev.mirror.non_rotating_seek_inc1
vfs.zfs.vdev.mirror.non_rotating_inc    0
vfs.zfs.vdev.mirror.rotating_seek_offset1048576
vfs.zfs.vdev.mirror.rotating_seek_inc   5
vfs.zfs.vdev.mirror.rotating_inc        0
vfs.zfs.vdev.trim_on_init               1
vfs.zfs.vdev.cache.bshift               16
vfs.zfs.vdev.cache.size                 10485760
vfs.zfs.vdev.cache.max                  16384
vfs.zfs.vdev.metaslabs_per_vdev         200
vfs.zfs.txg.timeout                     5
vfs.zfs.space_map_blksz                 4096
vfs.zfs.spa_min_slop                    134217728
vfs.zfs.spa_slop_shift                  5
vfs.zfs.spa_asize_inflation             24
vfs.zfs.deadman_enabled                 1
vfs.zfs.deadman_checktime_ms            5000
vfs.zfs.deadman_synctime_ms             1000000
vfs.zfs.debug_flags                     0
vfs.zfs.debugflags                      0
vfs.zfs.recover                         0
vfs.zfs.spa_load_verify_data            1
vfs.zfs.spa_load_verify_metadata        1
vfs.zfs.spa_load_verify_maxinflight     10000
vfs.zfs.ccw_retry_interval              300
vfs.zfs.check_hostid                    1
vfs.zfs.mg_fragmentation_threshold      85
vfs.zfs.mg_noalloc_threshold            0
vfs.zfs.condense_pct                    200
vfs.zfs.metaslab.bias_enabled           1
vfs.zfs.metaslab.lba_weighting_enabled  1
vfs.zfs.metaslab.fragmentation_factor_enabled1
vfs.zfs.metaslab.preload_enabled        1
vfs.zfs.metaslab.preload_limit          3
vfs.zfs.metaslab.unload_delay           8
vfs.zfs.metaslab.load_pct               50
vfs.zfs.metaslab.min_alloc_size         33554432
vfs.zfs.metaslab.df_free_pct            4
vfs.zfs.metaslab.df_alloc_threshold     131072
vfs.zfs.metaslab.debug_unload           0
vfs.zfs.metaslab.debug_load             0
vfs.zfs.metaslab.fragmentation_threshold70
vfs.zfs.metaslab.gang_bang              16777217
vfs.zfs.free_bpobj_enabled              1
vfs.zfs.free_max_blocks                 -1
vfs.zfs.no_scrub_prefetch               0
vfs.zfs.no_scrub_io                     0
vfs.zfs.resilver_min_time_ms            3000
vfs.zfs.free_min_time_ms                1000
vfs.zfs.scan_min_time_ms                1000
vfs.zfs.scan_idle                       50
vfs.zfs.scrub_delay                     4
vfs.zfs.resilver_delay                  2
vfs.zfs.top_maxinflight                 32
vfs.zfs.zfetch.array_rd_sz              1048576
vfs.zfs.zfetch.max_distance             8388608
vfs.zfs.zfetch.min_sec_reap             2
vfs.zfs.zfetch.max_streams              8
vfs.zfs.prefetch_disable                0
vfs.zfs.delay_scale                     500000
vfs.zfs.delay_min_dirty_percent         60
vfs.zfs.dirty_data_sync                 67108864
vfs.zfs.dirty_data_max_percent          10
vfs.zfs.dirty_data_max_max              4294967296
vfs.zfs.dirty_data_max                  851961036
vfs.zfs.max_recordsize                  1048576
vfs.zfs.send_holes_without_birth_time   1
vfs.zfs.mdcomp_disable                  0
vfs.zfs.nopwrite_enabled                1
vfs.zfs.dedup.prefetch                  1
vfs.zfs.l2c_only_size                   0
vfs.zfs.mfu_ghost_data_esize            0
vfs.zfs.mfu_ghost_metadata_esize        41795584
vfs.zfs.mfu_ghost_size                  41795584
vfs.zfs.mfu_data_esize                  0
vfs.zfs.mfu_metadata_esize              2670592
vfs.zfs.mfu_size                        2719744
vfs.zfs.mru_ghost_data_esize            1622016
vfs.zfs.mru_ghost_metadata_esize        6819840
vfs.zfs.mru_ghost_size                  8441856
vfs.zfs.mru_data_esize                  42278912
vfs.zfs.mru_metadata_esize              4341760
vfs.zfs.mru_size                        52709376
vfs.zfs.anon_data_esize                 0
vfs.zfs.anon_metadata_esize             0
vfs.zfs.anon_size                       147456
vfs.zfs.l2arc_norw                      1
vfs.zfs.l2arc_feed_again                1
vfs.zfs.l2arc_noprefetch                1
vfs.zfs.l2arc_feed_min_ms               200
vfs.zfs.l2arc_feed_secs                 1
vfs.zfs.l2arc_headroom                  2
vfs.zfs.l2arc_write_boost               8388608
vfs.zfs.l2arc_write_max                 8388608
vfs.zfs.arc_meta_limit                  13369344
vfs.zfs.arc_free_target                 14047
vfs.zfs.compressed_arc_enabled          1
vfs.zfs.arc_shrink_shift                7
vfs.zfs.arc_average_blocksize           8192
vfs.zfs.arc_min                         52428800
vfs.zfs.arc_max                         53477376

------------------------------------------------------------------------



SSD performance testing completed @ Sat Jun 17 03:00:19 UTC 2017

-------------------------------------------------------------------------------

Testing SSD performance @ Sat Jun 17 04:00:00 UTC 2017

FreeBSD xyz.test 10.3-STABLE FreeBSD 10.3-STABLE #0 r319701: Mon Jun 12 19:23:44 UTC 2017     root at releng1.nyi.freebsd.org:/usr/obj/usr/src/sys/GENERIC  amd64

 4:00AM  up  9:12, 0 users, load averages: 0.00, 0.00, 0.00

Starting 'dd' test of large file...please wait
16000+0 records in
16000+0 records out
16777216000 bytes transferred in 268.165167 secs (62562995 bytes/sec)


------------------------------------------------------------------------
ZFS Subsystem ReportSat Jun 17 04:04:28 2017
------------------------------------------------------------------------

System Information:

Kernel Version:1003514 (osreldate)
Hardware Platform:amd64
Processor Architecture:amd64

ZFS Storage pool Version:5000
ZFS Filesystem Version:5

FreeBSD 10.3-STABLE #0 r319701: Mon Jun 12 19:23:44 UTC 2017 root
 4:04AM  up  9:17, 0 users, load averages: 1.05, 0.67, 0.30

------------------------------------------------------------------------

System Memory:

0.04%2.80MiB Active,0.60%47.30MiB Inact
5.90%465.19MiB Wired,0.07%5.36MiB Cache
93.40%7.20GiB Free,0.00%4.00KiB Gap

Real Installed:8.00GiB
Real Available:99.18%7.93GiB
Real Managed:97.11%7.71GiB

Logical Total:8.00GiB
Logical Used:9.40%769.96MiB
Logical Free:90.60%7.25GiB

Kernel Memory:108.62MiB
Data:74.51%80.93MiB
Text:25.49%27.69MiB

Kernel Memory Map:7.71GiB
Size:2.68%211.31MiB
Free:97.32%7.50GiB

------------------------------------------------------------------------

ARC Summary: (HEALTHY)
Memory Throttle Count:0

ARC Misc:
Deleted:20.64m
Recycle Misses:0
Mutex Misses:682.38k
Evict Skips:2.33b

ARC Size:141.09%71.95MiB
Target Size: (Adaptive)100.00%51.00MiB
Min Size (Hard Limit):98.04%50.00MiB
Max Size (High Water):1:151.00MiB

ARC Size Breakdown:
Recently Used Cache Size:66.20%47.63MiB
Frequently Used Cache Size:33.80%24.32MiB

ARC Hash Breakdown:
Elements Max:11.93k
Elements Current:41.76%4.98k
Collisions:204.57k
Chain Max:3
Chains:7

------------------------------------------------------------------------

ARC Efficiency:40.75m
Cache Hit Ratio:49.30%20.09m
Cache Miss Ratio:50.70%20.66m
Actual Hit Ratio:49.29%20.08m

Data Demand Efficiency:99.96%19.58m
Data Prefetch Efficiency:0.00%20.47m

CACHE HITS BY CACHE LIST:
  Anonymously Used:0.02%3.46k
  Most Recently Used:99.63%20.02m
  Most Frequently Used:0.33%66.92k
  Most Recently Used Ghost:0.01%2.68k
  Most Frequently Used Ghost:0.00%604

CACHE HITS BY DATA TYPE:
  Demand Data:97.43%19.57m
  Prefetch Data:0.00%639
  Demand Metadata:2.54%510.40k
  Prefetch Metadata:0.03%6.10k

CACHE MISSES BY DATA TYPE:
  Demand Data:0.04%8.56k
  Prefetch Data:99.10%20.47m
  Demand Metadata:0.08%16.61k
  Prefetch Metadata:0.78%160.96k

------------------------------------------------------------------------

L2ARC is disabled

------------------------------------------------------------------------

File-Level Prefetch: (HEALTHY)

DMU Efficiency:775.18k
Hit Ratio:20.75%160.83k
Miss Ratio:79.25%614.36k

Colinear:0
  Hit Ratio:100.00%0
  Miss Ratio:100.00%0

Stride:0
  Hit Ratio:100.00%0
  Miss Ratio:100.00%0

DMU Misc:
Reclaim:0
  Successes:100.00%0
  Failures:100.00%0

Streams:0
  +Resets:100.00%0
  -Resets:100.00%0
  Bogus:0

------------------------------------------------------------------------

VDEV Cache Summary:156.84k
Hit Ratio:1.80%2.83k
Miss Ratio:97.53%152.96k
Delegations:0.67%1.05k

------------------------------------------------------------------------

ZFS Tunables (sysctl):
kern.maxusers                           843
vm.kmem_size                            8273297408
vm.kmem_size_scale                      1
vm.kmem_size_min                        0
vm.kmem_size_max                        1319413950874
vfs.zfs.trim.max_interval               1
vfs.zfs.trim.timeout                    30
vfs.zfs.trim.txg_delay                  32
vfs.zfs.trim.enabled                    0
vfs.zfs.vol.unmap_enabled               1
vfs.zfs.vol.mode                        1
vfs.zfs.version.zpl                     5
vfs.zfs.version.spa                     5000
vfs.zfs.version.acl                     1
vfs.zfs.version.ioctl                   7
vfs.zfs.debug                           0
vfs.zfs.super_owner                     0
vfs.zfs.sync_pass_rewrite               2
vfs.zfs.sync_pass_dont_compress         5
vfs.zfs.sync_pass_deferred_free         2
vfs.zfs.zio.dva_throttle_enabled        1
vfs.zfs.zio.exclude_metadata            0
vfs.zfs.zio.use_uma                     1
vfs.zfs.cache_flush_disable             0
vfs.zfs.zil_replay_disable              0
vfs.zfs.min_auto_ashift                 9
vfs.zfs.max_auto_ashift                 13
vfs.zfs.vdev.trim_max_pending           10000
vfs.zfs.vdev.bio_delete_disable         0
vfs.zfs.vdev.bio_flush_disable          0
vfs.zfs.vdev.queue_depth_pct            1000
vfs.zfs.vdev.write_gap_limit            4096
vfs.zfs.vdev.read_gap_limit             32768
vfs.zfs.vdev.aggregation_limit          131072
vfs.zfs.vdev.trim_max_active            64
vfs.zfs.vdev.trim_min_active            1
vfs.zfs.vdev.scrub_max_active           2
vfs.zfs.vdev.scrub_min_active           1
vfs.zfs.vdev.async_write_max_active     10
vfs.zfs.vdev.async_write_min_active     1
vfs.zfs.vdev.async_read_max_active      3
vfs.zfs.vdev.async_read_min_active      1
vfs.zfs.vdev.sync_write_max_active      10
vfs.zfs.vdev.sync_write_min_active      10
vfs.zfs.vdev.sync_read_max_active       10
vfs.zfs.vdev.sync_read_min_active       10
vfs.zfs.vdev.max_active                 1000
vfs.zfs.vdev.async_write_active_max_dirty_percent60
vfs.zfs.vdev.async_write_active_min_dirty_percent30
vfs.zfs.vdev.mirror.non_rotating_seek_inc1
vfs.zfs.vdev.mirror.non_rotating_inc    0
vfs.zfs.vdev.mirror.rotating_seek_offset1048576
vfs.zfs.vdev.mirror.rotating_seek_inc   5
vfs.zfs.vdev.mirror.rotating_inc        0
vfs.zfs.vdev.trim_on_init               1
vfs.zfs.vdev.cache.bshift               16
vfs.zfs.vdev.cache.size                 10485760
vfs.zfs.vdev.cache.max                  16384
vfs.zfs.vdev.metaslabs_per_vdev         200
vfs.zfs.txg.timeout                     5
vfs.zfs.space_map_blksz                 4096
vfs.zfs.spa_min_slop                    134217728
vfs.zfs.spa_slop_shift                  5
vfs.zfs.spa_asize_inflation             24
vfs.zfs.deadman_enabled                 1
vfs.zfs.deadman_checktime_ms            5000
vfs.zfs.deadman_synctime_ms             1000000
vfs.zfs.debug_flags                     0
vfs.zfs.debugflags                      0
vfs.zfs.recover                         0
vfs.zfs.spa_load_verify_data            1
vfs.zfs.spa_load_verify_metadata        1
vfs.zfs.spa_load_verify_maxinflight     10000
vfs.zfs.ccw_retry_interval              300
vfs.zfs.check_hostid                    1
vfs.zfs.mg_fragmentation_threshold      85
vfs.zfs.mg_noalloc_threshold            0
vfs.zfs.condense_pct                    200
vfs.zfs.metaslab.bias_enabled           1
vfs.zfs.metaslab.lba_weighting_enabled  1
vfs.zfs.metaslab.fragmentation_factor_enabled1
vfs.zfs.metaslab.preload_enabled        1
vfs.zfs.metaslab.preload_limit          3
vfs.zfs.metaslab.unload_delay           8
vfs.zfs.metaslab.load_pct               50
vfs.zfs.metaslab.min_alloc_size         33554432
vfs.zfs.metaslab.df_free_pct            4
vfs.zfs.metaslab.df_alloc_threshold     131072
vfs.zfs.metaslab.debug_unload           0
vfs.zfs.metaslab.debug_load             0
vfs.zfs.metaslab.fragmentation_threshold70
vfs.zfs.metaslab.gang_bang              16777217
vfs.zfs.free_bpobj_enabled              1
vfs.zfs.free_max_blocks                 -1
vfs.zfs.no_scrub_prefetch               0
vfs.zfs.no_scrub_io                     0
vfs.zfs.resilver_min_time_ms            3000
vfs.zfs.free_min_time_ms                1000
vfs.zfs.scan_min_time_ms                1000
vfs.zfs.scan_idle                       50
vfs.zfs.scrub_delay                     4
vfs.zfs.resilver_delay                  2
vfs.zfs.top_maxinflight                 32
vfs.zfs.zfetch.array_rd_sz              1048576
vfs.zfs.zfetch.max_distance             8388608
vfs.zfs.zfetch.min_sec_reap             2
vfs.zfs.zfetch.max_streams              8
vfs.zfs.prefetch_disable                0
vfs.zfs.delay_scale                     500000
vfs.zfs.delay_min_dirty_percent         60
vfs.zfs.dirty_data_sync                 67108864
vfs.zfs.dirty_data_max_percent          10
vfs.zfs.dirty_data_max_max              4294967296
vfs.zfs.dirty_data_max                  851961036
vfs.zfs.max_recordsize                  1048576
vfs.zfs.send_holes_without_birth_time   1
vfs.zfs.mdcomp_disable                  0
vfs.zfs.nopwrite_enabled                1
vfs.zfs.dedup.prefetch                  1
vfs.zfs.l2c_only_size                   0
vfs.zfs.mfu_ghost_data_esize            1933312
vfs.zfs.mfu_ghost_metadata_esize        45613056
vfs.zfs.mfu_ghost_size                  47546368
vfs.zfs.mfu_data_esize                  0
vfs.zfs.mfu_metadata_esize              475136
vfs.zfs.mfu_size                        1127936
vfs.zfs.mru_ghost_data_esize            4771840
vfs.zfs.mru_ghost_metadata_esize        1060864
vfs.zfs.mru_ghost_size                  5832704
vfs.zfs.mru_data_esize                  0
vfs.zfs.mru_metadata_esize              0
vfs.zfs.mru_size                        26115584
vfs.zfs.anon_data_esize                 0
vfs.zfs.anon_metadata_esize             0
vfs.zfs.anon_size                       147456
vfs.zfs.l2arc_norw                      1
vfs.zfs.l2arc_feed_again                1
vfs.zfs.l2arc_noprefetch                1
vfs.zfs.l2arc_feed_min_ms               200
vfs.zfs.l2arc_feed_secs                 1
vfs.zfs.l2arc_headroom                  2
vfs.zfs.l2arc_write_boost               8388608
vfs.zfs.l2arc_write_max                 8388608
vfs.zfs.arc_meta_limit                  13369344
vfs.zfs.arc_free_target                 14047
vfs.zfs.compressed_arc_enabled          1
vfs.zfs.arc_shrink_shift                7
vfs.zfs.arc_average_blocksize           8192
vfs.zfs.arc_min                         52428800
vfs.zfs.arc_max                         53477376

------------------------------------------------------------------------


SSD performance testing completed @ Sat Jun 17 04:04:28 UTC 2017

-------------------------------------------------------------------------------


Hopefully, the above proves useful to help track down this issue.

--
Aaron
This message may contain confidential and privileged information. If it has been sent to you in error, please reply to advise the sender of the error and then immediately delete it. If you are not the intended recipient, do not read, copy, disclose or otherwise use this message. The sender disclaims any liability for such unauthorized use. PLEASE NOTE that all incoming e-mails sent to Weatherford e-mail accounts will be archived and may be scanned by us and/or by external service providers to detect and prevent threats to our systems, investigate illegal or inappropriate behavior, and/or eliminate unsolicited promotional e-mails (spam). This process could result in deletion of a legitimate e-mail before it is read by its intended recipient at our organization. Moreover, based on the scanning results, the full text of e-mails and attachments may be made available to Weatherford security and other personnel for review and appropriate action. If you have any concerns about this process, please contact us at dataprivacy at weatherford.com.


More information about the freebsd-hackers mailing list