展示HN:nfsdiag - 一个NFS诊断应用程序
Show HN: nfsdiag – A NFS diagnostic application

原始链接: https://github.com/lsferreira42/nfsdiag

## nfs-doctor:一个客户端NFS调试工具 `nfs-doctor`是一个用于从客户端诊断NFS服务器问题的命令行工具。它检查网络连接、RPC服务、NFS版本、mountd、导出、权限(包括root squash、ACL和UID/GID映射)、锁定和基本性能。 该工具执行一系列测试——包括在`/tmp`下挂载导出、模拟用户访问和运行基本的I/O——以识别潜在问题。它以紧凑的文本、详细的冗长输出、JSON(用于自动化)或独立的HTML报告的形式输出结果。 主要功能包括:检测Kerberos配置、捕获RPC统计信息、运行文件系统检查和模拟陈旧句柄。性能测试可以使用内部循环或`fio`基准测试。选项允许进行模拟运行、速率限制以及通过UDP或特定IP版本进行测试。 虽然不能替代服务器端分析,但`nfs-doctor`可以快速将问题缩小到网络、NFS配置、权限或其他客户端因素。它还包括用于可重复测试的Docker fixtures。

Sorry.
相关文章

原文

nfs-doctor is a small command line tool written in C to help debug NFS servers from the client side.

The idea is simple: you give one IP or hostname, and the tool checks the things that usually break in NFS: network, rpcbind, NFS versions, mountd, exports, permissions, root squash, locking, stale handles and some basic performance.

It is not magic, and it will not replace a good server side analysis. But it helps a lot to understand if the problem is network, NFS config, permissions, UID/GID mapping, or something more strange.


Today nfs-doctor can do these checks:

  • test if rpcbind TCP port 111 is reachable
  • test if NFS TCP port 2049 is reachable
  • query the RPC service map from rpcbind (supports IPv6 fallback)
  • detect registered NFS, mountd, lockd/NLM and statd/NSM services
  • test NFS v2, v3 and v4 with RPC NULLPROC (including v4.1 and v4.2 hints)
  • test mountd v1, v2 and v3
  • optionally test RPC over UDP with --udp
  • enumerate exports using mountd
  • check client prerequisite daemons (nfs-client.target, rpc.gssd, nfs-idmapd)
  • detect Kerberos tickets and configuration with --krb5
  • mount exports automatically under /tmp/nfsdoctor-*
  • try NFS v4.2 first, then fallback to v4.1, v4, and v3
  • parse and verify effective mount options from /proc/self/mountinfo
  • capture RPC stats (retransmissions, auth refreshes) before and after tests
  • extract deep latency metrics from /proc/self/mountstats
  • run filesystem checks after mount (close-to-open consistency, special files, quotas)
  • test read/traverse permission
  • test directory listing
  • test POSIX ACLs, NFSv4 ACLs, generic xattrs, and SELinux contexts
  • test create/write/read/fsync when allowed (with configurable timeouts)
  • test advanced I/O operations: copy_file_range, fallocate, and O_DIRECT
  • test advisory locks with fcntl
  • detect practical root_squash behavior
  • simulate UID/GID access with prctl termination safety
  • simulate supplemental groups with --groups
  • run metadata latency test with create/rename/unlink
  • run stale file handle loop looking for ESTALE
  • check for pNFS layouts and NFSoRDMA connectivity
  • run external fio benchmarks alongside internal smoke tests
  • safely handle temporary files, mounts and folders using O_NOFOLLOW
  • perform safe audits with --dry-run and rate limiting (--delay-ms)
  • generate hierarchical JSON reports for automation
  • generate standalone HTML reports with inline CSS and Base64 (--html)
  • output colored text and progress bars on interactive terminals
  • run Docker fixture tests for regression checks

By default the output is compact. If you want all details, use --verbose.


NFS problems are very environment dependent. The result can change because of:

  • firewall rules
  • server export options
  • NFS version
  • kernel client state
  • UID/GID mapping
  • root squash
  • ACLs
  • SELinux/AppArmor on server
  • server load
  • stale file handles that only happen during real use

So, if the tool says no ESTALE happened, it means only that the tool did not reproduce it during the test window. It does not mean the problem can never happen.


On Debian or Ubuntu:

sudo apt-get update
sudo apt-get install -y build-essential pkg-config libtirpc-dev

On Fedora/RHEL style distros:

sudo dnf install -y gcc make pkgconf-pkg-config libtirpc-devel

For live mount tests you also need NFS client tools.

Debian or Ubuntu:

sudo apt-get install -y nfs-common

Fedora/RHEL style distros:

sudo dnf install -y nfs-utils

Normal build:

Clean and rebuild:

Small self-check:

Install:

Install in another prefix:

make PREFIX=/opt/nfs-doctor install

Uninstall:

Manual compile, if you want:

gcc -O2 -Wall -Wextra -I/usr/include/tirpc nfsdiag.c -ltirpc -o nfsdiag

Full diagnostic:

sudo ./nfsdiag 192.168.1.10

Verbose mode:

sudo ./nfsdiag --verbose 192.168.1.10

Only network and RPC checks, without mounting anything:

./nfsdiag --no-mount 192.168.1.10

Test only one export:

sudo ./nfsdiag --export /data 192.168.1.10

Pass mount options:

sudo ./nfsdiag --mount-options soft,timeo=30,retrans=2 192.168.1.10

Do not create/write test files:

sudo ./nfsdiag --read-only 192.168.1.10

Keep the temp folder for manual inspection:

sudo ./nfsdiag --keep-temp 192.168.1.10

Default output is clean and short. For example, in a healthy server it can be something like:

nfsdiag: 192.168.0.21
[OK] 1 export(s) discovered
summary: ok=13 warn=0 fail=0

If you want to see all probe steps, use:

./nfsdiag --verbose 192.168.0.21

Warnings and failures always appear in normal mode. Informational and low-level OK messages only appear in verbose mode.


For automation, use JSON.

JSON to stdout, without human text mixed together:

./nfsdiag --json 192.168.1.10

JSON to file, keeping stdout empty:

./nfsdiag --json=report.json 192.168.1.10

The JSON includes:

  • tool name
  • host
  • timestamp
  • system information (kernel, hostname, arch)
  • summary (ok, warn, fail)
  • options used
  • exports (hierarchical list of mount tests with performance metrics, ACLs, etc)
  • global events
  • recommendations

For human-readable reports that can be easily shared or attached to tickets, use HTML:

./nfsdiag --html=report.html 192.168.1.10

The generated HTML file is fully standalone.


UID/GID and permission tests

Simulate one UID/GID:

sudo ./nfsdiag --uid 1000 --gid 1000 192.168.1.10

Simulate more than one identity:

sudo ./nfsdiag --uid 1000 --gid 1000 --uid 65534 --gid 65534 192.168.1.10

Simulate supplemental groups:

sudo ./nfsdiag --uid 1000 --gid 1000 --groups 10,20,30 192.168.1.10

This is useful because many NFS problems are not really NFS protocol problems. Many times it is UID, GID, groups, ACL, or root squash.


Performance and stale handle tests

Change write/read test size:

sudo ./nfsdiag --bench-bytes 167772160 192.168.1.10

Change metadata latency iterations:

sudo ./nfsdiag --bench-iterations 500 192.168.1.10

Change stale handle loop:

sudo ./nfsdiag --stale-iterations 1000 192.168.1.10

Run benchmarks using fio instead of the internal C loop (requires fio installed):

sudo ./nfsdiag --bench-type=fio 192.168.1.10

The performance test is only a smoke test. It is not a replacement for full benchmarking, though enabling fio provides more accurate storage baseline metrics.


Timeout for external commands like mount and umount:

sudo ./nfsdiag --command-timeout 15 192.168.1.10

Delay between testing each export (rate limiting):

sudo ./nfsdiag --delay-ms 500 192.168.1.10

Simulate the tool execution without actually mounting or modifying anything:

./nfsdiag --dry-run 192.168.1.10

Try to isolate live mounts in a private mount namespace:

sudo ./nfsdiag --mount-namespace 192.168.1.10

This needs root or CAP_SYS_ADMIN.


Probe UDP RPC too:

./nfsdiag --no-mount --udp 192.168.1.10

Force IPv4 direct TCP checks:

./nfsdiag --ipv4-only --no-mount 192.168.1.10

Force IPv6 direct TCP checks:

./nfsdiag --ipv6-only --no-mount nfs-server.example.com

Disable NFSv4 pseudo-root fallback:

sudo ./nfsdiag --no-nfs4-discovery 192.168.1.10

The NFSv4 fallback is useful when the server is NFSv4-only and mountd is not available.


Usage: ./nfsdiag [OPTIONS] <server-ip-or-hostname>

Options:
  -e, --export PATH          Test only this export
  -o, --mount-options OPTS   Extra mount options
      --no-mount             Run network/RPC checks only
      --keep-temp            Do not remove /tmp/nfsdoctor-* after tests
      --read-only            Do not create/write test files
      --uid UID              Simulate access as UID
      --gid GID              GID paired with last --uid
      --groups G1,G2         Supplemental groups for simulation
      --timeout SEC          Network/RPC timeout
      --command-timeout SEC  Timeout for mount/umount commands
      --fs-timeout SEC       Timeout for filesystem operations in benchmarks
      --mount-namespace      Use private mount namespace when possible
      --json[=PATH]          Write JSON report
      --udp                  Probe RPC over UDP too
      --ipv4-only            Force IPv4 direct TCP checks
      --ipv6-only            Force IPv6 direct TCP checks
      --no-nfs4-discovery    Disable NFSv4 pseudo-root fallback
      --krb5                 Check Kerberos prerequisites (ticket, gssd)
      --bench-iterations N   Metadata latency iterations
      --stale-iterations N   ESTALE loop iterations
      --bench-bytes BYTES    Bytes used in read/write smoke test
  -v, --verbose              Show detailed output
  -h, --help                 Show help

  • 0: no warnings or failures
  • 1: warning or failure found
  • 2: usage error or local runtime error

Warnings return 1 because in automation they usually need attention.


The project has Docker fixtures to reproduce bad NFS situations.

List fixtures:

Build all fixtures:

Build one fixture:

make docker-build-read-only-export

Run automated fixture tests:

Run only one test:

make test-fixture-rpcbind-unreachable

Some tests need root because they do real NFS mounts from the host. If the host cannot run kernel NFS inside Docker, the test runner skips those cases instead of failing everything.

The current fixture set includes:

  • rpcbind-unreachable
  • nfs-port-unreachable
  • rpc-map-missing-nfs
  • mountd-unavailable
  • empty-exports
  • mount-denied
  • permission-denied
  • acl-unsupported
  • identity-denied
  • read-only-export
  • root-squash
  • locking-missing
  • stale-handle
  • slow-performance

What was added in the last big update

This version has these improvements:

  1. fully modular C architecture
  2. hierarchical per-export JSON output
  3. timeouts for filesystem operations (--fs-timeout)
  4. robust FD leak prevention and poll() migration
  5. deep metrics via /proc/self/mountstats and mountinfo
  6. RPC stats monitoring (/proc/net/rpc/nfs) for retransmissions
  7. client daemon prerequisite checks (rpcbind, nfs-client.target, idmapd)
  8. Kerberos detection and support (--krb5, gssd checks)
  9. NFSv4 ACL detection (system.nfs4_acl)
  10. NFSv4.1 and NFSv4.2 cascading mount support
  11. real IPv6 RPC support

Be careful when running against production exports.

By default the tool may create hidden .nfsdoctor-* files to test write/read behavior. If you do not want this, use:

Also, UID/GID simulation requires root because the tool uses setgid, setgroups, and setuid in child processes.


Some things are impossible to guarantee from the client side:

  • ESTALE only appears if the handle becomes stale during the test
  • SELinux/AppArmor problems can look only like generic permission denied
  • ACL info depends on what the NFS client exposes
  • performance numbers are only smoke-test values
  • Docker NFS fixtures depend on host kernel and Docker privileges

So use this tool as a fast diagnostic helper, not as the only source of truth.

联系我们 contact @ memedata.com