NFS fails when using custom TCP/IP stack
search cancel

NFS fails when using custom TCP/IP stack

book

Article ID: 313941

calendar_today

Updated On:

Products

VMware VMware vCenter Server VMware vSphere ESXi

Issue/Introduction

Symptoms:
  • NFS traffic goes through the default stack even though default stack was created with a dedicated subnet.
  • NFS traffic fails to use custom TCP/IP stack.
  • Unable to add new NFS mounts with custom TCP/IP stack.


Environment

VMware Update Manager 6.0 Beta
VMware vCenter Converter Standalone 6.x
VMware vCenter Server 6.0.x
VMware Update Manager 6.0
VMware vCenter Server Appliance 6.0.x
VMware vSphere ESXi 6.0
VMware vCenter Server 6.0 Beta

Resolution

To resolve this issue, specify the Custom Netstack in NFS services for NFS to use custom TCP/IP stack.

To specify the Custom Netstack in NFS services:


1. Run this command to create the netstack:

# localcli network ip netstack add --netstack=nfs

2. Run this command to unmount any existing NFS shares:

# localcli storage nfs remove -v volume-name

3. Run this cmmnand to stop the NFS service:

# /etc/init.d/nfsgssd stop

4. Run this command to unload the NFS modules:

# vmkload_mod -u nfs41client
# vmkload_mod -u nfsclient
# vmkload_mod -u sunrpc

5. Run this command to configure the sunrpc module with the specified netstack:

# localcli system module parameters set -m sunrpc -p netstackName=nfs

6. Run this command to reload the NFS modules:

# localcli system module load -m sunrpc
# localcli system module load -m nfsclient
# localcli system module load -m nfs41client

7. Run this command to restart the NFS service:

# /etc/init.d/nfsgssd start

8. Run this command to remount the NFS shares:

# localcli storage nfs add -v volume name -H host -s /share