您的位置:首页 > 其它

Ubuntu12.04 安装配置NFSv4的官方文档

2012-08-21 17:05 169 查看
下面的文档虽然是英文的,但是很简单易懂,步骤很清晰,当然由于每个人需要配置的环境不宜样,可能出现各种问题。这一篇文章没有说对于NFSv4的一些操作,以及配置时候的注意点,可能会导致server无法关机。在下一篇文档里面会介绍,注意点。

原文链接:https://help.ubuntu.com/community/SettingUpNFSHowTo

Introduction

NFS (Network File System) allows you to 'share' a directory located on one networked computer with other computers/devices on that network. The computer 'sharing' the directory is called the server and the computers or devices connecting
to that server are called clients. The clients 'mount' the shared directory, it becomes part of their own directory structure.

NFS is perfect for a NAS (Networked Attached Storage) deployment in a Linux/Unix environment. It is a native Linux/Unix protocol as opposed to Samba which uses the SMB protocol developed by Microsoft. The Apple OS has good support for NFS.
Windows 7 has some support for NFS.

NFS is perhaps best for more 'permanent' network mounted directories such as /home directories or regularly accessed shared resources. If you want a network share that guest users can easily connect to, Samba is more suited. This is because
tools exist more readily across operating systems to temporarily mount and detach from Samba shares.

Before deploying NFS you should be familiar with:

Linux file and directory permissions
Mounting and detaching (unmounting) filesystems

NFSv4 quick start

Providing you understand what you are doing, use this brief walk-through to set up an NFSv4 server on Ubuntu (with no authentication security). Then mount the share on an Ubuntu client. It has been tested on Ubuntu 10.04 Lucid Lynx.

NFSv4 server

Install the required packages...

# apt-get install nfs-kernel-server


NFSv4 exports exist in a single pseudo filesystem, where the
real directories are mounted with the
--bind option.

Let's say we want to export our users' home directories in
/home/users. First we create the export filesystem:

# mkdir -p /export/users

It's important that /export and /export/users have 777 permissions as we will be accessing the NFS share from the client without LDAP/NIS authentication. This will not apply if
using authentication (see below). Now mount the real users directory with:

# mount --bind /home/users /export/users

To save us from retyping this after every reboot we add the following

line to /etc/fstab

/home/users    /export/users   none    bind  0  0


There are three configuration files that relate to an NFSv4 server:
/etc/default/nfs-kernel-server, /etc/default/nfs-common and
/etc/exports.

Those config files in our example would look like this:

In /etc/default/nfs-kernel-server we set:

NEED_SVCGSSD=no # no is default

because we are not activating NFSv4 security this time.

In /etc/default/nfs-common we set:

NEED_IDMAPD=yes
NEED_GSSD=no # no is default

because we want UID/GUID to be mapped from names.

In order for the ID names to be automatically mapped, both the client and server require the
/etc/idmapd.conf file to have the same contents with the correct domain names. Furthermore, this file should have the following lines in the
Mapping section:

[Mapping]

Nobody-User = nobody
Nobody-Group = nogroup


However, the client may have different requirements for the
Nobody-User and Nobody-Group. For example on
RedHat variants, it's nfsnobody for both. cat /etc/passwd and
cat /etc/group should show the "nobody" accounts.

This way, server and client do not need the users to share same UID/GUID.

For those who use LDAP-based authentication, add the following lines to your client's idmapd.conf:

[Translation]

Method = nsswitch


This will cause idmapd to know to look at nsswitch.conf to determine where it should look for credential information (and if you have LDAP authentication already working, nsswitch shouldn't require further explanation).

To export our directories to a local network 192.168.1.0/24

we add the following two lines to /etc/exports

/export       192.168.1.0/24(rw,fsid=0,insecure,no_subtree_check,async)
/export/users 192.168.1.0/24(rw,nohide,insecure,no_subtree_check,async)


Portmap Lockdown

optional

Add the following line to /etc/hosts.deny:

rpcbind mountd nfsd statd lockd rquotad : ALL


By blocking all clients first, only clients in /etc/hosts.allow below will be allowed to access the server.

Now add the following line to /etc/hosts.allow:

rpcbind mountd nfsd statd lockd rquotad : list of IP addresses


Where the "list of IP addresses" string is, you need to make a list
of IP addresses that consists of the server and all clients. These have
to be IP addresses because of a limitation in portmap (it doesn't like
hostnames). Note that if you have NIS set up, just add these to the
same line.

Now restart the service

# /etc/init.d/nfs-kernel-server restart


NFSv4 client

Install the required packages...

# apt-get install nfs-common


The client needs the same changes to /etc/default/nfs-common to connect to an NFSv4 server.

In /etc/default/nfs-common we set:

NEED_IDMAPD=yes
NEED_GSSD=no # no is default


because we want UID/GUID to be mapped from names. This way, server and client do not need the users to share same UID/GUID. Remember that mount/fstab defaults to NFSv3, so "mount -t nfs4" is necessary to make this work.

On the client we can mount the complete export tree with one command:

# mount -t nfs4 -o proto=tcp,port=2049 nfs-server:/ /mnt


Note that nfs-server:/export is not necessary in NFSv4, as it is in NFSv3. The root export
:/ defaults to export with fsid=0.

It can fail sometimes with the message

mount.nfs4: No such device


You have to load the nfs module by giving the command

# modprobe nfs


To make sure that the module is loaded at each boot, simply add nfs on the last line of
/etc/modules. We can also mount an exported
subtree with:

# mount -t nfs4 -o proto=tcp,port=2049 nfs-server:/users /home/users


To save us from retyping this after every reboot we add the following line to
/etc/fstab:

nfs-server:/   /mnt   nfs4    _netdev,auto  0  0


The auto option mounts on startup and the _netdev option waits until system network devices are loaded. However this will not work with
WiFi as
WiFi is set up at the user level (after login) not at system startup. If you use _netdev with
WiFi the boot process will pause waiting for the server to become available.

Note that _netdev only works with nfs version 3 and before. nfs4 ignores this option. Depending on how fast the network comes up on boot the mount entry may fail and the system will just keep booting. It can still be useful if you make your
own script to wait for the network to come up and then mount -a -O _netdev

Ubuntu Server doesn't come with any init.d/netfs or other scripts to do this for you.

Portmap Lockdown

optional

Add the following line to /etc/hosts.deny:

rpcbind : ALL


By blocking all clients first, only clients in /etc/hosts.allow below will be allowed to access the server.

Now add the following line to /etc/hosts.allow:

rpcbind : NFS server IP address


Where "NFS server IP address" is the IP address of the server.
This must be numeric! It's the way portmap works.

NFS Server

Pre-Installation Setup

None of the following pre-installation steps are strictly necessary.

User Permissions

NFS user permissions are based on user ID (UID). UIDs of any users on the client must match those on the server in order for the users to have access. The typical ways of doing this are:

Manual password file synchronization

Use of
LDAP

Use of
NIS

It's also important to note that you have to be careful on
systems where the main user has root access - that user can change
UID's on the system to allow themselves access to anyone's files. This
page assumes that the administrative team is the only group with root
access and that they are all trusted. Anything else represents a more
advanced configuration, and will not be addressed here.

Group Permissions

With NFS, a user's access to files is determined by his/her membership of groups on the client, not on the server. However, there is an important limitation: a maximum of 16 groups are passed from the client to the server, and, if a user
is member of more than 16 groups on the client, some files or directories might be unexpectedly inaccessible.

Host Names

optional if using DNS

Add any client name and IP addresses to /etc/hosts. The real (not 127.0.0.1) IP address of the server should already be here. This ensures that NFS will still work even if
DNS goes down. You could rely on DNS if you wanted, it's up to you.

NIS

optional - perform steps only if using NIS

Note: This only works if using NIS. Otherwise, you can't use netgroups, and should specify individual IP's or hostnames in
/etc/exports. Read the BUGS section in man netgroup.

Edit /etc/netgroup and add a line to classify your clients. (This
step is not necessary, but is for convenience).

myclients (client1,,) (client2,,)


Obviously, more clients can be added. myclients can be anything
you like; this is a netgroup name.

Run this command to rebuild the YP database:

sudo make -C /var/yp


Portmap Lockdown

optional

Add the following line to /etc/hosts.deny:

rpcbind mountd nfsd statd lockd rquotad : ALL


By blocking all clients first, only clients in /etc/hosts.allow below will be allowed to access the server.

Now add the following line to /etc/hosts.allow:

rpcbind mountd nfsd statd lockd rquotad : list of IP addresses


Where the "list of IP addresses" string is, you need to make a list
of IP addresses that consists of the server and all clients. These have
to be IP addresses because of a limitation in portmap (it doesn't like
hostnames). Note that if you have NIS set up, just add these to the
same line.

Installation and Configuration

Install NFS Server

sudo apt-get install rpcbind nfs-kernel-server


Shares

Edit /etc/exports and add the shares:

/home @myclients(rw,sync,no_subtree_check)
/usr/local @myclients(rw,sync,no_subtree_check)


The above shares /home and /usr/local to all clients in the
myclients netgroup.

/home 192.168.0.10(rw,sync,no_subtree_check) 192.168.0.11(rw,sync,no_subtree_check)
/usr/local 192.168.0.10(rw,sync,no_subtree_check) 192.168.0.11(rw,sync,no_subtree_check)


The above shares /home and /usr/local to two clients with fixed ip addresses.
Best used only with machines that have static ip addresses.

/home 192.168.0.0/255.255.255.0(rw,sync,no_subtree_check)
/usr/local 192.168.0.0/255.255.255.0(rw,sync,no_subtree_check)


The above shares /home and /usr/local to all clients in the
private network falling within the designated ip address range.

rw makes the share read/write, and sync requires
the server to only reply to requests once any changes have been flushed
to disk. This is the safest option (async is faster, but
dangerous. It is strongly recommended that you read
man exports.

After setting up /etc/exports, export the shares:

sudo exportfs -ra


You'll want to do this command whenever /etc/exports is modified.

Restart Services

By default, portmap only binds to the loopback interface. To enable access to portmap from remote machines, you need to change /etc/default/portmap to get rid of either "-l" or "-i 127.0.0.1".

If /etc/default/portmap was changed, portmap will need to be restarted:

sudo /etc/init.d/portmap restart


The NFS kernel server will also require a restart:

sudo /etc/init.d/nfs-kernel-server restart


Security Note

Aside from the UID issues discussed above, it should be noted that an
attacker could potentially masquerade as a machine that is allowed to
map the share, which allows them to create arbitrary UIDs to access
your files. One potential solution to this is
IPSec, see also the NFS and IPSec section below. You can set up all your domain members to talk only to each other over IPSec, which will effectively authenticate that your client is who it says it
is.

IPSec works by encrypting traffic to the server with the server's key,
and the server sends back all replies encrypted with the client's key.
The traffic is decrypted with the respective keys. If the client doesn't have the keys that the client is supposed to have, it can't send or receive data.

An alternative to IPSec is physically separate networks. This requires a separate network switch and separate ethernet cards, and physical security of that network.

NFS Client

Installation

sudo apt-get install rpcbind nfs-common


Portmap Lockdown

optional

Add the following line to /etc/hosts.deny:

rpcbind : ALL


By blocking all clients first, only clients in /etc/hosts.allow below will be allowed to access the server.

Now add the following line to /etc/hosts.allow:

rpcbind : NFS server IP address


Where "NFS server IP address" is the IP address of the server.
This must be numeric! It's the way portmap works.

Host Names

optional if using DNS

Add the server name to /etc/hosts. This ensures the NFS mounts will still work even if
DNS goes down. You could rely on DNS if you wanted, it's up to you.

Mounts

Check to see if everything works

You should try and mount it now. The basic template you will use is:

sudo mount ServerIP:/folder/already/setup/to/be/shared /home/username/folder/in/your/local/computer


so for example:

sudo mount 192.168.1.42:/home/music /home/poningru/music


Mount at startup

NFS mounts can either be automatically mounted when accessed using autofs or can be setup with static mounts using entries in /etc/fstab. Both are explained below.

Automounter

Install autofs:

sudo apt-get install autofs


The following configuration example sets up home directories to automount off an NFS server upon logging in. Other directories can be setup to automount upon access as well.

Add the following line to the end of /etc/auto.master:

/home         /etc/auto.home


Now create /etc/auto.home and insert the following:

*             solarisbox1.company.com.au,solarisbox2.company.com.au:/export/home/&


The above line automatically mounts any directory accessed at /home/[username] on the client machine from either solarisbox1.company.com.au:/export/home/[username] or solarisbox2.company.com.au:/export/home/[username].

Restart autofs to enable the configuration:

sudo /etc/init.d/autofs start


Static Mounts

Prior to setting up the mounts, make sure the directories that will act as mountpoints are already created.

In /etc/fstab, add lines for shares such as:

servername:dir /mntpoint nfs rw,hard,intr 0 0


The rw mounts it read/write. Obviously, if the server is sharing
it read only,
the client won't be able to mount it as anything more than that. The
hard mounts the share such that if the server becomes
unavailable, the program will wait until it is available. The
alternative is soft. intr allows you to interrupt/kill the
process. Otherwise, it will ignore you. Documentation for these can be
found in the Mount options for nfs section of
man mount.

The filesystems can now be mounted with mount /mountpoint, or
mount -a to mount everything that should be mounted at boot.

Notes

Minimalistic NFS Set Up

The steps above are very comprehensive. The minimum number of steps required to set up NFS are listed here:

http://www.ubuntuforums.org/showthread.php?t=249889

Using Groups with NFS Shares

When using groups on NFS shares (NFSv2 or NFSv3), keep in mind that this might not work if a user is a member of more than 16 groups. This is due to limitations in the NFS protocol. You can find more information on Launchpad ("Permission
denied when user belongs to group that owns group writable or setgid directories mounted via nfs") and in this article:

"What's the deal on the 16 group id limitation in NFS?"

IPSec Notes

If you're using IPSec, the default shutdown order in Breezy/Dapper causes the client to hang as it's being shut down because IPSec goes down before NFS does. To fix it, do:

sudo update-rc.d -f setkey remove
sudo update-rc.d setkey start 37 0 6 S .


A bug has been filed here:
https://launchpad.net/distros/ubuntu/+source/ipsec-tools/+bug/37536

Troubleshooting

Mounting NFS shares in encrypted home won't work on boot

Mounting an NFS share inside an encrypted home directory will only work after you are successfully logged in and your home is decrypted. This means that using
/etc/fstab to mount NFS shares on boot will not work - because your home has not been decrypted at the time of mounting. There is a simple way around this using Symbolic links:

Create an alternative directory to mount the NFS shares in:

$ sudo mkdir /nfs
$ sudo mkdir /nfs/music


Edit /etc/fstab to mount the NFS share into that directory instead:

nfsServer:music /nfs/music nfs4 _netdev,auto 0 0

Create a symbolic link inside your home, pointing to the actual mount location (in our case delete the 'Music' directory already existing there first):

$ rmdir /home/user/Music
$ ln -s /nfs/music/ /home/user/Music


Other resources

http://www.nfsv4.org/
内容来自用户分享和网络整理,不保证内容的准确性,如有侵权内容,可联系管理员处理 点击这里给我发消息
标签: