The NetApp Docker Volume Plugin (nDVP) provides direct integration with the Docker ecosystem for NetApp's ONTAP, E-Series, and SolidFire storage platforms. The nDVP package supports the provisioning and management of storage resources from the storage platform to Docker hosts, with a robust framework for adding additional platforms in the future.
Multiple instances of the nDVP can run concurrently on the same host. The allows simultaneous connections to multiple storage systems and storage types, with the ablity to customize the storage used for the Docker volume(s).
- Quick Start
- Running Multiple nDVP Instances
- Configuring your Docker host for NFS or iSCSI
- Global Config File Variables
- ONTAP Config File Variables
- E-Series Config File Variables
- SolidFire Config File Variables
-
Ensure you have Docker version 1.10 or above.
docker --version
If your version is out of date, update to the latest.
curl -fsSL https://get.docker.com/ | sh
-
After ensuring the correct version of Docker is installed, install and configure the NetApp Docker Volume Plugin. Note, you will need to ensure that NFS and/or iSCSI is configured for your system. See the installation instructions below for detailed information on how to do this.
# download and unpack the application wget https://github.com/NetApp/netappdvp/releases/download/v1.3/netappdvp-1.3.tar.gz tar zxf netappdvp-1.3.tar.gz # move to a location in the bin path sudo mv netappdvp /usr/local/bin sudo chown root:root /usr/local/bin/netappdvp sudo chmod 755 /usr/local/bin/netappdvp # create a location for the config files sudo mkdir -p /etc/netappdvp # create the configuration file, see below for more configuration examples cat << EOF > /etc/netappdvp/ontap-nas.json { "version": 1, "storageDriverName": "ontap-nas", "managementLIF": "10.0.0.1", "dataLIF": "10.0.0.2", "svm": "svm_nfs", "username": "vsadmin", "password": "netapp123", "aggregate": "aggr1" } EOF
-
After placing the binary and creating the configuration file(s), start the nDVP daemon using the desired configuration file.
Note: Unless specified, the default name for the volume driver will be "netapp".
sudo netappdvp --config=/etc/netappdvp/ontap-nas.json
-
Once the daemon is started, create and manage volumes using the Docker CLI interface.
docker volume create -d netapp --name ndvp_1
Provision Docker volume when starting a container:
docker run --rm -it --volume-driver netapp --volume ndvp_2:/my_vol alpine ash
Destroy docker volume:
docker volume rm ndvp_1 docker volume rm ndvp_2
-
Launch the plugin with an NFS configuration using a custom driver ID:
sudo netappdvp --volume-driver=netapp-nas --config=/path/to/config-nfs.json
-
Launch the plugin with an iSCSI configuration using a custom driver ID:
sudo netappdvp --volume-driver=netapp-san --config=/path/to/config-iscsi.json
-
Provision Docker volumes each driver instance:
NFS
docker volume create -d netapp-nas --name my_nfs_vol
iSCSI
docker volume create -d netapp-san --name my_iscsi_vol
Install the following system packages:
sudo yum install -y nfs-utils
sudo apt-get install -y nfs-common
-
Install the following system packages:
sudo yum install -y lsscsi iscsi-initiator-utils sg3_utils device-mapper-multipath
-
Start the multipathing daemon:
sudo mpathconf --enable --with_multipathd y
-
Ensure that
iscsid
andmultipathd
are enabled and running:sudo systemctl enable iscsid multipathd sudo systemctl start iscsid multipathd
-
Discover the iSCSI targets:
sudo iscsiadm -m discoverydb -t st -p <DATA_LIF_IP> --discover
-
Login to the discovered iSCSI targets:
sudo iscsiadm -m node -p <DATA_LIF_IP> --login
-
Start and enable
iscsi
:sudo systemctl enable iscsi sudo systemctl start iscsi
-
Install the following system packages:
sudo apt-get install -y open-iscsi lsscsi sg3-utils multipath-tools scsitools
-
Enable multipathing:
sudo tee /etc/multipath.conf <<-'EOF' defaults { user_friendly_names yes find_multipaths yes } EOF sudo service multipath-tools restart
-
Ensure that
iscsid
andmultipathd
are running:sudo service open-iscsi start sudo service multipath-tools start
-
Discover the iSCSI targets:
sudo iscsiadm -m discoverydb -t st -p <DATA_LIF_IP> --discover
-
Login to the discovered iSCSI targets:
sudo iscsiadm -m node -p <DATA_LIF_IP> --login
Option | Description | Example |
---|---|---|
version | Config file version number | 1 |
storageDriverName | ontap-nas , ontap-san , eseries-iscsi , or solidfire-san |
ontap-nas |
debug | Turn debugging output on or off | false |
storagePrefix | Optional prefix for volume names. Default: "netappdvp_" | netappdvp_ |
A new config file variable has been added in v1.2 called "storagePrefix" that allows you to modify the prefix
applied to volume names by the plugin. By default, when you run docker volume create
, the volume name
supplied is prepended with "netappdvp_". ("netappdvp-" for SolidFire.)
If you wish to use a different prefix, you can specify it with this directive. Alternatively, you can use
pre-existing volumes with the volume plugin by setting storagePrefix
to an empty string, "".
A note of caution: docker volume rm
will delete these volumes just as it does volumes created by the
plugin using the default prefix. Be very careful when using pre-existing volumes!
In addition to the global configuration values above, when using clustered Data ONTAP, these options are avaialble.
Option | Description | Example |
---|---|---|
managementLIF | IP address of clustered Data ONTAP management LIF | 10.0.0.1 |
dataLIF | IP address of protocol lif; will be derived if not specified | 10.0.0.2 |
svm | Storage virtual machine to use (req, if management LIF is a cluster LIF) | svm_nfs |
username | Username to connect to the storage device | vsadmin |
password | Password to connect to the storage device | netapp123 |
aggregate | Aggregate to use for volume/LUN provisioning | aggr1 |
NFS Example for ontap-nas driver
{
"version": 1,
"storageDriverName": "ontap-nas",
"managementLIF": "10.0.0.1",
"dataLIF": "10.0.0.2",
"svm": "svm_nfs",
"username": "vsadmin",
"password": "netapp123",
"aggregate": "aggr1"
}
iSCSI Example for ontap-san driver
{
"version": 1,
"storageDriverName": "ontap-san",
"managementLIF": "10.0.0.1",
"dataLIF": "10.0.0.3",
"svm": "svm_iscsi",
"username": "vsadmin",
"password": "netapp123",
"aggregate": "aggr1"
}
In addition to the global configuration values above, when using E-Series, these options are available.
Option | Description | Example |
---|---|---|
webProxyHostname | Hostname or IP address of Web Services Proxy | localhost |
webProxyPort | Port number of the Web Services Proxy (optional) | 8443 |
webProxyUseHTTP | Use HTTP instead of HTTPS for Web Services Proxy (default = false) | true |
webProxyVerifyTLS | Verify server's certificate chain and hostname (default = false) | true |
username | Username for Web Services Proxy | rw |
password | Password for Web Services Proxy | rw |
controllerA | IP address of controller A | 10.0.0.5 |
controllerB | IP address of controller B | 10.0.0.6 |
passwordArray | Password for storage array if set | blank/empty |
hostData_IP | Host iSCSI IP address (if multipathing just choose either one) | 10.0.0.101 |
Example for eseries-iscsi driver
{
"version": 1,
"storageDriverName": "eseries-iscsi",
"debug": true,
"webProxyHostname": "localhost",
"webProxyPort": "8443",
"webProxyUseHTTP": false,
"webProxyVerifyTLS": true,
"username": "rw",
"password": "rw",
"controllerA": "10.0.0.5",
"controllerB": "10.0.0.6",
"passwordArray": "",
"hostData_IP": "10.0.0.101"
}
The E-Series Docker driver assumes that you have a volume group or a DDP pool pre-configured (N number of drives; segment size; RAID type; ...). The driver then allocates Docker volumes out of this volume group or DDP pool. The volume group and/or DDP pool must be given a specific name and there must be two groups allocated. For example, you could create a volume group/DDP pool named 'netappdvp_hdd' and another named 'netappdvp_ssd'.
When creating a docker volume you can specify the volume size as well as the allocation group/DDP pool using the '-o' option and the tags 'size' and 'mediaType'. Note that these are optional; if unspecified, the defaults will be a 1 GB volume allocated from the HDD pool. An example of using these tags to create a 2 GB volume from the SSD volume group/DDP pool:
docker volume create -d netapp --name my_vol -o size=2g -o mediaType=ssd
Note that the current driver is meant to be used with iSCSI.
Compatibility note: When assigning LUN numbers for volumes attached to Linux hosts, start with LUN 1 rather than LUN 0. If an array is connected to the host before a device is mapped to LUN 0, the Linux host will detect the REPORT LUNS well known logical unit as LUN 0, so that it can complete discovery. LUN 0 might not immediately map properly with a simple rescan, depending on the version of the host operating system in use.
See SANtricity Storage Manager 11.20 SAS Configuration and Provisioning for Linux Express Guide for more details.
In addition to the global configuration values above, when using SolidFire, these options are avaialble.
Option | Description | Example |
---|---|---|
Endpoint | Ex. https://:@/json-rpc/ | |
SVIP | iSCSI IP address and port | 10.0.0.7:3260 |
TenantName | SF Tenant to use (created if not found) | "docker" |
DefaultVolSz | Volume size in GiB | 1 |
InitiatorIFace | Specify interface when restricting iSCSI traffic to non-default interface | "default" |
Types | QoS specifications | See below |
{
"version": 1,
"storageDriverName": "solidfire-san",
"debug": false,
"Endpoint": "https://admin:admin@192.168.160.3/json-rpc/7.0",
"SVIP": "10.0.0.7:3260",
"TenantName": "docker",
"DefaultVolSz": 1,
"InitiatorIFace": "default",
"Types": [
{
"Type": "Bronze",
"Qos": {
"minIOPS": 1000,
"maxIOPS": 2000,
"burstIOPS": 4000
}
},
{
"Type": "Silver",
"Qos": {
"minIOPS": 4000,
"maxIOPS": 6000,
"burstIOPS": 8000
}
},
{
"Type": "Gold",
"Qos": {
"minIOPS": 6000,
"maxIOPS": 8000,
"burstIOPS": 10000
}
}
]
}
##Known Issues and Limitations
-
Volume names must be a minimum of 2 characters in length
This is a Docker client limitation. The client will interpret a single character name as being a Windows path. Bug 25773