Install a basic cluster
From the CLI
To start a basic test cluster with three journal nodes, two name nodes, and three data nodes, run the following command on the DC/OS CLI.
dcos package install hdfs
This command creates a new instance with the default name hdfs
. Two instances cannot share the same name, so installing additional instances beyond the default instance requires customizing the name
at install time for each additional instance.
All dcos hdfs
CLI commands have a --name
argument allowing you to specify which instance to query. If you do not specify a service name, the CLI assumes a default value matching the package name, hdfs
. The default value for --name
can be customized via the DC/OS CLI configuration:
dcos hdfs --name=hdfs <cmd>
From the web interface
Alternatively, you can install from the DC/OS web interface. If you install Apache HDFS from the DC/OS web interface, the dcos hdfs
CLI commands are not automatically installed to your workstation. They may be manually installed using the DC/OS CLI:
dcos package install hdfs --cli
After running the package install
command, the service will begin installing.
Enterprise DC/OS installation
Depending on the security mode of the Enterprise DC/OS cluster, Enterprise DC/OS users may need to create a custom .json file and use it to install Apache HDFS.
Create a Configuration File
Create a custom configuration file that will be used to install Apache HDFS, and save it as config.json
. Specify the service account (<service_account_id>
) and a secret path (hdfs/<secret-name>
) .
{
"service": {
"service_account": "<service_account_id>",
"service_account_secret": "hdfs/<secret-name>"
}
}
Installing with a custom config file
Use the custom configuration file you just created to install Apache HDFS with this command:
dcos package install --options=config.json hdfs
Installing with external volumes enabled
Create a custom configuration file that will be used to install Apache HDFS, and save it as config.json
.
{
"service": {
"user": "root"
},
"node": {
"external_volume": {
"enabled": true,
"driver_name": "pxd",
}
}
}
Use the custom configuration file you just created to install Apache HDFS with this command:
dcos package install --options=config.json hdfs
Service Deployment
To monitor the deployment of your test instance, install the package cli (see command above) and run the command:
dcos hdfs plan status deploy
Once the deploy plan has a status of Complete
, the service is fully deployed.
Service Discovery
To connect a client, query the service for its endpoints.
dcos hdfs endpoints
Select an endpoint from the list to see the available connections.
dcos hdfs endpoints <endpoint>
Working with the Service
Using the endpoint information, you can connect a client to the service from within the DC/OS cluster (for example, a Marathon app running a client application). See the other sections of the documentation for more details on configuration, operation, and service capabilities.