Installing the Data Fabric Client on Windows
Installing the HPE Ezmeral Data Fabric client makes it possible to access the file system from a Windows workstation.
Compatibility with Network Address Translation (NAT) Adapters
In VM environments, the data-fabric client on Windows works with a single NAT virtual adapter as long as it is the only virtual adapter configured for the VM. If you want to use more than one adapter, you must use other types of virtual adapters. If you use multiple NAT adapters in your VM environment, your jobs and file-system operations will fail.
Use these steps to install the client:
-
To use the client with Release 6.2.0, make sure that a supported distribution of Java 11 is installed on the Windows computer. See JDK Support Matrix. To check the Java version, use this command in the Windows command prompt:
java -version
-
Create the
\opt\mapr
directory on yourc:
drive (or on another hard drive of your choosing). You can use Windows Explorer, or type the following at the command prompt:mkdir c:\opt\mapr
- Add the following environment variables:
System Variable Value JAVA_HOME
JAVA_HOME=C:\jdk-11
NOTE The path that you set for the JAVA_HOME environment variable must not include spaces.MAPR_HOME
MAPR_HOME=C:\opt\mapr
PATH
%JAVA_HOME%\bin %MAPR_HOME%\bin %MAPR_HOME%\hadoop\hadoop-2.7.4\bin
-
After adding environment variables, exit and reopen the command prompt.
- Download the client package archive:
- Navigate to the Internet
repository:
https://package.ezmeral.hpe.com/releases/v<version>/windows/<package name>
- Download the
mapr-client-6.2.0.0
package toC:\opt\mapr
. - Extract the archive by right-clicking the file and selecting Extract All....
- Specify
C:\opt\mapr\
as the folder where the files are extracted. If you extract the files to a subfolder ofC:\opt\mapr\
, such asC:\opt\mapr\mapr-client-6.2.0.0.<timestamp>
, theconfigure.bat
command can return errors.
- Navigate to the Internet
repository:
-
At the command prompt, run
configure.bat
to configure the client.In the following examples:-N
specifies the cluster name.-c
(lowercase) specifies a client configuration.-secure
is added if the cluster is secure.-C
(uppercase) specifies the CLDB nodes.-HS
specifies the HistoryServer node.- 7222 is the default port for the CLDB node.
Secure cluster exampleserver\configure.bat -N <cluster_name> -c -secure -C mynode01:7222,mynode02:7222,mynode03:7222
Non-secure cluster exampleserver\configure.bat -N <cluster_name> -c -C mynode01:7222,mynode02:7222,mynode03:7222
-
To use this client with a secure cluster or clusters, copy the
If this client will connect to multiple clusters, you must merge thessl_truststore
andssl-client.xml
files from the/opt/mapr/conf
directory on the cluster to thec:\opt\mapr\conf
directory on the client.ssl_truststore
files on the server by using the/opt/mapr/server/manageSSLKeys.sh
tool, and then copy the merged file toc:\opt\mapr\conf
on the client. For an example of merging thessl_truststore
files, see step 3 in Configuring Secure Clusters for Running Commands Remotely.For more information about connecting to a secure cluster, see Managing Secure Clusters.
- On the Windows computer, create a
ticket:
This command creates a ticket formaprlogin password -user <DataFabricUserName>
<DataFabricUserName>
, usually as:C:\Users\<WindowsUserName>\AppData\Local\Temp\maprticket_<WindowsUserName>
NOTE If you intend to run MapReduce jobs as<DataFabricUserName>
, set theMAPR_TICKETFILE_LOCATION
system variable toC:\Users\<WindowsUserName>\AppData\Local\Temp\maprticket_<DataFabricUserName>
. - Use the
hadoop fs -ls /
command to check for connectivity to the cluster. For example:hadoop fs -ls / Found 6 items drwxr-xr-x - mapr / 3 2020-08-24 00:39 /apps drwxr-xr-x - mapr / 0 2020-08-24 00:42 /hbase drwxr-xr-x - mapr / 0 2020-08-24 00:40 /opt drwxrwxrwx - mapr / 0 2020-08-26 20:36 /tmp drwxr-xr-x - mapr / 0 2020-08-24 00:40 /user drwxr-xr-x - mapr / 0 2020-08-24 00:39 /var