Integrating PolyBase with Cloudera using Active Directory Authentication

Published Mar 23 2019 02:59 PM 179 Views
First published on MSDN on Oct 21, 2016
This article outlines the steps to use PolyBase in SQL 2016(including R-Services) with a Cloudera Cluster and setup authentication using Active Directory in both SQL 2016 and Cloudera.

Prerequisites



  1. Cloudera Cluster

  2. Active Directory with Domain Controller

  3. SQL Server 2016 with PolyBase and R-Services installed


NOTE: We have tested the configuration using the Cloudera Cluster 5.5 running on CentOS 6.6, SQL Server 2016 running on Windows Server 2012 R2 and Active Directory with Domain Controller running on Windows Server 2012 R2. Other Windows Server and CentOS operating systems might also work in this configuration.

All the prerequisites above must be in the same network and domain say ( CORP.CONTOSO.COM ). After the prerequisites are completed, we will follow the steps listed below in order:

  1. Connecting SQL to AD

  2. Connecting Cloudera to AD

  3. Connecting PolyBase to Cloudera



Connecting SQL 2016 with AD


Since SQL 2016 and DC are in the same domain CORP.CONTOSO.COM – you should be able to create a new login in SQL Server from an existing user in CORP.CONTOSO.COM


Connecting Cloudera with AD


For all usernames and principals, we will use the suffixes like Cluster14 for name-scalability.

  1. Active Directory setup:




  1. Install OpenLDAP utilities (openldap-clients on RHEL/Centos) on the host of Cloudera Manager server. Install Kerberos client (krb5-workstation on RHEL/Centos) on all hosts of the cluster. This step requires internet connection in Hadoop server. If there is no internet connection in the server, you can download the rpm and install.


sudo yum -y install openldap-clients krb5-workstation
sudo yum -y install krb5-workstation

  1. Apply the JCE Unlimited Strength Jurisdiction Policy Files. Download the Java Cryptography Extension (JCE) Unlimited Strength Jurisdiction Policy Files from Oracle. Be sure to download the correct policy file updates for your version of Java 7 or Java 8 . Uncompress and extract the downloaded file. The download includes a Readme.txt and two .jar files with the same names as the existing policy files. Locate the two existing policy files: local_policy.jar, US_export_policy.jar. Look in JAVA_HOME/lib/security/ and replace the existing policy files with the unlimited strength policy files you extracted.


We will use the wizard in Cloudera Manager to enable Active Directory Authentication. The 9 steps involved in the "Enable Kerberos" Wizard are provided through the following screenshots (use relevant values for your own cluster and AD):























You can view the credentials generated by Cloudera in the Active Directory OU=Hadoop, OU=CORP, DC=CONTOSO, DC=COM ( we gave the prefix “cluster14” in step 2)



Once Kerberos is successfully enabled, let us use kinit to obtain a ticket in cache and then list the directories in HDFS
kinit hdfsCluster14@CORP.CONTOSO.COM
hadoop fs -ls /
If the above command is successful, then we have configured AD Authentication for Cloudera!

Create a folder in hdfs for PolyBase tables (Say cdh)
hadoop fs -mkdir /cdh
NOTE: Make sure the hadoop.rpc.protection setting in HDFS is set to Authentication:



Currently there is a known issue when setting this to “integrity” or “privacy” will result in failures to connect from PolyBase to HDFS. You will see error message like the following:
EXTERNAL TABLE access failed due to internal error: 'Java exception raised on call to HdfsBridge_IsDirExist: Error
[Failed on local exception: java.io.IOException: Couldn't setup connection] occurred while accessing external file.'


Connecting PolyBase to Cloudera


Run the following command to confirm that PolyBase has been successfully installed. If PolyBase is installed, returns 1; otherwise, 0
SELECT SERVERPROPERTY ('IsPolybaseInstalled') AS IsPolybaseInstalled;
Run sp_configure (Transact-SQL) ‘hadoop connectivity’ and set an appropriate value. To find the value, see PolyBase Connectivity Configuration (Transact-SQL) .
sp_configure 'hadoop connectivity', 6;
sp_configure 'allow polybase export', 1;
reconfigure
You must restart SQL Server using services.msc . Restarting SQL Server restarts these services:

  • SQL Server PolyBase Data Movement Service

  • SQL Server PolyBase Engine




In the following location, set the appropriate values in the configuration files from the Cloudera Cluster settings:
C:\Program Files\Microsoft SQL Server\MSSQL13.MSSQLSERVER\MSSQL\Binn\Polybase\Hadoop\conf

core-site.xml


<property>
<name>polybase.kerberos.realm</name>
<value>CORP.CONTOSO.COM</value>
</property>
<property>
<name>polybase.kerberos.kdchost</name>
<value>ACTIVEDIRECTORY.CORP.CONTOSO.COM</value>
</property>
<property>
<name>hadoop.security.authentication</name>
<value>KERBEROS</value>
</property>

hdfs-site.xml


<property>
<name>dfs.namenode.kerberos.principal</name>
<value>hdfsCluster14/_HOST@CORP.CONTOSO.COM</value>
</property>

mapred-site.xml


<property>
<name>mapreduce.jobhistory.principal</name>
<value>mapred/_HOST@CORP.CONTOSO.COM</value>
</property>
<property>
<name>mapreduce.jobhistory.address</name>
<value><HOSTNAME and port of YARN JobHistory Server></value>
</property>

yarn-site.xml


<property>
<name>yarn.application.classpath</name>
<value>$HADOOP_CLIENT_CONF_DIR,$HADOOP_CONF_DIR,$HADOOP_COMMON_HOME/*,$HADOOP_COMMON_HOME/lib/*,
$HADOOP_HDFS_HOME/*,$HADOOP_HDFS_HOME/lib/*,$HADOOP_YARN_HOME/*,$HADOOP_YARN_HOME/lib/*</value>
</property>
<property>
<name>yarn.resourcemanager.principal</name>
<value>yarnCluster14/_HOST@CORP.CONTOSO.COM</value>
</property>
Now, we are ready to use PolyBase – let’s try creating an external table:
-- 1: Create a database scoped credential.
-- Create a master key on the database. This is required to encrypt the credential secret.
CREATE MASTER KEY ENCRYPTION BY PASSWORD = 'Pas5w0rd_';

-- 2: Create a database scoped credential  for Kerberos-secured Hadoop clusters.
-- IDENTITY: the Kerberos user name.
-- SECRET: the Kerberos password
CREATE DATABASE SCOPED CREDENTIAL myCredObject WITH IDENTITY = 'myHdfsUser', Secret = 'P455w0rd!#' ;

-- 3:  Create an external data source.
-- LOCATION (Required) : Hadoop Name Node IP address and port.
-- RESOURCE MANAGER LOCATION (Optional): Hadoop Resource Manager location to enable pushdown computation.
-- CREDENTIAL (Optional):  the database scoped credential, created above.
CREATE EXTERNAL DATA SOURCE clouderaCluster14 WITH (
TYPE = HADOOP,
LOCATION ='hdfs://CLUSTER14.CORP.CONTOSO.COM:8020',
RESOURCE_MANAGER_LOCATION = 'CLUSTER14.CORP.CONTOSO.COM:8032',
CREDENTIAL = myCredObject
);

-- 4: Create an external file format.
CREATE EXTERNAL FILE FORMAT CsvFileFormat WITH (
FORMAT_TYPE = DELIMITEDTEXT,
FORMAT_OPTIONS (FIELD_TERMINATOR =',', USE_TYPE_DEFAULT = TRUE))

-- 5:  Create an external table pointing to data stored in Hadoop.
-- LOCATION: path to file or directory that contains the data (relative to HDFS root).
CREATE EXTERNAL TABLE [dbo].[CarSensor_Data] (
[SensorKey] int NOT NULL,
[CustomerKey] int NOT NULL,
[GeographyKey] int NULL,
[Speed] float NOT NULL,
[YearMeasured] int NOT NULL
)
WITH (LOCATION='/cdh/',
DATA_SOURCE = clouderaCluster14,
FILE_FORMAT = CsvFileFormat
);

-- 6: Insert some data into external table and view the data
INSERT INTO [dbo].[CarSensor_Data] VALUES (1,1,1,40,2011)
SELECT * FROM [dbo].[CarSensor_Data]




The above data will be stored in CSV format in hdfs, you can browse the demo folder in hdfs to find the contents.



Now you can work with the table [dbo].[CarSensor_Data] as a normal table in SQL, but the data storage will be in HDFS.

You can also import existing data in HDFS as a SQL table to leverage awesome SQL features like Columnstore Indexes , R-services (In-Database)

Here is a simple example of using rxSummary on the external table [dbo].[CarSensor_Data]
exec sp_execute_external_script
@language =N'R',
@script=N'print(rxSummary(~.,InputDataSet))',
@input_data_1 =N'select * from [dbo].[CarSensor_Data]'














REFERENCES


Enabling Kerberos Authentication Using the Wizard


Create the HDFS Superuser


Direct Active Directory Integration for Kerberos Authentication


Quickly Configure Kerberos for Your Apache Hadoop Cluster


Integrating Cloudera cluster with Active Directory


PolyBase Guide


PolyBase Installation


Get Started with PolyBase

PolyBase Connectivity Configuration

PolyBase Configuration

PolyBase Setup Errors and Possible Solutions
Version history
Last update:
‎Mar 23 2019 02:59 PM
Updated by: