Skip to content

Latest commit

 

History

History
436 lines (350 loc) · 28.3 KB

documentdb-run-hadoop-with-hdinsight.md

File metadata and controls

436 lines (350 loc) · 28.3 KB
title description services author manager editor documentationcenter ms.assetid ms.service ms.workload ms.tgt_pltfrm ms.devlang ms.topic ms.date ms.author
Run a Hadoop job using DocumentDB and HDInsight | Microsoft Docs
Learn how to run a simple Hive, Pig, and MapReduce job with DocumentDB and Azure HDInsight.
documentdb
dennyglee
jhubbard
mimig
06f0ea9d-07cb-4593-a9c5-ab912b62ac42
documentdb
data-services
na
java
article
09/20/2016
denlee

Run a Hadoop job using DocumentDB and HDInsight

This tutorial shows you how to run Apache Hive, Apache Pig, and Apache Hadoop MapReduce jobs on Azure HDInsight with DocumentDB's Hadoop connector. DocumentDB's Hadoop connector allows DocumentDB to act as both a source and sink for Hive, Pig, and MapReduce jobs. This tutorial will use DocumentDB as both the data source and destination for Hadoop jobs.

After completing this tutorial, you'll be able to answer the following questions:

  • How do I load data from DocumentDB using a Hive, Pig, or MapReduce job?
  • How do I store data in DocumentDB using a Hive, Pig, or MapReduce job?

We recommend getting started by watching the following video, where we run through a Hive job using DocumentDB and HDInsight.

[!VIDEO https://channel9.msdn.com/Blogs/Windows-Azure/Use-Azure-DocumentDB-Hadoop-Connector-with-Azure-HDInsight/player]

Then, return to this article, where you'll receive the full details on how you can run analytics jobs on your DocumentDB data.

Tip

This tutorial assumes that you have prior experience using Apache Hadoop, Hive, and/or Pig. If you are new to Apache Hadoop, Hive, and Pig, we recommend visiting the Apache Hadoop documentation. This tutorial also assumes that you have prior experience with DocumentDB and have a DocumentDB account. If you are new to DocumentDB or you do not have a DocumentDB account, please check out our Getting Started page.

Don't have time to complete the tutorial and just want to get the full sample PowerShell scripts for Hive, Pig, and MapReduce? Not a problem, get them here. The download also contains the hql, pig, and java files for these samples.

Newest Version

Hadoop Connector Version 1.2.0
Script Uri https://portalcontent.blob.core.windows.net/scriptaction/documentdb-hadoop-installer-v04.ps1
Date Modified 04/26/2016
Supported HDInsight Versions 3.1, 3.2
Change Log Updated DocumentDB Java SDK to 1.6.0
Added support for partitioned collections as both a source and sink

Prerequisites

Before following the instructions in this tutorial, ensure that you have the following:

Warning

In order to avoid the creation of a new collection during any of the jobs, you can either print the results to stdout, save the output to your WASB container, or specify an already existing collection. In the case of specifying an existing collection, new documents will be created inside the collection and already existing documents will only be affected if there is a conflict in ids. The connector will automatically overwrite existing documents with id conflicts. You can turn off this feature by setting the upsert option to false. If upsert is false and a conflict occurs, the Hadoop job will fail; reporting an id conflict error.

Step 1: Create a new HDInsight cluster

This tutorial uses Script Action from the Azure Portal to customize your HDInsight cluster. In this tutorial, we will use the Azure Portal to create your HDInsight cluster. For instructions on how to use PowerShell cmdlets or the HDInsight .NET SDK, check out the Customize HDInsight clusters using Script Action article.

  1. Sign in to the Azure Portal.

  2. Click + New on the top of the left navigation, search for HDInsight in the top search bar on the New blade.

  3. HDInsight published by Microsoft will appear at the top of the Results. Click on it and then click Create.

  4. On the New HDInsight Cluster create blade, enter your Cluster Name and select the Subscription you want to provision this resource under.

    Cluster nameName the cluster.
    DNS name must start and end with an alpha numeric character, and may contain dashes.
    The field must be a string between 3 and 63 characters.
    Subscription Name If you have more than one Azure Subscription, select the subscription that will host your HDInsight cluster.
  5. Click Select Cluster Type and set the following properties to the specified values.

    Cluster typeHadoop
    Cluster tierStandard
    Operating SystemWindows
    Versionlatest version

    Now, click SELECT.

    Provide Hadoop HDInsight initial cluster details

  6. Click on Credentials to set your login and remote access credentials. Choose your Cluster Login Username and Cluster Login Password.

    If you want to remote into your cluster, select yes at the bottom of the blade and provide a username and password.

  7. Click on Data Source to set your primary location for data access. Choose the Selection Method and specify an already existing storage account or create a new one.

  8. On the same blade, specify a Default Container and a Location. And, click SELECT.

    [!NOTE] Select a location close to your DocumentDB account region for better performance

  9. Click on Pricing to select the number and type of nodes. You can keep the default configuration and scale the number of Worker nodes later on.

  10. Click Optional Configuration, then Script Actions in the Optional Configuration Blade.

    In Script Actions, enter the following information to customize your HDInsight cluster.

    PropertyValue
    Name Specify a name for the script action.
    Script URI Specify the URI to the script that is invoked to customize the cluster.

    Please enter:
    https://portalcontent.blob.core.windows.net/scriptaction/documentdb-hadoop-installer-v04.ps1.
    Head Click the checkbox to run the PowerShell script onto the Head node.

    Check this checkbox.
    Worker Click the checkbox to run the PowerShell script onto the Worker node.

    Check this checkbox.
    Zookeeper Click the checkbox to run the PowerShell script onto the Zookeeper.

    Not needed.
    Parameters Specify the parameters, if required by the script.

    No Parameters needed.
  11. Create either a new Resource Group or use an existing Resource Group under your Azure Subscription.

  12. Now, check Pin to dashboard to track its deployment and click Create!

Step 2: Install and configure Azure PowerShell

  1. Install Azure PowerShell. Instructions can be found here.

    [!NOTE] Alternatively, just for Hive queries, you can use HDInsight's online Hive Editor. To do so, sign in to the Azure Portal, click HDInsight on the left pane to view a list of your HDInsight clusters. Click the cluster you want to run Hive queries on, and then click Query Console.

  2. Open the Azure PowerShell Integrated Scripting Environment:

    • On a computer running Windows 8 or Windows Server 2012 or higher, you can use the built-in Search. From the Start screen, type powershell ise and click Enter.
    • On a computer running a version earlier than Windows 8 or Windows Server 2012, use the Start menu. From the Start menu, type Command Prompt in the search box, then in the list of results, click Command Prompt. In the Command Prompt, type powershell_ise and click Enter.
  3. Add your Azure Account.

    1. In the Console Pane, type Add-AzureAccount and click Enter.
    2. Type in the email address associated with your Azure subscription and click Continue.
    3. Type in the password for your Azure subscription.
    4. Click Sign in.
  4. The following diagram identifies the important parts of your Azure PowerShell Scripting Environment.

    Diagram for Azure PowerShell

Step 3: Run a Hive job using DocumentDB and HDInsight

Important

All variables indicated by < > must be filled in using your configuration settings.

  1. Set the following variables in your PowerShell Script pane.

     # Provide Azure subscription name, the Azure Storage account and container that is used for the default HDInsight file system.
     $subscriptionName = "<SubscriptionName>"
     $storageAccountName = "<AzureStorageAccountName>"
     $containerName = "<AzureStorageContainerName>"
    
     # Provide the HDInsight cluster name where you want to run the Hive job.
     $clusterName = "<HDInsightClusterName>"
    
  2. Let's begin constructing your query string. We'll write a Hive query that takes all documents' system generated timestamps (_ts) and unique ids (_rid) from a DocumentDB collection, tallies all documents by the minute, and then stores the results back into a new DocumentDB collection.

    First, let's create a Hive table from our DocumentDB collection. Add the following code snippet to the PowerShell Script pane after the code snippet from #1. Make sure you include the optional DocumentDB.query parameter t trim our documents to just _ts and _rid.

    [!NOTE] Naming DocumentDB.inputCollections was not a mistake. Yes, we allow adding multiple collections as an input:

     '*DocumentDB.inputCollections*' = '*\<DocumentDB Input Collection Name 1\>*,*\<DocumentDB Input Collection Name 2\>*' A1A</br> The collection names are separated without spaces, using only a single comma.
    
     # Create a Hive table using data from DocumentDB. Pass DocumentDB the query to filter transferred data to _rid and _ts.
     $queryStringPart1 = "drop table DocumentDB_timestamps; "  +
                         "create external table DocumentDB_timestamps(id string, ts BIGINT) "  +
                         "stored by 'com.microsoft.azure.documentdb.hive.DocumentDBStorageHandler' "  +
                         "tblproperties ( " +
                             "'DocumentDB.endpoint' = '<DocumentDB Endpoint>', " +
                             "'DocumentDB.key' = '<DocumentDB Primary Key>', " +
                             "'DocumentDB.db' = '<DocumentDB Database Name>', " +
                             "'DocumentDB.inputCollections' = '<DocumentDB Input Collection Name>', " +
                             "'DocumentDB.query' = 'SELECT r._rid AS id, r._ts AS ts FROM root r' ); "
    
  3. Next, let's create a Hive table for the output collection. The output document properties will be the month, day, hour, minute, and the total number of occurrences.

    [!NOTE] Yet again, naming DocumentDB.outputCollections was not a mistake. Yes, we allow adding multiple collections as an output:
    'DocumentDB.outputCollections' = '<DocumentDB Output Collection Name 1>,<DocumentDB Output Collection Name 2>'
    The collection names are separated without spaces, using only a single comma.

    Documents will be distributed round-robin across multiple collections. A batch of documents will be stored in one collection, then a second batch of documents will be stored in the next collection, and so forth.

    # Create a Hive table for the output data to DocumentDB.
    $queryStringPart2 = "drop table DocumentDB_analytics; " +
                          "create external table DocumentDB_analytics(Month INT, Day INT, Hour INT, Minute INT, Total INT) " +
                          "stored by 'com.microsoft.azure.documentdb.hive.DocumentDBStorageHandler' " +
                          "tblproperties ( " +
                              "'DocumentDB.endpoint' = '<DocumentDB Endpoint>', " +
                              "'DocumentDB.key' = '<DocumentDB Primary Key>', " +  
                              "'DocumentDB.db' = '<DocumentDB Database Name>', " +
                              "'DocumentDB.outputCollections' = '<DocumentDB Output Collection Name>' ); "
    
  4. Finally, let's tally the documents by month, day, hour, and minute and insert the results back into the output Hive table.

     # GROUP BY minute, COUNT entries for each, INSERT INTO output Hive table.
     $queryStringPart3 = "INSERT INTO table DocumentDB_analytics " +
                           "SELECT month(from_unixtime(ts)) as Month, day(from_unixtime(ts)) as Day, " +
                           "hour(from_unixtime(ts)) as Hour, minute(from_unixtime(ts)) as Minute, " +
                           "COUNT(*) AS Total " +
                           "FROM DocumentDB_timestamps " +
                           "GROUP BY month(from_unixtime(ts)), day(from_unixtime(ts)), " +
                           "hour(from_unixtime(ts)) , minute(from_unixtime(ts)); "
    
  5. Add the following script snippet to create a Hive job definition from the previous query.

     # Create a Hive job definition.
     $queryString = $queryStringPart1 + $queryStringPart2 + $queryStringPart3
     $hiveJobDefinition = New-AzureHDInsightHiveJobDefinition -Query $queryString
    

    You can also use the -File switch to specify a HiveQL script file on HDFS.

  6. Add the following snippet to save the start time and submit the Hive job.

     # Save the start time and submit the job to the cluster.
     $startTime = Get-Date
     Select-AzureSubscription $subscriptionName
     $hiveJob = Start-AzureHDInsightJob -Cluster $clusterName -JobDefinition $hiveJobDefinition
    
  7. Add the following to wait for the Hive job to complete.

     # Wait for the Hive job to complete.
     Wait-AzureHDInsightJob -Job $hiveJob -WaitTimeoutInSeconds 3600
    
  8. Add the following to print the standard output and the start and end times.

     # Print the standard error, the standard output of the Hive job, and the start and end time.
     $endTime = Get-Date
     Get-AzureHDInsightJobOutput -Cluster $clusterName -JobId $hiveJob.JobId -StandardOutput
     Write-Host "Start: " $startTime ", End: " $endTime -ForegroundColor Green
    
  9. Run your new script! Click the green execute button.

  10. Check the results. Sign into the Azure Portal.

    1. Click Browse on the left-side panel.
    2. Click everything at the top-right of the browse panel.
    3. Find and click DocumentDB Accounts.
    4. Next, find your DocumentDB Account, then DocumentDB Database and your DocumentDB Collection associated with the output collection specified in your Hive query.
    5. Finally, click Document Explorer underneath Developer Tools.

    You will see the results of your Hive query.

    Hive query results

Step 4: Run a Pig job using DocumentDB and HDInsight

Important

All variables indicated by < > must be filled in using your configuration settings.

  1. Set the following variables in your PowerShell Script pane.

     # Provide Azure subscription name.
     $subscriptionName = "Azure Subscription Name"
    
     # Provide HDInsight cluster name where you want to run the Pig job.
     $clusterName = "Azure HDInsight Cluster Name"
    
  2. Let's begin constructing your query string. We'll write a Pig query that takes all documents' system generated timestamps (_ts) and unique ids (_rid) from a DocumentDB collection, tallies all documents by the minute, and then stores the results back into a new DocumentDB collection.

    First, load documents from DocumentDB into HDInsight. Add the following code snippet to the PowerShell Script pane after the code snippet from #1. Make sure to add a DocumentDB query to the optional DocumentDB query parameter to trim our documents to just _ts and _rid.

    [!NOTE] Yes, we allow adding multiple collections as an input:
    '<DocumentDB Input Collection Name 1>,<DocumentDB Input Collection Name 2>'
    The collection names are separated without spaces, using only a single comma.

    Documents will be distributed round-robin across multiple collections. A batch of documents will be stored in one collection, then a second batch of documents will be stored in the next collection, and so forth.

     # Load data from DocumentDB. Pass DocumentDB query to filter transferred data to _rid and _ts.
     $queryStringPart1 = "DocumentDB_timestamps = LOAD '<DocumentDB Endpoint>' USING com.microsoft.azure.documentdb.pig.DocumentDBLoader( " +
                                                     "'<DocumentDB Primary Key>', " +
                                                     "'<DocumentDB Database Name>', " +
                                                     "'<DocumentDB Input Collection Name>', " +
                                                     "'SELECT r._rid AS id, r._ts AS ts FROM root r' ); "
    
  3. Next, let's tally the documents by the month, day, hour, minute, and the total number of occurrences.

    # GROUP BY minute and COUNT entries for each.
    $queryStringPart2 = "timestamp_record = FOREACH DocumentDB_timestamps GENERATE `$0#'id' as id:int, ToDate((long)(`$0#'ts') * 1000) as timestamp:datetime; " +
                        "by_minute = GROUP timestamp_record BY (GetYear(timestamp), GetMonth(timestamp), GetDay(timestamp), GetHour(timestamp), GetMinute(timestamp)); " +
                        "by_minute_count = FOREACH by_minute GENERATE FLATTEN(group) as (Year:int, Month:int, Day:int, Hour:int, Minute:int), COUNT(timestamp_record) as Total:int; "
    
  4. Finally, let's store the results into our new output collection.

    [!NOTE] Yes, we allow adding multiple collections as an output:
    '<DocumentDB Output Collection Name 1>,<DocumentDB Output Collection Name 2>'
    The collection names are separated without spaces, using only a single comma.
    Documents will be distributed round-robin across the multiple collections. A batch of documents will be stored in one collection, then a second batch of documents will be stored in the next collection, and so forth.

     # Store output data to DocumentDB.
     $queryStringPart3 = "STORE by_minute_count INTO '<DocumentDB Endpoint>' " +
                         "USING com.microsoft.azure.documentdb.pig.DocumentDBStorage( " +
                             "'<DocumentDB Primary Key>', " +
                             "'<DocumentDB Database Name>', " +
                             "'<DocumentDB Output Collection Name>'); "
    
  5. Add the following script snippet to create a Pig job definition from the previous query.

     # Create a Pig job definition.
     $queryString = $queryStringPart1 + $queryStringPart2 + $queryStringPart3
     $pigJobDefinition = New-AzureHDInsightPigJobDefinition -Query $queryString -StatusFolder $statusFolder
    

    You can also use the -File switch to specify a Pig script file on HDFS.

  6. Add the following snippet to save the start time and submit the Pig job.

     # Save the start time and submit the job to the cluster.
     $startTime = Get-Date
     Select-AzureSubscription $subscriptionName
     $pigJob = Start-AzureHDInsightJob -Cluster $clusterName -JobDefinition $pigJobDefinition
    
  7. Add the following to wait for the Pig job to complete.

     # Wait for the Pig job to complete.
     Wait-AzureHDInsightJob -Job $pigJob -WaitTimeoutInSeconds 3600
    
  8. Add the following to print the standard output and the start and end times.

     # Print the standard error, the standard output of the Hive job, and the start and end time.
     $endTime = Get-Date
     Get-AzureHDInsightJobOutput -Cluster $clusterName -JobId $pigJob.JobId -StandardOutput
     Write-Host "Start: " $startTime ", End: " $endTime -ForegroundColor Green
    
  9. Run your new script! Click the green execute button.

  10. Check the results. Sign into the Azure Portal.

    1. Click Browse on the left-side panel.
    2. Click everything at the top-right of the browse panel.
    3. Find and click DocumentDB Accounts.
    4. Next, find your DocumentDB Account, then DocumentDB Database and your DocumentDB Collection associated with the output collection specified in your Pig query.
    5. Finally, click Document Explorer underneath Developer Tools.

    You will see the results of your Pig query.

    Pig query results

Step 5: Run a MapReduce job using DocumentDB and HDInsight

  1. Set the following variables in your PowerShell Script pane.

     $subscriptionName = "<SubscriptionName>"   # Azure subscription name
     $clusterName = "<ClusterName>"             # HDInsight cluster name
    
  2. We'll execute a MapReduce job that tallies the number of occurrences for each Document property from your DocumentDB collection. Add this script snippet after the snippet above.

     # Define the MapReduce job.
     $TallyPropertiesJobDefinition = New-AzureHDInsightMapReduceJobDefinition -JarFile "wasb:///example/jars/TallyProperties-v01.jar" -ClassName "TallyProperties" -Arguments "<DocumentDB Endpoint>","<DocumentDB Primary Key>", "<DocumentDB Database Name>","<DocumentDB Input Collection Name>","<DocumentDB Output Collection Name>","<[Optional] DocumentDB Query>"
    

    [!NOTE] TallyProperties-v01.jar comes with the custom installation of the DocumentDB Hadoop Connector.

  3. Add the following command to submit the MapReduce job.

     # Save the start time and submit the job.
     $startTime = Get-Date
     Select-AzureSubscription $subscriptionName
     $TallyPropertiesJob = Start-AzureHDInsightJob -Cluster $clusterName -JobDefinition $TallyPropertiesJobDefinition | Wait-AzureHDInsightJob -WaitTimeoutInSeconds 3600  
    

    In addition to the MapReduce job definition, you also provide the HDInsight cluster name where you want to run the MapReduce job, and the credentials. The Start-AzureHDInsightJob is an asynchronized call. To check the completion of the job, use the Wait-AzureHDInsightJob cmdlet.

  4. Add the following command to check any errors with running the MapReduce job.

     # Get the job output and print the start and end time.
     $endTime = Get-Date
     Get-AzureHDInsightJobOutput -Cluster $clusterName -JobId $TallyPropertiesJob.JobId -StandardError
     Write-Host "Start: " $startTime ", End: " $endTime -ForegroundColor Green
    
  5. Run your new script! Click the green execute button.

  6. Check the results. Sign into the Azure Portal.

    1. Click Browse on the left-side panel.

    2. Click everything at the top-right of the browse panel.

    3. Find and click DocumentDB Accounts.

    4. Next, find your DocumentDB Account, then DocumentDB Database and your DocumentDB Collection associated with the output collection specified in your MapReduce job.

    5. Finally, click Document Explorer underneath Developer Tools.

      You will see the results of your MapReduce job.

      MapReduce query results

Next Steps

Congratulations! You just ran your first Hive, Pig, and MapReduce jobs using Azure DocumentDB and HDInsight.

We have open sourced our Hadoop Connector. If you're interested, you can contribute on GitHub.

To learn more, see the following articles: