title | description | services | documentationcenter | author | manager | editor | tags | ms.assetid | ms.service | ms.workload | ms.tgt_pltfrm | ms.devlang | ms.topic | ms.date | ms.author |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Install and use Giraph on Linux-based HDInsight (Hadoop) | Microsoft Docs |
Learn how to install Giraph on Linux-based HDInsight clusters using Script Actions. Script Actions allow you to customize the cluster during creation, by changing cluster configuration or installing services and utilities. |
hdinsight |
Blackmist |
jhubbard |
cgronlun |
azure-portal |
9fcac906-8f06-4002-9fe8-473e42f8fd0f |
hdinsight |
big-data |
na |
na |
article |
10/17/2016 |
larryfr |
You can install Giraph on any type of cluster in Hadoop on Azure HDInsight by using Script Action to customize a cluster.
In this topic, you will learn how to install Giraph by using Script Action. Once you have installed Giraph, you'll also learn how to use Giraph for most typical applications, which is to process large-scale graphs.
Note
The information in this article is specific to Linux-based HDInsight clusters. For information on working with Windows-based clusters, see Install Giraph on HDInsight Hadoop clusters (Windows)
Apache Giraph allows you to perform graph processing by using Hadoop, and can be used with Azure HDInsight. Graphs model relationships between objects, such as the connections between routers on a large network like the Internet, or relationships between people on social networks (sometimes referred to as a social graph). Graph processing allows you to reason about the relationships between objects in a graph, such as:
- Identifying potential friends based on your current relationships.
- Identifying the shortest route between two computers in a network.
- Calculating the page rank of webpages.
Warning
Components provided with the HDInsight cluster are fully supported and Microsoft Support will help to isolate and resolve issues related to these components.
Custom components, such as Giraph, receive commercially reasonable support to help you to further troubleshoot the issue. This might result in resolving the issue OR asking you to engage available channels for the open source technologies where deep expertise for that technology is found. For example, there are many community sites that can be used, like: MSDN forum for HDInsight, http://stackoverflow.com. Also Apache projects have project sites on http://apache.org, for example: Hadoop.
This script performs the following actions:
- Installs Giraph to
/usr/hdp/current/giraph
- Copies the
giraph-examples.jar
file to default storage (WASB) for your cluster:/example/jars/giraph-examples.jar
A sample script to install Giraph on an HDInsight cluster is available at the following location.
https://hdiconfigactions.blob.core.windows.net/linuxgiraphconfigactionv01/giraph-installer-v01.sh
This section provides instructions on how to use the sample script while creating the cluster by using the Azure Portal.
Note
Azure PowerShell, the Azure CLI, the HDInsight .NET SDK, or Azure Resource Manager templates can also be used to apply script actions. You can also apply script actions to already running clusters. For more information, see Customize HDInsight clusters with Script Actions.
-
Start creating a cluster by using the steps in Create Linux-based HDInsight clusters, but do not complete creation.
-
On the Optional Configuration blade, select Script Actions, and provide the information below:
- NAME: Enter a friendly name for the script action.
- SCRIPT URI: https://hdiconfigactions.blob.core.windows.net/linuxgiraphconfigactionv01/giraph-installer-v01.sh
- HEAD: Check this option
- WORKER: Leave this unchecked
- ZOOKEEPER: Leave this unchecked
- PARAMETERS: Leave this field blank
-
At the bottom of the Script Actions, use the Select button to save the configuration. Finally, use the Select button at the bottom of the Optional Configuration blade to save the optional configuration information.
-
Continue creating the cluster as described in Create Linux-based HDInsight clusters.
Once the cluster has finished creating, use the following steps to run the SimpleShortestPathsComputation example included with Giraph. This implements the basic Pregel implementation for finding the shortest path between objects in a graph.
-
Connect to the HDInsight cluster using SSH:
For more information on using SSH with HDInsight, see the following:
-
Use the following to create a new file named tiny_graph.txt:
nano tiny_graph.txt
Use the following as the contents of this file:
[0,0,[[1,1],[3,3]]] [1,0,[[0,1],[2,2],[3,1]]] [2,0,[[1,2],[4,4]]] [3,0,[[0,3],[1,1],[4,4]]] [4,0,[[3,4],[2,4]]]
This data describes a relationship between objects in a directed graph, by using the format [source_id, source_value,[[dest_id], [edge_value],...]]. Each line represents a relationship between a source_id object and one or more dest_id objects. The edge_value (or weight) can be thought of as the strength or distance of the connection between source_id and dest_id.
Drawn out, and using the value (or weight) as the distance between objects, the above data might look like this:
-
To save the file, use Ctrl+X, then Y, and finally Enter to accept the file name.
-
Use the following to store the data into primary storage for your HDInsight cluster:
hdfs dfs -put tiny_graph.txt /example/data/tiny_graph.txt
-
Run the SimpleShortestPathsComputation example using the following command.
yarn jar /usr/hdp/current/giraph/giraph-examples.jar org.apache.giraph.GiraphRunner org.apache.giraph.examples.SimpleShortestPathsComputation -ca mapred.job.tracker=headnodehost:9010 -vif org.apache.giraph.io.formats.JsonLongDoubleFloatDoubleVertexInputFormat -vip /example/data/tiny_graph.txt -vof org.apache.giraph.io.formats.IdWithValueTextOutputFormat -op /example/output/shortestpaths -w 2
The parameters used with this command are described in the following table.
Parameter What it does jar /usr/hdp/current/giraph/giraph-examples.jar
The jar file containing the examples. org.apache.giraph.GiraphRunner
The class used to start the examples. org.apache.giraph.examples.SimpleShortestPathsCoputation
The example that will be ran. In this case, it will compute the shortest path between ID 1 and all other IDs in the graph. -ca mapred.job.tracker=headnodehost:9010
The headnode for the cluster. -vif org.apache.giraph.io.formats.JsonLongDoubleFloatDoubleVertexInputFromat
The input format to use for the input data. -vip /example/data/tiny_graph.txt
The input data file. -vof org.apache.giraph.io.formats.IdWithValueTextOutputFormat
The output format. In this case, ID and value as plain text. -op /example/output/shortestpaths
The output location. -w 2
The number of workers to use. In this case, 2. For more information on these, and other parameters used with Giraph samples, see the Giraph quickstart.
-
Once the job has finished, the results will be stored in the wasbs:///example/out/shotestpaths directory. The files created will begin with part-m- and end with a number indicating the first, second, etc. file. Use the following to view the output:
hdfs dfs -text /example/output/shortestpaths/*
The output should appear similar to the following:
0 1.0 4 5.0 2 2.0 1 0.0 3 1.0
The SimpleShortestPathComputation example is hard coded to start with object ID 1 and find the shortest path to other objects. So the output should be read as
destination_id distance
, where distance is the value (or weight) of the edges traveled between object ID 1 and the target ID.Visualizing this, you can verify the results by traveling the shortest paths between ID 1 and all other objects. Note that the shortest path between ID 1 and ID 4 is 5. This is the total distance between ID 1 and 3, and then ID 3 and 4.
- Install and use Hue on HDInsight clusters. Hue is a web UI that makes it easy to create, run and save Pig and Hive jobs, as well as browse the default storage for your HDInsight cluster.
- Install R on HDInsight clusters: Instructions on how to use cluster customization to install and use R on HDInsight Hadoop clusters. R is an open-source language and environment for statistical computing. It provides hundreds of built-in statistical functions and its own programming language that combines aspects of functional and object-oriented programming. It also provides extensive graphical capabilities.
- Install Solr on HDInsight clusters. Use cluster customization to install Solr on HDInsight Hadoop clusters. Solr allows you to perform powerful search operations on data stored.