Hops (Hadoop Open Platform-as-a-Service) is a next generation distribution of Apache Hadoop with scalable, highly available, customizable metadata. Hops consists internally of two main sub projects, HopsFs and HopsYarn. HopsFS is a new implementation of the the Hadoop Filesystem (HDFS), that supports multiple stateless NameNodes, where the metadata is stored in MySQL Cluster, an in-memory distributed database. HopsFS enables more scalable clusters than Apache HDFS (up to ten times larger clusters), and enables NameNode metadata to be both customized and analyzed, because it can now be easily accessed via a SQL API. HopsYARN introduces a distributed stateless Resource Manager, whose state is migrated to MySQL Cluster. This enables our YARN architecture to have no down-time, with failover of a ResourceManager happening in a few seconds. Together, HopsFS and HopsYARN enable Hadoop clusters to scale to larger volumes and higher throughput.
You can find the latest Hops documentation, including a programming guide, on the project web page. This README file only contains basic setup and compilation instructions.
For compiling the Hops Hadoop Distribution you will need the following software.
- Java 1.7 or higher
- Maven
- cmake for compiling the native code
- Google Protocol Buffer Version 2.5
- MySQL Cluster NDB native client library
We combine Apache and GPL licensed code, from Hops and MySQL Cluster, respectively, by providing a DAL API (similar to JDBC). We dynamically link our DAL implementation for MySQL Cluster with the Hops code. Both binaries are distributed separately.
Perform the following steps in the following order to compile the Hops Hadoop Distribution.
git clone https://github.com/hopshadoop/hops-metadata-dal
The master
branch represents most stable version of Hops Hadoop and the develop
branch contains all the newly developed features and bug fixes. If you choose to use master
branch then also checkout the master
branch in the other Hops Projects and vice versa.
git checkout master
mvn clean install -DskipTests
git clone https://github.com/hopshadoop/hops-metadata-dal-impl-ndb
git checkout master
mvn clean install -DskipTests
This project also contains c++ code that requires NDB libndbclient.so
library. Download the MySQL Cluster Distribution and extract the libndbclient.so
library. Alternatively you can download a custom MySQL Cluster library from our servers. Our custom library supports binding I/O threads to CPU cores for better performance.
cd tmp
wget http://kompics.sics.se/maven/repository/com/mysql/ndb/clusterj-native/7.5.4/clusterj-native-7.5.4-natives-linux.jar
unzip clusterj-native-7.5.4-natives-linux.jar
cp libndbclient.so /usr/lib
See this section for specifying the database URI
and username/password
.
git clone https://github.com/hopshadoop/hops
git checkout master
mvn package -Pdist,native,docs -DskipTests -Dtar
mvn clean generate-sources install -Pndb -DskipTests
mvn test-compile
For running tests use the ndb
profile to add the the database access layer driver to the class path. The driver is loaded at run time. For example,
mvn test -Dtest=TestFileCreation -Pndb
Set the ndb
profile in the IDE if you are running the unit test in an IDE. For example, in IntelliJ you can do this by
View --> Tool Windows --> Maven Projects
Expand the "Profiles"
Check the "ndb" profile
There are two way to configure the NDB data access layer driver
-
**Hard Coding The Database Configuration Parameters: ** While compiling the database access layer all the required configuration parameters can be written to the
./hops-metadata-dal-impl-ndb/src/main/resources/ndb-config.properties
file. When the diver is loaded it will try to connect to the database specified in the configuration file. -
**hdfs-site.xml: ** Add
dfs.storage.driver.configfile
parameter to hdfs-site.xml to read the configuration file from a sepcified path. For example, to read the configuration file in the current directory add the following the hdfs-site.xml
<property>
<name>dfs.storage.driver.configfile</name>
<value>hops-ndb-config.properties</value>
</property>
The Hops stack includes a number of services also requires a number of third-party distributed services:
- Java 1.7 (OpenJDK or Oracle JRE/JDK)
- NDB 7.4+ (MySQL Cluster)
- J2EE7 web application server (default: Glassfish)
- ElasticSearch 1.7+
Due to the complexity of installing and configuring all Hops’ services, we recommend installing Hops using the automated installer Karamel/Chef. Detailed documentation on the steps for installing and configuring all services in Hops is not discussed here. Instead, Chef cookbooks contain all the installation and configuration steps needed to install and configure Hops. The Chef cookbooks are available at: https://github.com/hopshadoop.
- Download and install Karamel (http://www.karamel.io).
- Run Karamel.
- Click on the “Load Cluster Definition” menu item in Karamel. You are now prompted to select a cluster definition YAML file. Go to the examples/stable directory, and select a cluster definition file for your target cloud platform for one of the following cluster types:
- Amazon Web Services EC2 (AWS)
- Google Compute Engine (GCE)
- OpenStack
- On-premises (bare metal)
For more information on how to configure cloud-based installations, go to help documentation at http://www.karamel.io. For on-premises installations, we provide some additional installation details and tips later in this section.
For on-premises (bare-metal) installations, you will need to prepare for installation by:
-
Identifying a master host, from which you will run Karamel;
- the master must have a display for Karamel’s user interface;
- the master must be able to ping (and connect using ssh) to all of the target hosts.
-
Identifying a set of target hosts, on which the Hops software and 3rd party services will be installed.
- the target nodes should have http access to the open Internet to be able to download software during the installation process. (Cookbooks can be configured to download software from within the private network, but this requires a good bit of configuration work for Chef attributes, changing all download URLs).
The master must be able to connect using SSH to all the target nodes, on which the software will be installed. If you have not already copied the master’s public key to the .ssh/authorized_keys file of all target hosts, you can do so by preparing the machines as follows:
- Create an openssh public/private key pair on the master host for your user account. On Linux, you can use the ssh-keygen utility program to generate the keys, which will by default be stored in the $HOME/.ssh/id_rsa and $HOME/.ssh/id_rsa.pub files. If you decided to enter a password for the ssh keypair, you will need to enter it again in Karamel when you reach the ssh dialog, part of Karamel’s Launch step. We recommend no password (passwordless) for the ssh keypair.
- Create a user account USER on the all the target machines with full sudo privileges (root privileges) and the same password on all target machines.
- Copy the $HOME/.ssh/id_rsa.pub file on the master to the /tmp folder of all the target hosts. A good way to do this is to use pscp utility along with a file ( hosts.txt ) containing the line-separated hostnames (or IP addresss) for all the target machines. You may need to install the pssh utility programs ( pssh ), first.
sudo apt-get install pssh
or
yum install pssh
vim hosts.txt
# Enter the row-separated IP addresses of all target nodes in hosts.txt
128.112.152.122
18.31.0.190
128.232.103.201
.....
pscp -h hosts.txt -P PASSWORD -i USER ~/.ssh/id_rsa.pub /tmp
pssh -h hosts.txt -i USER -P PASSWORD mkdir -p /home/USER/.ssh
pssh -h hosts.txt -i USER -P PASSWORD cat /tmp/id_rsa.pub >> /home/USER/.ssh/authorized_keys
Update your Karamel cluster definition file to include the IP addresses of the target machines and the USER account name. After you have clicked on the launch menu item, you will come to a ssh dialog. On the ssh dialog, you need to open the advanced section. Here, you will need to enter the password for the USER account on the target machines ( sudo password text input box). If your ssh keypair is password protected, you will also need to enter it again here in the keypair password text input box.
You can install Hops on your laptop/desktop with Vagrant. You will need to have the following software packages installed:
- chef-dk, version >0.5+ (but not >0.8+)
- git
- vagrant
- vagrant omnibus plugin
- virtualbox
You can now run vagrant, using:
sudo apt-get install virtualbox vagrant
vagrant plugin install vagrant-omnibus
git clone https://github.com/hopshadoop/hopsworks-chef.git
cd hopsworks-chef
berks vendor cookbooks
vagrant up
This distribution includes cryptographic software. The country in which you currently reside may have restrictions on the import, possession, use, and/or re-export to another country, of encryption software. BEFORE using any encryption software, please check your country's laws, regulations and policies concerning the import, possession, or use, and re-export of encryption software, to see if this is permitted. See http://www.wassenaar.org/ for more information.
The U.S. Government Department of Commerce, Bureau of Industry and Security (BIS), has classified this software as Export Commodity Control Number (ECCN) 5D002.C.1, which includes information security software using or performing cryptographic functions with asymmetric algorithms. The form and manner of this Apache Software Foundation distribution makes it eligible for export under the License Exception ENC Technology Software Unrestricted (TSU) exception (see the BIS Export Administration Regulations, Section 740.13) for both object code and source code.
The following provides more details on the included cryptographic software: Hadoop Core uses the SSL libraries from the Jetty project written by mortbay.org.
#Contact
- Follow our Twitter account.
- Chat with Hops developers in Gitter.
- Join our developer mailing list.
- Checkout the current build status.
Hops is released under an Apache 2.0 license.