15 Minute Guide To Install Hadoop Cluster - Edureka

1y ago
4 Views
1 Downloads
1.02 MB
19 Pages
Last View : Today
Last Download : 3m ago
Upload by : Cannon Runnels
Transcription

15-MinuteGuide to InstallApache Hadoop Cluster 2.0With single Data Node Configuration on Ubuntu

15-Minute Guide to install Apache Hadoop Cluster 2.0Let’s know what is Apache Hadoop 2.0 all about?Apache Hadoop 2.0 is now generally available! Thanks to the longawaited announcement done by Apache Software Foundation (ASF).The elephant has gone bigger with great features, all set to manage BigData even better than before!What’s new in Hadoop 2.0? YARN Framework (MapReduce 2.0)HDFS High Availability (NameNode HA)HDFS FederationData SnapshotSupport for WindowsNFS AccessBinary CompatibilityExtensive Testing2.0All these remarkable attributes and many more will increase Hadoopadoption tremendously in the industry to solve Big Data problems.Hadoop is now very much enterprise-ready with crucial security abilities!According to Chris Douglas, Vice President of Apache Hadoop,“With the release of stable Hadoop 2, the community celebrates not only an iterationof the software, but an inflection point in the project's development. We believe thisplatform is capable of supporting new applications and research in large-scale,commodity computing.The Apache Software Foundation creates the conditions for innovative,community-driven technology like Hadoop to evolve. When that processconverges, the result is inspiring."Share this ebook!02

15-Minute Guide to install Apache Hadoop Cluster 2.0An Intro to this Guide This setup and configuration document is a guide to setup a Single-NodeApache Hadoop 2.0 cluster on an Ubuntu Virtual Machine (VM) on yourPC. If you are new to both Ubuntu and Hadoop, this guide comes handyto quickly setup a Single-Node Apache Hadoop 2.0 Cluster on Ubuntuand start your Big Data and Hadoop learning journey!The Guide describes the whole process in two steps:Step 1: Setting up the Ubuntu OS for Hadoop 2.0This section describes step by step guide to download, configure anUbuntu Virtual Machine image in VMPlayer, and provides steps to installpre-requisites for Hadoop Installation on Ubuntu.Step 2: Installing Apache Hadoop 2.0 and Setting up the SingleNode ClusterThis section explains primary Hadoop 2.0 configuration files, SingleNode cluster configuration and Hadoop daemons start and stop processin detail.Note:The configuration described here is intended for learning purposes only.Share this ebook!03

15-Minute Guide to install Apache Hadoop Cluster 2.0Follow these simple steps Step 1:Setting up the Ubuntu ServerThis section describes the steps to downloadand create an Ubuntu image on VMPlayer.1.1 Creating an Ubuntu VMPlayer instanceThe first step is to download an Ubuntu image and create an UbuntuVMPlayer instance.1.1.1 Download the VMware imageAccess the following link and download the 12.0.4 Ubuntu t.html1.1.2 Open the image fileExtract the Ubuntu VM image and Open it in VMware Player.Click open virtual machine and select path where you have extractedthe image. Select the ‘.vmx’ file and click ‘ok’.Share this ebook!04

15-Minute Guide to install Apache Hadoop Cluster 2.0The VM:Share this ebook!05

15-Minute Guide to install Apache Hadoop Cluster 2.01.1.3 Play the Virtual MachineYou would see the below screen in VMware Player after the VMimage creation completes. Double click on the link to run the machine.You will get the home screen of Ubuntu.The user details for the Virtual instance are:Username : userPassword : passwordOpen the Terminal to access the File System:1.1.4 Update the OS packages and their dependenciesThe first task is to run ‘apt-get update’ to download the package lists fromthe repositories and "update" them to get information on the newestversions of packages and their dependencies. sudo apt-get update1.1.5 Install Java for Hadoop 2.2.0Use apt-get to install JDK 6 on the server. sudo apt-get install openjdk-6-jdkCheck Java Version: java -versionShare this ebook!06

15-Minute Guide to install Apache Hadoop Cluster 2.01.2 Download the Apache Hadoop 2.0 Binaries:1.2.1 Download the Hadoop Package:Download the binaries to your home directory. Use the default user ‘user’for the installation.In Live production instances a dedicated Hadoop user account forrunning Hadoop is used. Though, it’s not mandatory to use a dedicatedHadoop user account but is recommended because this helps toseparate the Hadoop installation from other software applications anduser accounts running on the same machine (separating for security,permissions, backups, etc.). /common/stable2/hadoop2.2.0.tar.gzUnzip the files and review the package content and configurationfiles. tar -xvf hadoop-2.2.0.tar.gzHadoop Package Content:Share this ebook!07

15-Minute Guide to install Apache Hadoop Cluster 2.0Share this ebook!08

15-Minute Guide to install Apache Hadoop Cluster 2.0Review the Hadoop Configuration Files:After creating and configuring your virtual servers, the Ubuntu instanceis now ready to start installation and configuration of Apache Hadoop2.0 Single Node Cluster. This section describes the steps in detail toinstall Apache Hadoop 2.0 and configure a Single-Node ApacheHadoop cluster.Some High-end Hadoop Users Share this ebook!09

15-Minute Guide to install Apache Hadoop Cluster 2.0Step 2:Configure the Apache Hadoop 2.0 Single Node ServerThis section explains the steps to configure Single Node Apache Hadoop2.0 Server on Ubuntu.2.1 Update the Configuration Files2.1.1 Update “.bashrc” file for user ‘ubuntu’.Move to ‘user’ HOME directory and edit ‘.bashrc’ file.FILE ‘.BACHRC’ LOCATIONUpdate the ‘.bashrc’ file to add important Apache Hadoopenvironment variables for user.Change directory to home - cdEdit the file - vi .bashrcShare this ebook!10

15-Minute Guide to install Apache Hadoop Cluster 2.0Set Hadoop Environment Variables - Begin# Set Hadoop-related environment variablesexport HADOOP HOME HOME/hadoop-2.2.0export HADOOP CONF DIR HOME/hadoop-2.2.0/etc/hadoopexport HADOOP MAPRED HOME HOME/hadoop-2.2.0export HADOOP COMMON HOME HOME/hadoop-2.2.0export HADOOP HDFS HOME HOME/hadoop-2.2.0export YARN HOME HOME/hadoop-2.2.0# Set JAVA HOME (we will also configure JAVA HOME for Hadoopexecution later on)export JAVA HOME /usr/lib/jvm/java-6-openjdk-amd6# Add Hadoop bin/ directory to PATHexport PATH PATH: HOME/hadoop-2.2.0/binSet Hadoop Environment Variables - EndEDIT .BASHRCShare this ebook!11

15-Minute Guide to install Apache Hadoop Cluster 2.0c)Source the .bashrc file to set the Hadoop environment variableswithout having to invoke a new shell: . /.bashrcExecute all the steps of this section on all the remaining cluster servers.2.2 Setup the Hadoop Cluster:This section describes the detail steps needed for setting up the HadoopCluster and configuring the core Hadoop configuration files.2.2.1 Configure JAVA HOMEConfigure JAVA HOME in ‘hadoop-env.sh’. This file specifiesenvironment variables that affect the JDK used by Apache Hadoop 2.0daemons started by the Hadoop start-up scripts. cd HADOOP CONF DIR vi hadoop-env.shUpdate the JAVA HOME to:export JAVA HOME /usr/lib/jvm/java-6-openjdk-amd64Java Home Set-upShare this ebook!12

15-Minute Guide to install Apache Hadoop Cluster 2.02.2.2 Create NameNode and DataNode directoryCreate DataNode and NameNode directories to store HDFS data. mkdir -p HOME/hadoop2 data/hdfs/namenode mkdir -p HOME/hadoop2 data/hdfs/datanode2.2.3 Configure the Default File systemThe ’core-site.xml’ file contains the configuration settings for ApacheHadoop Core such as I/O settings that are common to HDFS, YARN andMapReduce. Configure default files-system (Parameter: fs.default.name)used by clients in core-site.xml cd HADOOP CONF DIR vi core-site.xmlAdd the following lines in between the configuration tag:Configuring the Default File System:Here the Hostname and Port are the machine and port on whichName Node daemon runs and listens. It also informs the NameNode as to which IP and port it should bind. The commonly usedport is 9000 and you can also specify IP address rather thanhostname.Share this ebook!13

15-Minute Guide to install Apache Hadoop Cluster 2.0Note:For the simplicity of understanding the cluster setup, we have updatedonly the necessary parameters to start a cluster. You can researchmore on Apache Hadoop 2.0 page and experiment the configuration fordifferent features.2.2.4 Configure the HDFS:This file contains the cconfiguration settings for HDFS daemons; theName Node and the data nodes.Configure hdfs-site.xml and specify default block replication, andNameNode and DataNode directories for HDFS. The actual number ofreplications can be specified when the file is created. The default is usedif replication is not specified in create time. cd HADOOP CONF DIR vi hdfs-site.xmlAdd the following lines in between the configuration tag:Configuring the Default File System:Share this ebook!14

15-Minute Guide to install Apache Hadoop Cluster 2.02.2.5 Configure YARN framework:This file contains the configuration settings for YARN; the NodeManager. cd HADOOP CONF DIR vi yarn-site.xmlAdd the following lines in between the configuration tag:Configuring the Default FilesystemDo You Know? HDFS is a File system, not a database management system(DBMS), as commonly perceived! Hadoop is an ecosystem consisting of multiple products, not asingle product! Hadoop enables several kinds of analytics, apart from Webanalytics!Share this ebook!15

15-Minute Guide to install Apache Hadoop Cluster 2.02.2.6 Configure MapReduce frameworkThis file contains the configuration settings for MapReduce.So, Configure mapred-site.xml and specify framework details. cd HADOOP CONF DIRYou need to copy the mapred-site.xml template. cp mapred-site.xml template mapred-site.xml vi mapred-site.xmlAdd the following line in between the configuration tag:Configuring the JobTracker Details:2.2.7 Start the DFS services:The first step in starting up your Hadoop installation is formatting theHadoop file-system, which is implemented on top of the local file-systemsof your cluster. This is required on the first time Hadoop installation.Do not format a running Hadoop file-system, this will cause all your datato be erased.To format the file-system, run the command: hadoop namenode –formatYou are now all set to start the HDFS services i.e. Name Node,Resource Manager, Node Manager and Data Nodes on yourApache Hadoop Cluster!Share this ebook!16

15-Minute Guide to install Apache Hadoop Cluster 2.0Starting the Services:Start the YARN Daemons i.e. Resource Manager and Node Manager.Cross check the service start-up using JPS (Java Process MonitoringTool).Starting the YARN Daemons:Share this ebook!17

15-Minute Guide to install Apache Hadoop Cluster 2.0Starting the History Server:2.2.8 Finally, perform the Health Check!a) Check the NameNode status:http://localhost:50070/dfshealth.jspb) JobHistory status:http://localhost:19888/jobhistory.jspAnd You are DONE!Share this ebook!18

Click Here to learn more aboutBig Data & Hadoop Contact us at: learn@edureka.in

Ubuntu. and . Hadoop, this guide comes handy to quickly setup a Single- Node Apache Hadoop 2.0 Cluster on Ubuntu and start your Big Data and Hadoop learning journey! The Guide describes the whole process in two steps: Step 1: Setting up the Ubuntu OS for Hadoop 2.0. This section describes step by step guide to download, configure an

Related Documents:

Newari Traditional Massage 60 minute 5749 90 minute 7499 Hot Stone Massage 60 minute 5999 90 minute 7999 Pregnancy Massage (Pre-Post Natal) 60 minute 6999 90 minute 8999 Traditional Herbal Nepali Massage 60 minute 5749 90 minute 7499 Indian Head M

Radio-paging services (starting 076) 45.84p call Personal numbering services (starting 070) 0p 1 minute minute Call forwarding services (e.g. 07744, 07755) 45.84p 1 minute minute ACCESS CHARGE FOR: Directory enquiries numbers (118 numbers) # 54.16p 1 minute minute Premium Rate call charges Cost per minute Cost Per Call Minimum Call Charge .

30 Ten-Minute Plays from the Actors Theatre of Louisville for 4, 5, and 6 Actors 2004: The Best 10-Minute Plays for Two Actors 2004: The Best 10-Minute Plays for Three or More Actors 2005: The Best 10-Minute Plays for Two Actors 2005: The Best 10-Minute Plays for Three or More Actors 2006: The Best 10-Minute Plays for Two Actors 2006: The Best .

Workout #4 - 4, 3, 2, 1 Cardio Core Blast Equipment needed: Mat or towel, dumbbells (optional), timer. Mission: Complete 2 rounds. First 5 Minutes 1 minute Jump Jacks 30 second Squat Jumps 1 minute Mountain Climbers 1 minute Jump Jacks 30 seconds Squat Jumps 1 minute Mountain Climbers Next 4 Minutes 1 minute Walking Lunges 1 minute Sumo Squats

Y to schedule your massage appointment with our Massage Therapist. Member: Nonmember: 30 - 30 minute massage 35 - 30 minute massage 50 - 60 minute massage 60 - 60 minute massage 85 - 90 minute massage 95 - 90 minute massage PERSONAL WELLNESS TRAINING (YOUTH & ADULT) One-on-on

The New One Minute Manager Explains 49 Why One Minute Goals Work 53 Why One Minute Praisings Work 64 Why One Minute Re-Directs Work 71 Another New One Minute Manager 86 The New One Minute Manager’s Game Plan 88 A Gift to Yourself 89 A Gift to Others 92 Acknowledgments 95 About the Authors

AQUILA COMMERCIAL FOR SALE VISTA RIDGE AT STEINER RANCH ACCESSIBILITY & NEARBY AMENITIES Q A N P A R K R D 2222 & Mopac 17-minute Drive 2222 & 360 10-minute Drive 620 & 183 8-minute Drive CBD 26-minute Drive Austin-Bergstrom International Airport 42-minute Drive 620 & 2222 2-minute Drive STEINER RANCH JOLLYVILLE AVERY RANCH ANDERSON MILL NORTHWEST

work/products (Beading, Candles, Carving, Food Products, Soap, Weaving, etc.) ⃝I understand that if my work contains Indigenous visual representation that it is a reflection of the Indigenous culture of my native region. ⃝To the best of my knowledge, my work/products fall within Craft Council standards and expectations with respect to