Hortonworks DataFlow - Installing HDF Services On A New .

3y ago
30 Views
2 Downloads
540.63 KB
48 Pages
Last View : 9d ago
Last Download : 3m ago
Upload by : Mya Leung
Transcription

Hortonworks DataFlowInstalling HDF Services on a New HDP Cluster(June 6, 2018)docs.cloudera.com

Hortonworks DataFlowJune 6, 2018Hortonworks DataFlow: Installing HDF Services on a New HDP ClusterCopyright 2012-2018 Hortonworks, Inc. Some rights reserved.Except where otherwise noted, this document is licensed underCreative Commons Attribution ShareAlike 4.0 4.0/legalcodeii

Hortonworks DataFlowJune 6, 2018Table of Contents1. Installing Ambari . 11.1. Getting Ready for an Ambari Installation . 11.1.1. Reviewing System Requirements . 11.1.2. Set Up Password-less SSH . 11.1.3. Set Up Service User Accounts . 21.1.4. Enable NTP on the Cluster and on the Browser Host . 21.1.5. Check DNS and NSCD . 31.1.6. Configuring iptables . 41.1.7. Disable SELinux and PackageKit and check the umask Value . 51.2. Download the Ambari Repository . 51.2.1. RHEL/CentOS/Oracle Linux 6 . 61.2.2. RHEL/CentOS/Oracle Linux 7 . 71.2.3. SLES 12 . 81.2.4. SLES 11 . 91.2.5. Ubuntu 14 . 101.2.6. Debian 7 . 111.3. Install the Ambari Server . 121.3.1. RHEL/CentOS/Oracle Linux 6 . 121.3.2. RHEL/CentOS/Oracle Linux 7 . 131.3.3. SLES 12 . 141.3.4. SLES 11 . 151.3.5. Ubuntu 14 . 161.3.6. Debian 7 . 171.4. Set Up the Ambari Server . 171.4.1. Setup Options . 191.5. Start the Ambari Server . 202. Deploying an HDP Cluster Using Ambari . 222.1. Installing an HDP Cluster . 222.2. Customizing Druid Services . 232.3. Configure Superset . 232.4. Deploy the Cluster Services . 242.5. Access the Stream Insight Superset UI . 253. Installing the HDF Management Pack . 274. Update the HDF Base URL . 285. Add HDF Services to an HDP Cluster . 296. Configure HDF Components . 306.1. Configure NiFi . 306.2. Configure NiFi for Atlas Integration . 306.3. Configure Kafka . 326.4. Configure Storm . 326.5. Configure Log Search . 326.6. Deploy the Cluster Services . 336.7. Access the UI for Deployed Services . 337. Configuring Schema Registry and SAM for High Availability . 348. Install the Storm Ambari View . 359. Using a Local Repository . 379.1. Setting Up a Local Repository . 379.1.1. Preparing to Set Up a Local Repository . 37iii

Hortonworks DataFlowJune 6, 20189.1.2. Setting up a Local Repository with Temporary Internet Access . 389.1.3. Setting Up a Local Repository with No Internet Access . 409.2. Preparing the Ambari Repository Configuration File to Use the LocalRepository . 4210. Navigating the HDF Library . 44iv

Hortonworks DataFlowJune 6, 20181. Installing AmbariPerform the following tasks to install Ambari.1. Getting Ready for an Ambari Installation [1]2. Download the Ambari Repository [5]3. Install the Ambari Server [12]4. Set Up the Ambari Server [17]5. Start the Ambari Server [20]NoteThis document describes how to install Ambari and HDF on Intel x86 hardware.To install Ambari and HDF on IBM Power Systems, review your deploymentoptions using Planning Your Deployment for IBM Power Systems.1.1. Getting Ready for an Ambari InstallationThis section describes the information and materials you should get ready to install a clusterusing Ambari. Ambari provides an end-to-end management and monitoring solution foryour cluster. Using the Ambari Web UI and REST APIs, you can deploy, operate, manageconfiguration changes, and monitor services for all nodes in your cluster from a centralpoint.1.1.1. Reviewing System RequirementsYour first task in installing Ambari is to review the Hortonworks DataFlow (HDF)support matrices for system requirements, supported operating systems, componentinteroperability, and similar information. HDF Support Matrices1.1.2. Set Up Password-less SSHAbout This TaskTo have Ambari Server automatically install Ambari Agents on all your cluster hosts, youmust set up password-less SSH connections between the Ambari Server host and all otherhosts in the cluster. The Ambari Server host uses SSH public key authentication to remotelyaccess and install the Ambari Agent.NoteYou can choose to manually install an Ambari Agent on each cluster host. Inthis case, you do not need to generate and distribute SSH keys.Steps1. Generate public and private SSH keys on the Ambari Server host.1

Hortonworks DataFlowJune 6, 2018ssh-keygen2. Copy the SSH Public Key (id rsa.pub) to the root account on your target hosts.ssh/id rsa.ssh/id rsa.pub3. Add the SSH Public Key to the authorized keys file on your target hosts.cat id rsa.pub authorized keys4. Depending on your version of SSH, you may need to set permissions on the .ssh directory(to 700) and the authorized keys file in that directory (to 600) on the target hosts.chmod 700 /.sshchmod 600 /.ssh/authorized keys5. From the Ambari Server, make sure you can connect to each host in the cluster usingSSH, without having to enter a password.ssh root@ remote.target.host where remote.target.host has the value of each host name in your cluster.6. If the following warning message displays during your first connection: Are you sureyou want to continue connecting (yes/no)? Enter Yes.7. Retain a copy of the SSH Private Key on the machine from which you will run the webbased Ambari Install Wizard.NoteIt is possible to use a non-root SSH account, if that account can execute sudowithout entering a password.1.1.3. Set Up Service User AccountsEach service requires a service user account. The Ambari Cluster Install wizard creates newand preserves any existing service user accounts, and uses these accounts when configuringHadoop services. Service user account creation applies to service user accounts on the localoperating system and to LDAP/AD accounts.1.1.4. Enable NTP on the Cluster and on the Browser HostThe clocks of all the nodes in your cluster and the machine that runs the browser throughwhich you access the Ambari Web interface must be able to synchronize with each other.To install the NTP service and ensure it's ensure it's started on boot, run the followingcommands on each host:RHEL/CentOS/Oracle 6yum install -y ntpchkconfig ntpd onRHEL/CentOS/Oracle 7yum install -y ntpsystemctl enable ntpd2

Hortonworks DataFlowJune 6, 2018SLESzypper install ntpchkconfig ntp onUbuntuapt-get install ntpupdate-rc.d ntp defaultsDebianapt-get install ntpupdate-rc.d ntp defaults1.1.5. Check DNS and NSCDAll hosts in your system must be configured for both forward and and reverse DNS.If you are unable to configure DNS in this way, you should edit the /etc/hosts file on everyhost in your cluster to contain the IP address and Fully Qualified Domain Name of eachof your hosts. The following instructions are provided as an overview and cover a basicnetwork setup for generic Linux hosts. Different versions and flavors of Linux might requireslightly different commands and procedures. Please refer to the documentation for theoperating system(s) deployed in your environment.Hadoop relies heavily on DNS, and as such performs many DNS lookups during normaloperation. To reduce the load on your DNS infrastructure, it's highly recommended to usethe Name Service Caching Daemon (NSCD) on cluster nodes running Linux. This daemonwill cache host, user, and group lookups and provide better resolution performance, andreduced load on DNS infrastructure.1.1.5.1. Edit the Host File1. Using a text editor, open the hosts file on every host in your cluster. For example:vi /etc/hosts2. Add a line for each host in your cluster. The line should consist of the IP address and theFQDN. For example:1.2.3.4 fully.qualified.domain.name ImportantDo not remove the following two lines from your hosts file. Removing orediting the following lines may cause various programs that require networkfunctionality to fail.127.0.0.1 localhost.localdomain localhost::1 localhost6.localdomain6 localhost61.1.5.2. Set the Hostname1. Confirm that the hostname is set by running the following command:hostname -fThis should return the fully.qualified.domain.name you just set.2. Use the "hostname" command to set the hostname on each host in your cluster. Forexample:3

Hortonworks DataFlowJune 6, 2018hostname fully.qualified.domain.name 1.1.5.3. Edit the Network Configuration File1. Using a text editor, open the network configuration file on every host and set thedesired network configuration for each host. For example:vi /etc/sysconfig/network2. Modify the HOSTNAME property to set the fully qualified domain name.NETWORKING yesHOSTNAME fully.qualified.domain.name 1.1.6. Configuring iptablesFor Ambari to communicate during setup with the hosts it deploys to and manages, certainports must be open and available. The easiest way to do this is to temporarily disableiptables, as follows:RHEL/CentOS/Oracle Linux 6chkconfig iptables off/etc/init.d/iptables stopRHEL/CentOS/Oracle Linux 7systemctl disable firewalldservice firewalld stopSLESrcSuSEfirewall2 stopchkconfig SuSEfirewall2 setup offUbuntusudosudosudosudosudosudosudosudosudoufw disableiptables -Xiptables -tiptables -tiptables -tiptables -tiptables -Piptables -Piptables -Pnat -Fnat -Xmangle -Fmangle -XINPUT ACCEPTFORWARD ACCEPTOUTPUT esnat -Fnat -Xmangle -Fmangle -XINPUT ACCEPTFORWARD ACCEPTOUTPUT ACCEPTDebian-X-t-t-t-t-P-P-PYou can restart iptables after setup is complete. If the security protocols in yourenvironment prevent disabling iptables, you can proceed with iptables enabled, if allrequired ports are open and available.Ambari checks whether iptables is running during the Ambari Server setup process. Ifiptables is running, a warning displays, reminding you to check that required ports are openand available. The Host Confirm step in the Cluster Install Wizard also issues a warning foreach host that has iptables running.4

Hortonworks DataFlowJune 6, 20181.1.7. Disable SELinux and PackageKit and check the umaskValue1. You must disable SELinux for the Ambari setup to function. On each host in your cluster,enter:setenforce 0NoteTo permanently disable SELinux set SELINUX disabled in /etc/selinux/config This ensures that SELinux does not turn itself on after you rebootthe machine .2. On an installation host running RHEL/CentOS with PackageKit installed, open /etc/yum/pluginconf.d/refresh-packagekit.conf using a text editor. Make thefollowing change:enabled 0NotePackageKit is not enabled by default on Debian, SLES, or Ubuntu systems.Unless you have specifically enabled PackageKit, you may skip this step for aDebian, SLES, or Ubuntu installation host.3. UMASK (User Mask or User file creation MASK) sets the default permissions or basepermissions granted when a new file or folder is created on a Linux machine. Most Linuxdistros set 022 as the default umask value. A umask value of 022 grants read, write,execute permissions of 755 for new files or folders. A umask value of 027 grants read,write, execute permissions of 750 for new files or folders.Ambari, HDP, and HDF support umask values of 022 (0022 is functionally equivalent),027 (0027 is functionally equivalent). These values must be set on all hosts.UMASK Examples:Setting the umask for your current login session:umask 0022Checking your current umask:umask 0022Permanently changing the umask for all interactive users:echo umask 0022 /etc/profile1.2. Download the Ambari RepositoryFollow the instructions in the section for the operating system that runs your installationhost.5

Hortonworks DataFlowJune 6, 2018 RHEL/CentOS/Oracle Linux 6 [6] RHEL/CentOS/Oracle Linux 7 [7] SLES 12 [8] SLES 11 [9] Ubuntu 14 [10] Debian 7 [11]Use a command line editor to perform each instruction.1.2.1. RHEL/CentOS/Oracle Linux 6On a server host that has Internet access, use a command line editor to perform thefollowing:Steps1. Log in to your host as root.2. Download the Ambari repository file to a directory on your installation host.wget -nv 6/2.x/updates/2.6.2.2/ambari.repo -O /etc/yum.repos.d/ambari.repoImportantDo not modify the ambari.repo file name. This file is expected to beavailable on the Ambari Server host during Agent registration.3. Confirm that the repository is configured by checking the repo list.yum repolistYou should see values similar to the following for Ambari repositories in the list.repo idambari-2.6.2.2-1baseextrasupdatesrepolist: 7,746repo nameambari Version - ambari-2.6.2.2-1CentOS-6 - BaseCentOS-6 - ExtrasCentOS-6 - Updatesstatus126,69664974Version values vary, depending on the installation.NoteWhen deploying a cluster having limited or no Internet access, you shouldprovide access to the bits using an alternative method.Ambari Server by default uses an embedded PostgreSQL database. Whenyou install the Ambari Server, the PostgreSQL packages and dependencies6

Hortonworks DataFlowJune 6, 2018must be available for install. These packages are typically available as part ofyour Operating System repositories. Please confirm you have the appropriaterepositories available for the postgresql-server packages.Next Step Install the Ambari Server [12] Set Up the Ambari Server [17]1.2.2. RHEL/CentOS/Oracle Linux 7On a server host that has Internet access, use a command line editor to perform thefollowingSteps1. Log in to your host as root.2. Download the Ambari repository file to a directory on your installation host.wget -nv 7/2.x/updates/2.6.2.2/ambari.repo -O /etc/yum.repos.d/ambari.repoImportantDo not modify the ambari.repo file name. This file is expected to beavailable on the Ambari Server host during Agent registration.3. Confirm that the repository is configured by checking the repo list.yum repolistYou should see values similar to the following for Ambari repositories in the list.repo idstatusambari-2.6.2.2-1epel/x86 6411,387ol7 UEKR4/x86 64repo nameambari Version - ambari-2.6.2.2-112Extra Packages for Enterprise Linux 7 - x86 64Latest Unbreakable Enterprise Kernel Release 4for Oracle Linux 7Server (x86 64)295Oracle Linux 7Server Latest (x86 64)ol7 latest/x86 6418,642puppetlabs-deps/x86 64Puppet Labs Dependencies El 7 - x86 6417puppetlabs-products/x86 64 Puppet Labs Products El 7 - x86 64225repolist: 30,578Version values vary, depending on the installation.NoteWhen deploying a cluster having limited or no Internet access, you shouldprovide access to the bits using an alternative method.7

Hortonworks DataFlowJune 6, 2018Ambari Server by default uses an embedded PostgreSQL database. Whenyou install the Ambari Server, the PostgreSQL packages and dependenciesmust be available for install. These packages are typically available as part ofyour Operating System repositories. Please confirm you have the appropriaterepositories available for the postgresql-server packages.Next Step Install the Ambari Server [12] Set Up the Ambari Server [17]1.2.3. SLES 12On a server host that has Internet access, use a command line editor to perform thefollowing:Steps1. Log in to your host as root.2. Download the Ambari repository file to a directory on your installation host.wget -nv /2.x/updates/2.6.2.2/ambari.repo -O /etc/zypp/repos.d/ambari.repoImportantDo not modify the ambari.repo file name. This file is expected to beavailable on the Ambari Server host during Agent registration.3. Confirm the downloaded repository is configured by checking the repo list.zypper reposYou should see the Ambari repositories in the list.# AliasEnabled Refresh-- ------------------------ --------- -------1 ambari-2.6.2.2-1No2 http-demeter.uni-regensburg.de-c997c8f9 Yes3 opensuse Yes Name ------------------------------------- ambari Version - ambari-2.6.2.2-1 Yes SUSE-Linux-Enterprise-Software -Development-Kit-12-SP1 12.1.1-1.57 Yes OpenSuse YesVersion values vary, depending on the installation.NoteWhen deploying a cluster having limited or no Internet access, you shouldprovide access to the bits using an alternative method.8

Hortonworks DataFlowJune 6, 2018Ambari Server by default uses an embedded PostgreSQL database. Whenyou install the Ambari Server, the PostgreSQL packages and dependen

Hortonworks DataFlow June 6, 2018 3 SLES zypper install ntp chkconfig ntp on Ubuntu apt-get install ntp update-rc.d ntp defaults Debian apt-get install ntp update-rc.d ntp defaults 1.1.5. Check DNS and NSCD All hosts in your system must be configured for both forward and and reverse DNS.

Related Documents:

HDF Makes Big Data Ingest Easy Complicated, messy, and takes weeks to months to move the right data into Hadoop HDP HORTONWORKS DATA PLATFORM Streamlined, Efficient, Easy HDP HORTONWORKS DATA PLATFORM Powered by Apache Hadoop

Hortonworks DataFlow August 9

various Big Data tools like Apache Hadoop, Apache Spark, Apache Flume, Apache Impala, Apache Kudu and Apache HBase needed by data scientists. In 2011, Hortonworks was founded by a group of engineers from Yahoo! Hortonworks released HDP (Hortonworks Data Platform), a competitor to CDH. In 2019, Cloudera and Hortonworks merged, and the two .

Hortonworks: A Leader In Hadoop The Forrester Wave : Big Data Hadoop Solutions, Q1 2014 "Hortonworks loves and lives open source innovation" Vision & Execution for Enterprise Hadoop. Hortonworks leads with a strong strategy and roadmap for open source innovation with Hadoop and a strong delivery of that innovation in Hortonworks Data .

standard HDF datatypes and is supported by a special application programming interface (API) which aids the data product user or producer in the application of the conventions. The APIs allow data products to be created and manipulated in ways appropriate to each datatype, without regard to the actual HDF objects and conventions underlying them.

The Hortonworks Data Platform, powered by Apache Hadoop, is a massively scalable and 100% open source platform for storing, processing and analyzing large volumes of data. It is designed to deal with . 6. Installing, Configuring, and Deploying a Cluster . 51. Hortonworks Data Platform November 15, 2018 iv 6.1. Start the Ambari Server .

Dataflow process networks are shown to be a special case of Kahn process networks, a model of computation where a number of concurrent processes communicate through unidirectional FIFO channels, where writes to the channel are nonblocking, and reads are blocking. In dataflow process networks, each process consists

Copyright National Literacy Trust (Alex Rider Secret Mission teaching ideas) Trademarks Alex Rider ; Boy with Torch Logo 2010 Stormbreaker Productions Ltd .