Hadoop

Hadoop 集群安装(hadoop-3.1.2)

环境准备

  • 3台centos7虚拟机(hadoop01、hadoop02、hadoop03)

  • 在三台机器上分别安装好jdk-1.8、ssh等

  • 还需要hadoop安装包 hadoop-3.1.2.tar.gz

开始安装

三台机器的职责

IP SSH port 职责 名称
192.168.127.128 22 NameNode、SecondaryName、NodeResourceManager hadoop01
192.168.127.129 22 DataNode、NodeManager hadoop02
192.168.127.130 22 DataNode、NodeManager hadoop03
  1. 首先修改三台机器的名称
    1. hostnamectl set-hostname hadoop01;
    2. hostnamectl set-hostname hadoop02;
    3. hostnamectl set-hostname hadoop03;
  1. 配置hosts 文件
    三台机器上都要在hosts添加一下内容:
    1. 192.168.127.128 hadoop01
    2. 192.168.127.129 hadoop02
    3. 192.168.127.130 hadoop03
  1. 配置免密登录
    1. ssh-keygen -t rsa -P ""
    2. cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys


因为hadoop01为master节点,所以hadoop01机器可以免密登录到hadoop02、hadoop03
将hadoop01生成的公钥分别存到hadoop02、hadoop03 的~/.ssh/authorized_keys文件中

  1. 安装hadoop

    1. 将hadoop-3.1.2.tar.gz文件解压
      1. tar -zxvf hadoop-3.1.2.tar.gz
  1. 配置环境变量
    1. JAVA_HOME=/usr/java/jdk1.8.0_192-amd64
    2. CLASSPATH=%JAVA_HOME%/lib:%JAVA_HOME%/jre/lib
    3. HADOOP_HOME=/opt/hadoop-3.1.2
    4. PATH=$PATH:$JAVA_HOME/bin:$JAVA_HOME/jre/bin:$HADOOP_HOME/bin
    5. export PATH CLASSPATH JAVA_HOME HADOOP_HOME
  1. 修改hadoop中文件

    • 配置hadoop-env.sh ```shell #

      Licensed to the Apache Software Foundation (ASF) under one

      or more contributor license agreements. See the NOTICE file

      distributed with this work for additional information

      regarding copyright ownership. The ASF licenses this file

      to you under the Apache License, Version 2.0 (the

      “License”); you may not use this file except in compliance

      with the License. You may obtain a copy of the License at

      #

      http://www.apache.org/licenses/LICENSE-2.0

      #

      Unless required by applicable law or agreed to in writing, software

      distributed under the License is distributed on an “AS IS” BASIS,

      WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.

      See the License for the specific language governing permissions and

      limitations under the License.

Set Hadoop-specific environment variables here.

#

THIS FILE ACTS AS THE MASTER FILE FOR ALL HADOOP PROJECTS.

SETTINGS HERE WILL BE READ BY ALL HADOOP COMMANDS. THEREFORE,

ONE CAN USE THIS FILE TO SET YARN, HDFS, AND MAPREDUCE

CONFIGURATION OPTIONS INSTEAD OF xxx-env.sh.

#

Precedence rules:

#

{yarn-env.sh|hdfs-env.sh} > hadoop-env.sh > hard-coded defaults

#

{YARN_xyz|HDFS_xyz} > HADOOP_xyz > hard-coded defaults

#

Many of the options here are built from the perspective that users

may want to provide OVERWRITING values on the command line.

For example:

#

JAVA_HOME=/usr/java/testing hdfs dfs -ls

#

Therefore, the vast majority (BUT NOT ALL!) of these defaults

are configured for substitution and not append. If append

is preferable, modify this file accordingly.

#

Generic settings for HADOOP

#

Technically, the only required environment variable is JAVA_HOME.

All others are optional. However, the defaults are probably not

preferred. Many sites configure these options outside of Hadoop,

such as in /etc/profile.d

The java implementation to use. By default, this environment

variable is REQUIRED on ALL platforms except OS X!

export JAVA_HOME=/usr/java/jdk1.8.0_192-amd64

Location of Hadoop. By default, Hadoop will attempt to determine

this location based upon its execution path.

export HADOOP_HOME=/opt/hadoop-3.1.2

Location of Hadoop’s configuration information. i.e., where this

file is living. If this is not defined, Hadoop will attempt to

locate it based upon its execution path.

#

NOTE: It is recommend that this variable not be set here but in

/etc/profile.d or equivalent. Some options (such as

—config) may react strangely otherwise.

# export HADOOP_CONF_DIR=${HADOOP_HOME}/etc/hadoop

The maximum amount of heap to use (Java -Xmx). If no unit

is provided, it will be converted to MB. Daemons will

prefer any Xmx setting in their respective _OPT variable.

There is no default; the JVM will autoscale based upon machine

memory size.

export HADOOP_HEAPSIZE_MAX=

The minimum amount of heap to use (Java -Xms). If no unit

is provided, it will be converted to MB. Daemons will

prefer any Xms setting in their respective _OPT variable.

There is no default; the JVM will autoscale based upon machine

memory size.

export HADOOP_HEAPSIZE_MIN=

Enable extra debugging of Hadoop’s JAAS binding, used to set up

Kerberos security.

export HADOOP_JAAS_DEBUG=true

Extra Java runtime options for all Hadoop commands. We don’t support

IPv6 yet/still, so by default the preference is set to IPv4.

export HADOOP_OPTS=”-Djava.net.preferIPv4Stack=true”

For Kerberos debugging, an extended option set logs more invormation

export HADOOP_OPTS=”-Djava.net.preferIPv4Stack=true -Dsun.security.krb5.debug=true -Dsun.security.spnego.debug”

Some parts of the shell code may do special things dependent upon

the operating system. We have to set this here. See the next

section as to why….

export HADOOP_OS_TYPE=${HADOOP_OS_TYPE:-$(uname -s)}

Extra Java runtime options for some Hadoop commands

and clients (i.e., hdfs dfs -blah). These get appended to HADOOP_OPTS for

such commands. In most cases, # this should be left empty and

let users supply it on the command line.

export HADOOP_CLIENT_OPTS=””

#

A note about classpaths.

#

By default, Apache Hadoop overrides Java’s CLASSPATH

environment variable. It is configured such

that it sarts out blank with new entries added after passing

a series of checks (file/dir exists, not already listed aka

de-deduplication). During de-depulication, wildcards and/or

directories are NOT expanded to keep it simple. Therefore,

if the computed classpath has two specific mentions of

awesome-methods-1.0.jar, only the first one added will be seen.

If two directories are in the classpath that both contain

awesome-methods-1.0.jar, then Java will pick up both versions.

An additional, custom CLASSPATH. Site-wide configs should be

handled via the shellprofile functionality, utilizing the

hadoop_add_classpath function for greater control and much

harder for apps/end-users to accidentally override.

Similarly, end users should utilize ${HOME}/.hadooprc .

This variable should ideally only be used as a short-cut,

interactive way for temporary additions on the command line.

export HADOOP_CLASSPATH=”/some/cool/path/on/your/machine”

Should HADOOP_CLASSPATH be first in the official CLASSPATH?

export HADOOP_USER_CLASSPATH_FIRST=”yes”

If HADOOP_USE_CLIENT_CLASSLOADER is set, the classpath along

with the main jar are handled by a separate isolated

client classloader when ‘hadoop jar’, ‘yarn jar’, or ‘mapred job’

is utilized. If it is set, HADOOP_CLASSPATH and

HADOOP_USER_CLASSPATH_FIRST are ignored.

export HADOOP_USE_CLIENT_CLASSLOADER=true

HADOOP_CLIENT_CLASSLOADER_SYSTEM_CLASSES overrides the default definition of

system classes for the client classloader when HADOOP_USE_CLIENT_CLASSLOADER

is enabled. Names ending in ‘.’ (period) are treated as package names, and

names starting with a ‘-‘ are treated as negative matches. For example,

export HADOOP_CLIENT_CLASSLOADER_SYSTEM_CLASSES=”-org.apache.hadoop.UserClass,java.,javax.,org.apache.hadoop.”

Enable optional, bundled Hadoop features

This is a comma delimited list. It may NOT be overridden via .hadooprc

Entries may be added/removed as needed.

export HADOOP_OPTIONAL_TOOLS=”hadoop-aliyun,hadoop-aws,hadoop-azure-datalake,hadoop-azure,hadoop-kafka,hadoop-openstack”

#

Options for remote shell connectivity

#

There are some optional components of hadoop that allow for

command and control of remote hosts. For example,

start-dfs.sh will attempt to bring up all NNs, DNS, etc.

Options to pass to SSH when one of the “log into a host and

start/stop daemons” scripts is executed

export HADOOP_SSH_OPTS=”-o BatchMode=yes -o StrictHostKeyChecking=no -o ConnectTimeout=10s”

The built-in ssh handler will limit itself to 10 simultaneous connections.

For pdsh users, this sets the fanout size ( -f )

Change this to increase/decrease as necessary.

export HADOOP_SSH_PARALLEL=10

Filename which contains all of the hosts for any remote execution

helper scripts # such as workers.sh, start-dfs.sh, etc.

export HADOOP_WORKERS=”${HADOOP_CONF_DIR}/workers”

#

Options for all daemons

#

#

#

Many options may also be specified as Java properties. It is

very common, and in many cases, desirable, to hard-set these

in daemon _OPTS variables. Where applicable, the appropriate

Java property is also identified. Note that many are re-used

or set differently in certain contexts (e.g., secure vs

non-secure)

#

Where (primarily) daemon log files are stored.

${HADOOP_HOME}/logs by default.

Java property: hadoop.log.dir

export HADOOP_LOG_DIR=${HADOOP_HOME}/logs

A string representing this instance of hadoop. $USER by default.

This is used in writing log and pid files, so keep that in mind!

Java property: hadoop.id.str

export HADOOP_IDENT_STRING=$USER

How many seconds to pause after stopping a daemon

export HADOOP_STOP_TIMEOUT=5

Where pid files are stored. /tmp by default.

export HADOOP_PID_DIR=/tmp

Default log4j setting for interactive commands

Java property: hadoop.root.logger

export HADOOP_ROOT_LOGGER=INFO,console

Default log4j setting for daemons spawned explicitly by

—daemon option of hadoop, hdfs, mapred and yarn command.

Java property: hadoop.root.logger

export HADOOP_DAEMON_ROOT_LOGGER=INFO,RFA

Default log level and output location for security-related messages.

You will almost certainly want to change this on a per-daemon basis via

the Java property (i.e., -Dhadoop.security.logger=foo). (Note that the

defaults for the NN and 2NN override this by default.)

Java property: hadoop.security.logger

export HADOOP_SECURITY_LOGGER=INFO,NullAppender

Default process priority level

Note that sub-processes will also run at this level!

export HADOOP_NICENESS=0

Default name for the service level authorization file

Java property: hadoop.policy.file

export HADOOP_POLICYFILE=”hadoop-policy.xml”

#

NOTE: this is not used by default! <——-

You can define variables right here and then re-use them later on.

For example, it is common to use the same garbage collection settings

for all the daemons. So one could define:

#

export HADOOP_GC_SETTINGS=”-verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps”

#

.. and then use it as per the b option under the namenode.

#

Secure/privileged execution

#

#

Out of the box, Hadoop uses jsvc from Apache Commons to launch daemons

on privileged ports. This functionality can be replaced by providing

custom functions. See hadoop-functions.sh for more information.

#

The jsvc implementation to use. Jsvc is required to run secure datanodes

that bind to privileged ports to provide authentication of data transfer

protocol. Jsvc is not required if SASL is configured for authentication of

data transfer protocol using non-privileged ports.

export JSVC_HOME=/usr/bin

#

This directory contains pids for secure and privileged processes.

export HADOOP_SECURE_PID_DIR=${HADOOP_PID_DIR}

#

This directory contains the logs for secure and privileged processes.

Java property: hadoop.log.dir

export HADOOP_SECURE_LOG=${HADOOP_LOG_DIR}

#

When running a secure daemon, the default value of HADOOP_IDENT_STRING

ends up being a bit bogus. Therefore, by default, the code will

replace HADOOP_IDENT_STRING with HADOOP_xx_SECURE_USER. If one wants

to keep HADOOP_IDENT_STRING untouched, then uncomment this line.

export HADOOP_SECURE_IDENT_PRESERVE=”true”

#

NameNode specific parameters

#

Default log level and output location for file system related change

messages. For non-namenode daemons, the Java property must be set in

the appropriate _OPTS if one wants something other than INFO,NullAppender

Java property: hdfs.audit.logger

export HDFS_AUDIT_LOGGER=INFO,NullAppender

Specify the JVM options to be used when starting the NameNode.

These options will be appended to the options specified as HADOOP_OPTS

and therefore may override any similar flags set in HADOOP_OPTS

#

a) Set JMX options

export HDFS_NAMENODE_OPTS=”-Dcom.sun.management.jmxremote=true -Dcom.sun.management.jmxremote.authenticate=false -Dcom.sun.management.jmxremote.ssl=false -Dcom.sun.management.jmxremote.port=1026”

#

b) Set garbage collection logs

export HDFS_NAMENODE_OPTS=”${HADOOP_GC_SETTINGS} -Xloggc:${HADOOP_LOG_DIR}/gc-rm.log-$(date +’%Y%m%d%H%M’)”

#

c) … or set them directly

export HDFS_NAMENODE_OPTS=”-verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -Xloggc:${HADOOP_LOG_DIR}/gc-rm.log-$(date +’%Y%m%d%H%M’)”

this is the default:

export HDFS_NAMENODE_OPTS=”-Dhadoop.security.logger=INFO,RFAS”

#

SecondaryNameNode specific parameters

#

Specify the JVM options to be used when starting the SecondaryNameNode.

These options will be appended to the options specified as HADOOP_OPTS

and therefore may override any similar flags set in HADOOP_OPTS

#

This is the default:

export HDFS_SECONDARYNAMENODE_OPTS=”-Dhadoop.security.logger=INFO,RFAS”

#

DataNode specific parameters

#

Specify the JVM options to be used when starting the DataNode.

These options will be appended to the options specified as HADOOP_OPTS

and therefore may override any similar flags set in HADOOP_OPTS

#

This is the default:

export HDFS_DATANODE_OPTS=”-Dhadoop.security.logger=ERROR,RFAS”

On secure datanodes, user to run the datanode as after dropping privileges.

This MUST be uncommented to enable secure HDFS if using privileged ports

to provide authentication of data transfer protocol. This MUST NOT be

defined if SASL is configured for authentication of data transfer protocol

using non-privileged ports.

This will replace the hadoop.id.str Java property in secure mode.

export HDFS_DATANODE_SECURE_USER=hdfs

Supplemental options for secure datanodes

By default, Hadoop uses jsvc which needs to know to launch a

server jvm.

export HDFS_DATANODE_SECURE_EXTRA_OPTS=”-jvm server”

#

NFS3 Gateway specific parameters

#

Specify the JVM options to be used when starting the NFS3 Gateway.

These options will be appended to the options specified as HADOOP_OPTS

and therefore may override any similar flags set in HADOOP_OPTS

#

export HDFS_NFS3_OPTS=””

Specify the JVM options to be used when starting the Hadoop portmapper.

These options will be appended to the options specified as HADOOP_OPTS

and therefore may override any similar flags set in HADOOP_OPTS

#

export HDFS_PORTMAP_OPTS=”-Xmx512m”

Supplemental options for priviliged gateways

By default, Hadoop uses jsvc which needs to know to launch a

server jvm.

export HDFS_NFS3_SECURE_EXTRA_OPTS=”-jvm server”

On privileged gateways, user to run the gateway as after dropping privileges

This will replace the hadoop.id.str Java property in secure mode.

export HDFS_NFS3_SECURE_USER=nfsserver

#

ZKFailoverController specific parameters

#

Specify the JVM options to be used when starting the ZKFailoverController.

These options will be appended to the options specified as HADOOP_OPTS

and therefore may override any similar flags set in HADOOP_OPTS

#

export HDFS_ZKFC_OPTS=””

#

QuorumJournalNode specific parameters

#

Specify the JVM options to be used when starting the QuorumJournalNode.

These options will be appended to the options specified as HADOOP_OPTS

and therefore may override any similar flags set in HADOOP_OPTS

#

export HDFS_JOURNALNODE_OPTS=””

#

HDFS Balancer specific parameters

#

Specify the JVM options to be used when starting the HDFS Balancer.

These options will be appended to the options specified as HADOOP_OPTS

and therefore may override any similar flags set in HADOOP_OPTS

#

export HDFS_BALANCER_OPTS=””

#

HDFS Mover specific parameters

#

Specify the JVM options to be used when starting the HDFS Mover.

These options will be appended to the options specified as HADOOP_OPTS

and therefore may override any similar flags set in HADOOP_OPTS

#

export HDFS_MOVER_OPTS=””

#

Router-based HDFS Federation specific parameters

Specify the JVM options to be used when starting the RBF Routers.

These options will be appended to the options specified as HADOOP_OPTS

and therefore may override any similar flags set in HADOOP_OPTS

#

export HDFS_DFSROUTER_OPTS=””

#

#

Advanced Users Only!

#

#

When building Hadoop, one can add the class paths to the commands

via this special env var:

export HADOOP_ENABLE_BUILD_PATHS=”true”

#

To prevent accidents, shell commands be (superficially) locked

to only allow certain users to execute certain subcommands.

It uses the format of (command)_(subcommand)_USER.

#

For example, to limit who can execute the namenode command,

export HDFS_NAMENODE_USER=hdfs

  1. -
  2. 配置core-site.xml
  3. ```xml
  4. <configuration>
  5. <property>
  6. <name>fs.defaultFS</name>
  7. <!--如果不想写端口只写localhost,hadoop会给默认端口8020-->
  8. <value>hdfs://hadoop01:9000</value>
  9. </property>
  10. <property>
  11. <name>hadoop.tmp.dir</name>
  12. <value>/data/hadoop</value>
  13. </property>
  14. </configuration>
  1. -

配置hdfs-site.xml

  1. <configuration>
  2. <property>
  3. <name>dfs.namenode.name.dir</name>
  4. <value>file:///data/hadoop/dfs/name</value>
  5. </property>
  6. <property>
  7. <name>dfs.datanode.data.dir</name>
  8. <value>file:///data/hadoop/dfs/data</value>
  9. </property>
  10. <property>
  11. <!--副本数-->
  12. <name>dfs.replication</name>
  13. <value>2</value>
  14. </property>
  15. <property>
  16. <name>dfs.namenode.secondary.http-address</name>
  17. <value>hadoop01:50090</value>
  18. </property>
  19. <property>
  20. <name>dfs.namenode.secondary.https-address</name>
  21. <value>hadoop01:50091</value>
  22. </property>
  23. </configuration>
  1. -

配置yarn-site.xml

  1. <configuration>
  2. <property>
  3. <name>yarn.resourcemanager.hostname</name>
  4. <value>hadoop01</value>
  5. </property>
  6. <property>
  7. <name>yarn.nodemanager.aux-services</name>
  8. <value>mapreduce_shuffle</value>
  9. </property>
  10. <!--<property>
  11. <name>yarn.nodemanager.env-whitelist</name>
  12. <value>JAVA_HOME,HADOOP_HOME,HADOOP_HDFS_HOME,HADOOP_CONF_DIR,CLASSPATH_PREPEND_DISTCACHE,HADOOP_YARN_HOME,HADOOP_MAPRED_HOME</value>
  13. </property>-->
  14. </configuration>
  1. -

配置mapred-site.xml

  1. <configuration>
  2. <property>
  3. <name>mapreduce.framework.name</name>
  4. <value>yarn</value>
  5. </property>
  6. </configuration>
  1. -

修改works文件

  1. hadoop02
  2. hadoop03
  1. 在另外两台机器上安装hadoop

  2. 通过scp命令将hadoop01上修改的hadoop文件发送到hadoop02、hadoop03文件相同的位置

    1. scp /opt/hadoop-3.1.2/etc/hadoop/* hadoop02:/opt/hadoop-3.1.2/etc/hadoop/
    2. scp /opt/hadoop-3.1.2/etc/hadoop/* hadoop03:/opt/hadoop-3.1.2/etc/hadoop/


同样的不要忘了配置hadoop的环境变量

  1. 启动hadoop集群

    • 格式化hadoop 集群master节点
      1. hdfs namenode -format
  • 启动
    1. ./start-all.sh
  • 分别在三台机器上查看否启动成功
    1. jps
  1. 访问http://hadoop01:9870/dfshealth.html#tab-datanode查看DataNode

Java 对HDFS操作(spring boot 集成Hadoop)

创建spring boot集成项目(hadoop)

  1. pom.xml引入hadoop依赖
    1. <dependency>
    2. <groupId>org.apache.hadoop</groupId>
    3. <artifactId>hadoop-client</artifactId>
    4. <version>3.2.0</version>
    5. <exclusions>
    6. <exclusion>
    7. <groupId>org.slf4j</groupId>
    8. <artifactId>slf4j-log4j12</artifactId>
    9. </exclusion>
    10. <exclusion>
    11. <groupId>javax.servlet</groupId>
    12. <artifactId>servlet-api</artifactId>
    13. </exclusion>
    14. </exclusions>
    15. </dependency>
    16. <dependency>
    17. <groupId>org.apache.hadoop</groupId>
    18. <artifactId>hadoop-common</artifactId>
    19. <version>3.2.0</version>
    20. <exclusions>
    21. <exclusion>
    22. <groupId>org.slf4j</groupId>
    23. <artifactId>slf4j-log4j12</artifactId>
    24. </exclusion>
    25. <exclusion>
    26. <groupId>javax.servlet</groupId>
    27. <artifactId>servlet-api</artifactId>
    28. </exclusion>
    29. </exclusions>
    30. </dependency>
    31. <dependency>
    32. <groupId>org.apache.hadoop</groupId>
    33. <artifactId>hadoop-hdfs</artifactId>
    34. <version>3.0.0</version>
    35. <exclusions>
    36. <exclusion>
    37. <groupId>org.slf4j</groupId>
    38. <artifactId>slf4j-log4j12</artifactId>
    39. </exclusion>
    40. <exclusion>
    41. <groupId>javax.servlet</groupId>
    42. <artifactId>servlet-api</artifactId>
    43. </exclusion>
    44. </exclusions>
    45. </dependency>
  1. application.yml配置文件中有关hadoop配置
    1. hadoop:
    2. namenode: hdfs://192.168.127.128:9000
    3. namespace: /mydata
  1. 提供的方法

    • 向hdfs上传文件
    • 向从hdfs中下载文件
    • 删除hdfs中的文件夹或文件
    • 创建目录
  2. 详细内容请查看项目