博客
关于我
强烈建议你试试无所不能的chatGPT,快点击我
hadoop 开发环境搭建
阅读量:6076 次
发布时间:2019-06-20

本文共 3891 字,大约阅读时间需要 12 分钟。

一,安装java环境
添加java环境变量
vi /etc/profile
 
# add by tank
export JAVA_HOME=/data/soft/jdk/jdk1.7.0_71
export JRE_HOME=$JAVA_HOME/jre
export CLASSPATH=.:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar
export PATH=$PATH:$JAVA_HOME/bin:$JRE_HOME/bin
 
 
二,修改文件句柄数
vi  /etc/security/limits.conf
 
# add by tank
* soft nofile 65536
* hard nofile  65536
 
 
三,设置ssh无密码登录
 
  参考:http://www.cnblogs.com/tankaixiong/p/4172942.html
 
四,安装hadoop 环境
 
这里用的是hadoop2.20版本
目录结构:
修改配置文件:
[tank@192 hadoop]$ vi core-site.xml 
<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!--
  Licensed under the Apache License, Version 2.0 (the "License");
  you may not use this file except in compliance with the License.
  You may obtain a copy of the License at
    http://www.apache.org/licenses/LICENSE-2.0
  Unless required by applicable law or agreed to in writing, software
  distributed under the License is distributed on an "AS IS" BASIS,
  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  See the License for the specific language governing permissions and
  limitations under the License. See accompanying LICENSE file.
-->
<!-- Put site-specific property overrides in this file. -->
<configuration>
    <property>

        <name>hadoop.tmp.dir</name>

        <value>/usr/hadoop/tmp</value>

        (备注:请先在 /usr/hadoop 目录下建立 tmp 文件夹)

        <description>A base for other temporary directories.</description>

    </property>

  <property>
     <name>fs.default.name</name>
     <value>hdfs://
192.168.149.128:9000</value>
  </property>
</configuration>
备注:如没有配置hadoop.tmp.dir参数,此时系统默认的临时目录为:/tmp/hadoo-hadoop。而这个目录在每次重启后都会被干掉,必须重新执行format才行,否则会出错。
 
[tank@192 hadoop]$ vi hdfs-site.xml 
<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!--
  Licensed under the Apache License, Version 2.0 (the "License");
  you may not use this file except in compliance with the License.
  You may obtain a copy of the License at
    http://www.apache.org/licenses/LICENSE-2.0
  Unless required by applicable law or agreed to in writing, software
  distributed under the License is distributed on an "AS IS" BASIS,
  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  See the License for the specific language governing permissions and
  limitations under the License. See accompanying LICENSE file.
-->
<!-- Put site-specific property overrides in this file. -->
<configuration>
  <property>
    <name>dfs.replication</name>
    <value>1</value>
  </property>
  
<property>
      <name>dfs.namenode.name.dir</name>
      <value>file:/data/soft/hadoop/hadoop-2.2.0/hdfs/name</value>
        <final>true</final>
   </property>
   <property>
        <name>dfs.datanode.data.dir</name>
        <value>file:/data/soft/hadoop/hadoop-2.2.0/hdfs/data</value>
   </property>
</configuration>
 
文件必须已经存在
 
[tank@192 hadoop]$ vi mapred-site.xml
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!--
  Licensed under the Apache License, Version 2.0 (the "License");
  you may not use this file except in compliance with the License.
  You may obtain a copy of the License at
    http://www.apache.org/licenses/LICENSE-2.0
  Unless required by applicable law or agreed to in writing, software
  distributed under the License is distributed on an "AS IS" BASIS,
  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  See the License for the specific language governing permissions and
  limitations under the License. See accompanying LICENSE file.
-->
<!-- Put site-specific property overrides in this file. -->
<configuration>
   <property>
     <name>mapred.job.tracker</name>
     <value>
192.168.149.128:9001</value>
   </property>
</configuration>
~
~
 
 
注意上面一定要填Ip,不要填localhost,不然eclipse会连接不到!
 
 bin/hadoop namenode -format //格式化
启动:
sbin/start-all.sh
 
查看状态:
[tank@192 hadoop-2.2.0]$ jps
10680 ResourceManager //yarn
10784 NodeManager
10350 DataNode
10254 NameNode
10527 SecondaryNameNode
 
总共有5个hadoop线程 
 
访问地址查看:
http://192.168.149.128:50070/dfshealth.jsp
 
 
 
你可能感兴趣的文章
LightOJ - 1010 Knights in Chessboard
查看>>
杭电2122--Ice_cream's world III
查看>>
Lucene.Net介绍
查看>>
如何在sumbline运行python文件
查看>>
[转] babel的使用
查看>>
jquery对话框插件
查看>>
IE去掉链接虚线框的几个方法
查看>>
[Linux]ubuntu16.04 nginx彻底删除与重新安装
查看>>
Snackbar
查看>>
1074 Reversing Linked List
查看>>
织梦标签教程
查看>>
React Native 集成code push
查看>>
Android启动页重复启动问题
查看>>
c++ MFC 利用组码 自己生成DXF
查看>>
MVP模式在Android实际项目中的应用和优化
查看>>
3016: [Usaco2012 Nov]Clumsy Cows
查看>>
React 深度学习:ReactSideEffectTags
查看>>
dmidecode输出详解
查看>>
统计单词个数
查看>>
Vue.js 第二天: 组件
查看>>