Sie sind auf Seite 1von 3

To stop entire cluster:

clush -w 192.168.1.171 -b "/usr/local/hadoop/sbin/mr-jobhistory-daemon.sh stop


historyserver"
clush -g rm -b "/usr/local/hadoop/sbin/yarn-daemon.sh stop resourcemanager"
clush -g dn -b "/usr/local/hadoop/sbin/yarn-daemon.sh stop nodemanager"
clush -g nn -b "/usr/local/hadoop/sbin/hadoop-daemon.sh stop namenode"
clush -g dn -b "/usr/local/hadoop/sbin/hadoop-daemon.sh stop datanode"
clush -g jn -b "/usr/local/hadoop/sbin/hadoop-daemon.sh stop journalnode"
clush -g nn -b "/usr/local/hadoop/sbin/hadoop-daemon.sh stop zkfc"
clush -g zk -b "/usr/local/zookeeper/bin/zkServer.sh stop"
for i in 192.168.1.{171,172,173,174,175,176}; do echo -e $i"\n-------------"; ssh $i jps | grep -v Jps;
echo; done

To start entire cluster:

clush -g zk -b "/usr/local/zookeeper/bin/zkServer.sh start"


clush -g nn -b "/usr/local/hadoop/sbin/hadoop-daemon.sh start zkfc"
clush -g jn -b "/usr/local/hadoop/sbin/hadoop-daemon.sh start journalnode"
clush -g nn -b "/usr/local/hadoop/sbin/hadoop-daemon.sh start namenode"
clush -g dn -b "/usr/local/hadoop/sbin/hadoop-daemon.sh start datanode"
clush -g rm -b "/usr/local/hadoop/sbin/yarn-daemon.sh start resourcemanager"
clush -g dn -b "/usr/local/hadoop/sbin/yarn-daemon.sh start nodemanager"
clush -w 192.168.1.171 -b "/usr/local/hadoop/sbin/mr-jobhistory-daemon.sh start
historyserver"
for i in 192.168.1.{171,172,173,174,175,176}; do echo -e $i"\n-------------"; ssh $i jps | grep -v Jps;
echo; done

Apache Pig Installation:

Download latest version of Pig from below websites.

https://archive.apache.org/dist/

sudo mkdir -p /etc/hadoop


sudo unlink /etc/hadoop/pig
sudo tar -zxvf /var/www/html/hadoop_tools/pig-0.13.0.tar.gz -C /etc/hadoop/
sudo ln -s /etc/hadoop/pig-0.13.0 /etc/hadoop/pig
sudo chown -R hdpuser:hdpadmin /etc/hadoop/
ls -ltr /etc/hadoop/pig*

Create Log Directory:


sudo mkdir -p /etc/hadoop/pig/logs
sudo chmod -R 777 /etc/hadoop/pig/logs

Update bash variables:

sudo vi /etc/profile

export PIG_HOME=/etc/hadoop/pig

And add below property to the end of PATH:

:${PIG_HOME}/bin

Refresh bash variables:

source /etc/profile

Change log4j properties to ignore warnings:

sudo cp /etc/hadoop/pig/conf/log4j.properties.template /etc/hadoop/pig/conf/log4j.properties


sudo vi /etc/hadoop/pig/conf/log4j.properties

Sample Commands:

vi Student.csv

1,Venkat,5000,IT

2,Sundar,6000,IT

3,Satheesh,7000,CSC

4,Prasad,8000,CSC

5,Ravi,9000,MECH

6,Chenna,10000,MECH

7,PrasadK,11000,ECE

8,Sandeep,120000,ECE

hdfs dfs -put Student.csv /user/hdpuser/


A = LOAD '/user/hdpuser/Student.csv' USING PigStorage(',') AS (no:int, name:chararray, tot:float,
desg:chararray);

A = LOAD '/user/hdpuser/Student.csv' USING PigStorage(',') AS (no, name, tot, desg);

$0=no

$1=name

$2=tot

$3=desg

X = FOREACH A GENERATE name,$2;

out = GROUP A BY desg;

totfilt = FILTER X BY (tot>=10000);

DESCRIBE A;

ILLUSTRATE A;

DESCRIBE totfilt;

B = foreach A generate $1 as name;


dump B;
hdfs dfs -rm -r /pigoutput
store B INTO ' hdfs://mycluster/pigoutput ' using PigStorage(',');
hdfs dfs -cat /pigoutput/part-m-00000

Das könnte Ihnen auch gefallen