Beruflich Dokumente
Kultur Dokumente
minute
video
Watc
h the
2
2. Get organized
You start with "My Notebook" - everything lives in here
Remember everything
Add Tags to any notes
Make checklists and to-do lists
Create your own custom tags
From Outlook:
Also for maven projects all dev teams should use Hortonworks repo instead of public apache repo:
<repositories>
<repository>
<releases>
<enabled>true</enabled>
<updatePolicy>always</updatePolicy>
<checksumPolicy>warn</checksumPolicy>
</releases>
<snapshots>
<enabled>false</enabled>
<updatePolicy>never</updatePolicy>
<checksumPolicy>fail</checksumPolicy>
</snapshots>
<id>HDPReleases</id>
<name>HDP Releases</name>
<url>http://repo.hortonworks.com/content/repositories/releases/</url>
<layout>default</layout>
</repository>
</repositories>
Hi All,
Pl find the attached script for taking backup and recreate the Hbase table with pre split. Pl follow the
below steps. Example command has used a HBase table named master:TEST_ADD with two column A, B.
o create 'master:TEST_ADD', {NAME => 'cf1', COMPRESSION => 'SNAPPY'}, {SPLITS => ['00000',
'00001', '00002', '00003', '00004', '00005', '00006', '00007', '00008',
'1','2','3','4','5','6','7','8','9','a','c','e','g', 'l', 'j', 'k', 'm', 'o', 'q', 's', 'u', 'w', 'y']}
create 'party:NPANXX', {NAME => 'cf1', COMPRESSION => 'SNAPPY'}, {SPLITS => ['00000', '00001',
'00002', '00003', '00004', '00005', '00006', '00007', '00008', '1','2','3','4','5','6','7','8','9','a','c','e','g', 'l', 'j',
'k', 'm', 'o', 'q', 's', 'u', 'w', 'y']}
Example:
snapshot 'DEV_RPX.PAG_ADM.NGP_REF', 'DEV_RPX.PAG_ADM.NGP_REF_SNAP'
clone_snapshot 'DEV_RPX.PAG_ADM.NGP_REF_SNAP' , 'DEV_RPX.PAG_ADM.NGP_REF_BKP'
HBase commands:
get 'DEV_NCS.CPN_OWN.SKU_REF','984216:1900-01-01 12:00:00','CF1:SKU_ID'
get '<tablename>','<rowkey>','CF1:<columnname>'
deleteall 'DEV_NCS.CPN_OWN.SKU_REF','rowkey'
mapred.job.queue.name=dev
export PIG_OPTS="-Dhive.metastore.uris=thrift://devehdp004.unix.gsm1900.org:9083 -
Dmapred.job.queue.name=dev"
export PIG_CLASSPATH=/usr/hdp/current/hive-webhcat/share/hcatalog/*:/usr/hdp/2.2.4.2-2/hive/lib/*
pig -useHCatalog
REGISTER /usr/hdp/2.2.4.2-2/hbase/lib/*.jar;
REGISTER /usr/hdp/current/hive-webhcat/share/hcatalog/*.jar;
REGISTER /usr/hdp/2.2.4.2-2/hive/lib/*.jar;
REGISTER /usr/local/share/eit_hadoop/applications/idw/Finance_Retrofit/piggybank.jar;
REGISTER /usr/local/share/eit_hadoop/applications/idw/Finance_Retrofit/idwudf-1.0.jar;
Connecting to mysql db
mysql --host=devehdp004.unix.gsm1900.org --user=sqoop --password=sqoop
Deleting a Sqoop job in DEV
sqoop job --meta-connect "jdbc:mysql://devehdp004.unix.gsm1900.org:3306/sqoop?
user=sqoop&password=sqoop" --delete hdp_event_account_transaction_event_disp_EXPORT
Below are the instructions for using Phoenix Client on Hadoop server. You are seeing error message as it is
unable to read Hbase configs.
export HBASE_CONF_PATH=/etc/hbase/conf:/etc/hadoop/conf
cd /usr/hdp/current/phoenix-client/bin
./sqlline.py devehdp001,devehdp002,devehdp003:2181:/hbase-secure
Note: Make sure you have valid kerberos ticket before starting sqlline client
#command
Kinit <<NTID>>@GSM1900.ORG
Commands:
!tables
!describe <tablename>
First:
Did you ran this job by adding below property? If any job is running beyond 24 hours it is failing as
delegation token is cancelled. If you are still seeing the same issue even after adding that property,
then it has to be made at cluster level to reflect the change.
<property>
<name>mapreduce.job.complete.cancel.delegation.tokens</name>
<value>false</value>
</property>
Second:
1. hbase.rpc.timeout
2. hbase.client.scanner.timeout.period
set below properties to 900000
Third:
Production details:
211 datanodes
24 core
126 GB ram
111 hbase region servers
Each data node size: 24TB (of which we use 2TB for Harddisk root)
Total hdfs storage is 3.6PB
Versions:
Horton works version: 2.2.4
Hadoop: 2.6
Pig: 0.14
Hive: 0.14
Oozie: 4.1
Hbase: 0.98.4
QA Cluster Details
IP Address: 10.158.163.15
Hostname: qatehdp005.unix.gsm1900.org
Dev Details:
Dev UI URL: http://devbeam002.unix.gsm1900.org:8080/abcapp/login#
Credentials: dadmin/abcuiadmin2016
MySql details:
Hostname : devehdp006.unix.gsm1900.org
Port : 3306
DB Name : demo_abc_platform
Username : readonlyabc
Password: readonly123
QA Details:
QA UI URL: http://qatbeam001.unix.gsm1900.org:8080/abc_qat
Credentials: dadmin/abcuiadmin2016
For Testing:
For control check service testing, you can look into following runtime tables
CONTROL_DETAILS
For balance check service testing, you can look into following runtime tables
BALANCE_CHECK_LOG
MISMATCH_CHECK_LOG
MISMATCH_JOB_STATUS_CHECK
You can use individual process id from table LOAD_RPOCESS_AUDIT_STATS and link to the runtime tables above.
Notes:
ABC ERD pdf:
https://tmobileusa.sharepoint.com/teams/da/DM/PR207463/PL/000 IDW Architecture Design and Development/Frameworks/ABC/design/abc_ERD_20151101.pdf
Sqoop Metastore
1) Login to devehdp004
2) mysql -u sqoop -p
3) password: sqoop
Git: https://git-scm.com/downloads
Git Commands:
Rt click -> Git Bash here
git config --global user.name aakula
git config --global user.email anusha.akula@T-Mobile.com
git pull
git sync
Jenkins:
URL: http://prdcicd005.unix.gsm1900.org:8080/login?from=%2F
Username: aakula
Password: NT password
Nexus:
URL: http://prdcicd003.unix.gsm1900.org:8081/nexus
ID: aakula
Password: password
IP Address: 10.158.31.206
Username: hdpsrvc
Password: G+j4y=z6s@ef-uh_che4ut&7bE?5xa
PIG http://blog.cloudera.com/blog/2015/07/how-to-tune-mapreduce-
performance parallelism-in-apache-pig-jobs/
Good PIG https://www.xplenty.com/blog/2014/05/improving-pig-data-integration-
performance performance-with-join/
PIG http://hortonworks.com/blog/pig-performance-and-optimization-analysis/
performance
Good PIG http://pig.apache.org/docs/r0.9.1/perf.html
performance
http://sanjivblogs.blogspot.com/2015/05/10-ways-to-optimizing-hive-queries.html - Hive
http://phoenix.apache.org/presentations/OC-HUG-2014-10-4x3.pdf - Phoenix
http://www.cs.brandeis.edu/~rshaull/cs147a-fall-2008/hadoop-troubleshooting/ - Troubleshooting
http://0x0fff.com/hadoop-mapreduce-comprehensive-description/ - MapReduce
Phoenix grammar:
https://phoenix.apache.org/language/index.html#create_view
https://phoenix.apache.org/faq.html - FAQs
https://phoenix.apache.org/Phoenix-in-15-minutes-or-less.html - Phoenix in 15 mins
http://kubilaykara.blogspot.com/2015/07/query-existing-hbase-tables-with-sql.html
https://github.com/JerryLead/SparkInternals
https://github.com/JerryLead/SparkInternals/blob/master/EnglishVersion/3-JobPhysicalPlan.md