Guest User

bhd

a guest
Sep 27th, 2018
60
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
Bash 1.85 KB | None | 0 0
  1. // Initialization Commands
  2.  
  3. start-dfs.sh
  4. start-yarn.sh
  5. jps
  6.  
  7. // Hadoop Basic FS Commands
  8.  
  9. hadoop fs -mkdir %folder_name%
  10. hadoop fs -ls
  11. hadoop fs -copyFromLocal %locallocation% %hadooplocation%
  12. hadoop fs -copyToLocal %hadooplocation% %locallocation%
  13.  
  14. // Word Count Map & Reduce
  15.  
  16. cd myc
  17. ls -l
  18. vi map.c
  19. cc -o mymapper map.c
  20. cc -o myreduce reduce.c
  21. ls -l
  22. hadoop jar $HADOOP_HOME/share/hadoop/tools/lib/hadoop-streaming-2.9.0.jar -input input -output out1 -mapper mymapper -reducer myreduce -file $HOME/myc/mymapper -file $HOME/myc/myreduce
  23. hadoop fs -ls out1
  24. hadoop fs -copyToLocal out1/part-00000 out1.txt
  25. nano out1.txt
  26.  
  27. // Start Hive - SQL based interface for map & reduce using hadoop
  28.  
  29. cd hive
  30. ls -l
  31. hive
  32.  
  33. // Hive Commands
  34.  
  35. create table student(name STRING, age INT, score FLOAT);
  36. show tables;
  37. describe student;
  38. load data local inpath 'studentdata.csv' overwrite into table student;
  39. select * from student;
  40.  
  41. // Start Pig
  42.  
  43. cd pig
  44. pig -x local
  45.  
  46. // Pig Commands
  47.  
  48. A = load './pg20417.txt';
  49. B = foreach A generate flatten(TOKENIZE((chararray)$0)) as word;
  50. c = group B by word;
  51. D = foreach C generate COUNT(B), group;
  52. store D into './wordcount';
  53. quit;
  54. ls -l
  55. cd wordcount
  56. nano part-r-00000
  57.  
  58. // Sqoop - get data from external database into hadoop
  59.  
  60. cd sqoop
  61. ls -l
  62. hadoop fs -ls
  63. sqoop import --table BhavData -- connect jdbc:mysql://54.70.8.241/StockData --username hduser --password hduser -m 1
  64. hadoop fs -ls
  65. hadoop fs -ls BhavData
  66.  
  67. // Flume - allows you to capture streaming data - general purpose tools not limited to hadoop
  68. cd flime
  69. cp ../data/*.txt
  70. ls -l
  71. flume-ng agent --conf $FLUME_HOME/conf/ -f $FLUME_HOME/conf/flume.conf -n a1 &
  72. ls -l sink
  73. cp pg20417.txt spool
  74. ls -l spool
  75. ls -l sink
  76. cp pg5000.txt spool
  77. ls -l spool
  78. ls -l sink
  79. jps
  80. kill -g 3667
  81.  
  82. // Clear spool and sink
  83.  
  84. rm spool/*
  85. rm sink/*
  86. ls -l spool
  87. ls -l sink
Add Comment
Please, Sign In to add comment