Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- [training@localhost ~]$ ls
- Desktop Downloads Pictures src workspace
- Documents eclipse Public udacity_training
- [training@localhost ~]$ cd udacity_training/
- [training@localhost udacity_training]$ ls
- code data
- [training@localhost udacity_training]$ cd data
- [training@localhost data]$ ls
- access_log.gz purchases.txt
- [training@localhost data]$ hadoop fs -ls
- [training@localhost data]$ hadoop fs -put purchases.txt
- [training@localhost data]$ hadoop fs -ls
- Found 1 items
- -rw-r--r-- 1 training supergroup 211312924 2014-10-31 17:44 purchases.txt
- [training@localhost data]$ hadoop fs -tail purchases.txt
- 31 17:59 Norfolk Toys 164.34 MasterCard
- 2012-12-31 17:59 Chula Vista Music 380.67 Visa
- 2012-12-31 17:59 Hialeah Toys 115.21 MasterCard
- 2012-12-31 17:59 Indianapolis Men's Clothing 158.28 MasterCard
- 2012-12-31 17:59 Norfolk Garden 414.09 MasterCard
- 2012-12-31 17:59 Baltimore DVDs 467.3 Visa
- 2012-12-31 17:59 Santa Ana Video Games 144.73 Visa
- 2012-12-31 17:59 Gilbert Consumer Electronics 354.66 Discover
- 2012-12-31 17:59 Memphis Sporting Goods 124.79 Amex
- 2012-12-31 17:59 Chicago Men's Clothing 386.54 MasterCard
- 2012-12-31 17:59 Birmingham CDs 118.04 Cash
- 2012-12-31 17:59 Las Vegas Health and Beauty 420.46 Amex
- 2012-12-31 17:59 Wichita Toys 383.9 Cash
- 2012-12-31 17:59 Tucson Pet Supplies 268.39 MasterCard
- 2012-12-31 17:59 Glendale Women's Clothing 68.05 Amex
- 2012-12-31 17:59 Albuquerque Toys 345.7 MasterCard
- 2012-12-31 17:59 Rochester DVDs 399.57 Amex
- 2012-12-31 17:59 Greensboro Baby 277.27 Discover
- 2012-12-31 17:59 Arlington Women's Clothing 134.95 MasterCard
- 2012-12-31 17:59 Corpus Christi DVDs 441.61 Discover
- [training@localhost data]$ hadoop fs -mkdir myinput
- hadoop[training@localhost data]$ hadoop fs -put purchases.txt myinput
- [training@localhost data]$ hadoop fs -ls
- Found 2 items
- drwxr-xr-x - training supergroup 0 2014-10-31 17:49 myinput
- -rw-r--r-- 1 training supergroup 211312924 2014-10-31 17:44 purchases.txt
- [training@localhost data]$ hadoop fs -ls myinput
- Found 1 items
- -rw-r--r-- 1 training supergroup 211312924 2014-10-31 17:49 myinput/purchases.txt
- [training@localhost data]$ ls
- access_log.gz purchases.txt
- [training@localhost data]$ cd
- [training@localhost ~]$ ls
- Desktop Downloads Pictures src workspace
- Documents eclipse Public udacity_training
- [training@localhost ~]$ cd udacity_training/
- [training@localhost udacity_training]$ ls
- code data
- [training@localhost udacity_training]$ cd code
- [training@localhost code]$ ls
- mapper.py reducer.py
- [training@localhost code]$ hadoop fs -ls
- Found 2 items
- drwxr-xr-x - training supergroup 0 2014-10-31 17:49 myinput
- -rw-r--r-- 1 training supergroup 211312924 2014-10-31 17:44 purchases.txt
- [training@localhost code]$ hadoop fs -ls myinput
- Found 1 items
- -rw-r--r-- 1 training supergroup 211312924 2014-10-31 17:49 myinput/purchases.txt
- [training@localhost code]$ hadoop jar /usr/lib
- lib/ libexec/
- [training@localhost code]$ hadoop jar /usr/lib
- lib/ libexec/
- [training@localhost code]$ hadoop jar /usr/lib
- lib/ libexec/
- [training@localhost code]$ hadoop jar /usr/lib/hadoop-0.20-mapreduce/contrib/streaming/hadoop-streaming-2.0.0-mr1-cdh4.1.1.jar -mapper mapper.py -reducer reducer.py -file mapper.py -file reducer.py -input myinput -output joboutput
- packageJobJar: [mapper.py, reducer.py, /tmp/hadoop-training/hadoop-unjar3607163364207776574/] [] /tmp/streamjob5522053172829658683.jar tmpDir=null
- 14/10/31 18:02:12 WARN mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same.
- 14/10/31 18:02:12 WARN snappy.LoadSnappy: Snappy native library is available
- 14/10/31 18:02:12 INFO snappy.LoadSnappy: Snappy native library loaded
- 14/10/31 18:02:12 INFO mapred.FileInputFormat: Total input paths to process : 1
- 14/10/31 18:02:12 INFO streaming.StreamJob: getLocalDirs(): [/var/lib/hadoop-hdfs/cache/training/mapred/local]
- 14/10/31 18:02:12 INFO streaming.StreamJob: Running job: job_201410311722_0001
- 14/10/31 18:02:12 INFO streaming.StreamJob: To kill this job, run:
- 14/10/31 18:02:12 INFO streaming.StreamJob: UNDEF/bin/hadoop job -Dmapred.job.tracker=0.0.0.0:8021 -kill job_201410311722_0001
- 14/10/31 18:02:12 INFO streaming.StreamJob: Tracking URL: http://0.0.0.0:50030/jobdetails.jsp?jobid=job_201410311722_0001
- 14/10/31 18:02:13 INFO streaming.StreamJob: map 0% reduce 0%
- 14/10/31 18:02:27 INFO streaming.StreamJob: map 8% reduce 0%
- 14/10/31 18:02:31 INFO streaming.StreamJob: map 13% reduce 0%
- 14/10/31 18:02:34 INFO streaming.StreamJob: map 17% reduce 0%
- 14/10/31 18:02:37 INFO streaming.StreamJob: map 22% reduce 0%
- 14/10/31 18:02:40 INFO streaming.StreamJob: map 26% reduce 0%
- 14/10/31 18:02:43 INFO streaming.StreamJob: map 31% reduce 0%
- 14/10/31 18:02:46 INFO streaming.StreamJob: map 36% reduce 0%
- 14/10/31 18:02:49 INFO streaming.StreamJob: map 41% reduce 0%
- 14/10/31 18:02:52 INFO streaming.StreamJob: map 43% reduce 0%
- 14/10/31 18:02:53 INFO streaming.StreamJob: map 45% reduce 0%
- 14/10/31 18:02:55 INFO streaming.StreamJob: map 47% reduce 0%
- 14/10/31 18:02:56 INFO streaming.StreamJob: map 49% reduce 0%
- 14/10/31 18:02:59 INFO streaming.StreamJob: map 50% reduce 0%
- 14/10/31 18:03:16 INFO streaming.StreamJob: map 73% reduce 17%
- 14/10/31 18:03:18 INFO streaming.StreamJob: map 78% reduce 17%
- 14/10/31 18:03:19 INFO streaming.StreamJob: map 81% reduce 17%
- 14/10/31 18:03:22 INFO streaming.StreamJob: map 86% reduce 25%
- 14/10/31 18:03:25 INFO streaming.StreamJob: map 91% reduce 25%
- 14/10/31 18:03:28 INFO streaming.StreamJob: map 96% reduce 25%
- 14/10/31 18:03:31 INFO streaming.StreamJob: map 100% reduce 25%
- 14/10/31 18:03:34 INFO streaming.StreamJob: map 100% reduce 33%
- 14/10/31 18:03:37 INFO streaming.StreamJob: map 100% reduce 69%
- 14/10/31 18:03:40 INFO streaming.StreamJob: map 100% reduce 76%
- 14/10/31 18:03:43 INFO streaming.StreamJob: map 100% reduce 82%
- 14/10/31 18:03:46 INFO streaming.StreamJob: map 100% reduce 87%
- 14/10/31 18:03:49 INFO streaming.StreamJob: map 100% reduce 93%
- 14/10/31 18:03:52 INFO streaming.StreamJob: map 100% reduce 100%
- 14/10/31 18:03:56 INFO streaming.StreamJob: Job complete: job_201410311722_0001
- 14/10/31 18:03:56 INFO streaming.StreamJob: Output: joboutput
- [training@localhost code]$ hadoop fs -ls
- Found 3 items
- drwxr-xr-x - training supergroup 0 2014-10-31 18:03 joboutput
- drwxr-xr-x - training supergroup 0 2014-10-31 17:49 myinput
- -rw-r--r-- 1 training supergroup 211312924 2014-10-31 17:44 purchases.txt
- [training@localhost code]$ hadoop fs -ls joboutput
- Found 3 items
- -rw-r--r-- 1 training supergroup 0 2014-10-31 18:03 joboutput/_SUCCESS
- drwxr-xr-x - training supergroup 0 2014-10-31 18:02 joboutput/_logs
- -rw-r--r-- 1 training supergroup 2296 2014-10-31 18:03 joboutput/part-00000
- [training@localhost code]$ hadoop fs -cat joboutput/part 0000 | less
- [training@localhost code]$ hadoop fs -cat joboutput/part-0000 | less
- [training@localhost code]$
- [training@localhost code]$ hadoop fs -cat joboutput/part-00000 | less
- [training@localhost code]$
- [training@localhost code]$ hadoop fs -cat joboutput/part-00000 mylocalfile.txt
- Albuquerque 10052311.42
- Anaheim 10076416.36
- Anchorage 9933500.4
- Arlington 10072207.97
- Atlanta 9997146.7
- Aurora 9992970.92
- Austin 10057158.9
- Bakersfield 10031208.92
- Baltimore 10096521.45
- Baton Rouge 10131273.23
- Birmingham 10076606.52
- Boise 10039166.74
- Boston 10039473.28
- Buffalo 10001941.19
- Chandler 9919559.86
- Charlotte 10112531.34
- Chesapeake 10038504.92
- Chicago 10062522.07
- Chula Vista 9974951.34
- Cincinnati 10139505.74
- Cleveland 10067835.84
- Colorado Springs 10061105.87
- Columbus 10035241.03
- Corpus Christi 9976522.77
- Dallas 10066548.45
- Denver 10031534.87
- Detroit 9979260.76
- Durham 10153890.21
- El Paso 10016409.97
- Fort Wayne 10132594.02
- Fort Worth 10120830.65
- Fremont 10053242.36
- Fresno 9976260.26
- Garland 10071043.92
- Gilbert 10062115.19
- Glendale 10044493.97
- Greensboro 10033781.39
- Henderson 10053416.05
- Hialeah 10047052.76
- Honolulu 10006273.49
- Houston 10042106.27
- Indianapolis 10090272.77
- Irvine 10084867.45
- Irving 10133944.08
- Jacksonville 10072003.33
- Jersey City 9920141.87
- Kansas City 9968118.73
- Laredo 10144604.98
- Las Vegas 10054257.98
- Lexington 10084510.95
- Lincoln 10069485.4
- Long Beach 10006380.25
- Los Angeles 10084576.8
- Louisville 10008566.47
- Lubbock 9958119.15
- Madison 10032035.54
- Memphis 10038565.32
- Mesa 10053642.6
- Miami 9947316.07
- Milwaukee 10064482.65
- Minneapolis 10011757.78
- Nashville 9961450.51
- New Orleans 9949257.75
- New York 10085293.55
- Newark 10144052.8
- Norfolk 10088563.17
- North Las Vegas 10029652.51
- Oakland 9947292.52
- Oklahoma City 10118986.25
- Omaha 10026642.34
- Orlando 10074922.52
- Philadelphia 10190080.26
- Phoenix 10079076.7
- Pittsburgh 10090124.82
- Plano 10046103.61
- Portland 10007635.77
- Raleigh 10061442.54
- Reno 10079955.16
- Richmond 9992941.59
- Riverside 10006695.42
- Rochester 10067606.92
- Sacramento 10123468.18
- Saint Paul 10057233.57
- San Antonio 10014441.7
- San Bernardino 9965152.04
- San Diego 9966038.39
- San Francisco 9995570.54
- San Jose 9936721.41
- Santa Ana 10050309.93
- Scottsdale 10037929.85
- Seattle 9936267.37
- Spokane 10083362.98
- St. Louis 10002105.14
- St. Petersburg 9986495.54
- Stockton 10006412.64
- Tampa 10106428.55
- Toledo 10020768.88
- Tucson 9998252.47
- Tulsa 10064955.9
- Virginia Beach 10086553.5
- Washington 10139363.39
- Wichita 10083643.21
- Winston–Salem 10044011.83
- cat: `mylocalfile.txt': No such file or directory
- [training@localhost code]$ hadoop fs -get joboutput/part-00000 mylocalfile.txt
- [training@localhost code]$ hs mapper.py reducer.py myinput joboutput
- packageJobJar: [mapper.py, reducer.py, /tmp/hadoop-training/hadoop-unjar3086953528091346500/] [] /tmp/streamjob5740223419707269269.jar tmpDir=null
- 14/10/31 18:10:16 WARN mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same.
- 14/10/31 18:10:16 INFO mapred.JobClient: Cleaning up the staging area hdfs://0.0.0.0:8020/var/lib/hadoop-hdfs/cache/mapred/mapred/staging/training/.staging/job_201410311722_0002
- 14/10/31 18:10:16 ERROR security.UserGroupInformation: PriviledgedActionException as:training (auth:SIMPLE) cause:org.apache.hadoop.mapred.FileAlreadyExistsException: Output directory hdfs://0.0.0.0:8020/user/training/joboutput already exists
- 14/10/31 18:10:16 ERROR streaming.StreamJob: Error launching job , Output path already exists : Output directory hdfs://0.0.0.0:8020/user/training/joboutput already exists
- Streaming Command Failed!
- [training@localhost code]$ hs mapper.py reducer.py myinput newoutputdir
- packageJobJar: [mapper.py, reducer.py, /tmp/hadoop-training/hadoop-unjar2291186339325694346/] [] /tmp/streamjob1552916525720226991.jar tmpDir=null
- 14/10/31 18:11:39 WARN mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same.
- 14/10/31 18:11:39 WARN snappy.LoadSnappy: Snappy native library is available
- 14/10/31 18:11:39 INFO snappy.LoadSnappy: Snappy native library loaded
- 14/10/31 18:11:40 INFO mapred.FileInputFormat: Total input paths to process : 1
- 14/10/31 18:11:40 INFO streaming.StreamJob: getLocalDirs(): [/var/lib/hadoop-hdfs/cache/training/mapred/local]
- 14/10/31 18:11:40 INFO streaming.StreamJob: Running job: job_201410311722_0003
- 14/10/31 18:11:40 INFO streaming.StreamJob: To kill this job, run:
- 14/10/31 18:11:40 INFO streaming.StreamJob: UNDEF/bin/hadoop job -Dmapred.job.tracker=0.0.0.0:8021 -kill job_201410311722_0003
- 14/10/31 18:11:40 INFO streaming.StreamJob: Tracking URL: http://0.0.0.0:50030/jobdetails.jsp?jobid=job_201410311722_0003
- 14/10/31 18:11:41 INFO streaming.StreamJob: map 0% reduce 0%
- 14/10/31 18:11:54 INFO streaming.StreamJob: map 9% reduce 0%
- 14/10/31 18:11:58 INFO streaming.StreamJob: map 12% reduce 0%
- 14/10/31 18:12:01 INFO streaming.StreamJob: map 16% reduce 0%
- 14/10/31 18:12:04 INFO streaming.StreamJob: map 20% reduce 0%
- 14/10/31 18:12:08 INFO streaming.StreamJob: map 23% reduce 0%
- 14/10/31 18:12:11 INFO streaming.StreamJob: map 27% reduce 0%
- 14/10/31 18:12:14 INFO streaming.StreamJob: map 32% reduce 0%
- 14/10/31 18:12:17 INFO streaming.StreamJob: map 37% reduce 0%
- 14/10/31 18:12:20 INFO streaming.StreamJob: map 41% reduce 0%
- 14/10/31 18:12:23 INFO streaming.StreamJob: map 47% reduce 0%
- 14/10/31 18:12:26 INFO streaming.StreamJob: map 50% reduce 0%
- 14/10/31 18:12:40 INFO streaming.StreamJob: map 53% reduce 17%
- 14/10/31 18:12:41 INFO streaming.StreamJob: map 71% reduce 17%
- 14/10/31 18:12:43 INFO streaming.StreamJob: map 81% reduce 17%
- 14/10/31 18:12:46 INFO streaming.StreamJob: map 86% reduce 17%
- 14/10/31 18:12:47 INFO streaming.StreamJob: map 86% reduce 25%
- 14/10/31 18:12:49 INFO streaming.StreamJob: map 91% reduce 25%
- 14/10/31 18:12:53 INFO streaming.StreamJob: map 96% reduce 25%
- 14/10/31 18:12:56 INFO streaming.StreamJob: map 100% reduce 25%
- 14/10/31 18:12:59 INFO streaming.StreamJob: map 100% reduce 33%
- 14/10/31 18:13:02 INFO streaming.StreamJob: map 100% reduce 71%
- 14/10/31 18:13:05 INFO streaming.StreamJob: map 100% reduce 78%
- 14/10/31 18:13:08 INFO streaming.StreamJob: map 100% reduce 84%
- 14/10/31 18:13:11 INFO streaming.StreamJob: map 100% reduce 90%
- 14/10/31 18:13:14 INFO streaming.StreamJob: map 100% reduce 97%
- 14/10/31 18:13:16 INFO streaming.StreamJob: map 100% reduce 100%
- 14/10/31 18:13:18 INFO streaming.StreamJob: Job complete: job_201410311722_0003
- 14/10/31 18:13:18 INFO streaming.StreamJob: Output: newoutputdir
- [training@localhost code]$ ls
- mapper.py mylocalfile.txt reducer.py
- [training@localhost code]$ hadoop fs -ls
- Found 4 items
- drwxr-xr-x - training supergroup 0 2014-10-31 18:03 joboutput
- drwxr-xr-x - training supergroup 0 2014-10-31 17:49 myinput
- drwxr-xr-x - training supergroup 0 2014-10-31 18:13 newoutputdir
- -rw-r--r-- 1 training supergroup 211312924 2014-10-31 17:44 purchases.txt
- [training@localhost code]$ cd newoputputdir
- bash: cd: newoputputdir: No such file or directory
- [training@localhost code]$ hadoop fs -cat newoutputdir
- cat: `newoutputdir': Is a directory
- [training@localhost code]$ cd newoutputdir
- bash: cd: newoutputdir: No such file or directory
- [training@localhost code]$ newoutputdir
- bash: newoutputdir: command not found
- [training@localhost code]$
Add Comment
Please, Sign In to add comment