Advertisement
Guest User

Untitled

a guest
Nov 14th, 2019
110
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
Bash 0.53 KB | None | 0 0
  1. #! /usr/bin/env bash
  2.  
  3.  
  4. OUT_DIR="local1"
  5. NUM_REDUCERS=5
  6.  
  7. hadoop dfs -rm -r -skipTrash ${OUT_DIR} > /dev/null
  8.  
  9. yarn jar /opt/cloudera/parcels/CDH/lib/hadoop-mapreduce/hadoop-streaming.jar \
  10.     -D mapreduce.job.reducer=${NUM_REDUCERS} \
  11.     -D mapred.job.name="MR1" \
  12.     -files mapper1.py,reducer1.py \
  13.     -mapper mapper1.py \
  14.     -reducer reducer1.py \
  15.     -input /data/ids_part \
  16.     -output $OUT_DIR > /dev/null
  17.  
  18.  
  19. for num in 'seq 0 (($NUM_REDUCERS - 1))'
  20. do
  21.     hdfs dfs -cat ${OUT_DIR}/part-0000$num | head
  22. done
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement