MapReduce streaming example will help you running word count program using Hadoop streaming. We use Python for writing mapper and reducer logic. Data is stored as sample.txt file. mapper, reducer and data can be downloaded in a bundle from the link provided.
Log in with your Hadoop user
Working directory should be set to /usr/local/hadoop.
Make sure you're in /usr/local/hadoop, if not use:
Start HDFS: start-dfs.sh
Start YARN: start-yarn.sh
Check if everything is up (6 services should be running): jps
Download data and code files to be used in this tutorial from here.
Unzip contents of streaming.zip:unzip stream