Wednesday, September 17, 2014

MapReduce Plugin for Eclipse

Any eclipse before 3.6 version is compatible with MapReduce plugin. So be sure to download the right version for proper functioning.
 
You can download the eclipse version from http://archive.eclipse.org/eclipse/downloads/. Installing Eclipse is pretty straightforward. Download the packaged .zip file into any directory on your computer. Unzip the compressed file and from the uncompressed folder you can directly run eclipse.exe file under eclipse directory with no modifications or other installation procedure.

After downloading eclipse,
 
1. Move mapreduce plugin's jar file, hadoop-0.18.0-eclipse-plugin.jar to plugins directory under eclipse folder.
2. Run eclipse.exe file to start Eclipse.
3. Please select the workspace directory. This is where Eclipse stores all your source projects and their related settings.
4. After choosing the workspace directory, if you are opening Eclipse for the first time  you are presented with welcome screen, click the button that says "workbench". This is the main view of Eclipse where you write source code, manage your projects etc.
5. On the upper right hand corner of the workbench, click the "Open Prespective" button which is next to "Java Prespective" button.
6. Select "Other" and then Map/Reduce and press OK.
7. Once this is done, you will see a Map/Reduce Locations panel next to Problems, Tasks, Javadoc panels on the bottom of the workbench.
8. Select the Map/Reduce Locations panel, and in that select the elephant logo which says "New Hadoop Location".
9. You will be asked to fill number of parameters identifying the server.
Under General Tab:
  • Location Name: Any descriptive name
  • Map/Reduce Master Host:  The IP address of the hadoop server
  • Map/Reduce Master Port: 9001
  • DFS Master Port: 9000
  • UserName: username used to connect to hadoop server.
Under Advanced Paramerters Tab:
  •  hadoop.job.ugi : It contains your current Windows login credentials. Highlight the first comma-separated value in this list (your username) and replace it with hadoop user.
  • mapred.system.dir : Erase the current value and set it to /hadoop/mapred/system.
When you are done, your server will appear in Map/Reduce Locations panel. In the project explorer, mapreduce plugin has added the ability to browse the HDFS. Any files inserted into HDFS file system can be viewed by expanding the tree by pressing [+] buttons.

No comments:

Post a Comment

Amazon S3: Basic Concepts

Amazon S3 is an reliable, scalable, online object storage that stores files. Bucket: A bucket is a container in Amazon S3 where the fil...