We provide IT Staff Augmentation Services!

Java Hadoop Developer Resume Profile

5.00/5 (Submit Your Rating)

Mclean, VA

Professional Summary

  • 7 years of IT experience with the last 1.5 years spent working on Hadoop/Big Data.
  • Good understanding/knowledge of Hadoop Architecture.
  • Hands-on experience on major components in Hadoop Ecosystem including Hive, PIG, Sqoop, Flume and knowledge of Mapper/Reduce/HDFS Frame work.
  • Experience on Hadoop CDH4
  • Set up standards and processes for Hadoop based application design and implementation.
  • Experience in analyzing data using HIVEQL, PIG Latin and custom MapReduce programs in JAVA.
  • Extending HIVE and PIG core functionality by using custom UDF's.
  • Good experience in writing Pig Latin scripts and Hive queries
  • Experienced in developing MapReduce programs using Apache Hadoop for working with Big Data.
  • Experience in importing and exporting the data using Sqoop from HDFS to Relational Database systems/mainframe and vice-versa.
  • Hands on experience in application development using Java, RDBMS, and Linux shell scripting.
  • Team player with excellent analytical, communication and project documentation skills

Technical Skills

  • Hardware / Platforms: Ubuntu, Windows 98/XP/2000, Win 7, Mainframe
  • Technical skills: Core Java, Map reduce, PIG, Hive, Shell scripting, Sqoop, Flume, Oozie
  • Database Tools: DB2, SQL, knowledge on NOSQL
  • Testing Tools: Selenium 2.0 web driver, Junit4, Quality Center 11

Project profile

Confidential

Role: Java Hadoop Developer

Confidential

Project Description

The scope of the project was to read the streaming data from twitter firehose and perform analytics on it. The data was read continuously by Flume and analytics like finding the count of number of times a particular mobile model, company was tweeted. The aggregated data was stored into a flat file or to a database like MySQL.

Role and Responsibilities

  • Utilized Flume to filter out the JSON input data read from the twitter firehose to retrieve only the data needed to perform analytics.
  • Involved in writing Pig script to count the number of times a particular model of phone was tweeted in a particular duration. Later a comparison of the count of various models shows the relative popularity of the phone model in that duration.
  • Hive was used to find the location, phone pair to the popularity if phone with respect to location/State.
  • Another Hive script was written to find the user, phone pair to the popularity if phone with respect to twitter users. This shows the interests of users in phone models.

Technology Used: Hadoop, HDFS, Hive, Flume, PIG, Ubuntu, Java JDK 1.6 ,Cloudera

Confidential

Role: Hadoop Developer

Project Description The scope of the project was to provide automated access to real time in-store inventory information in order to efficiently provide actionable information to manage replenishment processes related to aged inventory promotions for client. Automatically establish and maintain a store list for aged inventory items which have locally optimized prices pushed by the Pricing Team.

Role and Responsibilities

  • Extract the data from different sources like main table from Teradata, marketing promotional history, and manual watch list from UI.
  • Develop Pig scripts to establish the data flow to achieve the desired watch list at store and item level exception reporting.
  • Develop shell script to fetch the store, max timestamp, date combinations for Hive tables to pass them as parameters to pig script and establish connection to MySQL database.
  • Created components like PIG UDFs for missing functionality
  • Involved in creating Hive Tables, loading with data and writing Hive queries to do analytics on the data

Technology Used: Hadoop, HDFS, Hive, Sqoop, PIG, Teradata, Ubuntu, Java JDK 1.6 ,Cloudera, Oozie

Role: Hadoop Developer

Project Description:

  • The purpose of the project is to store Terabytes of log information generated by the ecommerce website and extract meaning information out of it. The solution is based on the open source BigData software Hadoop. The data will be stored in Hadoop file system and processed using Map/Reduce jobs. Which intern includes getting the raw html data from the websites, Process the html to obtain product and pricing information, Extract various reports out of the product pricing information and Export the information for further processing.
  • This project is mainly for the re-plat forming of the current existing system which is running on WebHarvest a third party JAR and in MySQL DB to Hadoop which can able to process large data sets i.e. Tera Bytes and Peta Bytes of data in order to meet the client requirements with the increasing completion from his retailers.

Roles and Responsibilities:

  • Moved all web crawling data flat files generated from various retailers to HDFS for further processing
  • Designed and developed MapReduce Jobs using Java.
  • Written the Apache PIG scripts and UDF to process the HDFS data.
  • Created Hive tables to store the processed results in a tabular format.
  • Developed the Sqoop scripts in order to make the interaction between HDFS and MySQL Database.
  • Completely involved in the requirement analysis phase.

Technology Used: Hadoop, HDFS, Flume, PIG, MySQL, Ubuntu, Java JDK 1.6 ,Cloudera, Oozie

Role: Test lead

Project Description: It's a web based application in which Real Estate transactions and mortgage details will be maintained.

Roles and Responsibilities:

  • Requirement analysis, Test scenario design, Test suite preparation.
  • Involved in Test Data creation, Test case execution and defect logging for each sprint cycle.
  • Coordination with the Client, Onsite manager and Development team.
  • Organize weekly meetings with offshore managers, Onsite Coordinator Onsite manager.
  • Responsible for reporting test metrics to client partners and Onsite manager.
  • Automated Smoke and Regression suite using Selenium Webdriver and Frame work using Java.

Technology Used: Selenium 2.0 webdriver, Cognizant Automation framework, Junit4, ANT, Soap UI

Role: Test Analyst

Project Description:Part of data will populate in Java based screen from different interfaces. This is called KIP Key Intelligent Profile screen.

Roles and Responsibilities:

  • Requirement analysis, Test scenario design, Test suite preparation.
  • Automating test cases using Selenium Webdriver
  • Involved in Test Data creation, Test case execution and defect logging.
  • Assigning testing tasks to team members.
  • Coordination with the Client and Development team.
  • Conduct and participate in meetings with Client Management to discuss the Project goals and report the health of the overall project status.

Technology Used: Selenium 2.0 webdriver, Cognizant Automation framework, Junit4, ANT, Soap UI

Role: Test Analyst

Project Description:Actuarial funding is a technique that allows improving capital efficiency by more closely matching its assets to its unit linked liabilities. The Actuarial Funding supplied by factoring an adjustment to the unit or value being processed. The result then provides a figure to reduce the unit liability. A series of tables are updated on the daily basis to hold the current factor for every applicable policy.

Roles and Responsibilities:

  • Requirement analysis, Test scenario design, Test suite preparation.
  • Involved in Test Data creation, Test case execution and defect logging.
  • Assigning

We'd love your feedback!