We provide IT Staff Augmentation Services!

Hadoop Developer Resume

3.00/5 (Submit Your Rating)

Chicago, IL

SUMMARY:

  • Over 8 years of experience spread across Hadoop, Java and ETL that includes extensive experience into Big Data Technologies and in development of standalone and web applications in multi - tiered environment using Java, Hadoop, Hive, HBase, Pig, Sqoop, J2EE Technologies (Spring, Hibernate), Oracle, HTML, Java Script.
  • Extensive experience on Big Data Analytics with hands on experience in writing Map Reduce Jobs on Hadoop Ecosystem including Hive and Pig.
  • Excellent knowledge on Hadoop Architecture as in HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce programming paradigm.
  • Experience with distributed systems, large scale non-relational data stores, mapreduce systems, data modeling and big data systems.
  • Involved in developing solutions to analyze large data sets efficiently
  • Excellent hands on with importing and exporting data from different Relational Database Systems like MySql and Oracle into HDFS and Hive and vice-versa using Sqoop.
  • Hands-on experience in writing Pig Latin scripts working with grunt shells and job scheduling with Oozie.
  • Experience in analyzing data using Hive QL, Pig Latin, and custom Map Reduce programs in Java.
  • Experience with web-based UI development using jQuery, ExtJS, CSS, HTML, HTML5, XHTML and Javascript.
  • Knowledge of job workflow scheduling and monitoring tools like Oozie and Zookeeper.
  • Experience with databases like DB2, Oracle 9i, Oracle 10g, MySQL, SQL Server and MS Access.
  • Experience in creating complex SQL Queries and SQL tuning, writing PL/SQL blocks like stored procedures, Functions, Cursors, Index, triggers and packages.
  • Very good understanding on NOSQL databases like monody and HBase.
  • Have good Knowledge in ETL and hands on experience in Informatica ETL.
  • Extensive experience in creating Class Diagrams, Activity Diagrams, Sequence Diagrams using Unified Modeling Language(UML)
  • Experienced in SDLC, Agile (SCRUM) Methodology, Iterative Waterfall.
  • Experience in developing test cases, performing Unit Testing, Integration Testing, experience in QA with test methodologies and skills for manual/automated testing using tools like WinRunner, JUnit.
  • Experience with various version control systems Clear Case, CVS, and SVN.
  • Expertise in extending Hive and Pig core functionality by writing custom UDFs.
  • Development Experience with all aspects of software engineering and the development life cycle
  • Strong desire to work for a fast-paced, flexible environment
  • Proactive problem solving mentality that thrives in an agile work environment
  • Good Experience on SDLC (Software Development Life cycle).
  • Exceptional ability to learn new technologies and to deliver outputs in short deadlines.
  • Worked with developers, DBAs, and systems support personnel in elevating and automating successful code to production.
  • Possess strong Communication skills of written, oral, interpersonal and presentation.
  • Ability to perform at a high level, meet deadlines, adaptable to ever changing priorities.

TECHNICAL SKILLS:

Programming Languages: Java, C/C++,SQL, PL/SQL

Big Data Technologies: Hadoop (HDFS & Map Reduce), PIG, HIVE, HBASE, ZOOKEEPER,OOZIE

J2EE Technologies: JSP, SERVLETS, JDBC,EJB

Web Technologies: AJAX, HTML5,JavaScript,CSS3

Frameworks: Spring 3.5 - Spring MVC, Spring ORM, Spring Security, Spring ROO, Hibernate, Struts.

Application Servers: IBM WebSphere, JBoss WebLogic.

Web Servers: Apache Tomcat

Relational Databases: Oracle 10/11g, MS SQL Server, My SQL

NoSQL Databases: Monod, HBase

Designing Tools: UML, Visio, Visual Paradigm

IDEs: Eclipse

Operating System: Unix, Windows

Version Control: CVS, SVN, GIT

PROFESSIONAL EXPERIENCE:

Confidential, Chicago, IL

Hadoop Developer

Responsibilities:

  • Analyzed large data sets by running Hive queries and Pig scripts
  • Involved in creating Hive tables, and loading and analyzing data using hive queries
  • Developed Simple to complex Map Reduce Jobs using Hive and Pig
  • Involved in running Hadoop jobs for processing millions of records of text data
  • Load and transform large sets of structured, semi structured and unstructured data
  • Responsible to manage data coming from different sources
  • Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
  • Monitor System health and logs and respond accordingly to any warning or failure conditions.
  • Implemented the workflows using Apache Oozie framework to automate tasks.
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required
  • Developed multiple Map Reduce jobs in java for data cleaning and preprocessing
  • Involved in loading data from LINUX file system to HDFS
  • Experienced in running Hadoop streaming jobs to process terabytes of xml format data.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Assisted in exporting analyzed data to relational databases using Sqoop
  • Supported Map Reduce Programs those are running on the cluster
  • Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts

Environment: Hadoop, HDFS, Pig, Hive, Map Reduce, Sqoop, Oozie, and Big Data, Apache Java (jdk1.6), Data tax, Flat files, Oracle 11g/10g, MySQL, Toad, Windows NT, LINUX.

Confidential, Media, PA

Hadoop Developer

Responsibilities:

  • Handled importing of data from various data sources, performed transformations using Hive, Map Reduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Developed Simple to complex Map/reduce Jobs.
  • Analyzed the data by performing Hive queries and running Pig scripts to know user behavior.
  • Developed Simple to complex Map/reduce Jobs using Hive.
  • Created partitioned tables in Hive.
  • Worked on Installed and configured Hadoop 0.22.0 Map Reduce, HDFS, developed multiple Map Reduce jobs in java for data cleaning and preprocessing.
  • Importing and exporting data into HDFS and HIVE using Sqoop
  • Responsible to manage data coming from different sources
  • Monitoring the running Map Reduce programs on the cluster.
  • Responsible for loading data from UNIX file systems to HDFS.
  • Installed and configured Hive and also wrote Hive UDFs.
  • Involved in creating Hive Tables, loading with data and writing Hive queries which will invoke and run Map Reduce jobs in the backend.
  • Implemented the workflows using Apache Oozie framework to automate tasks.
  • Developed scripts and automated data management from end to end and sync up b/w all the clusters.

Environment: Apache Hadoop, Java (jdk1.6), Data tax, Flat files, Oracle 11g/10g, MySql, Toad 9.6, Windows NT, UNIX, Sqoop, Hive, Oozie.

Confidential, Lowell, MA

Hadoop Developer

Responsibilities:

  • Developed shell scripts to automate the cluster installation.
  • Played a major role in choosing the right configurations for Hadoop.
  • Developed Pig Latin scripts to extract and filter relevant data from the web server output files to load into HDFS.
  • Involved in start to end process of hadoop cluster installation, configuration and monitoring.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
  • Setup and benchmarked Hadoop/HBase clusters for internal use.
  • Developed Simple to complex Map/reduce Jobs using Hive and Pig.
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Analyzed the data by performing Hive queries and running Pig scripts to study customer behavior.
  • Used UDF's to implement business logic in Hadoop.
  • Implemented business logic by writing UDFs in Java and used various UDFs from Piggybanks and other sources.
  • Continuous monitoring and managing the Hadoop cluster using Cloudera Manager.
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
  • Installed Oozieworkflow engine to run multiple Hive and Pig jobs.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for BI team.

Environment: Hadoop, MapReduce, HDFS, Hive, Pig, Java, SQL, Cloudera Manager, Sqoop, Flume, Oozie, Java (jdk 1.6), Eclipse.

Confidential, Birmingham, MI

Hadoop Developer

Responsibilities:

  • Installed and configured Hadoop MapReduce, HDFS and developed multiple MapReduce jobs in Java for data cleansing and preprocessing.
  • Involved in loading data from UNIX file system to HDFS.
  • Installed and configured Hive and also written Hive UDFs.
  • Evaluated business requirements and prepared detailed specifications that follow project guidelines required to develop written programs.
  • Devised procedures that solve complex business problems with due considerations for hardware/software capacity and limitations, operating times and desired results.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Provided quick response to ad hoc internal and external client requests for data and experienced in creating adhoc reports.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and troubleshooting, manage and review data backups, manage and review Hadoop log files.
  • Worked hands on with ETL process.
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, and loaded data into HDFS.
  • Extracted the data from Teradata into HDFS using Sqoop.
  • Analyzed the data by performing Hive queries and running Pig scripts to know user behavior like shopping enthusiasts, travelers, music lovers etc.
  • Exported the patterns analyzed back into Teradata using Sqoop.
  • Continuous monitoring and managing the Hadoop cluster through Cloudera Manager.
  • Installed Oozie workflow engine to run multiple Hive.

Environment: Hadoop, MapReduce, HDFS, Hive, Ooozie, Java (jdk1.6), Cloudera, NoSQL, Oracle 11g, 10g, PL SQL, SQL*PLUS, Toad 9.6, Windows NT, UNIX Shell Scripting.

Confidential

Java/Technical Support Engineer

Responsibilities:

  • Extensively worked in acquiring the requirements from the business analysts and involved in all requirement clarification calls.
  • Understanding the design documents.
  • Involved in Detail level design and coding activities at offshore.
  • Involved in Code review.
  • Writing and testing the JUNIT test classes.
  • Provide support to client applications in production and other environments.
  • Working on tickets raised by the real time users and continuous interaction with end users.
  • Prepared the Technical Design Document, understanding document and test cases (UTCs and ITCs).
  • Provided Technical & Functional support to the end users during UAT & Production.
  • Continuous monitoring of application for 100% availability.

Environment: Java, Spring MVC, MIMA ORM, CVS, AQT, WebSphere, Oracle 10g and HPSM Ticketing tool.

Confidential

Software Engineer

Responsibilities:

  • Handling Complete for Resource Maintenance, Bill Generation Module Designing, Coding and Supporting.
  • Developing and Supporting Customer Management module functionalities.
  • Implemented Code for all modules.
  • Helping the team in solving the technical and functional issues.
  • Involved in discussing with Business Analyst for requirements gathering.

Environment: Core Java, JSP,Hibernate, EJB, Struts, XML, Web Services, Oracle 10G, RAD,soapUI, Oracle SQL Developer, WebLogic Server and UNIX.

We'd love your feedback!