Hadoop Developer Resume
Chicago, IL
SUMMARY:
- Over 8 years of experience spread across Hadoop, Java and ETL that includes extensive experience into Big Data Technologies and in development of standalone and web applications in multi - tiered environment using Java, Hadoop, Hive, HBase, Pig, Sqoop, J2EE Technologies (Spring, Hibernate), Oracle, HTML, Java Script.
- Extensive experience on Big Data Analytics with hands on experience in writing Map Reduce Jobs on Hadoop Ecosystem including Hive and Pig.
- Excellent knowledge on Hadoop Architecture as in HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce programming paradigm.
- Experience with distributed systems, large scale non-relational data stores, mapreduce systems, data modeling and big data systems.
- Involved in developing solutions to analyze large data sets efficiently
- Excellent hands on with importing and exporting data from different Relational Database Systems like MySql and Oracle into HDFS and Hive and vice-versa using Sqoop.
- Hands-on experience in writing Pig Latin scripts working with grunt shells and job scheduling with Oozie.
- Experience in analyzing data using Hive QL, Pig Latin, and custom Map Reduce programs in Java.
- Experience with web-based UI development using jQuery, ExtJS, CSS, HTML, HTML5, XHTML and Javascript.
- Knowledge of job workflow scheduling and monitoring tools like Oozie and Zookeeper.
- Experience with databases like DB2, Oracle 9i, Oracle 10g, MySQL, SQL Server and MS Access.
- Experience in creating complex SQL Queries and SQL tuning, writing PL/SQL blocks like stored procedures, Functions, Cursors, Index, triggers and packages.
- Very good understanding on NOSQL databases like monody and HBase.
- Have good Knowledge in ETL and hands on experience in Informatica ETL.
- Extensive experience in creating Class Diagrams, Activity Diagrams, Sequence Diagrams using Unified Modeling Language(UML)
- Experienced in SDLC, Agile (SCRUM) Methodology, Iterative Waterfall.
- Experience in developing test cases, performing Unit Testing, Integration Testing, experience in QA with test methodologies and skills for manual/automated testing using tools like WinRunner, JUnit.
- Experience with various version control systems Clear Case, CVS, and SVN.
- Expertise in extending Hive and Pig core functionality by writing custom UDFs.
- Development Experience with all aspects of software engineering and the development life cycle
- Strong desire to work for a fast-paced, flexible environment
- Proactive problem solving mentality that thrives in an agile work environment
- Good Experience on SDLC (Software Development Life cycle).
- Exceptional ability to learn new technologies and to deliver outputs in short deadlines.
- Worked with developers, DBAs, and systems support personnel in elevating and automating successful code to production.
- Possess strong Communication skills of written, oral, interpersonal and presentation.
- Ability to perform at a high level, meet deadlines, adaptable to ever changing priorities.
TECHNICAL SKILLS:
Programming Languages: Java, C/C++,SQL, PL/SQL
Big Data Technologies: Hadoop (HDFS & Map Reduce), PIG, HIVE, HBASE, ZOOKEEPER,OOZIE
J2EE Technologies: JSP, SERVLETS, JDBC,EJB
Web Technologies: AJAX, HTML5,JavaScript,CSS3
Frameworks: Spring 3.5 - Spring MVC, Spring ORM, Spring Security, Spring ROO, Hibernate, Struts.
Application Servers: IBM WebSphere, JBoss WebLogic.
Web Servers: Apache Tomcat
Relational Databases: Oracle 10/11g, MS SQL Server, My SQL
NoSQL Databases: Monod, HBase
Designing Tools: UML, Visio, Visual Paradigm
IDEs: Eclipse
Operating System: Unix, Windows
Version Control: CVS, SVN, GIT
PROFESSIONAL EXPERIENCE:
Confidential, Chicago, IL
Hadoop Developer
Responsibilities:
- Analyzed large data sets by running Hive queries and Pig scripts
- Involved in creating Hive tables, and loading and analyzing data using hive queries
- Developed Simple to complex Map Reduce Jobs using Hive and Pig
- Involved in running Hadoop jobs for processing millions of records of text data
- Load and transform large sets of structured, semi structured and unstructured data
- Responsible to manage data coming from different sources
- Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
- Monitor System health and logs and respond accordingly to any warning or failure conditions.
- Implemented the workflows using Apache Oozie framework to automate tasks.
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required
- Developed multiple Map Reduce jobs in java for data cleaning and preprocessing
- Involved in loading data from LINUX file system to HDFS
- Experienced in running Hadoop streaming jobs to process terabytes of xml format data.
- Load and transform large sets of structured, semi structured and unstructured data.
- Assisted in exporting analyzed data to relational databases using Sqoop
- Supported Map Reduce Programs those are running on the cluster
- Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts
Environment: Hadoop, HDFS, Pig, Hive, Map Reduce, Sqoop, Oozie, and Big Data, Apache Java (jdk1.6), Data tax, Flat files, Oracle 11g/10g, MySQL, Toad, Windows NT, LINUX.
Confidential, Media, PA
Hadoop Developer
Responsibilities:
- Handled importing of data from various data sources, performed transformations using Hive, Map Reduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
- Developed Simple to complex Map/reduce Jobs.
- Analyzed the data by performing Hive queries and running Pig scripts to know user behavior.
- Developed Simple to complex Map/reduce Jobs using Hive.
- Created partitioned tables in Hive.
- Worked on Installed and configured Hadoop 0.22.0 Map Reduce, HDFS, developed multiple Map Reduce jobs in java for data cleaning and preprocessing.
- Importing and exporting data into HDFS and HIVE using Sqoop
- Responsible to manage data coming from different sources
- Monitoring the running Map Reduce programs on the cluster.
- Responsible for loading data from UNIX file systems to HDFS.
- Installed and configured Hive and also wrote Hive UDFs.
- Involved in creating Hive Tables, loading with data and writing Hive queries which will invoke and run Map Reduce jobs in the backend.
- Implemented the workflows using Apache Oozie framework to automate tasks.
- Developed scripts and automated data management from end to end and sync up b/w all the clusters.
Environment: Apache Hadoop, Java (jdk1.6), Data tax, Flat files, Oracle 11g/10g, MySql, Toad 9.6, Windows NT, UNIX, Sqoop, Hive, Oozie.
Confidential, Lowell, MA
Hadoop Developer
Responsibilities:
- Developed shell scripts to automate the cluster installation.
- Played a major role in choosing the right configurations for Hadoop.
- Developed Pig Latin scripts to extract and filter relevant data from the web server output files to load into HDFS.
- Involved in start to end process of hadoop cluster installation, configuration and monitoring.
- Responsible for building scalable distributed data solutions using Hadoop.
- Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
- Setup and benchmarked Hadoop/HBase clusters for internal use.
- Developed Simple to complex Map/reduce Jobs using Hive and Pig.
- Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
- Analyzed the data by performing Hive queries and running Pig scripts to study customer behavior.
- Used UDF's to implement business logic in Hadoop.
- Implemented business logic by writing UDFs in Java and used various UDFs from Piggybanks and other sources.
- Continuous monitoring and managing the Hadoop cluster using Cloudera Manager.
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
- Installed Oozieworkflow engine to run multiple Hive and Pig jobs.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for BI team.
Environment: Hadoop, MapReduce, HDFS, Hive, Pig, Java, SQL, Cloudera Manager, Sqoop, Flume, Oozie, Java (jdk 1.6), Eclipse.
Confidential, Birmingham, MI
Hadoop Developer
Responsibilities:
- Installed and configured Hadoop MapReduce, HDFS and developed multiple MapReduce jobs in Java for data cleansing and preprocessing.
- Involved in loading data from UNIX file system to HDFS.
- Installed and configured Hive and also written Hive UDFs.
- Evaluated business requirements and prepared detailed specifications that follow project guidelines required to develop written programs.
- Devised procedures that solve complex business problems with due considerations for hardware/software capacity and limitations, operating times and desired results.
- Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
- Provided quick response to ad hoc internal and external client requests for data and experienced in creating adhoc reports.
- Responsible for building scalable distributed data solutions using Hadoop.
- Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and troubleshooting, manage and review data backups, manage and review Hadoop log files.
- Worked hands on with ETL process.
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, and loaded data into HDFS.
- Extracted the data from Teradata into HDFS using Sqoop.
- Analyzed the data by performing Hive queries and running Pig scripts to know user behavior like shopping enthusiasts, travelers, music lovers etc.
- Exported the patterns analyzed back into Teradata using Sqoop.
- Continuous monitoring and managing the Hadoop cluster through Cloudera Manager.
- Installed Oozie workflow engine to run multiple Hive.
Environment: Hadoop, MapReduce, HDFS, Hive, Ooozie, Java (jdk1.6), Cloudera, NoSQL, Oracle 11g, 10g, PL SQL, SQL*PLUS, Toad 9.6, Windows NT, UNIX Shell Scripting.
Confidential
Java/Technical Support Engineer
Responsibilities:
- Extensively worked in acquiring the requirements from the business analysts and involved in all requirement clarification calls.
- Understanding the design documents.
- Involved in Detail level design and coding activities at offshore.
- Involved in Code review.
- Writing and testing the JUNIT test classes.
- Provide support to client applications in production and other environments.
- Working on tickets raised by the real time users and continuous interaction with end users.
- Prepared the Technical Design Document, understanding document and test cases (UTCs and ITCs).
- Provided Technical & Functional support to the end users during UAT & Production.
- Continuous monitoring of application for 100% availability.
Environment: Java, Spring MVC, MIMA ORM, CVS, AQT, WebSphere, Oracle 10g and HPSM Ticketing tool.
Confidential
Software Engineer
Responsibilities:
- Handling Complete for Resource Maintenance, Bill Generation Module Designing, Coding and Supporting.
- Developing and Supporting Customer Management module functionalities.
- Implemented Code for all modules.
- Helping the team in solving the technical and functional issues.
- Involved in discussing with Business Analyst for requirements gathering.
Environment: Core Java, JSP,Hibernate, EJB, Struts, XML, Web Services, Oracle 10G, RAD,soapUI, Oracle SQL Developer, WebLogic Server and UNIX.