Hadoop Developer Resume
Louisville, KY
PROFESSIONAL SUMMARY:
- 10 years of Software Development and System Engineering experience, with a demonstrated ability to quickly learn and integrate new technologies in Retail, Telecom and supply chain domain using Java/J2EE technologies.
- 3 Years of experience in Big data using Hadoop, Hive, Pig, Sqoop, Hbase and MapReduce Programing.
- Strong knowledge in using Mapreduce programming model for analyzing the data stored in Hadoop.
- Hands - on experience on major components in Hadoop Ecosystem including Mapreduce, Hive, Sqoop and knowledge of Mapper/Reduce/HDFS Frame work.
- Extending HIVE core functionality by using custom UDF's, UDAF and UDTF.
- Extensive experience in Data migration from existing data stores to Hadoop.
- Experience in JAVA/J2ee, Struts, JSF, Spring, Hibernate and web services.
- Experience in working with web/applications servers JBOSS, WebLogic, Apache Tomcat and glass fish.
- Experience in developing front-end using jQuery, JavaScript, JSF, HTML, XHTML and CSS.
- Experience in analyzing data using HIVEQL and custom MapReduce programs in JAVA.
- Experience in SQL programming including complex SQL queries and dynamic queries.
- Experience in UNIX shells scripting.
- Supported ETL processes for large-scale data warehousing.
- Experience in designing highly transactional web sites using J2EE technologies and handling design/implementation using NetBeans/Eclipse/MyEclipse.
- Used MyEclipse for supporting a web site application using knowledge on Core Java.
- Supported QA team during the QA phase.
- Experience of end-to-end implementation of the SDLC; including finalization of specifications, design, development, implementation and quality.
- Experience in Agile, SDLC and waterfall methodologies
- Experience in DAO framework.
- Significant experience working with customers, Project Managers and technical teams for securing and executing concurrent multi-technology projects.
TECHNICAL SKILLS:
Big Data: Hadoop, HDFS, Map reduce, Hive, Pig, Sqoop, oozie and Hbase
Languages: Java, Python, Scala, SQL, Shell Scripting
J2EE Technologies: JSP, Servlets, JDBC
Servers: Jboss, Web Logic and Tomcat
Frameworks: Struts, Spring, TSF, Hibernate, Hadoop.
Java IDEs: NetBeans, RAD, Eclipse.
Version Control / Tracking Tools: SVN, CVS, ClearCase
Databases: Oracle, SQL (DDL, DML, DCL) and PL/SQL.
Design Skills: J2EE design patterns, Object Oriented Analysis and Design (OOAD)
Operating Systems: Windows7, Windows XP, 2000, 2003, Unix and Linux
PROFESSIONAL EXPERIENCE:
Confidential, Louisville, KY
Hadoop Developer
Responsibilities:
- Developed complex Map Reduce programs in Java for Data Analysis on different data formats.
- Developed Map Reduce programs that filter bad and un-necessary records and find out unique records based on different criteria.
- Developed Secondary sorting implementation to get sorted values at reduce side to improve map reduce performance.
- Implemented custom Data Types, Input Format, Record Reader, Output Format, Record Writer for Map Reduce computations to handle custom business requirements.
- Developed Shell scripts and schedule jobs on Cron to import data files from various sources to HDFS to archive.
- Develops and tests for extraction, transformation, and load processes.
- Develop Oozie workflows by integrating all tasks relating to a project and schedule the jobs as per requirements.
- Implemented Daily Oozi jobs that automate parallel tasks of loading the data into HDFS and pre-processing with Pig using Oozie co-coordinator jobs.
- Importing and exporting data into HDFS and Hive using Sqoop.
- Perform data analysis using Hive and Pig.
- Monitoring scheduled jobs on Oozie web console.
- Planning and conducting code reviews for changes and enhancements that ensure standards compliance.
- Experience maintaining code versions on SVN and migrate code from dev to test and prod.
- Experience gathering requirements and data producers from non-technical end users.
- Translate business user requirements into technical data documentation such as data models, process flows and other required documentation.
- Develop data mapping and data loading processes.
- Preparing and packaging production-ready code modules for staging.
- Worked on Sequence files, RC files, Map side joins, bucketing, partitioning for Hive performance enhancement and storage improvement.
- Worked on the proof-of-concept for Apache Spark framework initiation.
Environment: HDFS, MapReduce, Hive, Pig, Sqoop, oozie, Zookeeper, Java, Cloudera Manager Ubuntu, UNIX Shell Scripting.
Confidential, Minneapolis, MN
Hadoop Consultant
Responsibilities:
- Involved in gathering the business requirements and in turn converted the same into IT standard requirements.
- Analyzed the requirement to setup a 25 node cluster.
- Ensured NFS is configured for Name Node.
- Created two different users (hduser for performing hdfs operations and mapred user for performing map reduce operations only).
- Worked on setting password less Hadoop.
- Worked on setting up CRON job to delete Hadoop logs/local old job files/cluster temp files.
- Moved all crawl data flat files generated from various retailers to HDFS for further processing.
- Handled 7TB of data as a part of development.
- Written CLI commands using HDFS.
- Written the script files for processing data and loading to HDFS
- Written the Apache PIG scripts to process the HDFS data.
- Configured Hive with MySQL as a Remote Metastore
- Created Hive tables to store the processed results in a tabular format.
- Created External Hive Table on top of parsed data.
- Developed the SQOOP scripts in order to make the interaction between Pig and MySQL Database.
- Moved all log/text files generated by various products into HDFS location
- Written Map Reduce code that will take input as log files and parse the logs and structure them in tabular format to facilitate effective querying on the log data
- Preparation of status reports.
Environment: HDFS, MapReduce, Hive, Pig, Sqoop, oozie, Java, Cloudera Manager Ubuntu, UNIX Shell Scripting, Python.
Confidential, Denver, CO
Java/ Hadoop Developer
Responsibilities:
- Archiving seven years claims data, which requires complex processing to get into a normalized format.
- Logging terabytes of data generated from transactional systems daily, and storing them in CDH for analytical purposes.
- Importing and exporting data into HDFS from Relational Database systems using Sqoop.
- Worked with Hive to bulk load Data into HBase.
- Worked with creating tables and Analyzing data using Hive queries.
- Scheduled Recurrent Jobs, Packaged Multiple Jobs, Sequenced jobs in Oozie.
- Experienced in running Hadoop streaming jobs to process terabytes of xml format.
- Load and transform large sets of structured, semi structured and unstructured data.
- Experienced in writing MapReduce java programs to process huge data.
- Experience in Data Integration using Microsoft BI tools.
- Hands on Experience with installing, configuring, and using Hadoop ecosystem components like Hadoop Map Reduce, HDFS, HBase, Hive, Sqoop, Pig.
- Experience in writing java programs to interact with Hbase and Hive
Environment: Hadoop, MapReduce, HDFS, Hive, Java, jdbc, Pig, Cloudera, Ubuntu, UNIX Shell Scripting.
Confidential
Technical Lead
Responsibilities:
- Analyze existing application and finding out key areas that required changes to incorporate new requirements.
- Design new workflow based on the new requirements.
- Finalized requirements with regular meetings and discussions with users.
- Optimized many existing functionalities for better maintainability and quality of code.
- Provided mentoring and Technical training to new team members.
- Provided unit test cases for the application using Junit3.0
- Getting clarifications from customer,
- Supporting our team members technically, functionally to finish their work in time,
- Supporting deployments in SIT, UAT and PROD environments.
- Used JSP, JavaScript, JSP Tag Libraries and AJAX YUI data table for web presentation layer.
- Developed application on MVC architecture using Struts 1.2 Framework, Action Classes and performed client side validations using JavaScript.
- Implemented Tiles framework to develop JSP templates with headers and footer elements, and varying body contents.
- Build project using Apache Maven build scripts, and the applications Version controlling was accomplished using SVN.
Environment: Java, Servelts, JSP, Spring, Hibernate, UNIX Shell Scripting, Tomcat, CVS, Toad, Eclipse, SQL, Oracle.
Confidential
Module Lead
Responsibilities:
- Involved in the development of Functional Requirements Specifications, Technical Specifications, detailed design documents, user guides, test procedures, and test cases for the application components
- Followed Object Oriented Design and Analysis by preparing Use Cases, business domain model, Sequence Diagrams and Class Diagrams and designed the UML components for technical specification in Microsoft Visio.
- Implemented the Software Development Guidelines based on Agile Methodologies.
- Developed front end of application on MVC architecture employing Struts Framework.
- Responsible for setting up configuration files- web.xml, struts- config.xml, tiles-defs.xml, and validation.xml; developed UI layer using JSP, Struts Tag Libraries, JavaScript, AJAX, HTML/DHTML, CSS
- Developed Action classes, Action Forms performed form validations using Struts Validation Frameworks, and used Tiles Frame work.
- Used Hibernate in DAO layer to access and update information in Oracle database, developed Hibernate configuration files (hbm.xml) for object relational mapping with database, fine-tuned performance by optimizing query and data caching mechanisms.
- Developed HQL queries to retrieve data using Hibernate
- Developed SQL queries and Procedures using SQL and PL/SQL.
- Involved in Test data creation, Unit testing using JUNIT
- Used Subversion for code check-ins
- Used CSS for style sheets.
Environment: Java, J2EE, SQL Server, J2EE Web Services, Apache Tomcat Server, XML, Eclipse, Rational Rose, Ajax, HTML, JavaScript and WebLogic Application Server, subversion.
Confidential
Senior Java Developer
Responsibilities:
- Understanding business requirements and creating Detail Design Document.
- Extensively used Scrum Programming concept which is part of agile methodologies to plan work for every iteration and used continuous integration tool to make the build passes before deploying the code to other environments.
- Developed the application using Struts Framework that leverages the classical Model-View-Controller (MVC) architecture.
- Used Hibernate Mapping Files to connect to database and spring framework as IOC container.
- Involved in creating all Java Server Pages, JavaBeans and invoking EJB Components through JNDI.
- Implemented Design Patterns like Business Delegates, Session Facades, Data Access Objects, and Value Objects etc.
- Created many stored procedures, PL/SQL transactions to interact with oracle database and they in turn are called using EJB.
- Used Hibernate Mapping Files to connect to database
- Monitored the error logs using Log4J and fixed the problems.
- Involved in designing the user interfaces using JSPs.
- Used JDBC to invoke Stored Procedures and database connectivity to Oracle.
- Developed JUnit test cases for all the developed modules.
- Interact with testing team for IST and UAT.
- Used CVS for version control across common source code used by developers.
Environment: Java, JSP, Struts, Spring, EJB, JDBC, HTML, XML, Hibernate, JNDI, Log4j, CVS, ANT, Web logic, Eclipse, Oracle.
Confidential
Java Developer
Responsibilities:
- Designed and developed an interactive module in Java
- Implemented Presentation layer using JSP, Servlets
- Developed the application using Struts Framework that leverages the classical MVC architecture
- Worked on query handling, customer support, helpdesk
- Migrated a poorly performing and outdated application to Struts, Hibernate based system for Sprint.
- Maintained the interface of Oracle using JDBC
- Wrote procedures and queries to extract data from database
- Tested the flow of modules using JUNIT
- Monitored the error logs using Log4J and fixed the problems.
- Handled the JDBC backend operations of the respective modules also.
Environment: java, Struts, Hibernate, JSP, HTML, CSS, JDBC, JavaScript, xml, CVS, Log4J, JNDI, JUnit, Tomcat, My Eclipse.