We provide IT Staff Augmentation Services!

Hadoop Developer Resume

4.00/5 (Submit Your Rating)

St Louis, MO

SUMMARY

  • Over 8 years' of IT experience in software Development and support with experience in developing strategic methods for deploying big data technologies to efficiently solve Big Data processing requirement.
  • Experience with distributed systems, large - scale non-relational data stores, RDBMS, NoSQL map-reduce systems, data modeling, database performance, and multi-terabyte data warehouses.
  • Working experience in Hadoop framework, Hadoop Distributed file system and Parallel processing implementation.
  • 4 years of hands on experience in Hadoop Framework and its ecosystem including but not limited to HDFS Architecture, MapReduce Programming, Hive, Pig, Sqoop, Hbase, Oozieetc.
  • Working experience with large scale Hadoop environments build and support including design, configuration, installation, performance tuning and monitoring.
  • Experience in application development using the technologies Java, RDBMS, Linux/Unix shell scripting and Linux internals. Excellent knowledge on Hadoop Architecture and ecosystems such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce.
  • Experience in strong and analyzing data using HiveQL, Pig Latin, HBase and custom Map Reduce programs in Java.
  • Experience in writing custom UDFs in java for Hive and Pig.
  • Experience in writing map-reduce programs in java.
  • Experience in installation, configuration and management of development, testing and production Hadoop Cluster.
  • Performed Importing and exporting data into HDFS and Hive using Sqoop.
  • Experience in working with Flume to load the log data from multiple sources directly into HDFS.
  • Experience in designing both time driven and data driven automated workflows using Oozie.
  • Experience in writing UNIX shell scripts.
  • Experience working with JAVA, J2EE, JDBC, ODBC, JSP, Java Eclipse, Java Beans, EJB, Servlets, MSSQL Server.
  • Experience in using IDEs like Eclipse and NetBeans.
  • Experience with Client Side designing and validations using HTML, DHTML and Java Script.
  • Expert in writing SQL queries and database programming using PL/SQL.
  • Participated in an Agile SDLC to deliver new cloud platform services and components
  • Developing and Maintenance the Web Applications using the Web server Tomcat.
  • Exceptional ability to learn new technologies and to deliver outputs in short deadlines.
  • Quick learning skills and effective team spirit with good communication skills.
  • Strong analytical and Problem solving skills.
  • Good Inter personnel skills and ability to work as part of a team.
  • Exceptional ability to learn and master new technologies and to deliver outputs in short deadlines.

TECHNICAL SKILLS

Hadoop: HDFS, Map Reducing, HDFS, Oozie, Hive, Pig, Sqoop, Flume, Zookeeper and Hbase.

Web Tech Spec: DHTML, HTML, XHTML, XML, XSL (XSLT, XPATH), XSD, CSS, JavaScript

Server SideScripting: UNIX Shell Scripting

Database: Oracle 10g, Microsoft SQL Server, MySQL, DB2

Programming Languages: C, C++,Core Java, JSP, Impala, JDBC, SQL, PL/SQL, HQL, PIG Latin and Python

J2EE technologies: Web Tier (JSF, Struts, JQuery, DOJO, JSON, AJAX, HTML, CSS, Icefaces)J2EE Technologies Business Tier (EJB, JMS, Web services JAX-WS) J2EE Technologies Persistence (Hibernate, JPA, Spring JDBC)

Web Servers: WebLogic, Apache, Tomcat, WebSphere, JBoss

IDE: WSAD5.0, IRAD 6.0, Eclipse, Dreamweaver

OS/Platforms: Windows 2008/Vista/2003/XP/2000/NT, Linux, Unix.

Client side: JavaScript, CSS, HTML, JQuery

XML: XML, HTML, DTD, XML Schema

Methodologies: Agile, UML, Design Patterns

NOSQL Databases: Cassandra, MongoDb

PROFESSIONAL EXPERIENCE

Confidential, St Louis, MO

Hadoop Developer

Environment: Apache Hadoop, Map Reduce, HDFS, Hive, Java, SQL, PIG, Zookeeper, Oozie, Cassandra, Java (jdk1.6), Flat files, Oracle 11g/10g, MySQL, Windows NT, UNIX, Sqoop, Hive, Oozie.

Responsibilities:

  • Worked extensively on importing data using Sqoop and flume.
  • Installed and configured Hadoop, Map Reduce, HDFS, Developed multiple Map Reduce jobs in java for data cleaning and preprocessing.
  • Experience in installing, configuring and using Hadoop Ecosystem components.
  • Responsible for creating complex tables using hive.
  • Created partitioned tables in Hive for best performance and faster querying.
  • Transportation of data to Hbase using pig.
  • Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig.
  • Experience with professional software engineering practices and best practices for the full software development life cycle including coding standards, code reviews, source control management and build processes.
  • Worked collaboratively with all levels of business stakeholders to architect, implement and test Big Data based analytical solution from disparate sources
  • Involved in source system analysis, data analysis, data modeling to ETL (Extract, Transform and Load) and HiveQL
  • Written multiple MapReduce procedures to power data for extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV& other compressed file formats.
  • Handling structured and unstructured data and applying ETL processes.
  • Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS
  • Developed the Pig UDF'S to pre-process the data for analysis.
  • Develop Hive queries in Hive QLfor the analysts.
  • Prepare Developer (Unit) Test cases and execute Developer Testing.
  • Create/Modify shell scripts for scheduling various data cleansing scripts and ETL loading process.
  • Supports and assist QA Engineers in understanding, testing and troubleshooting.
  • Written build scripts using ant and participated in the deployment of one or more production systems.
  • Experience in Data Serialization formats for converting Complex objects into sequence bits by using AVRO, JSON, CSV formats.
  • Production Rollout Support which includes monitoring the solution post go-live and resolving any issues that are discovered by the client and client services teams.
  • Designed, documented operational problems by following standards and procedures using a software reporting tool JIRA.
  • Created Cassandra Advanced Data Modeling course for Data Stax.
  • Successfully loaded files to Hive and HDFS from Cassandra.
  • Work in a language agnostic environment with exposure to multiple web platforms such as AWS, databases like Cassandra.
  • Performed Data scrubbing and processing with Oozie.
  • Responsible for managing data coming from different sources.
  • Gained good experience with NOSQL database.
  • Experience inCassandraDatabase Configurations and its C++ client libQtCassandra
  • Experience working with large datasets in Hadoop using Hive.

Confidential, Atlanta, GA

Hadoop Developer

Environment: Java 6, Eclipse, Oracle 10g, Sub Version, Hadoop, Hive, Hbase, Linux,, MapReduce, HDFS,Hive, Java (JDK 1.6), Hadoop Distribution of HortonWorks, Cloudera, MapReduce, Data Stax, IBM DataStage8.1, Oracle 11g / 10g, PL/SQL, SQL*PLUS, Toad 9.6, Windows NT, UNIX Shell Scripting

Responsibilities:

  • Worked on analyzing Hadoop cluster and different big data analytic tools including Pig, Hbase database and Sqoop
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Implemented nine nodes CDH3 Hadoop cluster on Red hat LINUX.
  • Involved in loading data from LINUX file system to HDFS.
  • Analyzed large data sets by running Hive queries and Pig scripts
  • Involved in creating Hive tables, and loading and analyzing data using hive queries
  • Developed Simple to complex MapReduce Jobs using Hive and Pig
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig.
  • Mentored analyst and test team for writing Hive Queries.
  • Develop and maintains complex outbound notification applications that run on custom architectures, using diverse technologies including Core Java, J2EE, SOAP, XML, JMS, JBoss and Web Services.
  • Involved in running Hadoop jobs for processing millions of records of text data
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required
  • Developed multiple MapReduce jobs in java for data cleaning and preprocessing
  • Implemented a script to transmit sysprin information from Oracle to Hbase using Sqoop.
  • Implemented best income logic using Pig scripts and UDFs.
  • Implemented test scripts to support test driven development and continuous integration.
  • Worked on tuning the performance Pig queries.
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
  • Responsible to manage data coming from different sources.
  • Involved in loading data from UNIX file system to HDFS.
  • Load and transform large sets of structured, semi structured and unstructured data
  • Cluster coordination services through Zookeeper.
  • Experience in managing and reviewing Hadoop log files.
  • Job management using Fair scheduler.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and
  • Troubleshooting, manage and review data backups, manage and review Hadoop log files.
  • Installed Oozie workflow engine to run multiple Hive and pig jobs.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Supported in setting up QA environment and updating configurations for implementing scripts with Pig and Sqoop.
  • Crawled some websites usingPythonand collected information about users, questions asked and the answers posted.
  • Hands- on experience in developing web applications using Python on Linux and UNIX platform.
  • Experience inAutomation Testing, Software Development Life Cycle (SDLC)using theWaterfall Modeland good understanding ofAgileMethodology.

Confidential, Richmond, VA

Hadoop Developer

Environment: Hadoop, HDFS, Pig, Sqoop, HBase, Shell Scripting, Ubuntu, Linux Red Hat, Mongo DB

Responsibilities:

  • Exported data from DB2 to HDFS using Sqoop.
  • Developed Map Reduce jobs using Java API.
  • Installed and configured Pig and also wrote Pig Latin scripts.
  • Wrote Map Reduce jobs using Pig Latin.
  • Developed workflow using Oozie for running Map Reduce jobs and Hive Queries.
  • Worked on Cluster coordination services through Zookeeper.
  • Worked on loading log data directly into HDFS using Flume.
  • Involved in loading data from LINUX file system to HDFS.
  • Responsible for managing data from multiple sources.
  • Experienced in running Hadoop streaming jobs to process terabytes of xml format data.
  • Responsible to manage data coming from different sources.
  • Assisted in exporting analyzed data to relational databases using Sqoop.
  • ImplementedJMSfor asynchronous auditing purposes.
  • Involved in developingMessage Driven and Session beansfor claimant information integration with MQ based JMS queues.
  • Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts
  • Worked on installing cluster, commissioning & decommissioning of data node, name node recovery, capacity planning, and slots configuration.
  • Created Hbase tables to store variable data formats of PII data coming from different portfolios.
  • Experience in Architect and build Turn's multi-petabyte scale big data Hadoop infrastructure
  • Experience in Develop monitoring and performance metrics for Hadoop clusters.
  • Experience in Document designs and procedures for building and managing Hadoop clusters.
  • Experience in Partner with Hadoop developers in building best practices for Warehouse and Analytics environment.
  • Strong Experience in troubleshooting the operating system, maintaining the cluster issues and also java related bugs.
  • Experience in working with various kinds of data sources such as MongoDb, Solar and Oracle.
  • Successfully loaded files to Hive and HDFS from Mongo DB Solar.
  • Experience with installing and Configuring on Open LDAP with SSL/TLS on Debian.
  • Experience in configuring Kerberos with Open LDAP at back end.
  • Experience in Automate deployment, management and self-serve troubleshooting applications.
  • Define and evolve existing architecture to scale with growth data volume, users and usage.
  • Investigate emerging technologies relevant to our needs and also Mentoring Sr. Hadoop Administrators on Hadoop cluster management best practices.
  • Design and develop JAVA API (Commerce API) which provides functionality to connect to the Cassandra through Java services.
  • Installed and configured Hive and also written Hive UDFs.
  • Familiarity with a NoSQL database such as MongoDb Solar.
  • Experience in managing the CVS and migrating into Subversion.
  • Experience in managing development time, bug tracking, project releases, development speed, release forecast, scheduling and many more.
  • Using a custom framework of Nodes and MongoDb to take care of the back-end calls with a lightning fast speed.
  • Intensive Object-Oriented JavaScript, jQuery and plug-ins are used to work on dynamic user interface.

Confidential

Java Developer

Environment: Java 1.6, J2EE, Eclipse SDK 3.3.2, Java Spring 3.x, JQuery, Oracle 10i, Hibernate, JPA, Json,Apache Ivy, SQL, stored procedures, Shell Scripting, JQuery, XML, HTML and JUnit, TFS, Ant, Visual Studio, Premium 2010, Rational Clear quest

Responsibilities:

  • Worked with sprint planning, sprint demo, status and daily standup meeting.
  • Developed the application using SpringWeb MVC framework.
  • Worked with Spring Configuration files to add new content to the website.
  • Worked on the Spring DAO module and ORM using Hibernate. Used Hibernate Template and Hibernate Dao Support for Spring-Hibernate Communication.
  • Extensively used Spring's features such as Dependency Injection/Inversion of Control to allow loose coupling between business classes (POJOs)
  • Used Hibernate framework to retrieve and update information and dependency Injection is achieved by Spring MVC Framework.
  • Configured Association Mappings such as one-one and one-many in Hibernate
  • Worked with JavaScript calls as the Search is triggered through JS calls when a Search key is entered in the Search window
  • Worked on analyzing other Search engines to make use of best practices.
  • Collaborated with the Business team to fix defects.
  • Worked on XML, XSL and XHTML files.
  • Worked on Multi-Threaded environment.
  • Interacted with project management to understand, learn and to perform analysis of the Search Techniques.
  • Used Ivy for dependency management.
  • As part of the team to develop and maintain an advanced search engine, would be able to attain expertise on a variety of new software technologies.

Confidential

Java Developer

Environment: Java 1.6, J2EE, Eclipse SDK 3.3.2, Java Spring 3.x, JQuery, Oracle 10i, Hibernate, JPA, Json, Apache Ivy, SQL, stored procedures, Shell Scripting, JQuery, XML, HTML and JUnit, TFS, Ant, VisualStudio, Premium 2010, Rational Clear quest, J2EE, JDBC, Java 1.4, Servlets, JSP, Struts, Hibernate, Web services, SOAP, WSDL, Design Patterns, MVC, HTML, JavaScript 1.2, WebLogic 8.0, XML, JUnit, Oracle 10g, My Eclipse.

Responsibilities:

  • Played an active role in the team by interacting with welfare business analyst/program specialists and converted business requirements into system requirements.
  • Developed analysis level documentation such as Use Case, Business Domain Model, Activity & Sequence and Class Diagrams.
  • Conducted Design reviews and Technical reviews with other project stakeholders.
  • Implemented Services using Core Java.
  • Developed and deployed UI layer logics of sites using JSP.
  • Struts (MVC) is used for implementation of business model logic.
  • Worked with Struts MVC objects like Action Servlets, Controllers, and Validators, Web Application Context, Handler Mapping, Message Resource Bundles and JNDI for look-up for J2EE components.
  • Developed dynamic JSP pages with Struts.
  • Used built-in/custom Interceptors and Validators of Struts.
  • Developed the XML data object to generate the PDF documents and other reports.
  • Used Hibernate, DAO, and JDBC for data retrieval and medications from database.
  • Messaging and interaction of Web Services is done using SOAP.
  • Developed JUnit Test cases for Unit Test cases and as well as System and User test scenarios
  • Involved in Unit Testing, User Acceptance Testing and Bug Fixing.
  • Implemented mid-tier business services to integrate UI requests to DAO layer commands.

We'd love your feedback!