We provide IT Staff Augmentation Services!

Hadoop Developer Resume

3.00/5 (Submit Your Rating)

Chicago, IL

PROFESSIONAL SUMMARY:

  • Over 8+ years experience in Linux System Administration, Big Data Analytics, Hadoop, Java, Data Warehousing and Software development expertise.
  • Strong hands on experience in Hadoop Framework and its ecosystem including but not limited to HDFS Architecture, MapReduce Programming, Hive, Pig, Sqoop, Hbase, Oozie etc.
  • Potential experience in (SDLC) Analysis, Design, Development, Integration and Testing in diversified areas of Client - Server/Enterprise applications using Java, J2EE technologies.
  • Extensively Worked on Data Warehousing ETL using Informatica 9.1/8.6.1/8.5/ 8.1/7.1 PowerCenter Client tools - Mapping Designer, Repository manager, Workflow Manager/Monitor and Server tools - Informatica Server, Repository Server manager.
  • Installed, Configured & Administered CDH & Hortonworks clusters with scalability, fault tolerance, elasticity and high availability
  • Strong database development skills using Database Servers like Oracle, IBM DB2, MySQL and hands on experience with SQL, PL/SQL. Extensive experience of backend database programming in oracle environment using PL/SQL with tools such as TOAD.
  • Excellent knowledge on Hadoop Architecture and ecosystems such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce.
  • Good experience in Big Data Landscape Technologies like Hbase, Cassandra, Mapreduce, Mahout and Hbase.
  • Good Handson Expereince in Apache STORM in processing real-time data.
  • Used Apache STORM along with Hadoop for analyzing data stream in real time view.
  • Experience in strong and analyzing data using HiveQL, Pig Latin, HBase and custom Map Reduce programs in Java.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
  • Hands on experience in installing, configuring CloudEra's ApacheHadoop ecosystem components like Flume-ng, Hbase, Zoo Keeper, Oozie, Hive, Sqoop, Hue, Pig, Hue with CDH3&4 clusters and understanding of SQOOP and Puppet.
  • Extending Hive and Pig core functionality by writing custom UDFs.
  • Architected, Designed and maintained high performing ELT/ETL Processes.
  • Good work experience on JAVA, JDBC, Servlets, JSP
  • Proficient in Java, J2ee, JDBC, Collections, Servlets, JSP, Struts, Spring, Hibernate, JAXB, JSON, XML, XSLT, XSD, JMS, WSDL, WADL, REST, SOAP Web services, CXF, Groovy, Grails, Jersey, Gradle and Eclipse Link
  • Strong experience in ETL tools on Oracle, DB2 and SQL Server Databases.
  • Strong experience in Dimensional Modeling using Star and Snow Flake Schema, Identifying Facts and Dimensions, Physical and logical data modeling using ERwin and ER-Studio.Expertise in working with relational databases such as Oracle 11g/10g/9i/8x, SQL Server 2008/2005, DB2 8.0/7.0, UDB, MS Access and Teradata.
  • Hands on experience working with Java project build managers Apache MAVEN and ANT.
  • Hands on experience in implementing M.V.C Architecture using Struts, Spring, Jersey and Grails Frameworks.
  • Fair amount of experience with scripting in PERL and Python.
  • Good knowledge in integration of various data sources like RDBMS, Spreadsheets, Text files, JSON and XML files.
  • Participated in an Agile SDLC to deliver new cloud platform services and components
  • Developing and Maintenance the Web Applications using the Web server Tomcat.
  • 1 year of experience in MainFrame programming with various technologies like COBOL, JCL, CICS and DB2, SQL.
  • Exceptional ability to learn new technologies and to deliver outputs in short deadlines.
  • Having Experience on UNIX commands, shell scripting and Deployment of Applications in Server.

TECHNICAL SKILLS:

Hadoop: Hadoop 2.1, HDFS, MapReduce, PIG 0.8, Hive0.13, Hbase 0.94, Sqoop 1.4.4, Zoopkeeper 3.4.5, Yarn

Hadoop management & Security: Hortonworks Ambari, Cloudera Manager, Apache Knox, XA Secure

NoSQL Databases: MongoDB, Hbase, Redis and Cassandra

Web Technologies: DHTML, HTML, XHTML, XML, XSL (XSLT, XPATH), XSD, CSS, JavaScript

Server Side Scripting: UNIX Shell Scripting

Database: Oracle 11g/10g/9i/8i, MS SQL Server 2008/2005, DB2 v8.1, Teradata, MySQL.

Programming Languages: Java, J2EE, JSTL, JDBC 3.0/2.1, JSP 1.2/1.1, Java Servlets, JMS, JUNIT,Log4j, COBOL, C

Scripting Languages: Python, Perl, Javascript

Web Servers: Apache Tomcat 5.x, BEA Weblogic 8.x, IBM Websphere 6.0/5.1.1

IDE: WSAD5.0, IRAD 6.0, Eclipse3.5, Dreamweaver13.2.1

OS/Platforms: Windows2008/Vista/2003/XP/2000/NT,Linux(All major distributions, mainly Centos and Ubuntu), Unix.

Client side: JavaScript, CSS, HTML, JQuery

XML: XML, HTML, DTD, XML Schema, XPath

Build tools: Maven and ANT

ETL Tools: Informatica Power Center 9.1/8.6/8.5/8.1/7.1 (Designer, Workflow Manager, Workflow Monitor, Repository manager and Informatica Server)

Data Modeling tools: Erwin, MS Visio

Cloud Platforms: Amazon EC2

Methodologies: Agile, UML, Design Patterns, SDLC

Scripting Languages: R,Maplesoft, Rstudio

Office Tools: MSOFFICE: Excel, Word, PowerPoint.

Mainframe Technologies: COBOL, CICS, JCL

PROFESSIONAL EXPERIENCE:

Confidential, Chicago, IL

Hadoop Developer

Responsibilities:

  • Work with the Teradata analysis team to gather the business requirements.
  • Worked extensively on importing data using scoop and flume.
  • Responsible for creating complex tables using hive and developing Hive queries for the analysts.
  • Created partitioned tables in Hive for best performance and faster querying.
  • Transportation of data to Hbase using pig.
  • Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig.
  • Experience with professional software engineering practices and best practices for the full software development life cycle including coding standards, code reviews, source control management and build processes.
  • Worked collaboratively with all levels of business stakeholders to architect, implement and test Big Data based analytical solution from disparate sources
  • Involved in source system analysis, data analysis, data modeling to ETL
  • Written multiple MapReduce procedures to power data for extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV & other compressed file formats.
  • Handling structured and unstructured data and applying ETL processes.
  • Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS
  • Developed the Pig UDF'S to pre-process the data for analysis.
  • Prepare Developer (Unit) Test cases and execute Developer Testing.
  • Create/Modify shell scripts for scheduling various data cleansing scripts and ETL loading process.
  • Supports and assist QA Engineers in understanding, testing and troubleshooting.
  • Written build scripts using ant and participated in the deployment of one or more production systems
  • Analyzed the web input data in real time with the help of Apache STORM.
  • Production Rollout Support which includes monitoring the solution post go-live and resolving any issues that are discovered by the client and client services teams.
  • Analyzed the ad server data by using both Apache KAFKA and Apache STORM
  • Designed, documented operational problems by following standards and procedures using a software reporting tool JIRA.
  • Used Zookeeper to coordinate the clusters in the Apache STORM
  • Involved in loading and transforming large sets of Structured, Semi-Structured and Unstructured data and analyzed them by running Hive queries and Pig scripts
  • Responsible for tracking the jobs in the Apace STORM by using Nimbus.
  • Assisted in Cluster maintenance, Cluster Monitoring and Troubleshooting, Manage and review data backups and log files
  • Participated in requirement gathering form the Experts and Business Partners and converting the requirements into technical specifications

Environment: Apache Hadoop, Sqoop, Hive, Oozie, Java (jdk1.6), Flat files, Oracle 11g/10g, MySQL, Windows NT, UNIX, Zoo Keeper, Cloudera, FLUME, CentOS, Maven.

Confidential, Pleasanton, CA

Hadoop Consultant

Responsibilities:

  • Analyzed large data sets by running Hive queries and Pig scripts
  • Involved in creating Hive tables, and loading and analyzing data using hive queries
  • Developed Simple to complex MapReduce Jobs using Hive and Pig
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig.
  • Mentored analyst and test team for writing Hive Queries.
  • Develop and maintains complex outbound notification applications that run on custom architectures, using diverse technologies including Core Java, J2EE, SOAP, XML, JMS, JBoss and Web Services.
  • Involved in running Hadoop jobs for processing millions of records of text data
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required
  • Developed multiple MapReduce jobs in java for data cleaning and preprocessing
  • Involved in loading data from LINUX file system to HDFS
  • Responsible for managing data from multiple sources
  • Experienced in running Hadoop streaming jobs to process terabytes of xml format data.
  • Assisted in exporting analyzed data to relational databases using Sqoop
  • Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts
  • Exported data to RDBMS via Sqoop to check whether the power saving program is successful or not.
  • Extensively used Sqoop for importing the data from RDBMS to HDFS.
  • Used ZooKeeper to coordinate the clusters.
  • Responsible for sending messages using Apache STROM i.e. Storm-Signals which aims to provide a way to send messages to spouts/bolts.
  • Handled the imported data to perform transformations, cleaning and filtering using Hive and Map Reduce.
  • Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into the Hadoop Distributed File System (HDFS) and PIG to pre-process the data.
  • Installed and configured MapReduce, HIVE and the HDFS; implemented CDH3 Hadoop cluster on CentOS. Assisted with performance tuning and monitoring.

Environment: Hadoop, Java, JBoss, HDFS, Pig, Hive, MapReduce, Zoo Keeper, CentOS, Maven Sqoop, LINUX. Hadoop Ecosystem, Pig, Hive, Sqoop, Eclipse, Shell Scripting, RDBMS, Cloudera, FLUME, Zoo Keeper.

Confidential, Princeton, NJ

ETL/Informatica Consultant

Responsibilities:

  • Logical and Physical data modeling was done using Erwin for data warehouse database in STAR SCHEMA
  • Using Informatica PowerCenter Designer analyzed the source data to Extract & Transform from various source systems(oracle 10g,DB2,SQL server and flat files) by incorporating business rules using different objects and functions that the tool supports.
  • Using Informatica PowerCenter created mappings and mapplets to transform the data according to the business rules.
  • Used various transformations like Source Qualifier, Joiner, Lookup, sql, router, Filter, Expression and Update Strategy.
  • Implemented slowly changing dimensions (SCD) for some of the Tables as per user requirement.
  • Developed Stored Procedures and used them in Stored Procedure transformation for data processing and have used data migration tools
  • Documented Informatica mappings in Excel spread sheet.
  • Tuned the Informatica mappings for optimal load performance.
  • Have used BTEQ, FEXP, FLOAD, MLOAD Teradata utilities to export and load data to/from Flat files.
  • Created and Configured Workflows and Sessions to transport the data to target warehouse Oracle tables using Informatica Workflow Manager.
  • Have generated reports using OBIEE 10.1.3 for the future business utilities.
  • This role carries primary responsibility for problem determination and resolution for each SAP application system database server and application server.
  • Worked along with UNIX team for writing UNIX shell scripts to customize the server scheduling jobs.
  • Constantly interacted with business users to discuss requirements.
  • Used the PL/SQL procedures for Informatica mappings for truncating the data in target tables at run time.
  • Created the ETL exception reports and validation reports after the data is loaded into the warehouse database.
  • Prepared SQL Queries to validate the data in both source and target databases.
  • Worked on TOAD and Oracle SQL Developer to develop queries and create procedures and packages in Oracle.

Environment: Informatica PowerCenter Designer 8.6/8.1, Informatica Repository Manager, Oracle10g/9i, DB2 6.1, Erwin, TOAD, SAP Version: 3.1.H, Unix- SunOS, PL/SQL, SQL Developer.

Confidential, Kansas City, MO

Sr. Java/J2EE Consultant

Responsibilities:

  • Involved in Design the Project Structure, System Design and Every phase in the project.
  • Responsible for developing Platform related logic and Resource classes, controller classes to access the Domain and Service classes.
  • Involved in the Development of Interfaces and services for the App Builder functionality.
  • Involved in technical discussions, design and workflow.
  • Participate in the requirement gathering and analysis.
  • Used Jaxb to unmarshall XML into Java Objects.
  • Developed Unit test cases using Junit Framework.
  • Implemented the DataAccess Using Hibernate and Wrote the Domain Classes to generate the database tables.
  • Designed and developed the communication tier to exchange data to Xpress Services through JMS and XML over HTTP.
  • Involved in implementation of view pages based on XML Attributes using normal java classes.
  • Involved in Integration of App builder and UI modules with the platform.
  • Used GitHub as a code repository.
  • Used Gradle as a build tool.
  • Implemented JQuery and Ajax for form submissions and design.
  • Developed Maven scripts to build the application.
  • Executed unit test cases using JUnit and Mock Objects.

Environment: Jaxb, Junit Framework, XML, GitHub, Gradle, JQuery, Ajax, Maven, Mock Objects, Junit.

Confidential, Houston, TX

Java/J2EE Developer

Responsibilities:

  • Involved in the process Design, Coding and Testing phases of the software development cycle.
  • Designed use-case, sequence and class diagram (UML).
  • Developed rich web user interfaces using JavaScript (pre-developed library).
  • Created modules in Java and C++, python.
  • Developed JSP pages with Struts framework, Custom tags and JSTL.
  • Developed Servlets, JSP pages, Beans, JavaScript and worked on integration.
  • Developed SOAP/WSDL interface to exchange usage and Image and terrain information from Geomaps.
  • Developed Unit test cases for the classes using JUnit.
  • Developed stored procedures to extract data from Oracle database.
  • Developed and maintained Ant Scripts for the build purposes on testing and production environments.
  • Designed and developed user interface components using AJAX, JQuery, JSON, JSP, JSTL & Custom Tag library.
  • Involved in building and parsing XML documents using SAX parser.
  • Application developed with strict adherence to J2EE best practices.

Environment: Java, C++, Python, Ajax, JavaScript, Struts, Spring, Hibernate, SQL/PLSQL, Web Services, WSDL, Linux, Unix.

Confidential

Main Frame Developer

Responsibilities:

  • Responsible for Start and shutdown process of the operating system.
  • Responsible for Maintenance of system software products (Walker, TGS, Infopac).
  • Responsible for Problem Diagnosis and Resolution for system related anomalies that cause disruption of service to the customer environment.
  • Responsible for Making system modifications for enhancement of functions of the z/OS operating system and other system software.
  • Fileaid/MVS, Quickref, FDRIAM, FDR/ABR/CPK, and Strobe
  • Have Issued TSO, MVS commands and FTP operation.
  • Key role in planning and participation in the Disaster recovery exercise.
  • Performing Quarterly and Yearly Health checks for avoiding Data base (DB2) issues.
  • Quarterly and Yearly Health checks on Data base (DB2).
  • Day-to-day support of the Jobs running under Walker Application.
  • Problem determination and resolution of Job Abend related incidents.
Environment: COBOL, JCL, SQL, DB2, HTML & Web Services.

We'd love your feedback!