We provide IT Staff Augmentation Services!

Hadoop Developer Resume

4.00/5 (Submit Your Rating)

Naperville, Il

SUMMARY:

  • 8 Years of working experience in IT Development, Enhancements, Production Support, Administration and Development including 4 years of Big Data Ecosystem related technologies.
  • Java Programmer with 4 years of Extensive programming experience in developing web based applications and Client - Server technologies.
  • Full Understanding of utilizing JEE technology Stack.
  • Experience in different layers of Hadoop Framework - Storage (HDFS), Analysis (Pig and Hive), Engineering (Jobs and Workflows).
  • Experienced in using Integrated Development environments like Eclipse, Sql Developer and Toad.
  • Experience in dealing with Apache Hadoop components like HDFS, MapReduce, Hive, HBase, Pig, Sqoop, Oozie, Big Data and Big Data Analytics.
  • Excellent understanding / knowledge of Hadoop architecture and such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce programming paradigm.
  • Set up standards and processes for Hadoop based application design and implementation.
  • Experience in analyzing data using HiveQL, PIG Latin and custom MapReduce programs in JAVA.
  • Extending HIVE and PIG core functionality by using custom UDF's.
  • Good experience in Scheduling Workflows by using OOZIE and understanding of SQOOP.
  • Experienced in developing MapReduce programs using Apache Hadoop for working with Big Data.
  • Experience in designing and coding web applications using Core Java & web Technologies- JSP, Servlets and JDBC.
  • Extensive experience in Extraction, Transformation and Loading (ETL) of data from multiple sources into Data Warehouse and Data Mart.
  • Building database objects like Tables, Views, Procedures, Functions, Triggers and Packages.
  • Developed stored procedures using cursors, ref cursors, bulk collects and Collections.
  • Good experience on Analytical Function, UTL Files, DBMS Metadata, Sub Query Factoring.
  • Good experience and knowledge on SQL queries for manipulating data.
  • Hands-On experience on using DBMS PROFILER for performance tuning.
  • Good experience on PRAGMA compiler directive like PRAGMA AUTONOMOUS and EXCEPTION INIT.
  • Experience in developing Materialized Views and scheduling various refresh methods.
  • Experience in developing Partition tables for performance improvement.
  • Develop scripts for Data Extraction, Transformation and Loading using SQL, T-SQL and Pl/SQL.
  • Extensive experience in creating complex reports such as: Drill through reports and Cross tab reports, Conditional Formatting, Master Detail report and Cascading Prompts.
  • Expertise in developing Metadata (Packages) using Frame work manager.
  • Hands on knowledge in Dimensional Modeling to construct a warehouse.
  • Through analysis of requirements and building the reports as per client specifications.
  • Knowledge on Cognos 8/10 BI Suite (Report Studio, Query Studio, Analysis Studio).
  • Scheduling jobs using JAMS Scheduler.
  • Developing ER-Diagrams for database using CA Erwin data modeler.
  • Documentation in Confluence.
  • Creating, Tracking and updating tickets in Jira and Remedy.
  • Experience in developing Shell scripts and Power shell Scripts for system management.
  • Comprehensive knowledge on Oracle Architecture. Willing to relocate: Anywhere Authorized to work in the US for any employer

TECHNICAL SKILLS:

Big Data Ecosystems: Hadoop, Hive, Pig, Sqoop, Oozie, Flume.

Programming Languages: Java, C/C++.

Scripting Languages: JavaScript, XML, JSON, HTML and Shell.

Databases: Oracle, My SQl, SQL Server, MS Access.

ETL Tools: Datamanager( IBM), SSIS.

Reporting Tools: Cognos BI(Query Studio, Report Studio).

Data Modeling: Framework Manager, CA Erwin.

IDE Tools: Sql Developer, TOAD, Eclipse, MS Visual Studio, JAMS, Jira, Remedy.

Platforms: Windows, Linux/Unix

NoSql Technologies: Hbase.

WORK EXPERIENCE:

Confidential, Naperville, IL

Hadoop Developer

Roles and Responsibilities:

  • Primary responsibilities include building scalable distributed data solutions using Hadoop ecosystem
  • Installed and configured Hive on the Hadoop cluster (Hortonworks) distribution.
  • Work closely (face-to-face) with the accountants, financial analysts, data analysts, data scientists, statisticians, compliance, sales, marketing, pricing strategists, product development, and business analysts to create solutions for their issues.
  • Developed Simple to complex Map/reduce streaming jobs using Java language that are implemented Using Hive and Pig.
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from Oracle into HDFS using Sqoop
  • Analyzed the data by performing Hive queries (HiveQL) and running Pig scripts (Pig Latin) to study customer behavior.
  • Working as a Hadoop consultant for converting the Oracle Stored Procedures based Data Warehouse Solution to Hadoop based Solution.
  • Filtered, transformed and combined data from multiple providers based on payer filter criteria using custom Pig UDFs.
  • Used the JSON and Avro SerDe's for serialization and de-serialization packaged with Hive to parse the contents of streamed log data and implemented Hive custom UDF's.
  • Implemented business logic by writing UDFs in Java and used various UDFs from Piggybanks.
  • Continuous monitoring and managing the Hadoop cluster using Cloudera Manager
  • Experience in using Sqoop to migrate data to and fro from HDFS and My SQL or Oracle.
  • Exported the analyzed data to the relational databases using HIVE for visualization and to generate reports for the BI team
  • Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Written multiple UDF programs in Java for data extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV and other compressed file formats

Environment: Hadoop 0.20.2 - PIG, Hive, Hortonworks, 30 Node cluster with Linux-Ubuntu

Confidential, Morgan Stanely, NY

Hadoop Consultant

Responsibilities:

  • Understand the exact requirement of report from the Business groups and users.
  • Frequent interactions with Business partners.
  • Good Understanding of the Hadoop Distributed File System and Eco System (MapReduce, Pig, Hive, Sqoop)
  • Worked on to setup Hadoop ecosystem.
  • Managed and reviewed Hadoop Log files.
  • Responsible writing PIG Script and Hive queries for data processing
  • Running Sqoop for importing metadata from Oracle
  • Creation of shell script to collect raw logs from different machines.
  • Created schema for each folder to define schema.
  • Implemented Pig Latin scripts using operators such as LOAD, STORE, DUMP, FILTER, DISTINCT, FOREACH, GENERATE, GROUP, COGROUP, ORDER, LIMIT, AND UNION.
  • Written lots PIG UDF to process some complex data
  • Coded many MapReduce program to process unstructured logs file.
  • Worked on Import and export data into HDFS and Hive using Sqoop
  • Used parameterize pig script and optimized script using illustrate and explain.
  • Involved in the process of configuring HA, Kerberos security issues and name node failure restoration activity time to time as a part of zero downtime.
  • Implemented FAIR Scheduler as well.

Environment: Hadoop, Linux, MapReduce, HDFS, Hive, Pig, Shell Scripting, Sqoop, Java 6, Eclipse, Oracle 10g, Ant, Log4j and Junit.

Confidential

Java Developer

Responsibilities:

  • Involved in Analysis and Design of the Project, which is based on MVC (Model-View-Controller) Architecture and Design Patterns.
  • Involved in developing prototypes of the product.
  • Developed activity, sequence and class diagrams using Unified Modeling Language and Rational Rose.
  • Designed GUI Components using Tiles frame work and Validation frame work.
  • Developed Controller Servlets, Action and Form objects for process of interacting with Oracle database and retrieving dynamic data.
  • Responsible for coding SQL Statements and Stored procedures for back end communication using JDBC.
  • The application was designed on Struts Framework to make use of MVC pattern.
  • Used Apache Log4J logging API to log errors and messages.
  • Developed XML parser for File parsing.
  • Involved in writing Detail Design Documents with UML Specifications.
  • Involved in unit testing and system testing and also responsible for preparing test scripts for the system testing.
  • Developed backend components, DB Scripts for the backend communication.
  • Used Cognos to create reports, which were mailed to be mailed to manager every hour.
  • Responsible for performance tuning of the product and eliminating memory leakages in the product.
  • Conducting Code Review meetings with developers.

Environment: Java, JSP, Servlets, JDBC, LOG4J, Oracle, XML, HTML, Cognos.

Confidential

Java Developer

Roles and Responsibilities:

  • Was a part of the architecture team for design and implementation of site components using J2EE framework.
  • Developed User Interface using JavaScript and HTML.
  • Worked with Java, J2EE, SQL, JDBC, XML, JavaScript, web services.
  • Develops, builds and unit tests components of low level
  • Carries out unit testing on own developed code, developing
  • Test harnesses if necessary.
  • Completes incident management cycle under supervision.
  • Applies all relevant standards and procedures to own work.
  • Develops technical knowledge and awareness of those
  • Technical areas in which requested to code.
  • Developed many sql queries as per business requirements using Joins, sub-queries, hierarchical queries and Analytical functions etc.
  • Developing queries for complex business requirements using SQL.
  • Configuring databases and listeners through Oracle Enterprise Manager (OEM).
  • Implementing partitions to the tables, also implemented daily/weekly/monthly purge process for dropping/ adding partitions.
  • Creating logical and physical database design as per application requirement with E/R models and normalization using CA Erwin Data Modeler.
  • Conducting Code Review meetings with developers.

Environment: Windows, Java, JSP, Servlet, Maven, Log4j, Junit, Oracle, Windows/UNIX, CA Erwin,

We'd love your feedback!