We provide IT Staff Augmentation Services!

Senior Hadoop Developer Resume

5.00/5 (Submit Your Rating)

New, JerseY

SUMMARY

  • 10+ years of IT experience with strong strategic planning, design and development skills in various technologies including 3+ years of Big Data Ecosystem in Healthcare and Telecom sectors. And 6 plus years of experience in Oracle Technologies and Oracle E - Business Suit.
  • Expertise in the various Hadoop Distributions such as Hortonworks, Cloudera and Oracle Big Data Appliance.
  • Excellent understanding onHadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, Name and Data Node, Secondary Name Node and Map Reduce programming paradigm.
  • Expertise in Hadoop Ecosystem components HDFS, Map Reduce, Hive, Pig, Sqoop, NoSQL, Hbase, Flume and Kafka for Data Analytics.
  • Performed Extract, Transform and Load (ETL) using Hive on huge volume of Verizon Wireless data.
  • Experience in analyzing data using HiveQL, Pig Latin and custom MapReduce programs in Java and Python. Extending Hive and Pig core functionality by writing custom UDFs.
  • Extensive knowledge of creating manage tables and external tables in Hive Eco system
  • Worked extensively on Pig Latin and Hive Queries to generate daily/weekly/monthly and ad hoc reports on huge data sets. Integrated Tableau with Hive tables to visualize reports.
  • Experienced in writing MapReduce programs to process large sets of structured, semi-structured and unstructured data sets and store them in HDFS using Pig, Java and Python.
  • Expertise in writing MapReduce programs using JAVA
  • Worked closely with data scientists to find solutions to business problems using predictive analytics, pattern matching and data classifications.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
  • Experience in Spark, Yarn
  • Familiarize with Tableau reporting
  • Work experience with cloud infrastructure like Amazon Web Services (AWS)
  • Experience in installation, configuration, supporting and managing Hadoop Clusters using Apache, Cloudera (CDH3, CDH4) distributions and on Amazon web services (AWS)
  • Hands on experience in writing Spark SQL scripting.
  • Skilled on migrating the data from different databases to Hadoop HDFS and Hive using Sqoop.
  • Experience in job workflow scheduling and monitoring tools like Oozie.
  • Proficient in designing and querying the NoSQL databases like HBase.
  • Worked extensively in design and development of business process using SQOOP, PIG, HIVE, HBASE
  • Knowledge on Scala Programming Language.
  • Good knowledge in Software Development Life Cycle (SDLC) and Software Testing Life Cycle.
  • Extensive experience in developing stored procedures, functions, Views and Triggers, Complex queries using SQL Server, Oracle PL/SQL and SQL*Loader.
  • Excellent communication and inter-personal skills detail oriented, analytical, time bound, responsible team player and ability to coordinate in a team environment and possesses high degree of self-motivation and a quick learner.

TECHNICAL SKILLS

Big Data Technologies: Hadoop 2.X (YARN), HDFS, Map Reduce, Hive, Pig, Hbase, NoSQL, Sqoop, Oozie, Flume, Kafka, Apache Spark, Spark SQL, Scala

Databases: Oracle 9i/10g, Teradata

Coding Language: SQL, PL/SQL, C, Shell Script, Core Java, Python

Development Tools: Forms6i, Reports6i, XML Publisher, Oracle Alerts

Tools: Toad, SQL * Loader, SQL Developer, Filezilla, SVN, PVCS, HPPM

Methodologies: Agile, Waterfall

Operating Systems: Windows, Linux (Cent OS, Ubuntu)

ERP: Oracle Applications 11i & R12 (PO, AP, AR, GL, INV, AOL)

PROFESSIONAL EXPERIENCE

Confidential, New Jersey

Senior Hadoop Developer

Primary areas of responsibilities:

  • Worked closely with data scientists for data gathering enterprise level to predict consumer behavior, such as Call In Rate, Battery Drain, Call Drop, SMS Message Drop and made recommendations based on recognizing patterns.
  • Involved in end to end data processing like ingestion, processing, quality checks and splitting
  • Bringing the data into Big Data Lake using Pig, Sqoop and Hive
  • Handled the spark streaming of data into the HDFS using Kafka
  • Transformation of data using Scala in Spark
  • Wrote Pig scripts to process unstructured data and create structure data for use with Hive
  • Handling of RDDs
  • Creation of Hive tables using the Spark
  • Developed and Executed the queries using HiveQL and SparkQL
  • Importing Large Data Sets from Oracle, Teradata to Hive Table using Sqoop.
  • Created Hive Managed and External Tables as per the requirements
  • Designing and developing tables in HBase and storing aggregating data from Hive
  • Developing Hive Scripts for data aggregating and processing as per the Use Case.
  • Writing Java, Python Custom UDF's for processing data in Hive and Pig.
  • Developing and maintaining Workflow Scheduling Jobs in Oozie for importing data from RDBMS to Hive.
  • The Hive tables created as per requirement were managed or external tables defined with appropriate static and dynamic partitions, intended for efficiency.
  • Implemented Partitioning, Bucketing in Hive for better organization of the data
  • Optimized Hive queries for performance tuning.
  • Developed data pipeline using Flume, Sqoop, Pig and Java map reduce and Spark to ingest customer behavioral data and purchase histories into HDFS for analysis.
  • Involved with the team of fetching live stream data from Oracle to Hive, Hbase table using Apache Kafka.

Environment: Hadoop v2.6.0, HDFS, Map Reduce, HBase, Sqoop, Core Java, Hive,Pig, Oozie,Oracle, Spark,Teradata,Linux

Hadoop Developer

Confidential, East Hanover, NJ

Responsibilities:

  • Installed and configured Hadoop MapReduce, HDFS, developed multiple MapReduce jobs in Java for data cleaning and preprocessing.
  • Importing and exporting data into HDFS from Oracle 10.2 database and vice versa using SQOOP.
  • Experienced in defining and coordination of job flows.
  • Work experience with cloud infrastructure like Amazon Web Services (AWS)
  • Experience in installation, configuration, supporting and managing Hadoop Clusters using Apache, Cloudera (CDH3, CDH4) distributions and on Amazon web services (AWS)
  • Gained experience in reviewing and managing Hadoop log files.
  • Extracted files from NoSQL database like CouchDB, HBase through sqoop and placed in HDFS for processing.
  • Involved in Writing Data Refinement Pig Scripts and Hive Queries
  • Good knowledge in running Hadoop streaming jobs to process terabytes of xml format data.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Used Flume to transport logs to HDFS
  • Configured connection between HDFS and Tableau using Impala for Tableau developer team.
  • Responsible to manage data coming from different sources.
  • Got good experience with various NoSQL databases.
  • Experienced with handling administration activations using Cloudera manager.
  • Supported MapReduce programs those are running on the cluster.
  • Involved in loading data from UNIX file system to HDFS.
  • Installed and configured Hive and also written Hive UDFs.
  • Involved in creating Hive tables, loading with data and writing Hive queries which will run internally in map reduce way.
  • Integrated Oozie with the rest of the Hadoop stack supporting several types of Hadoop jobs out of the box (like, Pig, Hive, and Sqoop) as well as system specific jobs (such as shell script).
  • Automated all the jobs, for pulling netflow data from relational databases to load data into Hive tables, using Oozie workflows and enabled email alerts on any failure cases.

Environment: Cloudera HadoopCDH4 (Classic Framework), Java, Linux, HBase, Hive, Pig, Sqoop, Flume, Oozie, NoSQL, Map-Reduce, Cloudera, HDFS, Impala, Tableau, My

SQLHadoop Developer/Oracle Technologies

Confidential

Responsibilities:

  • Working with business users and business analyst for requirements gathering and business analysis.
  • Converted business requirement into high level and low level design.
  • Worked on the proof-of-concept for Apache Hadoop 1.20.2 framework initiation.
  • Implemented Proof of concepts on Hadoop stack and different bigdata analytic tools, migration from different databases(Oracle, MySQL) to Hadoop
  • Installed and configured Hadoop clusters and eco-system.
  • Monitored Hadoop cluster job performance and capacity planning
  • Hands-on experience with Hadoop technology stack (HDFS, MapReduce, Hive,Oozie)
  • Automated all the jobs starting from pulling the Data from Oracle to pushing the result set data to Hadoop Distributed File System using Sqoop.
  • Distributing the work between offshore team member and tracking the development progress.
  • Discussing the technical way of development with onsite and offshore team lead.
  • Developed JavaScript behavior code for user interaction.
  • Used HTML, JavaScript, and JSP and developed UI
  • Used JDBC and managed connectivity, for inserting/querying& data management including stored procedures and triggers.
  • Involved in the design and coding of the data capture templates, presentation and component templates.
  • Developed an API to write XML documents from database.
  • Used JavaScript and designed user-interface and checking validations.
  • Part of a team which is responsible for metadata maintenance and synchronization of data from database.
  • Implemented the project according to the Software Development Life Cycle (SDLC).
  • Implemented JDBC for mapping an object-oriented domain model to a traditional relational database.
  • Created Stored Procedures to manipulate the database and to apply the business logic according to the user’s specifications.
  • Developed the Generic Classes, which includes the frequently used functionality, so that it can be reusable.
  • Exception Management mechanism using Exception Handling Application Blocks to handle the exceptions.
  • Designed and developed user interfaces using JSP, Java script and HTML.
  • Involved in Database design and developing SQL Queries, stored procedures on MySQL.
  • Used CVS for maintaining the Source Code.
  • Logging was done through log4j.
  • Involved in testing the modified reports and interface programs.
  • Developed Pig Latin scripts for the log data and stored the cleansed data in the Apache Hadoop.

Environment: Scoop, Hive, Oozie, Unix, Oracle9i, Oracle Applications11i & R12, JAVA, Java Script, HTML, XML, JDBC Drivers, Soap Web Services, Unix, Shell scripting

Confidential

Oracle Apps Technical Consultant

Responsibilities:

  • Gathering connectivity details from respective SME owners.
  • Converting the script from FTP to SFTP Connection.
  • Involved in testing the modified scripts.
  • Interacting with SME owners for their

Environment: Oracle Applications 11i, Unix, PVCS, HPPM

Confidential

Oracle Apps Technical Consultant

Responsibilities:

  • Review and Analysis of Functional Specifications
  • Interaction with Functional Specification owner on various functional issues during development
  • Development of Technical Specifications
  • Developed Bank Statements Load Interface program
  • Developed e-text template for payment process request program as per client required format.
  • Created Notifications using Oracle Alerts as per client requirement.
  • Developed XML reports as per client requirement.
  • Registration of Reports, Concurrent Programs.
  • Prepared MD120 and UTP for developed components.
  • Developed Validation scripts in PL/SQL for importing the Invoices from the legacy system into Oracle Payables module.
  • Developed GL Daily conversion rates program as per client requirements.
  • Developed and Customization of Reports and Forms.
  • Creation of Value Sets.
  • Prepared test cases for developed components.
  • Validation scripts in PL/SQL for importing the Item Category Assignments from the legacy system into Oracle Inventory module.
  • Developed Item Outbound Interface as per client requirements.
  • Validation scripts in PL/SQL for importing the Invoices from the legacy system into AP module
  • Registration of concurrent programs for developed reports and Interfaces.
  • Creation of request sets for developed interface.
  • Building Technical Design documents MD 070 based on MD 050.
  • Developed and Customized XML reports as per client requirement.
  • Creation of Value Sets and Assigning to concurrent programs
  • Developed interface to load data from legacy systems to oracle 11i system
  • Registration of concurrent programs for developed reports and Interfaces
  • Development of 6i and 10g reports.
  • Creation of Value Sets and Assigning to concurrent programs
  • Prepared Construction Work book and Unit Test plan.

Environment: Forms6i, Reports6i, XML Publisher, Oracle Alerts, Oracle Applications 11i & R12

Confidential

Oracle Apps Technical Consultant

Responsibilities:

  • Developed interface to load Purchase Orders from legacy systems to oracle apps systems.
  • Registration of concurrent programs for developed reports and Interfaces.
  • Developing Code from Technical Design.
  • Creation of Value Sets and Assigning to concurrent programs.
  • Involved in Support, Enhancements / Bug fixing.
  • Involved in development of GUI Screens using Forms 6i
  • Involved in development of Reports using Reports 6i.
  • Developed Procedures, Functions and Packages.
  • Modifications are undertaken and done as a part of change management.
  • Involved in testing.

Environment: Forms6i, Reports6i, Oracle9i, Oracle Applications 11i

We'd love your feedback!