We provide IT Staff Augmentation Services!

Hadoop Developer Resume

5.00/5 (Submit Your Rating)

SUMMARY:

  • Around 7 years of comprehensive experience in Hadoop ecosystem tools and Big Data technologies.
  • Currently working on Apache Kafka, Apache Spark, Apache Storm and Hadoop Map Reduce, Hbase and Phoenix.
  • Developed Storm topologies to read data from kafka topics, populated staging tables and stored the refined data in partitioned hive tables in Google Cloud Storage.
  • Developed application that serves a web application that performs a REST HBASE API calls to read and write data to hbase tables using PHOENIX
  • Worked on ETL tool Syncsort. Responsible for writing a DMX job that reads data from kafka topics and writes it to partitioned hive tables in HDFS, GCS and finally loading it into Teradata.
  • Experience in designing and developing real time big data processing applications using Kafka, Storm, HBase, Hive, Spark
  • Hands on experience in developing end to end solutions in the data pipeline and handling massive amounts of Data.
  • Actively Involved in User training for data analytics tool like Apache Zeppelin
  • Comprehensive experience in analyzing data using HiveQL, Pig Latin and custom MapReduce/Spark programs in Java.
  • Working experience in designing and implementing complete solutions using Hadoop Infrastructure including PIG, HIVE, Sqoop, Oozie
  • Hands on experience in working with various data sources like csv files, xml files, Json files, Oracle/Mysql to load data into Hive tables.
  • Active member of team in writing shell scripts for Unix OS for application deployments to production region.
  • Involved in maintaining hadoop cluster in development and test environment
  • Good knowledge in mining the data in hadoop file system for business insights using Hive, Pig
  • Expertise in Relational Database design, data extraction, data transformation from data sources using MySql and Oracle
  • Good working knowledge on Eclipse IDE for both developing and debugging java applications.
  • Involved in code performance improvement and query tuning activities.
  • Ability to take initiate, share ideas and skills and motivate co - workers from various backgrounds, creative problem solving and drive positive work environment by working collaboratively and in cohesion with rest of the team members.

PROFESSIONAL PROFILE:

Hadoop Developer

Confidential

Responsibilities:

  • Strong experience in Hadoop Administration & Big Data Technologies along with Linux & Database administration.
  • Experience with complete Software Design Lifecycle including design, development, testing and implementation of moderate to advanced complex systems.
  • Hands on experience in installation, configuration, supporting and managing Hadoop Clusters using Apache, Cloudera.
  • Expertise in Hadoop Cluster capacity planning, performance tuning, cluster Monitoring, Troubleshooting.
  • Design Big Data solutions for traditional enterprise businesses.
  • Backup configuration and Recovery from a Name Node failure.
  • Excellent command in creating Backups & Recovery and Disaster recovery procedures and Implementing BACKUP and RECOVERY strategies for off-line and on-line Backups.
  • Involved in bench marking Hadoop/Hbase cluster file systems various batch jobs and workloads
  • Expertise in Hadoop cluster ready for development team working on POCs.
  • Experience in minor and major upgrades of Hadoop and Hadoop eco system.
  • Experience monitoring and troubleshooting issues with Linux memory, CPU, OS, storage and network
  • Hands on experience in analyzing Log files for Hadoop and eco system services and finding root cause.
  • Experience on Commissioning, Decommissioning, Balancing, and Managing Nodes and tuning server for optimal performance of the cluster.
  • Involved in Administration of Cluster maintenance, trouble shooting, Monitoring and followed proper backup& Recovery strategies.
  • Created system security supporting multi-tier software delivery system by utilizing Active Directory and Kerberos.
  • Installed and configured Hadoop eco system like hive, sqoop, pig, and hive.
  • Handsome experience in Linux admin activities on RHEL.
  • Familiar with writing Oozie workflows and Job Controllers for job automation - shell, hive, scoop automation.

Hadoop Developer

Confidential

Responsibilities:

  • Contributed to Design of Application and Database Architecture.
  • Provided design recommendations and thought leadership to sponsors/stakeholders that improved review processes and resolved technical problems.
  • Managed and reviewed Hadoop log files.
  • Tested raw data and executed performance scripts.
  • Shared responsibility for administration of Hadoop, Hive and Pig. Created pig scripts and map reduce programs to filter the log files and aggregate the data.
  • Documented design document and conducted various POC’s to validate the design
  • Worked on Apache Kafka, Apache Spark, Apache Storm and Hadoop Map Reduce, Hbase and Phoenix.
  • Developed Storm topologies to read data from kafka topics, populated staging tables and stored the refined data in partitioned hive tables in Google Cloud Storage.
  • Developed application that serves a web application that performs a REST HBASE API calls to read and write data to hbase tables using PHOENIX.
  • Excellent understanding of Hadoop architecture and its components.
  • Developed a map reduce program to validate the raw data before loading it to database for analysis for the required columns and data format.
  • Loaded data from Linux file system to HDFS using Pentaho Data Integration
  • Created data flows in Pentaho Data Integration for aggregating the data and loading the data to Hive tables
  • Moved the data to Oracle tables from Hive using Sqoop for reporting
  • Created Bar graphs, Heat maps and Geo Maps from aggregated data using Pentaho User Console.
  • Completed Unit Testing and Integration Testing.
  • Documented the user manual and troubleshooting guides
  • Designed & developed several use cases. Created Oozie workflows to streamline the data flow and shell scripts to load the raw data to HDFS
  • Loaded log files data to Hive
  • Used Sqoop to move data between HDFS and MySql
  • Developed a spark code to migrate existing mapreduce code and pig scripts as part of proof of concept
  • Unit tested the application
  • Involved in making very important and major enhancements to the already existing Map Reduce programs and Pig scripts
  • Used SVN i.e. Subversion for version control and maintained different versions based on the release
  • Technically documented every design and development detail of each use case
  • Administered and Maintained the hadoop clusters in Development and Test environments
  • Provided support and maintenance
  • Provided business insights on purchase patterns during promo periods by mining data in Hive

Associate Hadoop Developer

Confidential

Responsibilities:

  • Installed and configured MapReduce, HIVE and the HDFS; implemented CDH3 Hadoop cluster on CentOS. Assisted with performance tuning and monitoring.
  • Created HBase tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
  • Supported code/design analysis, strategy development and project planning.
  • Created reports for the BI team using Sqoop to export data into HDFS and Hive.
  • Developed multiple MapReduce jobs in Java for data cleaning and preprocessing.
  • Assisted with data capacity planning and node forecasting.
  • Collaborated with the infrastructure, network, database, application and BI teams to ensure data quality and availability.
  • Administrator for Pig, Hive and Hbase installing updates, patches and upgrades.

We'd love your feedback!