We provide IT Staff Augmentation Services!

Hadoop Developer Resume

3.00/5 (Submit Your Rating)

Kansas City, MO

SUMMARY:

  • Over 5 years of IT experience in Application Development in Java, C, C++ and Big Data Hadoop.
  • Qualified as Industry - Recognized Hortonworks Data Platform Certified Developer (HDPCD) on Apache Hadoop Frameworks by performing actual hands-on big data tasks on a Hortonworks Data Platform (HDP) cluster.
  • Possesses strong knowledge & understanding of Apache Hadoop Frameworks and Hortonworks Data Platform 2.4 installed and managed with Ambari 2.2, which includes Pig 0.15.0, Hive 1.2.1 , Sqoop 1.4.6 , and Flume 1.5.2 , HDFS .
  • Attended Centriq Workshop on “What is Big Data Really?”
  • Hand on experience in IDE : Microsoft Visual Studio , Oracle ADE, Eclipse, and NetBeans.
  • Coded in Programming Language: C , C++ , Java , Data Structure, DBMS and Scripting Language: PERL , SHELL
  • Strong collaboration and a keen communicator with excellent interpersonal skills. Result oriented hard worker with initiative & energy. Team building skills with proficiency at grasping new concepts & utilize them in a productive manner.
  • Deep analytical and technical aptitude, client services skills, data analysis expertise and the ability to quickly learn/apply technologies and solve business issues in a time sensitive environment.
  • Successfully completed Nalanda Training, Gurgaon-India for duration of 3 months. Done a mini project on IP Fragmentation and Reassembly ( SDLC ). Attended ACF Training to improvise Behavioral competencies (Communication, Collaboration and Learning & Innovation) and Technical competencies (Design, Coding, Review and Quality).

SKILLS:

Big Data Frameworks: Apache Hadoop - HDFS, MapReduce, Yarn, Hive, Pig, Hive, Sqoop, Flume, Spark-Scala

Operating System: Windows, Linux, Android

Programming Language: Struct C, C++, Java, Data Structure

Scripting Language: PERL, SHELL

Web Technologies: HTML, CSS, JSP, XML, PHP, JSCRIPT

IDE: Microsoft Visual Studio, ADE, ECLIPSE, NETBEANS

Debugging Tool: Valgrind, splint, gdb, gcov, gprof

Database / Filesystem: MySQL, ORACLE ACFS, HDFS

Area of Interest: Business Intelligence Big Data, DBMS, Web Design

WORK EXPERIENCE:

Confidential

Hadoop Developer

Kansas City, MO

RESPONSIBILITIES :

  • Key team member involved in business analysis and requirements gathering, development, implementation, testing, and support including contributing to business requirements capture across cross-functional teams.
  • Developed simple, seamless, efficient and traceable Recruiters Office Portal to provide services for professional IT recruiters wanting to manage and track their activities.
  • Offers service plans for business users and aimed at reliable general consumers usage.
  • In-depth knowledge/experience in using BigData Apache Hadoop Frameworks and actively interacted with all teams and helped to achieve all project milestones in time and with best quality solutions.
  • Moved huge amount of big data from traditional RDBMS to Hadoop Distributed File System (HDFS)using Apache Hadoop Sqoop.
  • Transferred event logs into HDFS using Apache Flume.
  • Big Data are cleaned, filtered, identified and removed duplicates.
  • Carried out certain big data aggregation and transformed data into appropriate formats using Apache Pig.
  • Pre-processed big data which are ready to use are further analyzed using Apache Hive Query, an RDBMS service used in Hadoop.
  • Big Data are moved IN/OUT of Hadoop Framework using Apache Sqoop import and export tools based on need.
  • Good knowledge in Apache Spark - Scala, Python, Java.
  • Involved in team work, debug and monitored the code changes for each version of code development to catch the bugs at the early stage.
  • Carried out runtime big data analysis to improve the performance and deliver best quality services.
  • Got appreciation from Team for proven competency.

Environment : Hortonworks Data Platform HDP, Big Data Hadoop, Apache HDFS, MapReduce, Yarn, Sqoop, Flume, Pig, Hive, Spark, Web Interface Ambari and UNIX Shell Script. Linux and Windows.

Confidential

Java Developer / Software Engineer

RESPONSIBILITIES:

  • Work closely with business technology analysts to understand the design and business requirements from team. Wide experienced in estimating time for each requirement, design object models and class diagrams, create reusable components.
  • Expertise on setting up the system for various software and tools needed for development. Setup standalone and portable Oracle cluster-ware PCW. Start ASM instance to use ASM disks. Create volume for mounting Filesystem.
  • Extensive experience in developing java program adhere to coding standards specified by technical management. Proficient in using Object Oriented Analysis and applying proven design patterns in building high performance applications.
  • Worked in ACFS Functionalities like ACFS Compression and Replication. ACFS Compression: cost-effective way to cut down the need of disk storage in varies clientele ecosystems. ACFS Replication: provides disaster recovery capability for the file system.
  • Developed Java application to determine and manage the system capacity and stability. Designed highly responsive application on a Java EE platform. Experienced in Oracle Application Development Framework(ADF) that simplifies Java EE Development.
  • Expertise in application development using Java, Java Development Kit (JDK), Java Runtime Environment(JRE), JavaScript, JDBC, Servlets, J2EE, JSP. Experienced in client side design and validation using JavaScript, HTML, CSS, XML.
  • Involved in the design of use case diagrams, sequence diagrams, and class diagrams of Java application.
  • Multi-threaded, Multi-Process and FOST application has been specifically designed, coded and run to ensure that ACFS capable enough to scale without degrading its performance, stability for handling large files and performance in Port Platforms including Linux, Windows, Solaris and AIX Linux (x 86 64- bit).
  • Written Perl Scripts for validating the use cases that runs across different Operating Systems under different loads. In line Source Viewer for test failures, and triaging of the issue was done to increase the productivity. Used Shell Scripts performing complex tasks for analyzing the issue.
  • Use of IDE for developing environment like Eclipse, NetBeans, ADF. Experienced in using source code change management and version control tools ADE Oracle Advance Developed Environment as a repository for managing/ deploying application code and to check in/out files, etc. also responsible for branching, merging the branch to main repository.
  • Talented at managing requirements analysis, functional specifications, and requirement documentation. Capable of Learning new technologies quickly and adapting to a new environment. Excellent planning, monitoring and troubleshooting skills to solve issues in Front-end and Back-end.
  • Received appreciation from analysts and users for developing a highly interactive java application that determine each file exact storage need, compression ratio and rate of fragmentation of each file.

ENVIRONMENT: Linux, Windows, Solaris and AIX Linux (x 86 64- bit), Java, PERL, SHELL, Oracle ADE (Advanced Development Environment), ADF, Eclipse, NetBeans.

Confidential

Hadoop Developer

RESPONSIBILITIES :

  • Apache Hadoop Frameworks , involved in processing large sets of structured, semi- structured, unstructured big data and supporting systems application architecture.
  • Familiar with big data architecture including Hadoop data ingestion, data transformation and data analysis. Apache Hadoop supports numerous mechanisms to ingest big data from external sources.
  • Using Apache Sqoop , moved bulk of big data to Hadoop Distributed File System ( HDFS ) and Apache Hive for data processing and made the interaction between Apache Pig and Database.
  • Involved to fetch, transfer and load the result of certain process back from HDFS to RDBMS using Apache Sqoop export and satisfy the scaling business needs.
  • Using Apache Flume reliable service, efficiently collected, accumulated and moved huge volumes of log data from varies kind of sources to consolidated warehouse. Transferred all needed text/log data generated by various source into Hadoop HDFS location.
  • Experienced in optimization to achieve higher throughput by tuning and configuring Apache Flume memory channels.
  • Collaborated with application, database, network and optimization team to analyze the requisite for Hadoop environmental setup, anticipate the scope/degree of node and data capacity.
  • Integrated logging methods to trace errors, analysis problems, fix bugs and ensured greater service connections and quality. Expertise to deep dive into technical holes and rectify with best solutions.
  • Developed pig script to populate data from Hadoop HDFS or staging Hive tables, parse the raw data which are not ready for data analysis, carry out data cleaning/preprocessing, performed data transformation to match a given Hive schema and store the refined data in appropriate placeholders.
  • Good working knowledge in tuning the degree of Apache MapReduce (MR) jobs to achieve higher parallelism based on needs.
  • Facilitated insightful analysis of cleaned and common format data collected by external sources using Apache Hive .
  • Based on necessity, created managed/external Hive tables with optimized partition/bucket that facilitate effective data querying and store the processed results in a tabular format.
  • Executed Hive queries that improve back-end stability, suit best for business intelligence and market analysis and enriched consumer satisfaction.
  • Attended internal Hadoop training program.
  • Used Concurrent Versions System CVS as version control.
  • Added the ID3 support in Mpeg2tsparser by meeting all the requirement and followed Aricent Coding guidelines to maintain consistency and completeness in developed codes. Included SRT, SSA/ASS subtitle support in MKV Parser and all testing was done successfully.
  • Implemented Mpeg2Ts Streaming parser and Mpeg2TSParser for codec type AAC and H264 fulfilling all the requirements and tested with many clips to verify the proper working of code.
  • Mpeg2TsWriter Plugin was developed for integrating MPEG2TS writer into transcoding solution for H264 video and AAC audio. Added Codec enhancement support LPCM in Mpeg2Ts Streaming parser.
  • Fixed the bugs raised by the stack team after analyzing the code. Curiosity to explore new ideas and concepts making me to be a keen communicator with excellent interpersonal skills. Got appreciation from Team and Customers for proven competency.
  • Guided new joiner with product/technical training in the project.

Environment : Big Data Hadoop, Hortonworks Data Platform HDP, Apache Flume, HDFS, MapReduce, Yarn, Pig, Hive, Sqoop, Web Interface Ambari and UNIX Shell Script. Linux, Android and Windows, C, C++ and Data Structure, CVS, Valgrind, splint, gdb, gcov, gprof and Microsoft Visual Studio.

We'd love your feedback!