We provide IT Staff Augmentation Services!

Hadoop Developer Resume

0/5 (Submit Your Rating)

Sanjose, CA

SUMMARY

  • Around 8 years of experience in full software development life cycle from concept through the delivery of applications and customizable solutions with emphasis on Object Oriented Programming, Java/J2EE, SQL and Hadoop / Big Data technologies.
  • Hands on experience on Hadoop Ecosystem components. (HDFS,MapReduce,PIG,HIVE,HBASE,SQOOP,Flume,Oozie,ZooKeeper,Kafka,Spark(Scala),Storm,MongoDB, Cassandra)
  • Involved in Installation and configuration of Hadoop Ecosystem components along with Hadoop Admin.
  • Hands on experience on Teradata Migration to Hadoop Platform.
  • Extensive knowledge on Hive,Spark - Sql Development.
  • Exposure to Teradata development.
  • Working on implementing Spark and Strom frame work.
  • Automated Spark-SQL scripts using Unix shell script and Involved in Optimization of Hive Queries.Also used Optimization techniques for better performance.
  • Good understanding of Hadoop architecture and components of Hadoop Clusters (Job Tracker, Task Tracker, Name Node and Data Node).
  • Performed Data Ingestion to Hadoop file system from different data sources.
  • Analyzed different file formats and large data sets by running Hive queries and Pig scripts.
  • Experienced in working with senior level managers, business people and developers across multiple disciplines.
  • Using Apache Flume,collected and stored streaming data(log data) in HDFS.
  • Extensively used Apache Sqoop for efficiently transferring bulk data between Apache Hadoop and relational databases (Teradata).
  • Automated sqoop,hive and pig jobs using Oozie scheduling.
  • Extensive knowledge in NoSQL databases like Hbase,MangoDB,Cassandra.
  • Have good knowledge on writing and using the user defined functions in HIVE,PIG and MapReduce.
  • Configured & deployed and maintained multi-node Dev and Test Kafka Clusters.
  • Developed multiple Kafka Producers and Consumers from scratch as per the business requirements.
  • Responsible for creating, modifying and deleting topics (Kafka Queues) as and when required by the Business team.
  • Developed tests cases and POC’s to benchmark and verify data flow through the Kafka clusters.
  • Extensive exposure to all aspects of Software Development Life Cycle (SDLC) i.e. Requirements Definition for customization, Prototyping,Coding and Testing.
  • Experience in resolving on-going production and maintenance issues and bug fixes.
  • Proficiency in developing SQL queries.
  • Exposure to Java development projects.
  • Excellent communication skills.

TECHNICAL SKILLS

Big Data/Hadoop Ecosystems: (Hive,Pig,Sqoop,Flume,Zookeeper,oozie,MR,Hbase),Kafka,Storm andSpark(Spark SQL,Scala)

Scripting Language: Unix Shell Scripting, Python

Methodologies: Agile, Waterfall model

Technologies: Hadoop,Spark,Scala,Db2,CoreJava,JDBC,JavaScript,SQL,COBOL,CICS,JCL,VSAM,ISPF,FILEAID,SPUFI,IDCAMS,ENDEVOR

Database: Teradata, SQL, My SQL, DB2,HBase, Cassandra

Servers: Tomcat

IDE: Eclipse, Net Beans

PROFESSIONAL EXPERIENCE

Confidential, Sanjose, CA

Hadoop Developer

Responsibilities:

  • Involved in complete SDLC life cycle of big data project that includes requirement analysis, design, coding, testing and production
  • Extensively Used Sqoop to import/export data between RDBMS and hive tables, incremental imports and created Sqoop jobs for last saved value.
  • Extensive knowledge on Spark.
  • Worked on Redesigning and Coding in Hive and Spark-Sql.
  • Extensively worked with Partitions, Bucketing tables in Hive and designed both Managed and External table.
  • Worked on optimization of Hive Queries. Optimized Spark-Sql Parameters to work effectively.
  • Worked on Teradata understanding and migrating it to hadoop end to end.
  • Worked on developing Unix Shell scripts to automate Spark-Sql.
  • Involved in all phases of Project.
  • Worked with huge volumes of data and migrated effectively with no post production defects.
  • Involved in requirement analysis.
  • Involved in giving KT to other team members.
  • Involved in preparing Project documentation.
  • Designing Test Plans, Test Cases and performed System Testing.
  • Involved in daily SCRUM meetings to discuss the development/progress.
  • Sprints and was active in making scrum meetings more productive.
  • Experience in integrating RHadoop for categorization and statistical analysis to generate reports.
  • Involved in monitoring, testing Hadoop cluster by using Hadoop bench marks.
  • Moving data from HDFS to Cassandra using Map Reduce and Bulk Output Format class.

Environment: BigData, Hadoop, MapReduce, Pig, Hive, Sqoop, Oozie, Crunch, Scala, Spark, Strom, kafka, Cassandra, Linux, Python.

Confidential, Chicago, IL

Hadoop Developer

Responsibilities:

  • Worked with Data Ingestion techniques to move data from various sources to HDFS.
  • Analyzed different formats of Data.
  • Worked on writing Map Reduce programs using Java.
  • Extensively worked with Partitions, Bucketing tables in Hive and designed both Managed and External table.
  • Worked on optimization of Hive Queries.
  • Created and worked with Sqoop jobs with full refresh and incremental load to populate Hive External tables.
  • Worked on Pig to do data transformations.
  • Developed UDF’s in Map/reduce, Hive and Pig.
  • Worked on Hbase and Its integration with Strom.
  • Worked on Apache flume for getting data from Twitter to HDFS.
  • Presently implementing KAFKA.
  • Worked with MongoDB and Cassandra NoSql Db’s.
  • Presently implementing Strom, Spark.
  • Designing and creating Oozie workflows to schedule and manage Hadoop, Hive, pig and sqoop jobs.
  • Worked with RDBMS import and export to HDFS.
  • Involved in requirement analysis.
  • Involved in giving KT to other team members.
  • Involved in preparing Project documentation.

Environment: Hadoop, HDFS, Hive, HBase, Map Reduce, Pig, Cassandra Hive, Sqoop, Oozie, UNIX, Shell Scripting, MS Visio

Confidential, Atlanta,GA

Java Developer

Responsibilities:

  • Involved in design and development phases of Software Development Life Cycle (SDLC).
  • Developed code as per the rquirements.
  • Managed projects and meet milestones.
  • Implemented Multithread concepts inJavaclasses to avoid deadlocking.
  • Involved in High Level Design and prepared Logical view of the application.
  • Involved in designing and developing of Object Oriented methodologies using UML and created Use Case, Class, Sequence diagrams and also in complete development, testing and maintenance process of the application.
  • Created CorejavaInterfaces and Abstract classes for different functionalities.
  • Responsible for Analysis, Design, Development and Integration of UI components with backend usingJ2EEtechnologies such as Servlets, JSP, JDBC.

Environment: Eclipse, Core java.

Confidential

Application Developer

Responsibilities:

  • Worked on Db2 and Sql.
  • Involved in source analysis and Inventory Phase of the Project.
  • Involved in distribution & management of project work with other vendors like Accenture.
  • As this project needs to convert Assembler to C, I have prepared understanding documents which includes entire logic of Assembler programming modules and approach documents.
  • Involved in analysis of all the CICS modules to get test data as there is no test data available.
  • Involved in converting Mainframe CICS programs to Tx series (UNIX).
  • Involved in all phases of CICS conversion (both development and testing).
  • Involved in Unit testing, Bug fixing of CICS modules.
  • Involved in preparation of project report and took KT for new members of the team.
  • Actively participated in giving demo to clients during their visit to offshore.
  • Involved in Coding new modules, bug fixing, testing of Jobs and ABEND handling.
  • Involved in Unit testing, System Testing, UAT, Integration Testing, Regression Testing and Deployments.

Environment: Db2, Endeavor, Cobol, Cics, JCL,REXX

We'd love your feedback!