Hadoop Developer Resume
Sanjose, CA
SUMMARY
- Around 8 years of experience in full software development life cycle from concept through the delivery of applications and customizable solutions with emphasis on Object Oriented Programming, Java/J2EE, SQL and Hadoop / Big Data technologies.
- Hands on experience on Hadoop Ecosystem components. (HDFS,MapReduce,PIG,HIVE,HBASE,SQOOP,Flume,Oozie,ZooKeeper,Kafka,Spark(Scala),Storm,MongoDB, Cassandra)
- Involved in Installation and configuration of Hadoop Ecosystem components along with Hadoop Admin.
- Hands on experience on Teradata Migration to Hadoop Platform.
- Extensive knowledge on Hive,Spark - Sql Development.
- Exposure to Teradata development.
- Working on implementing Spark and Strom frame work.
- Automated Spark-SQL scripts using Unix shell script and Involved in Optimization of Hive Queries.Also used Optimization techniques for better performance.
- Good understanding of Hadoop architecture and components of Hadoop Clusters (Job Tracker, Task Tracker, Name Node and Data Node).
- Performed Data Ingestion to Hadoop file system from different data sources.
- Analyzed different file formats and large data sets by running Hive queries and Pig scripts.
- Experienced in working with senior level managers, business people and developers across multiple disciplines.
- Using Apache Flume,collected and stored streaming data(log data) in HDFS.
- Extensively used Apache Sqoop for efficiently transferring bulk data between Apache Hadoop and relational databases (Teradata).
- Automated sqoop,hive and pig jobs using Oozie scheduling.
- Extensive knowledge in NoSQL databases like Hbase,MangoDB,Cassandra.
- Have good knowledge on writing and using the user defined functions in HIVE,PIG and MapReduce.
- Configured & deployed and maintained multi-node Dev and Test Kafka Clusters.
- Developed multiple Kafka Producers and Consumers from scratch as per the business requirements.
- Responsible for creating, modifying and deleting topics (Kafka Queues) as and when required by the Business team.
- Developed tests cases and POC’s to benchmark and verify data flow through the Kafka clusters.
- Extensive exposure to all aspects of Software Development Life Cycle (SDLC) i.e. Requirements Definition for customization, Prototyping,Coding and Testing.
- Experience in resolving on-going production and maintenance issues and bug fixes.
- Proficiency in developing SQL queries.
- Exposure to Java development projects.
- Excellent communication skills.
TECHNICAL SKILLS
Big Data/Hadoop Ecosystems: (Hive,Pig,Sqoop,Flume,Zookeeper,oozie,MR,Hbase),Kafka,Storm andSpark(Spark SQL,Scala)
Scripting Language: Unix Shell Scripting, Python
Methodologies: Agile, Waterfall model
Technologies: Hadoop,Spark,Scala,Db2,CoreJava,JDBC,JavaScript,SQL,COBOL,CICS,JCL,VSAM,ISPF,FILEAID,SPUFI,IDCAMS,ENDEVOR
Database: Teradata, SQL, My SQL, DB2,HBase, Cassandra
Servers: Tomcat
IDE: Eclipse, Net Beans
PROFESSIONAL EXPERIENCE
Confidential, Sanjose, CA
Hadoop Developer
Responsibilities:
- Involved in complete SDLC life cycle of big data project that includes requirement analysis, design, coding, testing and production
- Extensively Used Sqoop to import/export data between RDBMS and hive tables, incremental imports and created Sqoop jobs for last saved value.
- Extensive knowledge on Spark.
- Worked on Redesigning and Coding in Hive and Spark-Sql.
- Extensively worked with Partitions, Bucketing tables in Hive and designed both Managed and External table.
- Worked on optimization of Hive Queries. Optimized Spark-Sql Parameters to work effectively.
- Worked on Teradata understanding and migrating it to hadoop end to end.
- Worked on developing Unix Shell scripts to automate Spark-Sql.
- Involved in all phases of Project.
- Worked with huge volumes of data and migrated effectively with no post production defects.
- Involved in requirement analysis.
- Involved in giving KT to other team members.
- Involved in preparing Project documentation.
- Designing Test Plans, Test Cases and performed System Testing.
- Involved in daily SCRUM meetings to discuss the development/progress.
- Sprints and was active in making scrum meetings more productive.
- Experience in integrating RHadoop for categorization and statistical analysis to generate reports.
- Involved in monitoring, testing Hadoop cluster by using Hadoop bench marks.
- Moving data from HDFS to Cassandra using Map Reduce and Bulk Output Format class.
Environment: BigData, Hadoop, MapReduce, Pig, Hive, Sqoop, Oozie, Crunch, Scala, Spark, Strom, kafka, Cassandra, Linux, Python.
Confidential, Chicago, IL
Hadoop Developer
Responsibilities:
- Worked with Data Ingestion techniques to move data from various sources to HDFS.
- Analyzed different formats of Data.
- Worked on writing Map Reduce programs using Java.
- Extensively worked with Partitions, Bucketing tables in Hive and designed both Managed and External table.
- Worked on optimization of Hive Queries.
- Created and worked with Sqoop jobs with full refresh and incremental load to populate Hive External tables.
- Worked on Pig to do data transformations.
- Developed UDF’s in Map/reduce, Hive and Pig.
- Worked on Hbase and Its integration with Strom.
- Worked on Apache flume for getting data from Twitter to HDFS.
- Presently implementing KAFKA.
- Worked with MongoDB and Cassandra NoSql Db’s.
- Presently implementing Strom, Spark.
- Designing and creating Oozie workflows to schedule and manage Hadoop, Hive, pig and sqoop jobs.
- Worked with RDBMS import and export to HDFS.
- Involved in requirement analysis.
- Involved in giving KT to other team members.
- Involved in preparing Project documentation.
Environment: Hadoop, HDFS, Hive, HBase, Map Reduce, Pig, Cassandra Hive, Sqoop, Oozie, UNIX, Shell Scripting, MS Visio
Confidential, Atlanta,GA
Java Developer
Responsibilities:
- Involved in design and development phases of Software Development Life Cycle (SDLC).
- Developed code as per the rquirements.
- Managed projects and meet milestones.
- Implemented Multithread concepts inJavaclasses to avoid deadlocking.
- Involved in High Level Design and prepared Logical view of the application.
- Involved in designing and developing of Object Oriented methodologies using UML and created Use Case, Class, Sequence diagrams and also in complete development, testing and maintenance process of the application.
- Created CorejavaInterfaces and Abstract classes for different functionalities.
- Responsible for Analysis, Design, Development and Integration of UI components with backend usingJ2EEtechnologies such as Servlets, JSP, JDBC.
Environment: Eclipse, Core java.
Confidential
Application Developer
Responsibilities:
- Worked on Db2 and Sql.
- Involved in source analysis and Inventory Phase of the Project.
- Involved in distribution & management of project work with other vendors like Accenture.
- As this project needs to convert Assembler to C, I have prepared understanding documents which includes entire logic of Assembler programming modules and approach documents.
- Involved in analysis of all the CICS modules to get test data as there is no test data available.
- Involved in converting Mainframe CICS programs to Tx series (UNIX).
- Involved in all phases of CICS conversion (both development and testing).
- Involved in Unit testing, Bug fixing of CICS modules.
- Involved in preparation of project report and took KT for new members of the team.
- Actively participated in giving demo to clients during their visit to offshore.
- Involved in Coding new modules, bug fixing, testing of Jobs and ABEND handling.
- Involved in Unit testing, System Testing, UAT, Integration Testing, Regression Testing and Deployments.
Environment: Db2, Endeavor, Cobol, Cics, JCL,REXX