We provide IT Staff Augmentation Services!

Hadoop Developer Resume

0/5 (Submit Your Rating)

IL

PROFESSIONAL SUMMARY:

  • An accomplished, results - driven IT professional with around 7.8 years of experience in the IT industry and 2.4+ years of experience in the Hadoop ecosystem and Apache Spark.
  • Experienced with Hadoop distributions like Cloudera, Hortonworks and depth knowledge on Hadoop architecture including MapReduce, Yarn, Hive, Sqoop, HDFS, Spark, Oozie, Zookeeper in a range of industries such as Financing sector and Health care.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
  • Expertise in working with RDBMS databases like MySQL and SQL Server.
  • Rich experience in automating Sqoop and Hive queries using Oozie workflow.
  • Experience in strong and analysing data using HiveQL, SparkQL and MapReduce programs.
  • Experience with AWS components like Amazon EC2 instances, S3 buckets and Cloud Formation templates
  • Experienced in Developing Spark application using Spark Core, Spark SQL and Spark Streaming API’s.
  • Analysed the data by performing Hive queries and used HIVE UDF’s.
  • Tuning, and Monitoring Hadoop jobs and clusters in a production environment.
  • Experience in designing tables and views for reporting using Impala.
  • Experienced in working with different file formats - Avro, Parquet.
  • Experience in different compression techniques like Gzip, LZO, Snappy and Bzip2.
  • Experience in web development using Java and front-end using JavaScript, HTML, XML and CSS.
  • Proficient in Automation (Selenium IDE), Manual (Blackbox), Database, functional and non-functional, negative, regression, system integration (SIT), acceptance (UAT), security, browser compatibility and system testing.
  • Strong Communication skills of written, oral, interpersonal and presentation
  • Proven ability to interface with stakeholders to ensure solutions to meet customer requirements.
  • Adept at managing multiple projects simultaneously. Strategic planning through operational and technology design for optimizing Software development, architectural design, creating various proof of concepts.
  • Exceptional ability to learn new technologies and to deliver outputs in short deadlines.

TECHNICAL SKILLS:

Hadoop: Hadoop 2.2, HDFS, MapReduce, Hive, Sqoop, Zookeeper, Yarn, Spark, Kafka, Hbase, Flume, Impala and Hue. Hadoop management & Security Cloudera Manager, Hortonworks

Web Technologies: HTML, XHTML, XML, XSL (XSLT, XPATH), XSD, CSS, JavaScript

Database: Microsoft SQL Server, MySQL, SQL, RDBMS, Familiar with NoSQL- Hbase

Programming Languages: Java, J2EE

Web Servers: Apache Tomcat 5.x

OS/Platforms: Mac OS X 10.9.5, Windows2008/Vista/2003/XP/2000/NT, Linux, Unix

Client side: JavaScript, CSS, HTML

Testing & Case Tools: Junit, Rational Clear Case, JBuilder, Selenium

Cloud Components: AWS (S3 Buckets, Ec2, Cloud Formation)

PROFESSIONAL EXPERIENCE:

Confidential, IL

Hadoop Developer

Responsibilities:

  • Used SQOOP to import data from RDMS sources into HDFS using Avro File format.
  • Performed transformation, cleaning and filtering on imported data using Hive, Spark and loaded data into HDFS.
  • Used Hive to Partition and Bucketing, for optimizing data storage and compute-various metrics for reporting
  • Developed Spark core and Spark SQL scripts using Scala for faster data processing.
  • Converted Hive queries into Spark transformations using Spark RDDs.
  • Responsible for creating shell script to run Hive jobs.
  • Worked with different file formats like Avro and Parquet.
  • Worked with HBase to conduct quick look ups (Update, Inserts and Deletes) in Hadoop.
  • Used Oozie workflow to automate all the jobs.
  • Exported the analysed data into Tableau Server using SCOOP for visualization and to generate reports for BI team.
  • Used Git hub as code repository and version control.
  • Used Agile methodology in developing the application, participated in stand-up meetings and Sprints.
  • Working on a sample POC to collect large amount of log data using Kafka in HDFS for further analysis.
  • Continuous monitoring and managing the Hadoop cluster through Cloudera Manager.

Environment: HDFS, Yarn, MapReduce, Hive, Sqoop, Oozie, HBase, Kafka, Impala, Spark SQL, Eclipse, Oracle, Teradata, Cloudera.

Confidential, NC

Hadoop Developer

Responsibilities:

  • Involved in the Complete Software development life cycle (SDLC) to develop the application.
  • Worked with the Data Science team to gather requirements for various data mining projects.
  • Validating the source file for Data Integrity and Data Quality by reading header and trailer information and column validations.
  • Used Hive to do transformations, joins, filter and some pre-aggregations before storing the data onto HDFS.
  • Experience in importing data from various data sources like MySQL and Netezza using Sqoop, performed transformations using Hive, and loaded data back into HDFS.
  • Import and export data between the environments like MySQL, HDFS and deploying into productions.
  • Worked on partitioning and used bucketing in HIVE tables and setting tuning parameters to improve the performance.
  • Experience in Oozie workflow scheduler template to managevarious jobs like Sqoop, MR, Hive, Shell scripts.
  • Involved in importing and exporting data from Sqoop using Spark.
  • Actively participating in the code reviews, meetings and solving any technical issues.
  • Implemented SparkSQL to access hive table into Spark for faster processing of data.

Environment: Java API,Hadoop2.x, Spark, Cloudera CDH 5.x, Kafka, HDFS, Hive 0.12, MapReduce, Sqoop 1.4.3, Oozie, Agile, JIRA

Confidential

System Engineer

Responsibilities:

  • Writing Test cases using element locators, web driver methods, java programming features.
  • Responsible in Development, Testing, Maintenance, and Production Support of ALM application.
  • Involved in regression testing of RESTful web services at the different phases of this application.
  • Involved in architecting, developing and testing SOAP based web services.
  • Received appreciations for scaling up quickly and taking up extra modules after having completed the assigned modules before time.
  • Involved in Analysis of requirements, low level designing, implementation of the services, Bug Fixing, System & Integration Testing, Onsite Coordination and undertaking the complete responsibility of a few of requirements.
  • Involved in designing, coding, Peer Code Reviews and was also responsible for performing functional, technical walk through, code reviews and debugging of various modules.
  • Done Functional testing, Regression testing, Integration testing, Database testing, user interface and System testing.
  • Preparing test reports for all the releases.
  • Was one of the points of contact for any production issue for this application.
  • Worked on Agile methodologies.

Environment: Selenium, Java, MySQL, SVN, Javascript, SQL Server 2008, XML, Excel, HTML, SOAP web services, SOAP UI Tool, Quality Assurance

Confidential

Process Associate

Responsibilities:

  • Prepared test plans based on the requirement walk through given by the client.
  • Discussed issues with the major stakeholders on calls and raised defects whenever required.
  • Designed and executed test cases in ALM.
  • Prepared test execution summary report for the completed
  • Prepared in-house defect tracking tool using Java, HTML, CSS,

Environment: Selenium, Java, MySQL, SVN, Quality Assurance

We'd love your feedback!