We provide IT Staff Augmentation Services!

Hadoop Developer Resume

0/5 (Submit Your Rating)

San Francisco, CA

SUMMARY:

  • Have around 10 years of programming experience with skills in analysis, design, development, testing and deploying various software applications with emphasis on object oriented programming.
  • Have around 3.5 years of experience with Hadoop MapReduce, HDFS and Hadoop Ecosystems (PIG, HIVE, OOZIE, HBASE, SQOOP, FLUME, KAFKA, ZOOKEEPER)
  • Good Exposure to Hadoop distributions such CDH3, CDH4 & CDH5 and Hortnworks.
  • Experience in analyzing data using Hive QL, Pig Latin and custom MapReduce programs in Java.
  • Experience in working with NoSQL databases like HBase and Cassandra.
  • Experience in importing and exporting data from different databases like Mysql, Oracle into HDFS and Hive using Sqoop.
  • Knowledge in extending Hive and Pig core functionality by writing custom UDFs like UDAFs and UDTFs and developing custom SerDe(s) in HIVE.
  • Experience in understanding the security requirements for Hadoop and integrate with Kerberos authentication and authorization infrastructure.
  • Experience in using Apache Flume for collecting, aggregating and moving large amounts of data from application servers.
  • Experience in using Zookeeper and Oozie Operational Services for coordinating the cluster and scheduling workflows.
  • Knowledge of AVRO, JSON, XML and PARQUET.
  • Hands - on experience in writing Pig Latin scripts, working with grunt shells and job scheduling with Oozie.
  • Good knowledge of developing data flow pipeline analytics with Apache Storm, Kafka and Twitter APIs.
  • Expertise in writing shell scripts.
  • Expertise in developing MRUnit tests to test the MapReduce programs.
  • Good Knowledge in Amazon AWS concepts like EMR and EC2 web services which provides fast and efficient processing of Big Data.
  • Knowledge on Reporting tools like Tableau Software which is used to do analytics on data in cloud.
  • Analyzed partitioned and bucketed data using HIVE by computing various matrices for reporting through dashboards.
  • Experience using SNAPPY and BZIP2 compression techniques.
  • Expertise in Java/J2EE technologies such as Core Java, Spring, Hibernate, JDBC, JSP, JSTL, HTML, JavaScript, JSON.
  • Good understanding of service oriented architecture (SOA) and web services like XML, XSD, XSDL, SOAP.
  • Experience in Object Oriented Analysis and Design (OOAD) and development of software using UML Methodology, good knowledge of J2EE design patterns and Core Java design patterns.
  • Experience with iterative and incremental process like Rational unified process (RUP) and also agile software development methodologies.
  • Experience in using version control management tools like CVS, SVN and Rational Clear Case.
  • Strong knowledge of Software Development Life Cycle and expertise in detailed design documentation.
  • Highly motivated, self-starter with a positive attitude, willingness to learn new concepts and acceptance of challenges.
  • Major strengths are familiarity with multiple software systems, ability to learn quickly new technologies, adapt to new environments, self-motivated, focused adaptive and quick learner with excellent interpersonal, technical and communication skills.

TECHNICAL SKILLS:

Software SKILLS: Cloud & SaaS Platforms: Amazon EC2, Google App Engine, Force.com, Workday, Concur, ServiceNow etc. J2EE Platforms such as JBoss, Tomcat, WebLogic, AquaLogic and some WebSphere. Open Source & Collaboration Platforms such as LAMP, Alfresco, SharePoint, Drupal, Facebook, WordPress, Twitter etc. Cassandra deployment, Extensive SOA and Web Services experience on different technology stacks. Hands on exp on secure system with TLS/SSL security from end to end. Easily manageable system, using customized deployment and server management scripts. Having exp 4 years of exp on RabbitMQ deployment with support for clustering and federation. Web 2.0 Platforms such as Blogs, Mashups, Facebook API, and other Social Technology.Architectural Experience: Python, DJango, Ruby on Rails, .NET, RIA (Silverlight, Adobe Flex/Air), Blackberry, Android & iPhones mobile development.

Storage: EMC Vmax, DMX-3000/4000, symmetric, Clarion, HP XP, EVA, MSA, P7000, IBM DS 8300/8700/4700/8000 , Hitachi USP, AMS, NetApp FAS series arrays, Symantec

Cloud: Huawei, Cisco, VMware, Amazon, Eucalyptus, Hadoop

Operating Systems: WIN NT/2000/XP/2003/Vista/7, Solaris 8/9/10, HP-UX, Red Hat Linux

Database: MS SQL Server, Oracle, MS Access, Teradata

Performance monitoring tools: HP,BSM

NMS Tools: Nagios, Infovista

Backup Tool: Tivoli

Open source: Nagios, Apache Open stack, Infovista

PROFESSIONAL EXPERIENCE:

Confidential, San Francisco, CA

HADOOP DEVELOPER

Responsibilities:

  • Loading the data from the different Data sources like (Teradata and DB2) into HDFS using sqoop and load into Hive tables, which are partitioned.
  • Developed Hive UDF’s to bring all the customers email id into a structured format.
  • Developed bash scripts to bring the Tlog files from ftp server and then processing it to load into hive tables.
  • Using sqoop to load data from DB2 into HBASE environment.
  • Inserted Overwriting the HIVE data with Hbasedata daily to get fresh data every day.
  • All the bash scripts are scheduled using Resource Manager Scheduler.
  • Developed Oozie Workflows for daily incremental loads, which gets data from Teradata and then imported into hive tables.
  • Developed pig scripts to transform the data into structured format and it are automated through oozie coordinators.
  • Worked on loading the data from MySQL to HBase where necessary using Sqoop
  • Developed Hive queries for Analysis across different banners.

Environment: Windows 7, Hadoop, HDFS, MapReduce, Sqoop, Hive, pig,Hbase, Teradata, DB2, Oozie, MySQL, Eclipse

Confidential, Los Angeles, CA

HADOOP DEVELOPER

Responsibilities:

  • Evaluated suitability of Hadoop and its ecosystem to the project and implementing / validating with various proof of concept (POC) applications to eventually adopt them to benefit from the Big Data Hadoop initiative.
  • Extracted the needed data from the server into HDFS and Bulk Loaded the cleaned data into HBase using MapReduce.
  • Used MRUnit for unit testing.
  • Developed HIVE queries for the analysts.
  • Performed ETL using Pig, Hive and MapReduce to transform transactional data to de-normalized form.
  • Configured periodic incremental imports of data from DB2 into HDFS using Sqoop.
  • Worked extensively with importing metadata into Hive using Sqoop and migrated existing tables and applications to work on Hive.
  • Wrote Pig and Hive User Defined Functions to analyze the complex data to find specific user behavior.
  • Created an e-mail notification service upon completion of job for the particular team which requested for the data.
  • Defined job work flows as per their dependencies in Oozie.
  • Played a key role in productionizing the application after testing by BI analysts.
  • Involved in code reviews and peer reviews.
  • Maintain System integrity of all sub-components related to Hadoop.
  • Worked with the Data Science team to gather requirements for various data mining projects.
  • Involved in running Hadoop jobs for processing data coming from different sources.

Environment: Apache Hadoop, HDFS, Hive,Pig, Sqoop, HBase, Map Reduce, Java, Cloudera CDH4, Oozie, Flume,DB2,,Maven, shell, Eclipse, LINUX.

Confidential, SUNNYVALE, CA

SR. BIG DATA CONSULTANT

Responsibilities:

  • Responsible for building Hadoop cluster and integrate with Confidential Data Integration (PDI) server
  • Experienced in creating ETL transformations/Jobs using Spoon
  • Experience in developing visual MapReduce Applications using Confidential Spoon
  • Data loaded into impala tables after cleansing
  • Established database connection from Confidential to store in MySQL
  • Experienced in working with different Hadoop Ecosystems such as Hive, Impala, Hbase, Pig, Sqoop through Confidential
  • Developed various complex Mapper and Reduce transformations for Confidential MapReduce Jobs
  • Experience in using MongoDB
  • Extensively involved in Hadoop testing where scripts written in Python
  • Responsible for analyzing logs generated from various test cases identifying the reasons in case of failures
  • Involved in various debugging sessions with team
  • Responsible test cases reporting and documenting the test results
  • Good Knowledge in Report Designing using Confidential BA suite

Environment: Confidential PDI, Spoon, Java 6, Linux, Hadoop (CDH and Hortonworks), Impala, Hive, Sqoop, Hbase, Pig, MySQL

Confidential, Orem, UT

Sr Java Consultant

Responsibilities:

  • Using Core-Java to write Map-Reduce Jobs.
  • Using Amazon Elastic Map-Reduce for processing data.
  • Using Json to get data and store data into HDFS.
  • Using Amazon RDS as Database for processing relations.
  • Actively participated in the deployment of Hadoop Cluster into individual nodes.
  • Involved in the decision making of shifting from MapR framework to HortonWorks.
  • Extensively used Map-Reduce to process the available terabytes of data.
  • Worked on loading the data from MySql to HBase where necessary using Sqoop.
  • Using cassandra to store majority of data which needs to be divided based on region.
  • Using Hive to query the exported CSV data containing billions of data records.
  • Used Git Stash for source control.
  • Using oozie to schedule various jobs on Hadoop Cluster.
  • Java multi-threaded environment and quartz scheduler.
  • Written JUnits and integration tests for the class developed.
  • Developed projects using Maven to build and deploy.
  • Deployed and tested the application on UNIX based environments.

Environment: Hadoop, HBase, Hive, Map-Reduce, Storm,Eclipse, JUnit, MySql, java, json,, maven, AWS.

Confidential, Salt Lake City, UT

Sr Java Consultant with Hadoop

Responsibilities:

  • Using Core java Technologies to Manipulate the Data from HDFS and getting desired Results.
  • Using hibernate and spring in the Real-Time services using J2EE technologies using MVC design.
  • Hadoop Map-Reduce is used to analyze the un-organized text files, searching for data or pattern.
  • Using AMAZON RDS to manage the application Data.
  • Hadoop HBase was used to store Data in de-normalized way.
  • Using oozie to run regular jobs in timely manner.
  • Using AMAZON EMR for analysis of large files.
  • Hosting RESTful services in the intranet to allow other teams to get access.
  • Worked on spring components IOC, AOP and MVC.
  • The Back-End data has been maintained in Oracle Database.
  • Development of java web services using SOAP.
  • Java multi-threaded environment and quartz scheduler.
  • Written JUnits and integration tests for the class developed.
  • Developed projects using Maven to build and deploy.

Environment: Hadoop, Map-Reduce, HBase, Eclipse, JUnit, J2EE, ClearCase, AMAZON RDS, spring, hibernate.

Confidential, SANTA CLARA, CA

SENIOR SOFTWARE ENGINEER

Responsibilities:

  • Participate in project planning sessions with business analysts and team members to analyze business IT Requirements and translate business requirements into working model.
  • Involved in Planning, Defining and Designing data base on business requirement and provided documentation.
  • Involve in Initial designing and creating Use case diagrams, Sequence Diagrams and class diagrams using the MS Visio Tool.
  • Develop Java application using Java Spring framework.
  • Developed dynamic ASPX web pages for application using ASP.NET, C#, XML, HTML and JavaScript.
  • Developed RESTful Web services using Java, SpringBoot
  • Wrote complex SQL queries using joins, sub queries and correlated sub queries to retrieve data from the database.
  • Created/Updated database objects like tables, views, stored procedures, function, packages
  • Used JavaScript and struts validation framework for performing front end validations.
  • Develop DAO design pattern for hiding the access to data source objects.
  • Use Hibernate framework for the backend persistence
  • Use Eclipse as IDE tool to develop the application and JIRA for bug and issue tracking
  • Use CVS for software configuration management and version control
  • Optimized and modified the triggers, complex store functions, procedures, user base data type etc.
  • Added methods for performing CRUD operations in applications using JDBC and wrote several SQL queries.
  • Responsible for the dealing with the problems, bug fixing and troubleshooting.
  • Developing Oracle PL/SQL stored procedures, Functions, Packages, SQL scripts to facilitate the functionality for various modules.

Environment: Java, J2EE, EJB 1.1, JSF, XML, JDBC, Oracle 9i, Log4J 1.2.,PL/SQL Developer, REST framework, C#, .NET Framework 3.0, Spring framework.

Confidential, Seattle, WA

Java/J2EE Developer

Responsibilities:

  • Involved in various phases of Software Development Life Cycle (SDLC) as design development and unit testing.
  • Developed and deployed UI layer logics of sites using JSP, XML, JavaScript, HTML/DHTML, and Ajax.
  • CSS and JavaScript were used to build rich internet pages.
  • Agile Scrum Methodology been followed for the development process.
  • Designed different design specifications for application development that includes front-end, back-end using design patterns.
  • Developed proto-type test screens in HTML and JavaScript.
  • Involved in developing JSP for client data presentation and, data validation on the client side with in the forms.
  • Developed the application by using the Spring MVC framework.
  • Collection framework used to transfer objects between the different layers of the application.
  • Developed data mapping to create a communication bridge between various application interfaces using XML and XSL.
  • Spring IOC being used to inject the parameter values for the Dynamic parameters.
  • Developed JUnit testing framework for Unit level testing.
  • Actively involved in code review and bug fixing for improving the performance.
  • Documented application for its functionality and its enhanced features.
  • Created connection through JDBC and used JDBC statements to call stored procedures.

Environment: Spring MVC, Oracle 11g J2EE, Java, JDBC, Servlets, JSP, XML, Design Patterns, CSS, HTML, JavaScript 1.2, JUnit, Apache Tomcat, My SQL.

Confidential, San Francisco, CA

Java/J2EE Developer

Responsibilities:

  • Participated in detailed requirements gathering, Analysis, Development and Testing.
  • Produced UML diagrams (use case, class, and sequence diagrams) to document in design phase using Microsoft Visio tool.
  • Involved in CR & Enhancements estimations, designs and tracking.
  • Involved in developing Oracle SQL queries and PL SQL programming.
  • Deliver the application code using Struts
  • Integrate the GUI with back end Oracle data base.
  • Developed UI screens using HTML, Java Script, DHTML, CSS, AJAX, XML and XSLT.
  • Extensively used design patterns like Singleton, Factory, Abstract Factory, Façade, DAO, and MVC etc.
  • Used Restful web services to upload a request agreement to process to hit Credit Engine to get the approval agreement details.
  • Designed and implemented database abstraction layer using Hibernate DAO objects.
  • Involved in backing up database & in creating dump files and created DB schemas from dump files.
  • Written developer test cases & executed. Prepared corresponding scope & traceability matrix.
  • Involved in JUnit testing, smoke testing, integration testing, and system testing.

Environment: Java/J2EE, OOAD, Maven, Shell AJAX, Agile, WebServices, SOAP, WSDL, JSF, PL/SQL, XML, JDBC, JavaScript, HTML, Oracle 9i, UNIX, JUnit

We'd love your feedback!