We provide IT Staff Augmentation Services!

Sr. Hadoop/spark Developer Resume

5.00/5 (Submit Your Rating)

Ofallon, MO

SUMMARY:

  • 9+ years IT experience in design, development, testing, deployment & post production maintenance.
  • 3+ years of experience in Hadoop Ecosystem (MapReduce, Hive, Pig, Flume, Sqoop, Spark, HBase, Oozie, Yarn, Kafka, Impala, Zookeeper) as Developer.
  • Experience in analyzing data using HiveQL, Pig Latin, HBase and custom Map Reduce programs in Java.
  • Extending Hive and Pig core functionality by writing custom UDFs.
  • Hands on Experience with installation, configuration, supporting and managing of Big Data and underlying infrastructure of Hadoop Cluster.
  • Experience in managing and reviewing Hadoop log files.
  • Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa.
  • Extensive experience in Unix Shell scripting.
  • Experience with build tools like Jenkins.
  • Automated several batch process to eliminate manual intervention.
  • Good understanding of NoSQL databases and hands on experience in writing applications on NoSQL databases like HBase and Cassandra.
  • Experience in J2EE technologies including Servlets, JSP, spring, Struts, JMS, Hibernate, JDBC, XML, XSLT and JNDI for building various client-server applications and Unix Shell Scripting.
  • Working experience with Oracle, MYSQL, DB2 databases and writing SQL, PLSQL Stored Procedures, Triggers, Functions and Packages.
  • Experience in Middle Tier Development using Design Patterns, Servlets, JSPs, Custom Tag Libraries, XML, and EJB.
  • Experience in working with application servers like JBoss, Weblogic, IBM Websphere, and Apache Tomcat.
  • Experience in data processing and interfacing (Back-end Front-end connectivity) using Oracle 10g/11g (SQL, Stored procedures), and MySQL.
  • Hands on experience in using JUnit, Log4J for debugging and testing the applications.
  • Good knowledge in Financial, HealthCare and Insurance domains.
  • Solid experience in communicating with all levels of enterprise.
  • Following Agile methodology and SCRUM meetings to track optimize and resolve impediments to meet deadlines on time.
  • Effective leadership quality with good skills in strategy, business development, client management and project management.
  • Experience in Onsite - Offshore model.
  • Strong Problem Solving and Analytical skills and abilities to make Balanced and Independent Decisions.
  • Motivation, Initiative and Project Management Attributes.

TECHNICAL SKILLS:

Java Technologies: Java, J2EE, Servlets, Custom Tag Libraries, EJB, JDBC, JMSJSF, RMI, JNDI, JSTL.

Web Technologies: HTML/DHTML, JSPs, CSS, AJAX, NodeJS, JavaScript

XML Related: XML, XSLT, JAXP, JAXB, JAXRPC, DOM, SAX, XMLSchema, DTD, XMLBeans

Frameworks: Jakarta Struts, Spring, Enterprise Java Beans (EJB)

Languages: Java, Scala, Python, C, C++, SQL

Tools: /Software: Apache Axis, Log4j, Ant, Toad, Altova XMLSpy, Maven, UML, Rational Rose, MVC architecture, Unix Shell Script

Databases: Oracle, PostGres SQL, SQL Server 2005

ORM Tool: Hibernate

Application/Web Servers: WebLogic, IBM Web sphere, Apache Tomcat, JBoss

Development tools: RAD, Eclipse, Net Beans

Operating Systems: Windows 2000/XP, Linux, UNIX

PROFESSIONAL EXPERIENCE:

Confidential, OFallon, MO

Sr. Hadoop/Spark Developer

Responsibilities:

  • Involved in requirement sessions along with all the business stakeholders.
  • Developed multiple MapReduce programs in Java.
  • Developed UDF in Java
  • Defining Job flows and dependencies
  • Written Unix shell scripts to integrate all the job flows
  • Experience with continuous integration build tools like Jenkins
  • Involved in Unit testing, System integration testing and provided support for UAT testing.
  • Involved in deployments in Pre-prod & Production.
  • Responsible for ingesting data into HDFS and HBASE from different data sources.
  • Developed data ingestion scripts using Sqoop.
  • Handled initial loads and delta during the ingestion process
  • Wrote shell scripts to invoke MapReduce programs, Hive scripts, Sqoop scripts etc.
  • Worked on performance tunings.
  • Performed advanced procedures like text analytics and processing, using the in-memory computing capabilities of Spark using Scala.
  • Experience in streaming the data using Spark with Kafka.
  • Import data using Sqoop to load data from RDBMS(Oracle, PostGres SQL) to HDFS on regular basis.
  • Played a key role in installation and configuration of the various Hadoop ecosystem tools such as Solr, Kafka, Pig, HBase.
  • Implemented Daily Oozie jobs that automate parallel tasks of loading the data into HDFS and pre-processing with Pig using Oozie co-coordinator jobs.
  • Deployed and monitored scalable infrastructure on Amazon Web Services (AWS) & configuration management using Puppet. .
  • Created Hive tables and wrote queries to fetch the data from HDFS.
  • Created the developer Unit test plans and executed unit testing in the development cluster.
  • Involved in automating the test flows and integrating them in the continuous integration tools like Jenkins

Environment: Hadoop, HDFS,AWS, HBASE, Spark, Spark-SQL, Talend, Oozie, Flume, Sqoop, Pig, Hive, SQL, XML, Eclipse, Java Regex, Jira, Shell Scripting.

Confidential, Dallas, TX

Sr. Hadoop/Java Developer

Responsibilities:

  • Involved in functional requirement review. Worked closely with Risk & Compliance Team and BA.
  • Developed complex Map Reduce programs in Java for Data Analysis on different data formats.
  • Developed Unix shell scripts to invoke all the process flows.
  • Designed and configured Flume servers to collect data from the network proxy servers and store to HDFS and HBASE.
  • Involved in Unit testing, System integration testing and provided support for UAT testing.
  • Used agent E2E Chain for reliability and failover in flume.
  • Worked on documentation of all Extract, Transform and Load, designed, developed, validated and deploy the Talend ETL Processes for Data ware house team using PIG and HIVE.
  • Worked on Sequence files, RC files, Map side joins, partitioning, bucketing for Hive performance enhancement and storage improvement.
  • Designed and implemented Restful APIs to retrieve the data from Hadoop Platform to Employee Portal Web Application.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Extensively used Sqoop to get data from RDBMS sources like Teradata and Netezza.
  • Installed Oozie workflow engine to run multiple Hive and Pig Jobs.
  • Wrote MRUnit tests for unit testing the Map Reduce jobs.
  • Implemented Hive, Pig Generic UDF's to implement business logic.
  • Worked on installing cluster, commissioning & decommissioning of DataNodes, NameNode recovery, capacity planning, and slots configuration.
  • Implemented NameNode backup using NFS. This was done for High availability.
  • Involved in the installation of HDP 2.3 and up-gradation from HDP 2.3to HDP 2.4.0
  • Responsible for developing data pipeline using HD Insight, flume, Sqoop and pig to extract the data from weblogs and store in HDFS.
  • Used Apache Log4J for logging.
  • Facilitated Knowledge transfer sessions.
  • Use TDD and Worked in an agile environment.

Environment: Hadoop, Yarn, AWS, HDFS, HDP, Flume, Sqoop, Hive, Pig, Hbase, Java, Scala, solr, Teradata, Netezza, Spring, Eclipse, Maven, XML, GIT, JSON, Linux, Avro.

Confidential, Atlanta, GA

Hadoop Developer

Responsibilities:

  • Developed Map Reduce programs in java for parsing the raw data and populating staging tables.
  • Implemented Generic writable to in corporate multiple data sources into reducer to implement recommendation based reports using Map Reduce programs.
  • Responsible for ingesting data into HDFS and HBASE from different data sources.
  • Developed data ingestion scripts using Sqoop.
  • Handled initial loads and delta during the ingestion process
  • Importing and exporting data into HDFS and Hive using Sqoop
  • Proactively monitored systems and services, architecture design and implementation of Hadoop deployment, configuration management, backup, and disaster recovery systems and procedures
  • Load and transform large sets of structured, semi structured and unstructured data
  • Load log data into HDFS using Flume.
  • Wrote shell scripts to monitor the health check of Hadoop daemon services and respond accordingly to any warning or failure conditions
  • Involved in loading data from UNIX file system to HDFS, configuring Hive and writing Hive UDFs
  • Gained very good business knowledge on health insurance, claim processing, fraud suspect identification, appeals process, etc.

Environment: java, Hadoop, Linux, MapReduce, HDFS, Hive, Sqoop, Flume, pig, J2EE, MySQL, Shell Scripting, Eclipse, SVN, Maven, JIRA.

Confidential, Richardson, TX

Java/J2EE Developer

Responsibilities:

  • Involved in quickly understanding the requirements and reporting the manager about the daily status of this project as this was a very urgent need for the company.
  • Extensively working in CSS and NodeJS for formatting search results with number of columns congaing images and icons.
  • Extensively used design pattern like Application façade, Data Access object and MVC.
  • Worked with different layers like Java beans, Dao, transactions, service and presentation.
  • Developed a presentation layer using JSP, CSS and JavaScript.
  • Understood the existing code and involved in helping and explaining the team members about the flow of the code. Worked on Server Tuning for increase the heap size for UNIX machines.
  • Handled offshore team for implement Customer gateway design/architecture.
  • Worked on Pl/SQL for query and fetch the data from database.
  • Implemented web layer using JSF and Ice faces.
  • Implemented business layer using Spring MVC.
  • Involved in writing the POJO code and making it consistent with the existing code.
  • Involved in interacting and arranging a live meeting session with the team member to be clear and to be on the same page.
  • Involved in testing the code locally and fixing the bugs reported by the QA.
  • Involved in debugging the code and replicating the problems reported by the tester.
  • Wrote different UTPs for the team and the tester and also worked-through the project in the live meeting session.
  • Worked with the SERENA dimension to check-in and check-out the project.
  • Used RAD 7.0 to develop the Web Application and deployed on WebSphere application server.

Environment: RAD 7.0, WebSphere 6.1, JAVA, JSP, HTML, JavaScript, CSS, Serena Dimensions, DB2, JavaBeans, Design Pattern, ANT.

Confidential

Java/J2EE Developer

Responsibilities:

  • Used WebSphere, which has high performance and full-integrated Java platform for Enterprise Applications.
  • Actively involved in component development, deployment for the application interface.
  • Strongly followed the coding standards and implemented MVC Design Patterns.
  • Involved in creating EJBs that handle business logic and persistence of data.
  • Understood the existing code and involved in helping and explaining the team members about the flow of the code.
  • Involved in impact analysis of Change requests and Bug fixes.
  • Unit testing and integration testing of the modules.
  • Involved in testing the code locally and fixing the bugs reported by the QA.
  • Involved in the configuration management using ClearCase.
  • Detecting and resolving errors/defects in the quality control environment.
  • Using Ibatis for mapping Java classes with database.
  • Involved in Code review and integration testing.
  • Integrated the modules with the other modules of the system.
  • Wrote different UTPs for the team and the tester and also worked-through the project in the live meeting session.
  • Java Naming/Directory Interface (JNDI) to support transparent access to distributed components.

Environment: Sybase, Web Sphere Studio Application Developer WSAD, Enterprise Java Beans (EJB), Struts, Web Sphere Application Server, HTML, Java.

We'd love your feedback!