We provide IT Staff Augmentation Services!

Hadoop/spark Developer Resume

5.00/5 (Submit Your Rating)

Boca Raton, FL

SUMMARY:

  • 8 years of professional experience involving project development, implementation, deployment and maintenance using Java/J2EEand Big Data related technologies.
  • Hadoop Developer with 3+ years of working experience in designing and implementing complete end - to-end Hadoop Infrastructure using HDFS, MapReduce, Spark, Yarn, Kafka, PIG, HIVE, Sqoop, Storm, Flume, Oozie, Impala, HBase, Zookeeper, etc.
  • Good experience in creating data ingestion pipelines, data transformations, data management, data governance and real time streaming at an enterprise level.
  • Profound experience in creating real time data streaming solutions using Apache Spark/Spark Streaming, Kafka and Flume.
  • Expertise developing MapReduce jobs to scrub, sort, filter, join and query data.
  • Experience developing PigLatin and HiveQL scripts for Data Analysis and ETL purposes and also extended the default functionality by writing User Defined Functions (UDFs), (UDAFs) for custom data specific processing.
  • Good Hands-on experience on full life cycle implementation using MapR, CDH (Cloudera) and HDP (Hartonworks Data Platform).
  • In depth understanding of Hadoop Architecture and its various components such as Resource Manager, Application Master, Name Node, Data Node.
  • Profound understanding of Partitions and Bucketing concepts in Hive and designed both Managed and External tables in Hive to optimize performance.
  • Experience in handling messaging services using Apache Kafka.
  • Experience with migrating data to and from RDBMS and unstructured sources into HDFS using Sqoop.
  • Worked on NoSQL databases including HBase, Cassandra and Mongo DB.
  • Experience in working with Java HBase API for ingestion processed data to Hbase tables.
  • Experience with Oozie Workflow Engine to automate and parallelize Hadoop Map/Reduce, Hive and Pig jobs.
  • Analyzed data with Hue, using Apache Hive via Hue’s Beeswax and Catalog applications.
  • Experienced in using Agile methodologies including extreme programming, SCRUM and Test Driven Development (TDD).
  • Excellent global exposure to various work cultures and client interaction with diverse teams.
  • Proficient in developing Web based user interfaces using HTML5, CSS3, JavaScript, jQuery, AJAX, XML, JSON, jQuery UI, Bootstrap, AngularJS, Node JS, Ext JS.
  • Experienced in implementing dynamic functionality to screens using Ext JS, React and Experienced in working with XML technologies like XML, XSL, XSD, XSLT, XQuery, XPath and DTD.
  • Extensive experience with Frameworks in Struts, ORM (Hibernate) and Spring, (SpringMVC, SpringAOP, Spring Context Dependency Injection, SpringJDBC, SpringDAO, Spring ORM, Spring Security and Spring Boot).
  • Experience in Service Oriented Architecture (SOA), Web Services design development and deployment using WSDL,SOAP, UDDI, JAXP, JAXB, SAX, DOM, XSLT, Apache AXIS,REST web services.
  • Hands on experience writing web services both client side and end point using SOAP and RESTful (JAX-RS, JAX-WS, and JAXP).
  • Expertise in back-end/server- side java technologies such as Enterprise Java Beans (EJB), Java Persistence API (JPA), Java Messaging Service (JMS), Java Connector Architecture (JCA),Java Data Base Connectivity (JDBC).
  • Expertise in using ANT, Maven, Jenkins and Gradle for building and deploying application on the servers.
  • Strong experience in deployment and configuration of Application Servers such as IBM WebSphere, BEA WebLogic, JAX-WS, JBOSS and Apache Tomcat.

TECHNICAL SKILLS:

Programming Languages: C, C++, Java JDK 1.5/1.6/1.7/1.8, SQL, PL/SQL, Pig Latin, Python, HiveQL, Scala.

Hadoop/Big Data: HDFS, MapReduce, Spark, Yarn, Kafka, PIG, HIVE, Sqoop, Storm, Flume, Oozie, Impala, HBase, Hue, Zookeeper.

NoSQL Technologies: Cassandra, MongoDB, HBase

Distributed platforms: Hortonworks, Cloudera, MapR

Visualization: Tableau, Qulickview, Microstratergy and MS Excel

ETL Tools: Talend

JAVA/J2EE Technologies: Servlets, JSP, JDBC, EJB, JAXB, JMS, JAX-RPC, JAX- WS, JAX-RS, Apache CFX.

Frameworks: Struts, Spring, Hibernate, iBatis.

Web Technologies: HTML, CSS, JavaScript, jQuery, AngularJS, Ajax, Backbone.js, React, Node.js, Ext JS, Bootstrap.

Development Tools: Eclipse, Net Beans, IBM RAD, IntelliJ, Spring tool Suite.

Databases: MySQL, MS-SQL Server, IBM DB2, Oracle.

Operating Systems: Windows XP/Vista/7/8, 10, UNIX, Linux, Mac OS.

Build Tools: ANT, Gradle, Maven, npm, Bower.

Web/ Application Servers: WebSphere, Apache Tomcat, WebLogic, JBoss.

PROFESSIONAL EXPERIENCE:

Confidential,Boca Raton,FL

Hadoop/Spark Developer

Responsibilities:

  • Developed simple to complex Map Reduce streaming jobs using Java language for processing and validating the data
  • Developed data pipeline using Map Reduce, Flume, Sqoop and Pig to ingest customer behavioral data into HDFS for analysis
  • Developed Map Reduce and Spark jobs to discover trends in data usage by users
  • Implemented Spark using Python and Spark SQL for faster processing of data
  • Implemented algorithms for real time analysis in Spark
  • Imported data from AWS S3 in to Spark data frames, Performed transformations and actions on data frames
  • Used Spark for interactive queries, processing of streaming data and integration with popular NoSQL database for huge volume of data
  • Used the Spark - Cassandra Connector to load data to and from Cassandra.
  • Real time streaming the data using Spark with Kafka
  • Handled importing data from different data sources into HDFS using Sqoop and also performing transformations using Hive, Map Reduce and then loading data into HDFS.
  • Exported the analyzed data to the relational databases using Sqoop, to further visualize and generate reports for the BI team
  • Collecting and aggregating large amounts of log data using Flume and staging data in HDFS for further analysis
  • Analyzed the data by performing Hive queries (Hive QL) and running Pig scripts (Pig Latin) to study customer behavior
  • Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting
  • Developed Pig Latin scripts to perform Map Reduce jobs
  • Developed product profiles using Pig and commodity UDFs
  • Developed Hive scripts in Hive QL to de-normalize and aggregate the data
  • Created HBase tables and column families to store the user event data
  • Written automated HBase test cases for data quality checks using HBase command line tools
  • Created UDF’s to store specialized data structures in HBase and Cassandra
  • Scheduled and executed workflows in Oozie to run Hive and Pig jobs
  • Used Impala to read, write and query the Hadoop data in HDFS from HBase or Cassandra
  • Used Tez framework for building high performance jobs in Pig and Hive
  • Configured Kafka to read and write messages from external programs
  • Configured Kafka to handle real time data
  • Developed end to end data processing pipelines that begin with receiving data using distributed messaging systems Kafka through persistence of data into HBase
  • Installed Solr on web servers to index the search data and performed real time updates.
  • Developed core search component using Solr
  • Continuous monitoring and managing the Hadoop cluster using Cloudera Manager
  • Used JUnit framework to perform Unit testing of the application
  • Developed interactive shell scripts for scheduling various data cleansing and data loading process
  • Performed data validation on the data ingested using Map Reduce by building a custom model to filter all the invalid data and cleanse the data
  • Experience with data wrangling and creating workable datasets
  • Developed schemas to handle reporting requirements using Jaspersoft

Environment: Hadoop, Spark, Map Reduce, Pig, Hive, Sqoop, Oozie, HBase, Zoo keeper, Kafka, Flume, Solr, Tez, Impala, Mahout, Cassandra, Cloudera manager, AWS S3, MySQL, Jaspersoft, Multi-node cluster with Linux-Ubuntu, Windows, Unix.

Confidential,Northlake,IL

Hadoop Developer

Responsibilities:

  • Responsible for the analysis, documenting the requirements and architecting the application based on J2EE standards
  • Worked on automation of delta feeds from Teradata using Sqoop, also from FTP Servers to Hive
  • Implemented Hive tables and HQL Queries for the reports. Written and used complex data type in Hive
  • Developed Hive queries to analyze reducer output data
  • Designed workflow by scheduling Hive processes for Log file data, which is streamed into HDFS using Flume
  • Developed MapReduce (YARN) programs to cleanse the data in HDFS obtained from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis
  • Designed and developed Map Reduce jobs to process data coming in different file formats like XML, CSV, JSON
  • Used Sqoop to import the data from RDBMS to Hadoop Distributed File System (HDFS) and later analyzed the imported data using Hadoop Components
  • Involved in loading and transforming large sets of Structured, Semi-Structured and Unstructured data and analyzed them by running Hive queries and Pig scripts
  • Implemented daily workflow for extraction, processing and analysis of data with Oozie
  • Part of application development team for refactoring existing code and deliver upcoming new Change Requests for the ongoing live application which is developed using Spring MVC, Spring Batch, Hibernate and Spring Web Services
  • Designed and Developed UI using Angular.JS 2.0, HTML5, CSS3, Bootstrap, JavaScript and jQuery
  • Responsible for design and developing Persistence classes using Hibernate, and spring boot data Template frameworks to save data in database tables
  • Build REST web service by building Node.js Server in the back-end to handle requests sent from the front-end jQuery Ajax calls
  • Used Spring MVC Model View Controller to handle/intercept the user requests and used various controllers to delegate the request flow to the Backend tier of the application
  • Configured spring to manage Actions as beans and set their dependencies in a context file and integrated middle tier with Hibernate
  • Produced Web Services using SOAP and Apache CFX
  • Developed Spring Boot based Micro Services& implemented Spring cloud/Netflix API architecture patterns (Eureka Service discovery, Configuration server)
  • Configured the xml files for Spring Security injection to ensure the back-end security by discriminating the role of users as well as access authority for the URIs and Server-side assets
  • Built DAO package as well as DAO IMPL package to implement the data access logic, and using Hibernate Annotation to do the data mapping
  • Wrote Hibernate configuration files to enable the data transactions between POJO and Oracle Database
  • Implemented Agile practices for workflow process, epic/user stories, scrums, sprints, releases and reporting
  • Converted and automated builds using Maven and Make. Scheduled automated nightly builds using Jenkins

Environment: Java, J2EE, Hibernate, Spring, Web services, REST, HTML, JavaScript, jQuery, XML,XSD, Log4j, Oracle 11g, PL/SQL, JUnit, JIRA, MongoDB, Hadoop, Hive, HQL, HDFS, MapReduce, Sqoop, Flume, Oozie, Python, Java, Maven, Eclipse, Putty, Cloudera Manager 4 and CDH 4.

Confidential, Madison, WI

Sr. Java Developer/Hadoop Developer

Responsibilities:
  • Part of a team that designed and implemented the user interface for Eligibility Verification System using the 270/271 HIPAA guidelines. I gained invaluable experience in user experience issues for enterprise web applications
  • Responsible for successfully creating the Design document for all the modules of discipline of the application
  • Developed Custom Tags to simplify the JSP code. Designed UI screens using JSP and HTML tags. Used JavaScript for client side validation
  • Developed Service (EJB) components for middle tier and implementation of business logic using J2EE Design patterns on WebLogic App Server
  • Used Custom Framework like Struts 2.0 in the application, which is based on MVC design pattern
  • Designed and implemented distributed Java/J2EE enterprise-level applications that model clients’ processes, practices, and concepts using JMS (Java Messaging Service) to create, send messages to and receive messages from MQ Series
  • Developed jobs which were responsible for scheduling and running the application batch jobs
  • Developed the Maven pom.xml for building and deploying the application using Maven 2.0.7
  • Used DAO pattern to retrieve the data from database
  • Developed store procedures on DB2 to built agent hierarchy
  • Used Web sphere for deploying the application. Created the database connections pools in the Web Sphere for accessing the database
  • Testing the Webservices exposed using SOAP UI
  • Involved in the production release support calls for the application
  • Used JUnit framework for Unit testing of application
  • Monitored the error logs using Log4J and fixed the problems
  • Controlled the version management for the application with of Tortoise SVN
  • Developed scalable distributed data solutions using Hadoop
  • Installed and Configured Hive, Pig, Sqoop and Oozie on the Hadoop cluster
  • Designed and developed data integration programs in Hadoop Environment with RDBMS stores for data access and analysis
  • Load and transform large sets of structured, semi-structured and unstructured data
  • Analyzed large sets of data by running Hive queries and pig Latin scripts

Environment: Struts 2, Java, J2EE, EJB 3, JSP, JMS, RMI, Spring, Hibernate, Java Servlets, JavaScript, Java Mail, Log4J, Maven 2.0.7, Tortoise SVN, Design Patterns, DHTML, HTML, UNIX, Windows 7, IBM RAD, Web Sphere, DB2, Hadoop 1.x, Map-Reduce, HDFS, Hive, Pig, Sqoop, Oozie, Cloudera.

Confidential,Atlanta GA

Sr. Java Developer

Responsibilities:
  • Developed whole Confidential Staff application with more than 16 features from end to end
  • Involved in Requirements gathering with Business Teams
  • Developed Prototype of the application and getting the approval from business teams.
  • Responsible for Portal Design, Technical design documents
  • Developed the user interface using JSPs and validations using JavaScript and JQuery
  • Developed the Controllers using Spring MVC framework
  • Developed the Validators for handling form field validations
  • Developed the Adapters for Data processing
  • Developed DAOs and Model mappings used Hibernate ORM
  • Developed Hibernate Model Objects
  • Created Data Dictionary and Logical Data model
  • Implemented Single Sign on by using Spring Security framework with SAML extension
  • Java Persistent API is used for Transaction management
  • Responsible for moving the code from local to Dev environment
  • Responsible for moving the code from Dev environment to UAT environment
  • Responsible for resolving the defects raised by QA team

Environment: Java 1.7, J2EE, Spring MVC, Spring Security using IDP, Hibernate ORM, Eclipse 3.6, Tomcat 7.1, DB2, SQL, HTML, XML, Java Script, JQuery, AJAX, Log4j, Putty, winscp, SQuirreL, QMF

Confidential,Manchester,UK

Programmer Analyst

Responsibilities:
  • Involved in designing of the application
  • Carried out development of the website using IBM WSAD and J2EE patterns ( Value object and Business Delegates)
  • Developed JSP pages and Servlets to handle client requests and carry out server side Processing. Implemented client side validations using JavaScript
  • Developed the application using OOP
  • Implemented logging using Log4j
  • Used WSAD integrated VSS to maintain version controlled source code
  • Performed IBM WebSphere Application Server configuration - Setup the configuration files for deploying the application. Created data sources and configured connection pooling
  • Developed Java Batches using Custom Thread framework
  • Developed Online Applications using Struts Framework
  • Written implementation plan and job cards
  • Written JUnits and Unit Test Plans
  • Interacted with Business/Marketing in defining requirements, designed, provided technical support to QA and post-deployment support

Environment: Java, JSP, Jakarta Struts1.1, WSAD 5.1, RAD 6, UDB DB2 and WAS 6.1, VSS

Confidential

Trainee \ Programmer Analyst

Responsibilities:
  • Deriving the Use cases following RUP, UML & OOAD techniques
  • Involved in developing the Software design document using Rational Rose 2000 and Microsoft Visio
  • Analyzed and designed the application components by creating Object Model using UML and Rational Rose following the Rational Unified Process (RUP)
  • Extensively used JFC/Swing to developed Client and Admin screens, which contains various customized swing components
  • Used Struts 1.01/1.02 framework for MVC architecture
  • Developed a web-based reporting system with HTML, JavaScript and JSP using Struts framework
  • Used SQLJ to access the oracle database from CAM application
  • Developed the ANT scripts for preparing EAR and WAR files used to deploy J2EE components

Environment: Core Java, Struts, JSP, Servlets, Java Beans, UML/Rational Rose 2000, Java Script, SQLJ, JFC/Swing, JUnit, Log4J, HTML, MS Excel, and Oracle.

We'd love your feedback!