Sr. Hadoop/spark Developer Resume
Cary, NC
PROFESSIONAL SUMMARY:
- 8+ years of overall IT experience in a variety of industries,which includes hands on experience of 3+years in Big Data technologies and extensive experience of 5 years in Java.
- In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node, and MapReduce conceptsand experience in working with MapReduce programs using Apache Hadoop for working with Big Data to analyze large data sets efficiently
- Hands on experience in working with Ecosystems like Hive, Pig, Sqoop, Map Reduce, Flume, Oozie.Strong knowledge of Pig and Hive’s analytical functions, extending Hive and Pig core functionality by writing custom UDFs
- Experience in importing and exporting terra bytes of data using Sqoop from HDFS to Relational Database Systems and vice - versa
- Knowledge of job workflow scheduling and monitoring tools like Oozie and Zookeeper, ofNoSQL databases such as HBase, Cassandra, and of administrative tasks such as installing Hadoop, Commissioning and decommissioning, and its ecosystem components such as Flume, Oozie, Hive and Pig.
- Knowledge of Cloud technologies like AWS Cloud.
- Hands on experience implementing ApacheSparkusing Scala.
- Experience in design, development and testing of Distributed, Internet/Intranet/E-Commerce, Client/Server and Database applications mainly using technologies Java, EJB, Servlets, JDBC, JSP, Struts, Hibernate, Spring, JavaScript on WebLogic, Apache Tomcat Web/Application Servers and with Oracle and SQL Server Databases on Unix, Windows platforms
- Extensive work experience in Object Oriented Analysis and Design,Java/J2EE technologies including HTML, XHTML, DHTML, JavaScript, JSTL, CSS, AJAX and Oracle for developing server-side applications and user interfaces.
- Experience in developing Middle-tier components in distributed transaction management system using Java. Good understanding of XML methodologies (XML,XSL,XSD) including Web Services and SOAP
- Extensive experience in working with different databases such as Oracle, IBM DB, RDBMS, SQL Server, MySQL and writing Stored Procedures, Functions, Joins and Triggers for different Data Models.
- Experience with Agile Methodology, Scrum Methodology, software version control and release management
- Handled several techno-functional responsibilities including estimates, identifying functional and technical gaps, requirements gathering, designing solutions, development, developing documentation, and production support
- An individual with excellent interpersonal and communication skills, strong business acumen, creative problem-solving skills, technical competency, team-player spirit, and leadership skills
TECHNICAL SKILLS:
Database: Teradata, DB2, MySQL, Oracle, MS SQL Server, IMS/DB
Languages: Java, PIG Latin, SQL, HiveQL, Scala, Shell Scripting
API s/Tools: Mahout, Eclipse, Log4j, Maven
Web Technologies: HTML, XML, JavaScript
BigData Ecosystem: HDFS, PIG, MAPREDUCE, HIVE, SQOOP, FLUME, OOZIE, HBase, MongoDb, AWS,Solr search, Impala, Cassandra, Apache Storm, Flume, Spark, KafkaOperating System:Unix, Linux, Windows
BI Tools: Tableau, Pentaho
PROFESSIONAL EXPERIENCE:
Confidential, Cary, NC
Sr. Hadoop/Spark Developer
Responsibilities:
- Used Spring MVC framework to develop IOT web application
- Loading data from different servers to S3 bucket and setting appropriate bucket permissions.
- Developed Spark jobs and ran them on Amazon EMR.
- Develop Simple to complex MapReduce Jobs using Hive, Pig, Java Map reduce and spark using Scala
- Developed Spark code and Spark-SQL/Streaming for faster testing and processing of data.
- Used Spark RDD’s to convert Hive Queries into Spark transformation.
- Worked on Apache Spark for data lake creation for building RWI (Real World Intelligence) application
- Apache Kafka consumer to get click streamto get the customer insight
- Cassandra data modelling for storing and transformation in spark using datastax connector
- Installed Apache Solr cloud on cluster and configured it with Zookeeper
- Worked on Apache SolrCloudto indexdocuments using hive-solr storage handler to import different datasets including xml,csv,and json
- Created and maintained Technical documentation of Hadoop clusters
- Used Kerberos security for authentication of users and ssl/tls for encryption of data and sentry for role based access to DBCDH 5.4: Hadoop Cluster
- Used git repository and Jenkins to build the application
Environment: Hadoop, Amazon Cloud Service like EMR, S3, HDFS, Pig, Hive,MapReduce shell scripting, Spark and Big Data.
Confidential, Irving, TX
Hadoop Developer
Responsibilities:
- Analysed large data sets by running custom map reduce, Hive queries and Pig scripts
- Complex pig udf for business transformations
- Worked with the Data Science team,Teradata team, and business to gather requirements for various data sources like webscrapes, APIs
- Involved in creating Hive/Impala tables, and loading and analysing data using hive queries
- Involved in running Hadoop jobs for processing millions of records and compression techniques
- Developed multiple MapReduce jobs in java for data cleaning and pre-processing
- Involved in loading data from LINUX file system to HDFS, and wrote shell scripts for productionizing the MAP(Member Analytics Platform) project and automated using cronacle scheduler
- Load and transform large sets of structured, and semi structured data
- Loaded Golden collection to Apache Solr using Morphline code for Business team
- Assisted in exporting analysed data to relational databases using Sqoop
- Data Modelled for Hbase for large transaction sales data
- Proof of Concept on Strom for streaming the data from one of the sources
- Implementation of one of the data source transformations in spark using scala
- Cassandra DataModel and design to connect with spark
- Teradata Fast Export and Parallel Transporter utilities and Sqoop to extract data and load to hadoop
- Worked in Agile methodology and used iceScrum for Development and tracking the project
- Worked on Git hub repository, branching, merging, etc
Environment: Hadoop, HDFS, Pig, Hive, IMPALA, Solr, Morphline, MapReduce, Sqoop, HBase, shell Pentaho, Spark, scala, Teradata Parallel Transporter,github, ApacheStorm, spark, CDH5.0, HDP and Big Data
Confidential
Sr. Java Developer
Responsibilities:
- Developed web components using JSP, Servlets and JDBC
- Designed tables and indexes
- Created Design specification using UML Class Diagrams, Sequence & Activity Diagrams
- Developed the Web Application using MVC Architecture, Java, JSP, and Servlets & Oracle Database.
- Developed various Java classes, SQL queries and procedures to retrieve and manipulate the data from backend Oracle database using JDBC
- Extensively worked with Java Script for front-end validations.
- Designed, Implemented, Tested and Deployed Enterprise Java Beans both Session and Entity using WebLogic as Application Server
- Developed stored procedures, packages and database triggers to enforce data integrity. Performed data analysis and created crystal reports for user requirements
- Provided quick turn around and resolving issues within the SLA.
- Implemented the presentation layer with HTML, XHTML and JavaScript
- Used EJBs to develop business logic and coded reusable components in Java Beans
- Development of database interaction code to JDBC API making extensive use of SQL
- Query Statements and advanced Prepared Statements.
- Used connection pooling for best optimization using JDBC interface
- Used EJB entity and session beans to implement business logic and session handling and transactions. Developed user-interface using JSP, Servlets, and JavaScript
- Wrote complex SQL queries and stored procedures
- Actively involved in the system testing
- Prepared the Installation, Customer guide and Configuration document which were delivered to the customer along with the product
Environment: Windows NT 2000/2003, XP, and Windows 7/ 8 C, Java, UNIX, and SQL using TOAD Finacle Core banking, CRM 10209, Microsoft Office Suit, Microsoft project
Confidential
Java Developer
Responsibilities:
- Involved in the design and development phases of Rational Unified Process (RUP).
- Designed Class Diagrams, Sequence Diagrams and Object Diagrams using IBM Rational Rose to model
- Application was built on MVC architecture with JSP 1.2 acting as presentation layer, Servlets as controller and Developed the application using Jakarta Struts 1.1 Framework: developed action classes, form beans and
- Used Struts Validation Framework for validating front end forms.
- Extensively used XML Web Services for transferring/retrieving data between different providers.
- Developed complete Business tier with Session beans and CMP Entity beans with EJB 2.0 standards using JMS Queue communication in authorization module.
- Designed and implemented Business Delegate, Session Facade and DTO Design Patterns
- Involved in implementing the DAO pattern
- Used JAXB API to bind XML Schema to java classes
- Used the report generation in the databases written in PL/SQL
- Used Maven for building the enterprise application modules
- Used Log4J to monitor the error logs
- Used JUnit for unit testing
- Used SVN for Version control
- Deployed the applications on WebLogic Application Server.