Big Data Engineer/java Resume
HoeniX
SUMMARY:
- Big data Engineer with 7 Years of IT experience including 4 years in Big data domain and 3 years in Java Development.
- Extensively worked on Hadoop eco system tools - MapReduce, Hive, Pig, Sqoop, Oozie, Impala and Spark.
- Experience in understanding of data processing patterns, distributed computing and building applications for real-time and batch analytics.
- Experience in productionizing the code from ground up according industry leading best practices which includes setting up Git Repos and Integrating them to Jenkins CI/CD.
- Experience with Java8, Scala and Python languages used for automation and building frameworks for data processing in Spark, Hadoop and implement solution in Object Oriented pattern.
- Good knowledge on Amazon Web Services(AWS) cloud services like EC2, S3 and EMR.
- Coding and design skills in Core Java, OOPS Concepts, Servlets and JSP.
- Experience on OOP (Objective Oriented Programming) like Inheritance, polymorphism, Abstraction, Encapsulation.
- Strong Understanding of J2EE design Patterns like Session facade, Singleton, Front Controller, Business delegate, Service Locator, DTO, DAO.
- Experience programming in Scala with exposure to the most well-known frameworks such as Akka and Play.
- Extensive experience in implementing MVC framework using Spring in J2EE architecture
- Good exposure and experience with Service Oriented Architecture (SOA) such as Web Services including SOAP, WSDL, JMS (Java Messaging Service), AJAX and tools like ANT, Junit.
- Experience in implementing end-to-end solutions full life cycle (SDLC).
TECHNICAL SKILLS:
Big Data Ecosystem: AWS, HDFS, MapReduce, YARN, Hive, Sqoop, Impala, Spark.
Cloud Ecosystem: Amazon Web services (EC2, S3)
Languages: Scala, Java, SQL, HTML, DHTML, JavaScript
Operating systems: UNIX, Linux, and Windows Variants
Tools: Maven, Gradle, Jenkins, IntelliJ, Eclipse, GIT.
PROFESSIONAL EXPERIENCE:
Confidential - PHOENIX
Big Data Engineer/Java
Responsibilities:
- Migrated the existing data to Hadoop from Mainframes and Teradata to Hadoop Cluster and performed data processing through map-reduce framework.
- Designed architecture for end-to end workflow in migrating process from Mainframe to Hadoop environment.
- Developed map-reduce programs and custom Udfs in hive and pig for various data transformations.
- Implemented Map-reduce paradigm in Spark with Scala API.
- Developed Pig, Hive scripts from Teradata sql scripts to process data in hadoop.
- Used Oozie to automate the workflow and Autosys for scheduling the jobs in cluster in timely manner to meet the SLA for various portfolios existing in retail risk.
- Prepared spark scripts to replace existing pig scripts to process data in spark framework.
- Designed and implemented Data check and Data quality frameworks in spark during the initial load process and the final publish stages.
- Used Jenkins as build tool and GIT as code merge platform.
- Experienced in Agile Methodology, participated in Sprints and daily Scrums to deliver software tasks on-time and with good quality on basis with onsite and offshore teams.
- Extensively used ePass to integration with Enterprise and web applications.
- Core Service is developed as jar or image that can be included by other MicroServices as their dependency.
- Docker is used for creating images that are deployed on epass as MicroServices.
- Created auto deploy jobs using Jenkins, Gradle and Maven.
- Created and maintained various DevOps related tools for the team such as deployment tools and development and staging environments on ePass and Nexus.
- Used Jenkins tool and made customizations according to the local build to track out the build status.
Environment: Groovy, Java 1.8, ePass, Nexus, Docker, Kubernetes, MicroServices, Agile, Java, Beans, XML, Maven, Jenkins, Gradle,Spark,Hadoop, Hive,Sqoop,Oozie,aws.
Confidential
Big Data Engineer/Java
- Involved in project Life Cycle - from analysis to production implementation, with emphasis on identifying the source and source data validation, developing logic and transformation as per the requirement and creating mappings and loading the data into different targets.
- Extracted data from Oracle SQL server and MySQL databases to HDFS using Sqoop .
- Created Hive tables to store the processed results and written Hive scripts to transform and aggregate the disparate data.
- Driving the application from development phase to production phase using Continuous Integration and Continuous Deployment (CICD) model using Maven and Jenkins.
- Automated the process for extraction of data from warehouses and weblogs into Hive tables by developing workflows and coordinator jobs in Oozie .
- Exported the aggregated data onto RDBMS using Sqoop for creating dashboards in the Tableau .
- Utilized Agile Scrum Methodology to manage and organize the team with regular code review sessions.
- Worked with data in multiple file formats including Avro, Parquet, Sequence files, ORC, JSON and Text/CSV .
- Very good understanding of Partitions, bucketing concepts in Hive and designed both Managed and External tables in Hive to optimize performance.
- Used JIRA for incident creation, bug tracking and change management process.
Environment: Java, Eclipse, Hadoop, Map Reduce, HDFS, Sqoop, Oozie, WinScp, UNIX Shell Scripting, HIVE, Impala, Cloudera (Hadoop distribution), AWS, JIRA etc.
Confidential, Raleigh- NC
Sr. Big Data Engineer
Responsibilities:
- Designed and implemented Object Oriented Singleton Pattern for the Scala Application and created a library for sourcing and transformations in Spark.
- Created Projection step to implement Machine learning algorithm Linear Regression in spark for predictions for version 2.2.0.
- Implemented the Spark best practices to efficiently process data to meet ETAs by utilizing features like partitioning, resource tuning, memory management and Check pointing features.
- Implemented a testing framework in Scala by using ScalaTest and SparkTestingbase to test the application and automated through Jenkins.
- Established Git Lab Repository for the project and implemented CI/CD for the repository to Jenkins Server.
- Created Jenkins Pipeline to implement SBT to compile, test, assembly and publish.
- Created a remote debugger through IntelliJ to debug spark application.
- Extensively used Type Safe Configuration to Parameterize the Spark application using Scala.
Environment: Python, Java, Maven, Jenkins, Gradle, Spark, Hadoop, Hive, Sqoop, Oozie, aws.
Confidential, New York, NY
Software Engineer
Responsibilities:
- Effectively interacted with team members and business users for requirements gathering.
- Understanding functional specifications and documenting technical design documents.
- Used Spring MVC to decouple business logic and view components.
- Involved in the integration of spring for implementing Dependency Injection (DI/IOC).
- Developed the Business Logic using Plain Old Java Objects ( POJOs ).
- Developed Graphical User Interfaces using HTML and JSP’s for user interaction.
- Created Ext JS pages, used JavaScript for client-side validations, and AJAX to create interactive front-end GUI.
- Used JSON for data exchange between application modules along with XML.
- Used Direct Web Remoting, Spring JDBC Template and Spring DAO’s Support classes for JDBC.
- Created set of classes using DAO pattern to decouple the business logic and data.
- Used various Core Java concepts such as Multi Threading, Exception Handling, Collection APIs to implement various features and enhancements.
- Designed and Developed complex UI screens using Spring MVC, AJAX , jQuery , HTML5, CSS3, Python, Java Script for request management and search and filter modules.
- Involved in coding, maintaining, and administering Servlets , and JSP components to be deployed on a WebSphere Application server application server.
- Developed Web Services using XML messages that use SOAP . Developed Web Services for Payment Transaction and Payment Release.
- Extensively worked with VXML, GRXML in implementing the call flow for the IVR application. Implemented Speech-to-Text and Text-to-Speech using Nuance.
- Built scripts using Ant to build the J2EE application.
- Responsible for creation and consumption of Web services both SOAP and REST based.
- Developed the application using RAD as the IDE and used its features for editing, debugging, compiling, formatting, build automation and version control ( CVS ).
- Wrote test cases in Junit for unit testing of classes and implemented the logging using Log4j.
- Developed test cases for Unit Testing using JUnit and Mockito.
- Built NO SQL schemas and maintaining sessions by Mongo DB.
- Used CVS version control to track and maintain the different version of the application.
- Involved in offshore co-ordination, work assignment, query resolution and work product reviews.
- Extensive hands-on experience in Pivotal cloud foundry, other PaaS platforms and In-depth understanding of Domain Driven Design and 12 Factor principles.
Environment: Core Java 6, J2EE, JSP, JSTL, Servlets, Ext JS, Agile Methodology, Struts, MVC, Tomcat/JBoss, Python, XML, HTML, CSS, DHTML, DOM, Hibernate, SOAP, REST, JavaScript, Multithreading, Oracle, Junit, No SQL,Mongo DB, Junit, Web services, PLSQL, JDBC, ANT, Solaris and Windows 2000.
Confidential, Hartford, CT
Software Engineer
Responsibilities:
- Participated in development phases of the project with major emphasis on coding and unit testing of the modules.
- Development methodology used was the agile methodology. Process included iterative development, 2-week sprints, and daily stand.
- Developed the presentation layer on Struts framework using Servlets and JSP.
- Used Eclipse for developing web components such as JSP, Controller tier that includes action classes.
- Designed and developed the business logic components using facade design pattern.
- Created war/ear files and deployed them on WebSphere application server.
- Developed various stored procedures, functions, and packages and materialized views along with JDK and JDBC for backend and middle level data access layer between Java and Oracle in OLTP application.
- Developed PL/SQL stored procedures to perform complex database operations.
- Front-end designing and validation were done using the JavaScript.
- Prepared Test Case document to test the required functionalities.
- Performed Unit as well as integration testing for all the modules.
- Created Technical documentation as per the company standards.
- Used Ant scripts to fetch, build and deploy the application to development environment.
- Enhanced the system with change requests and coordinating with business and development team.
- Interacted with DBAs, Data Architects, Application Team, Infrastructure team.
Environment: Java, J2EE, JSPs, Servlets, JDBC, Web Services, SOAP/HTTP, Struts, WebSphere, XML, XSLT, JavaScript, HTML, NetBeans, Oracle, Windows.