Senior Hadoop Developer Resume
Herndon, VA
SUMMARY:
- Senior Hadoop Developer with 7+ years of experience in analysis, design, development and implementation of large - scale web-based applications using Big Data, Hadoop, Map-Reduce, Storm, HIVE, HBase, Core Java, J2EE and related Technologies.
- Good Exposure on Apache Hadoop Map Reduce programming, Hive, PIG scripting and HDFS.
- Strong knowledge of Big Data and provided Hadoop solutions to related use cases in Big data.
- Hands on experience in Import/Export of data using SQOOP Hadoop Data Management tool.
- Strong experience in writing Map Reduce programs, Hive and Pig for Data Analysis.
- Hands on experience in writing custom partitioners for Map Reduce.
- Excellent understanding and knowledge of NoSQL databases like HBase.
- Experience with installing, backup, recovery,configuring, DR and development on multiple Hadoop distribution platforms like Hortonworks Distribution Platform(HDP), Cloudera Distribution for Hadoop (CDH).
- Experience with integration of HDP and CDH to data integration tools and BI tools.
- Experience with HDP security implementations.
- Experienced in working with Struts, Hibernate and Spring MVC frameworks.
- Extensively worked with JBoss, IBM WebSphere application servers and Tomcat WebServer.
- Experience working with JAVA, J2EE, JDBC, ODBC, JSP, Java Beans, EJB, Servlets, Java Web Services and related technologies using IDEs like IntelliJ, Eclipse and NetBeans.
- Development experience in DBMS like Oracle, MS SQL Server and MYSQL.
- Experience in writing, diagnosing and solving performance queries in MySQL, DB2 and Stored Procedures in MySQL.
- Exposure to working on Oracle 10.x database using SQL (DML, DDL Queries)
- Experience in implementing projects using Waterfall, RUP, Agile Methodologies and exposure to SCRUM project implementation methodology.
- Hands on Design and development of the projects and mentoring teams technically while working on applications.
- Strong knowledge of algorithms and implementations.
- Thorough understanding of Data Warehouse and ETL concepts.
- Extensive experience working on CVS, Clear Case, SVN and GIT for Source Controlling.
- Experience in leading and managing Post Production Support teams.
- Vast experience working on all phases of System Development Life Cycle (SDLC) including, but not limited to Design, Development, Testing, Implementation and Post Production Support.
- Excellent analytical ability, Project Management and Presentation skills.
- Quick learner of business processes with excellent and proven analytical, trouble shooting and problem solving skills.
- Very good at development driven testing and working experience with QA’s.
- Strong Ability to handle multiple tasks and work independently as well as in a team.
- Have Strong analytical skills with ability to follow project standards and decision - making capability.
- Exceptional ability to quickly master new concepts and applications.
- Very good at communicating and addressing the customer issues.
TECHNICAL SKILLS:
Methods/Tools: Hadoop 2.x, Big Data, MapReduce, Pig, HDFS, Hive, ZooKeeper, Hbase 0.94.8, Sqoop, Oozie, Ambari, Java, J2EE, MySQL, HTML, Java Script, XML, UNIX Shell Scripting, C/C++, PL/SQL, Windows XP/Vista, Mac OS, UNIX, LINUX. HTML, XML, JSON, Web Services, Servlets, JSP.JSON, JBoss,AWS, Web Sphere 5.x. GIT, JIRA, Clear Quest, Clear Case, CVS, SVN, Apache Maven 2.2.1, Spring MVC, Struts, Hibernate, Log4j, Eclipse, IntelliJ, WebSphere Application Developer (WSAD - 5.x), NetBeans. MySQL 5.x, Oracle.
PROFESSIONAL EXPERIENCE:
Senior Hadoop Developer
Confidential, Herndon, VA
Responsibilities:
- Responsible for building scalable distributed data solutions using Hadoop.
- Involved in Design and Development of technical specification documents using Hadoop.
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
- Managed and reviewed Hadoop log files.
- Migrated the data from Oracle, MySQL in to HDFS using Sqoop and imported various formats of flat files into HDFS.
- Experienced on loading and transforming of large sets of structured, semi structured and unstructured data from HBase through Sqoop and placed in HDFS for further processing.
- Monitored Hadoop scripts which take the input from HDFS and load the data into Hive.
- Developed MapReduce programs to parse the raw data, populate tables and store the refined data in partitioned tables.
- Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
- Developed and written Apache PIG scripts and HIVE scripts to process the HDFS data.
- Working knowledge in writing Pig’s Load and Store functions.
- Installed Oozie workflow engine to run multiple Hive and Pig jobs.
- Defined job work flows as per their dependencies in Oozie.
- Maintain System integrity of all sub-components related to Hadoop.
- Worked on Apache and HDP clusters and integrated with BI tools.
- Worked on HDP security implementations,backup, recovery and DR.
- Loading log data directly into HDFS using Flume.
- Managed and monitored Apache Hadoop clusters using Apache Ambari.
- Developed and involved in the industry specific UDF (user defined functions).
- Migration of ETL processes from Oracle to Hive to test the easy data manipulation.
- Implemented test scripts to support test driven development and continuous integration
Environment: Apache Hadoop, HDFS, Hive, Map Reduce, Java, Eclipse, Hive, Pig,Avro, Oozie, MySQL, Hortonworks Distribution Platform (HDP).
Hadoop Developer
Confidential, Omaha, Nebraska
Responsibilities:
- Lead and managed team during Design, Development and Implementation phase of the application.
- As a Developer, worked directly with business partners discussing the requirements for new projects and enhancements to the existing applications.
- Wrote Java code to process streams for risk management analysis.
- Wrote extensive shell scripts to run appropriate programs
- Wrote multiple queries to pull data from Hbase
- Reporting on the project based on Agile-Scrum Method. Conducted daily Scrum meetings and updated JIRA with new details.
- Wrote Java to pull related data from Hbase.
- Developed a custom File System plug in for Hadoop so it can access files on Data Platform.
- This plugin allows Hadoop MapReduce programs, HBase, Pig and Hive to work unmodified and access files directly.
- Designed and implemented Mapreduce-based large-scale parallel relation-learning system
- Involved in review of functional and non-functional requirements.
- Facilitated knowledge transfer sessions.
- Installed and configured Hadoop Mapreduce, HDFS and developed multiple MapReduce jobs in java for data cleaning and pre-processing.
- Importing and exporting data into HDFS and Hive using Sqoop.
- Experienced in defining job flows.
- Worked on HDP security implementations
- Wrote Pig Scripts to perform ETL procedures on the data in HDFS.
- Analyzed the data by performing Hive queries and running Pigscripts and Python Scripts.
- Used Hive to partition and bucket data.
- Used Tableau for Data Visualization.
- Load and transform large sets of structured, semi structured and unstructured data.
- Responsible to manage data coming from different sources.
- Got good experience with NoSQL database.
Environment: Java 1.6, Hadoop 2.2.0 (Yarn), Map-Reduce, Hbase-0.94.8, Storm 0.9.1, Linux Centos 5.x, Agile, SVN, Maven, Jira, Hortonworks Distribution Platform (HDP).
Hadoop Developer
Confidential, Melville, NY
Responsibilities:
- Analyzed large data sets by running Hive queries and Pig scripts
- Worked with the Data Science team to gather requirements for various data mining projects
- Involved in creating Hive tables, and loading and analyzing data using hive queries
- Developed Simple to complex MapReduce Jobs using Hive and Pig
- Involved in running Hadoop jobs for processing millions of records of text data
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required
- Developed multiple MapReduce jobs in java for data cleaning and preprocessing
- Involved in loading data from LINUX file system to HDFS
- Responsible for managing data from multiple sources
- Extracted files from CouchDB through Sqoop and placed in HDFS and processed
- Experienced in runningHadoopstreaming jobs to process terabytes of xml format data
- Load and transform large sets of structured, semi structured and unstructured data
- Responsible to manage data coming from different sources
- Assisted in exporting analyzed data to relational databases using Sqoop
- Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts
- Responsible to manage data coming from different sources.
- Involved in loading data from UNIX file system to HDFS.
- Cluster coordination services through Zookeeper.
- Experience in managing and reviewing Hadoop log files.
- Job management using Fair scheduler.
- Involved in loading data from LINUX file system to HDFS.
- Worked on installing cluster, commissioning & decommissioning of datanode, namenode recovery, capacity planning, and slots configuration.
Environment: Hadoop, HDFS, Pig, Hive, MapReduce, Sqoop, LINUX, Big Data, Zookeeper, Cloudera Distribution for Hadoop (CDH)
J2EE Developer
Confidential, IL
Responsibilities:
- Involved in development of business domain concepts into Use Cases, Sequence Diagrams, Class Diagrams, Component Diagrams and Implementation Diagrams.
- Implemented various J2EE Design Patterns such as Model-View-Controller, Data Access Object, Business Delegate and Transfer Object.
- Responsible for analysis and design of the application based on MVC Architecture, using open source Struts Framework.
- Involved in configuring Struts, Tiles and developing the configuration files.
- Developed Struts Action classes and Validation classes using Struts controller component and Struts validation framework.
- Developed and deployed UI layer logics using JSP, XML, JavaScript, HTML /DHTML.
- Used Spring Framework and integrated it with Struts.
- Involved in Configuring web.xml and struts-config.xml according to the struts framework.
- Designed a lightweight model for the product using Inversion of Control principle and implemented it successfully using Spring IOC Container.
- Used transaction interceptor provided by Spring for declarative Transaction Management.
- The dependencies between the classes were managed by Spring using the Dependency Injection to promote loose coupling between them.
- Developed DAO using spring JDBC Template to run performance intensive queries.
- Developed ANT script for auto generation and deployment of the web service.
- Wrote stored procedure and used JAVA APIs to call these procedures.
- Developed various test cases such as unit tests, mock tests, and integration tests using the JUNIT.
- Experience in writing Stored Procedures, Functions and Packages.
- Used log4j to perform logging in the applications.
Environment: Java, J2EE, Struts MVC, Tiles, JDBC, JSP, JavaScript, HTML, Spring IOC, Spring AOP, JAX-WS, Ant, Web sphere Application Server, Oracle, JUNIT and Log4j, Eclipse.
J2EE Developer
Confidential, Newark, NJ
Responsibilities:
- Implemented J2EE Design Patterns such as Business Delegate, Front Controller, MVC, Session Facade, Value Object, DAO, Service Locator, Singleton.
- Lead Root Cause Analysis activities to successfully identify root causes of incidents.
- Assessment & Estimation of the change to be made in the application
- Implemented the database connectivity with JDBC to the database on DB2.
- Implemented the server side processing using Java Servlets.
- Implemented LDAP and role based application security
- Mentoring team members to develop use cases diagram, sequence diagrams, preliminary class diagrams using UML and for development of flexible, scalable and efficient java code.
- Maintenance and support of the existing application
- Production support done using Remedy Tickets
- Created development and test environment in WebSphere 5.1 and Apache Tomcat 4.1 web server.
- Actively involved in the integration of different use cases, code reviews etc.
- Created development and test environment in WebSphere 5.1.1.
- Client Interaction to get the requirements from the end user.
- Customizing application on J2EE to suit user’s requirements
- Performing Enhancements on the application based on the user’s requirements
- Maintaining the application using J2EE Components like Websphere
- Querying the database using SQL for related results
Environment: J2EE (Java 1.4, JSP), LDAP, DB2, WSAD 5.1
J2EE Developer
Confidential
Responsibilities:
- Involved in the design and development phases of Rational Unified Process (RUP).
- Designed Class Diagrams, Sequence Diagrams and Object Diagrams using IBM Rational Rose to model
- Application was built on MVC architecture with JSP 1.2 acting as presentation layer, Servlets as controller and developed the application using Jakarta Struts 1.1 Framework: developed action classes, form beans and Used Struts Validation Framework for validating front end forms.
- Extensively used XML Web Services for transferring/retrieving data between different providers.
- Developed complete Business tier with Session beans and CMP Entity beans with EJB 2.0 standards using JMS Queue communication in authorization module.
- Designed and implemented Business Delegate, Session Facade and DTO Design Patterns
- Involved in implementing the DAO pattern
- Used JAXB API to bind XML Schema to java classes
- Used the report generation in the databases written in PL/SQL
- Used Maven for building the enterprise application modules
- Used Log4J to monitor the error logs
- Used JUnit for unit testing
- Used Clearcase for Version control
- Deployed the applications on WebLogic Application Server.
Environment: Struts 1.1, EJB 2.0, Servlets 2.3, JSP 1.2, SQL, XML, XSLT, Web Services, JAXB, SOAP, WSDL, JMS1.1, JavaScript, TDD, JDBC, Oracle 9i, PL/SQL, Log4J, JUnit, WebLogic, Eclipse, RationalXDE, Clearcase, Linux.
J2EE Developer
Confidential
Responsibilities:
- Developed web components using JSP, Servlets and JDBC
- Designed tables and indexes
- Designed, Implemented, Tested and Deployed Enterprise Java Beans (Session and Entity) using WebLogic as Application Server
- Developed stored procedures, packages and database triggers to enforce data integrity.
- Performed data analysis and created crystal reports for user requirements
- Provided quick turn around and resolved issues within the SLA.
- Implemented the presentation layer with HTML, XHTML and JavaScript
- Used EJBs to develop business logic and coded reusable components in Java Beans
- Development of database interaction code to JDBC API making extensive use of SQL
- Query Statements and advanced Prepared Statements.
- Used connection pooling for best optimization using JDBC interface
- Used EJB entity and session beans to implement business logic and session handling and transactions. Developed user-interface using JSP, Servlets, and JavaScript
- Wrote complex SQL queries and stored procedures
- Prepared the Installation, Customer guide and Configuration document which were delivered to the customer along with the product
Environment: Windows NT 2000/2003, XP, and Windows 7/ 8 C, Java, UNIX, and SQL using TOAD, CRM 10209, Microsoft Office Suit, Microsoft project