Hadoop Developer/administrator Resume
River Woods, IL
SUMMARY:
- Resourceful and proactive Hadoop Developer with 8 years of IT experience encompassing a solid understating of technology and providing business solutions that enhance the individual and team productivity, for the desired objectives and achieve viable deliverables.
- Around 8 years of experience in Java/J2EE technologies and Big data Hadoop development
- 3+ years of experience as Hadoop Developer with good knowledge of Java Map Reduce, Hive, Pig Latin and Hbase.
- Organized data into tables, performing transformations, and simplifying complex queries with Hive.
- Performing real - time interactive analyses on massive data sets stored in HDFS
- In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS
- MapReduce, Hadoop GEN2 Federation, High Availability and YARN architecture and good understanding of workload management, schedulers, scalability and distributed platform architectures
- Developed Map/Reduce programs according to business requirements
- Experience in analyzing data using Hive QL, Pig Latin, and custom MapReduce programs in Java.
- Developed Customized UDF's in java to extend Hive and Pig Latin functionality.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems
- Good knowledge in using job scheduling tools like Oozie.
- Good experience in designing, developing database to create its Objects like Tables, Stored Procedures, Triggers and Cursors using PL/SQL.
- Experience in requirement gathering, analysis, planning, designing, coding and unit testing
- Strong problem solving skills, good communication, interpersonal skills and a good team player
- SDLC experience that includes requirements definition, analysis, design, development,, implementation and support.
- Experience inJava/J2EEdevelopment using JSP, Servlets, JDBC, JNDI, RMI, EJB, XML, HTML, CSS, JavaScript, JSON, Web Services, Hibernate, Struts, Spring, ADOBE FLEX, JSF-Rich faces-AJAX, and web servers including JBoss, Weblogic, WebSphere, and Tomcat.
- Experience in Object Oriented Design technologies developing use cases, activity diagrams, sequence diagrams and class diagrams using case tools, including Visio and UML.
- Implemented real-time messaging, batch processes, and web services client using XML-based protocols such as Simple Object Access Protocol (SOAP) and HTTP using XML artifacts with Axis and Web Service definition Language (WSDL).
- Expert in testing frameworks and testing processes, including JUnit, integration testing, and system testing.
- Experience in architecting and developing code usingJ2EEdesign patterns, developing component architectures, and implementing architecture frameworks
- Good experience in error handling, troubleshooting, and performance tuning in multi-tier J2EEapplications.
- Excellent Knowledge of IT Network Setup on various network platforms such as Linux, UNIX, Window Administration, Cisco Router/Switches, Virtualization, Backup, Saas, Recovery & Troubleshooting.
- Strong Knowledge of NAS (Network Attached Storage) & SAN (Storage Area Network)
- Extensive interaction with QA Analyst, configuration management while handling production issues with development and business analyst teams.
TECHNICAL SKILLS:
Hadoop Technologies: Hadoop, HDFS, Hadoop Map-Reduce, Hive, HBase, SQOOP, Oozie, AVRO, Pig-Latin, Hue, CDH, No SQL MongoDB, Column Oriented DB HBase, Eclipse
Programming languages: . NET, C, C++, Java, HTML, XML, Shell Scripting, UNIX
Testing Tools: QTP, Mercury Quality Center, JIRA, Selenium Web Driver, Eclipse
Operating Systems: Windows 95/98/NT/2000/2003/XP, MS-DOS, UNIX, LINUX
Ticketing System: Siebel, HEAT
Networking Technologies: TCP/IP, Ethernet, Token Ring, Routing & Switches Scripting Languages: TSL, Java Script, VB Scripts
RDBMS: Oracle 8i/9i/10g, DB2 & MySQL 4.x/5.x
Reporting Tools: Crystal Reports, Hyperion Interactive Reporting Studio, Cognos Imprompt
Project Management: MS Project
PROFESSIONAL EXPERIENCE:
Confidential, River Woods, IL
Hadoop Developer/Administrator
Responsibilities:
- This project was designed to extract raw data from different sources into Hadoop Eco system to create and populate the necessary Hive tables. The main aim of the project is to centralize the source of data for report generation using historical database which otherwise are generated from multiple sources. Implemented multiple Map Reduce Jobs in java for data cleansing and pre - processing.
- Performed big data processing using Hadoop, MapReduce, Sqoop, Oozie, Impala
- Architected data analytics and sentiment analysis practice
- Architected data received from multiple sources (RDBMS, Log files, Streaming data, Social media data)
- Modeled business and system processes based on findings through use case scenarios and data models
- Provided technical expertise to business analysts, data scientist and developers
- Mapped key data objects to business processes
- Facilitated information management practice, architectural guidelines, framework and policies
- Written system specifications document and technical design documents (low- and high-level)
- Cloudera Hadoop distribution Installation and Initial Configuration
- Deployed a multi-node Hadoop cluster
- Monitoring Hadoop Services
- Diagnostics, Recovery and Cluster Maintenance
- Service Monitoring and Log Management
- Auditing and Alerts
- Oozie, Hive, Impala Administration
- Hadoop cluster support for various projects
- Handled importing of data from various data sources, performed transformations using Hive Worked extensively with Sqoop for importing metadata from Oracle.
- Configured Sqoop and developed scripts to extract data from SQL Server into HDFS.
- Cluster co-ordination services through ZooKeeper.
- Gained experience in managing and reviewing Hadoop log files.
- Worked with business partners to gather business requirements.
- Installed and configured MapReduce, HIVE and the HDFS; implemented CDH4&5 Hadoop
- Analyzed Hadoop cluster and different big data analytic tools including Pig & Hbase NoSQL database.
- Extracted BSON files from MongoDB and placed in HDFS and processed.
- Designed and developed Map Reduce jobs to process data coming in BSON format
- Experience with NoSQL databases.
- Written Hive UDFs to extract data from staging tables.
- Involved in creating Hive tables, loading with data.
- Experience in creating integration between Hive and HBase.
- Used Oozie scheduler to submit workflows.
- Review QA test cases with the QA team.
Environment: Java 6, Eclipse, Hadoop, Pig0.12, Hive 0.13, Centos 6.4, Map Reduce, HDFS, My SQL, Sqoop 1.4.4, CDH4, Hue, Oozie, Mongodb, Toad, HBASE
Confidential, Chicago, IL
Hadoop Developer/Administrator
Responsibilities:
- Installed and Configured Apache Hadoop clusters for application development and Hadoop tools like Hive, Pig, HBase, Zookeeper and Sqoop.
- Implemented multiple Map Reduce Jobs in java for data cleansing and pre-processing.
- Worked with the team to increase cluster from 28 nodes to 42 nodes, the configuration for additional data nodes was done by Commissioning process in Hadoop.
- Responsible for Cluster maintenance, adding and removing cluster nodes, Cluster Monitoring and Troubleshooting, manage and review data backups and log files.
- Worked with systems engineering team to plan and deploy new Hadoop environments and expand existing Hadoop clusters.
- Managed and scheduled Jobs on a Hadoop cluster.
- Involved in defining job flows, managing and reviewing log files.
- Installed Oozie workflow engine to run multiple Map Reduce, Hive HQL and Pig jobs.
- Collected the log data from web servers and integrated into HDFS using Flume.
- Responsible to manage data coming from different sources.
- Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts.
- Participated in requirement gathering form the Experts and Business Partners and converting the requirements into technical specifications
- Constructed System components and developed server side part using Java, EJB, and Spring Frame work.
- Involved in designing the data model for the system.
- UsedJ2EEdesign patterns like DAO, MODEL, Service Locator, MVC and Business Delegate.
- Defined Interface Mapping between JDBC Layer and Oracle Stored Procedures.
- Developed/CustomizedJavaServer Pages (JSP) for Customer User Interface (UI).
- Developed web tier using Struts tag libraries, CSS, HTML, XML, JSP, and Servlet.
- Involved in writing test cases using JUnit.
- Used CVS tool for Version control.
- Involved in migrating code from development to QA, QA to UAT and UAT to Production environments.
- Provided release management of the code between and different environments.
Confidential, Deerfield, IL
Software Engineer
Responsibilities:
- Developed test cases and test plan for the application according to Design and Specifications
- Involved in manual testing, automated testing, system testing, functionality testing, regression testing, user acceptance testing, and integration testing in Agile methodology
- Developed and executed SQL Queries to perform database testing
- Reported bugs through Mercury Quality Center and categorized bugs based on the severity and interfaced with developers to resolve them.
- Interacted with developers, Business Analysts, and management in reviews and solving the problems encountered in the application
- Participated in defect review meetings with the team members
- Provided testing strategy directions to global team of 4 QA associates in absence of QA Lead
- Assisted QA Lead with test strategy approach and over all testing goal
Confidential, Pittsburgh, PA
Software Engineer
Responsibilities:
- Developed test cases and test plan for the application according to Design and Specifications
- Developed complex SQL queries to test data integrity during bank conversion using Hyperion Interactive Reporting Studio and Cognos Imprompt
- Executed test scripts in Mercury Quality Center
- Involved in manual testing, automated testing, system testing, functionality testing, regression testing, user acceptance testing, and integration testing
- Developed and executed automated test scripts using QTP for Regression Testing
- Involved in updating Object Repository with every new release by adding new objects, changing the object names and properties and merging old and new Object Repository in Quick Test Professional
- Performed automated testing using QTP Step Generator, Synchronization, Actions, Recovery Scenarios and Methods
- Parameterized the fixed values in checkpoint statements, created data tables for the parameters and wrote functions for the parameters to read new data from the table upon each iteration
- Reported bugs through Mercury Quality Center and categorized bugs based on the severity and interfaced with developers to resolve them.
- Interacted with developers and Business Analysts in reviews and solving the problems encountered in the application
Confidential, Des Plaines, IL
Software Engineer
Responsibilities:
- Involved in various Technical and Functional testing.
- Generated Test plans, Test cases and Test documentation.
- Recorded Test Cases and generated QTP Scripts for automation of the testing process
- Customized QTP test scripts using VB Script for enhancement
- Responsible for Regression Testing and Integration Testing
- Performed unit testing and integration
- Used SQL Queries to test the updates in SQL Server Tables
- Developed Requirement Traceability Matrix (RTM) to efficiently manage requirements
- Developed test cases and test scripts using Automated Tools for Functionality, Security and Regression Testing
- Mapped functional requirements to individual test case requirements.
- Reported and tracked the status of the bugs, assigning severity ratings
- Validated the application against the expected results by inserting Database, Bitmap and GUI checkpoints.
- Participated in defect review meetings with the team members.