Hadoop Developer Resume
Detroit, MI
SUMMARY
- Overall 8+ years of experience in Analysis, Architecture, Design, Development, Testing, maintenance and User training of software application which includes over 3 years in Big data, Hadoop & HDFS environment and around 3+ years of experience in JAVA, J2EE.
- A Qualified IT Professional with 8+ years of experience including 3+ years of extensive experience as a Hadoop developer
- Hands - on experience in Hadoop, MapReduce, YARN, Spark, Hive, Pig, HBase, Sqoop, Flume, Oozie, Zookeeper.
- Experienced in installing, configuring, and administrating Hadoop cluster of major Hadoop distributions Hortonworks, Cloudera.
- Knowledge of Hadoop architecture and various components such as HDFS, Job Tracker, Name Node, Data Node & Map-Reduce programming paradigm.
- Expertise in writing custom UDF’s in Pig & Hive Core Functionality.
- Having basic knowledge about real-time processing tools Storm, Spark
- Knowledge in writing Spark SQL scripts, PySpark, Scala & Java API’s.
- Expertise in creating Hive Internal/External Tables/Views using shared Meta store, writing scripts in HiveQL also data transformation & file processing using Pig Latin Scripts.
- Performed Hive optimization techniques using TEZ, cost-based optimization, data compression & partitioning.
- Expertise in Data load management, importing & exporting data using SQOOP & FLUME.
- Over 3+ years of programming experience in software development life cycle including requirements, design, development, deployment, testing, documentation, implementation & maintenance of Business-to-Business, Business-to-customer, Web applications and Client/Server build on Java and J2EE in Multitier environment.
- Maintenance/ Implementation of Commercial Software. Extensive work experience with Java/J2EE technologies such as Servlets, JSP, EJB, JDBC, JSF, Struts, Spring, SOA, AJAX, XML/XSL, DOJO, Web Services (REST, SOAP), UML, Design Patterns and XML Schemas.
- Have good knowledge and skills implementing Web/Enterprise architectures and in open source frameworks like Struts, Hibernate, Spring Framework (Aspect oriented programming and inversion of control).
- Good knowledge of HTML, XML, XSL, CSS, Power Builder, JavaScript, VBScript, Visual Basic, Tiles, ANT, Maven, JUnit and Log4J.
- Experience in using IDE like Eclipse, NetBeans for debugging and using java coding standards from the beginning of development.
- Experience in software configuration management using CVS, GIT and SVN.
- Excellent communication and interpersonal skills and is a proven team player.
- Strong experience in design and development of relational database concepts (Constraints, Integrity, Normalization) with multiple RDBMS databases including Oracle 10g, MySQL, MS SQL Server & PL/SQL.
- Experience with NoSQL databases such as MongoDb, HBase, Cassandra.
- Agile Management & Software Development with Test Driven Development (TDD), Iterative development, Scrum, pair and other agile practices and methodologies.
TECHNICAL SKILLS
Big Data Technologies: Pig, Hive, Sqoop, Flume, HBase, Spark 1.5, Oozie, Impala, ZooKeeper, Ambari, Kafka, Hadoop Distributions (Hortonworks, Cloudera, Pivotal)
Hadoop Paradigms: MapReduce, Yarn
Operating Systems: UNIX, Windows, LINUX
Java Technologies: J2SE, J2EE - JSP, Servlets, JNDI, JDBC, JSTL, EJB, Junit, JPA, RMI, JMS
Web Technologies: AJAX, JavaScript, JQuery, HTML 5, CSS 3, XML
Application/ Web servers: JBoss, Web Logic, Tomcat
Databases: Oracle, MySQL, MS SQL Server
Frameworks: Spring, Struts, JSF, Hibernate
Web Services: SOAP, REST
IDEs: Eclipse, NetBeans, TOAD
Version Control Systems: GIT, CVS, SVN
Tools: Maven, Ant, JUnit, Log4J
PROFESSIONAL EXPERIENCE
Confidential, Detroit, MI
Hadoop Developer
Responsibilities:
- Extracted data from Teradata to HDFS using Sqoop.
- Loaded the customer profiles data, customer spending data, credit from legacy warehouses onto HDFS using Sqoop.
- Provided structured data processing by using Spark SQL.
- Knowledge on handling Hive queries using Spark SQL that integrates with Spark environment.
- Created a Hive aggregator to update the Hive table after running the data profiling job.
- Analyzed large data sets by running hive queries.
- Involved in creating Hive tables, loading with data and writing Hive queries that will run internally in map reduce way.
- Analyzed the data by performing Hive queries.
- Implemented Partitioning, Dynamic Partitioning and Bucketing in Hive.
- Developed Hive queries to process the data and generate the data cubes for visualizing.
- Built reusable Hive UDF libraries for business requirements, which enabled users to use these UDF's in Hive Querying.
- Written Hive UDF to sort Structure fields and return complex data type.
- Modeled Hive partitions extensively for data separation and faster data processing and followed Pig and Hive best practices for tuning.
- Exported the patterns analyzed back to Teradata using Sqoop.
- Implemented a script to transmit sys print information from Oracle to HBase using Sqoop.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports.
- Involved in loading data from local file system (Linux) to HDFS.
Environment: Hortonworks, Hadoop, Hive, Spark, Sqoop, HBase, MapReduce, HDFS, Pig, Cassandra, Java, Oracle 11g/10g, FileZilla, Unix Shell Scripting.
Confidential, Kansas City, MO
Hadoop Developer
Responsibilities:
- Installed and configured Hadoop Map reduce, HDFS, developed multiple Map Reduce jobs in java for data cleaning and preprocessing.
- Importing and exporting data into HDFS and Hive using Sqoop.
- Experience in defining job flows using Oozie.
- Experience in managing and reviewing Hadoop log files.
- Extracted files from RDBMS through Sqoop and placed in HDFS and processed.
- Experience in running Hadoop streaming jobs to process Terabytes of xml format data.
- Got good experience with NOSQL database.
- Supported Map Reduce Programs those are running on the cluster.
- Involved in loading data from Unix file system to HDFS.
- Involved in creating Hive tables, loading with data and writing Hive queries which will run internally in map reduce way.
- Replaced default Derby metadata storage system for Hive with MySQL system.
- Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS.
- Developed the Pig UDF's to preprocess the data for analysis.
- Developed Hive queries for the analysts.
- Involved in loading data from Linux and Unix file system to HDFS.
- Load and transform large data sets of structured, semi structured and unstructured data.
- Worked with various Hadoop file formats, including TextFiles, SequenceFile, RCFile.
- Supported in setting up QA environment and updating configurations for implementing scripts with Pig.
- Developed a custom File System plug in for Hadoop. So, it can access files on Data Platform. This plugin allows Hadoop MapReduce programs, HBase, Pig and Hive to work unmodified and access files directly.
- Designed and implemented MapReduce based large scale parallel relation learning system.
Environment: Hadoop, Hive, HBase, MapReduce, HDFS, Pig, Cassandra, Java (JDK 1.6), Hadoop Distribution of Cloudera, IBM Data Stage 8.1, Oracle 11g/10g, PL/SQL, SQL*PLUS, Toad 9.6, Unix Shell Scripting.
Confidential, CA
Hadoop Developer
Responsibilities:
- Loading the data from the different Data sources like (Teradata and DB2) into HDFS using Sqoop and load into Hive tables, which are partitioned.
- Developed Hive UDF’s to bring all the customers email id into a structured format.
- Developed bash scripts to bring the log files from ftp server and then processing it to load into hive tables.
- Using Sqoop to load data from DB2 into HBASE environment.
- Inserted Overwriting the HIVE data with Hbasedata daily to get fresh data every day.
- All the bash scripts are scheduled using Resource Manager Scheduler.
- Developed Oozie Workflows for daily incremental loads, which gets data from Teradata and then imported into hive tables.
- Developed Pig scripts to transform the data into structured format and it are automated through oozie coordinators.
- Worked on loading the data from MySQL to HBase where necessary using Sqoop
- Developed Hive queries for Analysis across different banners.
Environment: Windows 7, Hadoop, HDFS, MapReduce, Sqoop, Hive, Pig, Hbase, Teradata, DB2, Oozie, MySQL, Eclipse
Confidential, Miami, FL
Senior Java/ J2EE Developer
Responsibilities:
- Analysis, test, debug documents and implements moderately complex software applications.
- Implemented the project by using Spring Web MVC module.
- Devise or modify procedures to solve complex problems considering computer equipment capacity and limitations, operating lime, and form of desired results.
- Implement, integrate and maintain the client server and web related applications.
- Implemented the project using Ajax,Java Script, and HTML as UI components.
- Implemented the project using JAXB to retrieve the data from the xml documents.
- Implemented the project using Web Services to communicate with other systems.
- Used spring framework for the business layer using Spring Core, DAO, Spring ORM, Spring Web MVC modules.
- Implemented the project using the Hibernate framework to communicate with database.
- Implemented Singleton, factory design pattern, DAO design Patterns based on the application requirements.
- Used SVN as version control tools to maintain the code repository.
- Developing Test cases using JUnit Testing Framework andLog4J for logging and audit trail
- Deployed the application using the JBOSS as application server.
Environment: Java 1.6, Servlets, JSP, JBOSS 6.1, Spring MVC, Hibernate, XML, JAXB, HTML, Ajax, JavaScript, JNDI, Web Services, WSDL, SOAP, SQL, PL/SQL, Oracle11g, Eclipse, SQL Developer, Log4j, JUnit, Maven, SVN.
Confidential
Java Developer
Responsibilities:
- Responsible for closing errors in ST (System Test) and AT (Acceptance Test) phases.
- Developed the Application and business logic using Java 1.5
- Developed, customized and delivered Java, Java Server Pages (JSP), Servlets for the application.
- Deployed the application in Weblogic 8.1.
- Used Oracle8i as the database for the application.
- Automated Development & Deployment process using ANT
- Used Mozilla Fire Bug tool for Java Script Debugging.
- Used VSS (Visual Source Safe) as Configuration Management Tool.
- Coordinated with Off-shore & On-site teams.
Environment: Java, JSP, Servlet, JDBC, JNDI, XML, HTML, JavaScript, Weblogic 8.1.4, PL/SQL, Oracle 8i, Windows NT, Linux, UNIX Shell Script, Eclipse, Apache ANT, dom4j.
Confidential
SQL Developer/Programmer Analyst
Responsibilities:
- Requirement gathering and worked according to the CR.
- Worked under the senior level guidance.
- Data validation/Reconciliation report generation.
- Code Development as per the client requirements.
- Involved in the development backend code, altered tables to add new columns, Constraints, Sequences and Indexes as per business requirements.
- Perform DML, DDL Operations as per the Business requirement.
- Creating views and prepares the Business Reports.
- Resolved production issues by modifying backend code as and when required.
- Used different joins, sub queries and nested query in SQL query.
- Involved in creation of sequences for automatic generation of Product ID.
- Created Database Objects like tables, Views, sequences, Synonyms, Stored Procedures, functions, Packages, Cursors, Ref Cursor and Triggers.
- Testing of code functionality using testing environment.
Environment: Oracle 9i, Windows, MS Excel, Reports, Java.