Sr. Hadoop Developer/hive Resume
Charlotte, NC
SUMMARY
- Around 8 years of experience in Business Analytics, Data Modeling and Data warehousing.
- Hands - on experience with Big Data architecture - Hadoop framework and its components.
- 4+ years Hadoop development experience with Hive.
- Extensive experience working with Hive, creating Hive tables, writing custom Hive UDF’s and Hive queries for analysis and reporting.
- Good knowledge with core components such as MapReduce, YARN, HDFS.
- In-depth knowledge of Hadoop EcoSystem - Pig, Hive, Spark, Scala, Avro, Parquet, Flume, Sqoop, Crunch, ZooKeeper, NoSQL databases like Cassandra, HBase.
- Good knowledge of Hadoop Development and various components such as HDFS, Job Tracker, Task Tracker, Data Node, Name Node and Map-Reduce concepts.
- Experienced working with Tableau software and other BI tools like Cognos.
- Strong experience in SQL, Python and Shell scripting.
- Working knowledge in Java development and its environment.
- Knowledge of data flows, data architecture, ETL and processing of unstructured data.
- Experience with Data Integration on traditional and Hadoop environments.
- Experience designing and implementing reporting and visualization for unstructured and structured data sets.
- Experience with Hadoop Cluster Administration and working with larger data sets.
- Experienced with Git, GIT hub, Perforce and other source code management tools.
- Experienced in all phases of Software Development Life Cycle (SDLC)
- Extensive knowledge in various branching/merging methodologies & release procedures.
- Excellent communication skills and experienced working with multiple teams gathering business requirements.
TECHNICAL SKILLS
Data Analytics: Hive, Pig, MapReduce, Hadoop 2.0, Apache Spark, Scala.
Databases: Cassandra, HBase,Oracle, MS SQL Server, DB2.
Reporting Technologies: Tableau v8.x, Cognos 10.x
ETL: Datastage 8.x
Languages: Python, C, C++, SQL, PL/SQL, Java, MATLAB.
Web Technologies: Java, Java Script, VBScript, HTML
Version Control SW: Subversion, Perforce, GIT, Jenkins, JIRA
IDE & Tools: Eclipse IDE, Android Studio, Maven
PROFESSIONAL EXPERIENCE
Confidential, Charlotte, NC
Sr. Hadoop Developer/Hive
Responsibilities:
- Extensively involved in creating Hive tables, loading data and writing Hive queries.
- Responsibilities included analyzing transformed data using HiveQL and Hive UDF's.
- Responsible for HIVE performance tuning and optimizing Hadoop Mapreduce jobs.
- Worked on HBase & Hive scripts to extract, transform & load data into HBase and Hive.
- Written MapReduce code to process the data from various sources and storing parsed data into HBase and Hive using HBase-Hive Integration.
- Worked in importing and exporting data into HDFS and Hive from RDBMS using Sqoop.
- Worked with Oozie workflow engine to run multiple Hive Jobs.
- Processed data into HDFS using MapReduce, Spark, Pig, Hive, Scala, Hive/Pig UDFs and Sqoop.
- Involved in development of multiple MapReduce jobs for data cleaning & preprocessing
- Managed and reviewed Hadoop log files to optimize MapReduce jobs performance.
- Installed and configured Hive and also written Hive UDFs.
- Involved in loading data from UNIX file system to HDFS.
- Performed test run of batch jobs and generated reports.
Environment: Hadoop, Cloudera CDH5, Hive, Pig, HBase, Sqoop, Flume, Spark, Scala.
Confidential, San Diego, CA
Hadoop Developer/Hive Developer
Responsibilities:
- Extensively worked with Hive creating tables, loading the data and writing hive queries.
- Involved in developing Hive DDLs to create, alter and drop Hive TABLES.
- Implemented various Hive UDF's as per business requirements
- Worked on Hive performance tuning and Hadoop Mapreduce operation optimization.
- Created partitioned tables in Hive.
- Performed analysis using Hive on the partitioned and bucketed data to compute various metrics for reporting using Tableau.
- Importing and exporting of data into HDFS and Hive using Sqoop.
- Worked with Oozie workflow engine to run multiple Hive Jobs.
- Developed shell scripts for processing of hive scripts and according to business need.
- Developed ETL code using different transformations to extract, transform the data from legacy sources and load data into target system.
- Worked on processing ingested raw data using MapReduce, Apache Pig, Scala and Hive.
- Worked extensively on Hadoop platform like installing cluster, commissioning & decommissioning of Datanode, capacity planning and slot configuration.
- Worked with NoSQL databases like Cassandra and HBASE.
- Gained experience working with Kafka-Storm on MAPR platform for real time analysis.
- Co-ordinated with Offshore team members to meet tight deadlines.
- Responsible for implementation and ongoing administration of Cloudera platform.
- Responsible for configuring and monitoring of a cluster of multiple nodes.
- Responsible for Installation and configuraton of Hadoop Map reduce, HDFS.
Environment: Hadoop, HDFS, Hive, Pig, Java, Python, Oracle, MS SQL Server.
Confidential, San Diego, CA
Sr. System Administrator
Responsibilities:
- Performing security management by maintaining roles, privileges, and user profiles. Add/Remove users from many different Security Groups in our domain (Auto groups)
- Run, Schedule, Enable SQL Jobs. Troubleshoot if the jobs fail, provide log files of jobs to dev teamsLook into the health of the Production Databases.
- Monitoring Mirroring, Log shipping & Replication. Regular monitoring of SQL Server logs for errors.
- Schedule backup jobs and object level recovery using maxima tool.
- Applying DML statements on different production databases as per the customer request. Monitoring of Online and scheduled jobs.
Confidential, Dulles, VA
System Administrator
Responsibilities:
- Configuring and maintenance various MySQL databases for applications like Drupal.
- Setup and Configured Oracle Golden Gate software on the client side
- Have installed Golden Gate for replication between RAC clustered Databases.
- Experience with Upgrading Oracle Databases Versions from 10g to 11gr2.
- Have worked with Oracle Data Guard for Disaster Recovery using Standby Databas
- Experience with installing Drupal Multi Site and Migration of data