We provide IT Staff Augmentation Services!

Hadoop Administrator Resume

5.00/5 (Submit Your Rating)

SUMMARY:

  • Having 11+years of overall experience in Big Data and DW technologies like Hadoop(Apache, Cloudera, Hortonworks Pivotal HD),Tableau, Splunk,Informatica.
  • Hadoop ( Apache,Cloudera, Hortonworks, Pivotal HD) Hive, Pig, Hbase, Sqoop, Flume, Oozie, Zookeeper, Kerberos, AWS - EC2,S3, DB2,Business Objects, Informtica
  • Around 4+ years of experience in Hadoop Administration & Big Data Technologies
  • Hands on experience in installation, configuration, supporting and managing Hadoop Clusters using Apache, Cloudera, and Hortonworks.
  • Hadoop Cluster capacity planning, performance tuning, cluster Monitoring, Troubleshooting.
  • Design Big Data solutions for traditional enterprise businesses.
  • Backup configuration and Recovery from a NameNode failure.
  • Decommissioning and commissioning the Node on running hadoop cluster.
  • Installation of various Hadoop Ecosystems tools like Hive, Pig, Sqoop, Hbase,Zookeeper, Impala, Spark.
  • Involved in bench marking Hadoop/HBase cluster file systems various batch jobs and workloads
  • Experience monitoring and troubleshooting issues with Linux memory, CPU, OS, storage and network
  • Good experience in Designing,configuring and managing the backup and disaster recovery for Hadoopdata.
  • Hands on experience in analyzing Log files for Hadoop and eco system services and finding root cause.
  • Experience on Commissioning, Decommissioning, Balancing, and Managing Nodes and tuning server for optimal performance of the cluster.
  • As a admin involved in Cluster maintenance, trouble shooting, Monitoring and followed proper backup& Recovery strategies.
  • Experience in HDFS data storage and support for running map reduce jobs.
  • Experience in importing and exporting the data using Sqoop from RDBMS to HDFS and viceversa.
  • Scheduling all hadoop/hive/sqoop/Hbase jobs using Oozie.
  • Experience in installing, administering, and supporting Linux operating systems and hardware in an enterprise environment. (CentOS/RHEL/Ubuntu).
  • Experience in writing complex queries in HIVE.
  • Experience in Database administration for DB2 in Z/OS environment.
  • Experienced in supporting critical production databases and worked on 24*7 production support
  • Expertise in Performance tuning & optimization.
  • Proficient in configuring, implementing, and monitoring different types of High Availability & disaster recovery scenarios like Replication, log shipping, Database mirroring and clustering.
  • SQL Server 2008 R2/2008/2005 Database Administration.
  • Creation, instance upgrade and version migration. Implement a library of homegrown scripts to automate the scheduling and notification of periodic maintenance jobs, including REORG, BACKUP, RUNSTATS, LOAD,RECOVERY
  • Fixing bind issues that will occur during the application testing in various phases.
  • Maintainedcreate/alter/drop/grant/revoke/rebind/free DB2 objectusing IBM DB2 Admin tool.
  • Migration of database objects between databases and DB2 subsystems using IBM DB2 Admin tool.
  • Automated the DB2 space reports generation for all the databases we support.
  • Created and managed production emergency fix process on the mainframes.
  • Executing DB2 reports/utilities for performance/database organization & database growth analysis.

TECHNICAL SKILLS:

Big Data: Apache Hadoop, Cloudera, Hortonworks

Hadoop Ecosystem: HDFS,YARN,MapReduce,Hive,Pig,HBase,Zookeeper, Kafka,Sqoop,Oozie,Flume,Spark,Kerberos,Ambari

Devops Tools: Chef, Puppet, Ansible, Jenkins, Git, SVN

Big Data Analytics: Hive

BI Tools: Business Objects, SSRS, Tableau

ETL Tools: Informatica8.6/9.1, Talend

Cloud Technologies: AWS-EC2,S3, Google Cloud

No SQL: Mongo DB, Cassandra

Database: DB2, IMS, Oracle11g,10g, SQL Server 2008R2

Scheduling Tools: Control-M,CA7

Languages: C,C++, Cobol, Python, Bash

Application Servers: Web logic, WebSphere.

Operating system: Linux(RedHat,CentOS,Debian,Ubuntu)/Unix, Windows

PROFESSIONAL EXPERIENCE:

Confidential

Hadoop Administrator

Responsibilities:

  • Involved in start to end process of hadoop cluster setup where in installation, configuration and monitoring the Hadoop Cluster.
  • Responsible for Cluster maintenance, commissioning and decommissioning Data nodes,Troubleshooting, Manage and review data backups, Manage & review Hadoop log files.
  • Monitoring systems and services, architecture design and implementation of hadoop deployment, configuration management, backup, and disaster recovery systems and procedures.
  • Importing and exporting data into HDFS using Sqoop.
  • Experienced in define the job flows with Oozie.
  • Loading log data directly into HDFS using Flume.
  • Experienced in managing and reviewing Hadoop log files.
  • Installation and configuration of Sqoop and Flume, Hbase,Hive,Hue,Kafka,Spark, Impala.
  • Managed and reviewed Hadoop Log files as a part of administration for troubleshooting purposes. Communicate and escalate issues appropriately.
  • As a admin followed standard Back up policies to make sure the high availability of cluster.
  • Involved in Analyzing system failures, identifying root causes, and recommended course of actions. Documented the systems processes and procedures for future references.
  • Worked with systems engineering team to plan and deploy new hadoop environments and expand existing hadoop clusters.
  • Monitored multiple hadoop clusters environments using Ganglia and Nagios. Monitored workload, job performance and capacity planning using Cloudera Manager.
  • Involved in Installing and configuring Kerberos for the authentication of users and hadoop daemons.
  • Experience in setup, configuration and management of security for hadoop clusters using Kerbero
  • Configuring HBASE indexer and Solr engines.
  • Reporting Nagios and Ganglia metrics for better cluster utilization.
  • Perform ongoing capacity management forecasts including timing and budget considerations.
  • Coordinate root cause analysis (RCA) efforts to minimize future system issues.
  • Installation, configuration, supporting and managing Hadoop Clusters using Apache, Cloudera (CDH4,CDH5), and Yarn acrhitecture.
  • Perform data analysis using Hive and Pig.
  • Supporting Sandbox Hadoop cluster users with basic Hadoop options
  • User/Disk quota Administration.
  • Communicated all issues and participated in weekly strategy meetings.
  • Implemented system wide monitoring and alerts.
  • Migration from traditional data center to AWS using EC2 and S3.
  • Upgrading hadoop cluster hbase/zookeeper from CDH3 to CDH4.

Environment: ApacheHadoop, Cloudera, Hive, Pig, Sqoop, Hbase,Zookeeper, Impala,Linux

Confidential

Hadoop Administrator

Responsibilities:

  • Work with Client’s internal support teams to Build/Troubleshoot/Fix Hadoop platforms
  • Work with the application teams to design and develop an effective Hadoop/Data Science solution.
  • Develop best practices for developing and deploying Hadoop applications and assist the team to manage compliance to the standards.
  • Manage the backup and disaster recovery for Hadoop data.
  • Optimize and tune the Hadoop environments to meet performance requirements.
  • Install and configure monitoring tools.
  • Work with big data developers and developers designing scalable supportable infrastructure.
  • Work with Linux server admin team in administering the server hardware and operating system
  • Assist with management of formal system runbooks.
  • Create and publish various production metrics including system performance and reliability information to systems owners and management.
  • Load log data into HDFS using Flume.
  • Monitoring Hadoop cluster using tools like Nagio's, Ganglia and Cloudera Manager.
  • Automating the script to monitor HDFS and HBase through cronjobs.
  • Administrative support for parallel computation research on a 24-node Fedora/ Linux cluster.
  • Integrate Kickstart, Kerberos with open source technologies.
  • Hadoop Cluster capacity planning, performance tuning, cluster Monitoring, Troubleshooting.
  • Design Big Data solutions for traditional enterprise businesses.
  • Used Network Monitoring Daemons like Ganglia and Service monitoring tools like Nagios.
  • Adding/removing new nodes to an existing hadoop cluster.
  • Backup configuration and Recovery from a NameNode failure.
  • Decommissioning and commissioning the Node on running hadoop cluster.
  • Installation of various Hadoop Ecosystems tools like Hive, Pig, Sqoop, Hbase, Zookeeper, Impala, SparkManaged and reviewed Hadoop log files.
  • Tested raw data and executed performance scripts.
  • Shared responsibility for administration of Hadoop, Hive and Pig.
  • Report development using splunk on top of the EDW
  • Designing and creating data sources for tableau reporting.
  • Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.
  • File system management and monitoring.
  • Performance tuning of Hadoop clusters and Hadoop MapReduce routines.
  • Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades when required.
  • Point of Contact for Vendor escalation
  • Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments.
  • Monitor Hadoop cluster connectivity and security
  • HDFS support and maintenance.
  • Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.

Environment: Apache Haddop, DB2, Tableau, Hadoop, HDFS, Hive, Sqoop, Flume, Zookeeper and HBase, Big Data Cloudera CDH Apache Hadoop, Toad, SQL plus, Redhat/Suse Linux

Confidential

DB2 Administrator

Responsibilities:

  • Analyzing and Gathering and documenting Requirements as per client by directly interacting with them and Physical database design and implementation.
  • Physical table and index design, referential integrity, check constraints based on application requirement. Test environment set-up, and production deployment
  • Defining stored procedures and creation of Joins, Views Triggers,etc.Experience in JCL, application programs performance monitoring and tuning
  • Check database status, monitor daily, monthly batch cycle, resolve any job abends in production Environment. Authorization and privileges for DB2 objects.
  • Created and altered the database objects tablespace, tables, indexes, views, stored procedures.Created the physical data modeling diagrams using ER-Studio.
  • Granted the appropriate privileges to the users on the database objects.
  • Prepared the housekeeping utility jobs image copy, reorg and runstat for new database objects and released these housekeeping jobs into control-M.
  • Re engineered the database objects into physical data model using ER-Studio.
  • Involved in improve the performance of the SQL queries.
  • Deployed the database objects from test region to system/production region using BMC change Manager.
  • Automation of database maintenance activities in production like finding the image copy gaps and alerts for space threshold.
  • Handled and administered the large data volumes in production tables.
  • Storage management including space allocation, space enhancement based on the growth, re-distribution of data or re-designs of database objects based on existing growth pattern or anticipated growth.
  • Application development support on creation of collection id, package, plan creation, bind/rebind, dclgen.
  • Technical support to the programmers and users.
  • Debugging application programs, resolution of database errors, support for application errors and bad application performance.
  • Supported performance issues of the database system.
  • Worked closely with the development teams and data architects.
  • Installed and scheduled backups, reorganization, recovery, and check point process.
  • Coordinated database changes in production to prevent outage, avoid application impacts, and smoothly execute Modifications.
  • Provided 24-hour 7-day production system support to ensure each service level agreement is achieved.
  • Granting and Revoking Utilities access on Table Spaces.
  • Creation, instance upgrade and version migration. Implement a library of homegrown scripts to automate the scheduling and notification of periodic maintenance jobs, including REORG, BACKUP, RUNSTATS, LOAD,RECOVERY.
  • Performance tuning of the Test and development databases using explain tools.
  • Schedule and execute reorg/runstat on the database tables.
  • Identifying and Streamlining the weak processes within the system

Confidential

DB2 Database Administrator

Responsibilities:

  • Created and managed production emergency fix process on the mainframes.
  • Expertise in Database design, Installation, Upgradations, Configuration, Backup, Recovery, Database Security, and Query o­ptimization
  • Experience in Data Modeling (logical modeling and physical modeling) using Erwin
  • Executed DB2 reports/utilities for performance/database organization & database growth analysis.
  • Generate the Image copy gaps reports and took the backup copy of an table spaces and alert the application DBA teams to fill the IC gaps
  • Involved in re-engineering the application to identify the highly CPU consuming jobs and improved the performance of those applications.
  • Facilitated execution of load/unload/DSN1COPY of data for application development teams and to migrate the data between different regions.
  • Altered the database objects like table spaces, tables, indexes, views, stored procedures.
  • Executed the database utility jobs like Image Copy, Unload, Load, Rebuild IMS database.
  • Runstats,Check Data, Rebuild and Recovery Utilities.
  • Rotational On-call production DBA support.
  • Expertise in Backup and restore of the databases.
  • Expertise in Performance tuning & optimization.
  • Responsible for database integrity & database backup and recovery.DB2 performance and problem determination,fixing production job abends in a 24 x 7 environment. Authorization and privileges for DB2objects.
  • Responsible for the evaluation and creation of appropriate backup and recovery strategies for new database requirements.
  • Bind plans & packages. Creation of Package, reviewing & providing DBA approval for production packagesUsing CHANGEMAN tool.
  • Removing the restriction of pending status against DB2 object. Analyze data & develop and execute appropriate DDL to support projects.
  • Responsible for database integrity & database backup and recovery.DB2 performance and problem determination, fixing production job abends in a 24 x 7 environment. Authorization and privileges for DB2objects.
  • Maintained create/alter/drop/grant/revoke/rebind/free DB2 Objects using IBM DB2 Admin Tool
  • Migration of database objects between databases and DB2 subsystems using IBM DB2 Admin tool.
  • Automated the DB2 space reports generation for all the databases we support.
  • Provided relations between tables Using Erwin Tool.
  • SQL Server 2008 R2/2008/2005 Database Administration.
  • Experienced in Performance tuning, Optimization, capacity planning.
  • Migrated databases from SQL Server 2000 to SQL Server 2005 and 2008 R2

Confidential

Application Developer and DBA

Responsibilities:

  • Created the new database objects like databases, table spaces, tables, index, views, and stored procedures. Altered the attributes of existing database objects.
  • Application development support on creation of collection id, package, plan creation, bind/rebind, dclgen. Loaded and unloaded the data using IBM load and unload utility.
  • Worked on writing and modifying native stored procedures as per business requirements.
  • Performed the unit testing, system testing on stored procedures.
  • Deployed the database objects from test region to system region/production region.
  • Prepared and executed the DB2 Utility jobs runstat, rebuild index, rebind, and check data utilities.
  • Fixed bind issues that will occur during the application testing in various phases.
  • Converting from plan bind to package bind.
  • Converting from simple Table Space to Segmented Table Space.
  • Converting from Partitioned by Range to Partition by Growth Table Space.
  • Granting and Revoking Utilities access on Table Spaces.
  • Experience in writing complex queries using SQL
  • Providing 24 hour on call support.
  • Analyzing Client requirements and preparation of low level design document.
  • Developing new COBOL, DB2, application programs as per the customer requirements and tested versions are delivered to the customer.
  • Creating batch jobs and procs to extract programs and moving the code to production.
  • Coding programs which writes data to XML file by reading DB2 tables.
  • Validating XML files According to schema. Provided technical guidance to the team.
  • Preparation of test plans, test cases, code reviews and delivery documents.

We'd love your feedback!