Hadoop Administrator Resume
SUMMARY:
- Having 11+years of overall experience in Big Data and DW technologies like Hadoop(Apache, Cloudera, Hortonworks Pivotal HD),Tableau, Splunk,Informatica.
- Hadoop ( Apache,Cloudera, Hortonworks, Pivotal HD) Hive, Pig, Hbase, Sqoop, Flume, Oozie, Zookeeper, Kerberos, AWS - EC2,S3, DB2,Business Objects, Informtica
- Around 4+ years of experience in Hadoop Administration & Big Data Technologies
- Hands on experience in installation, configuration, supporting and managing Hadoop Clusters using Apache, Cloudera, and Hortonworks.
- Hadoop Cluster capacity planning, performance tuning, cluster Monitoring, Troubleshooting.
- Design Big Data solutions for traditional enterprise businesses.
- Backup configuration and Recovery from a NameNode failure.
- Decommissioning and commissioning the Node on running hadoop cluster.
- Installation of various Hadoop Ecosystems tools like Hive, Pig, Sqoop, Hbase,Zookeeper, Impala, Spark.
- Involved in bench marking Hadoop/HBase cluster file systems various batch jobs and workloads
- Experience monitoring and troubleshooting issues with Linux memory, CPU, OS, storage and network
- Good experience in Designing,configuring and managing the backup and disaster recovery for Hadoopdata.
- Hands on experience in analyzing Log files for Hadoop and eco system services and finding root cause.
- Experience on Commissioning, Decommissioning, Balancing, and Managing Nodes and tuning server for optimal performance of the cluster.
- As a admin involved in Cluster maintenance, trouble shooting, Monitoring and followed proper backup& Recovery strategies.
- Experience in HDFS data storage and support for running map reduce jobs.
- Experience in importing and exporting the data using Sqoop from RDBMS to HDFS and viceversa.
- Scheduling all hadoop/hive/sqoop/Hbase jobs using Oozie.
- Experience in installing, administering, and supporting Linux operating systems and hardware in an enterprise environment. (CentOS/RHEL/Ubuntu).
- Experience in writing complex queries in HIVE.
- Experience in Database administration for DB2 in Z/OS environment.
- Experienced in supporting critical production databases and worked on 24*7 production support
- Expertise in Performance tuning & optimization.
- Proficient in configuring, implementing, and monitoring different types of High Availability & disaster recovery scenarios like Replication, log shipping, Database mirroring and clustering.
- SQL Server 2008 R2/2008/2005 Database Administration.
- Creation, instance upgrade and version migration. Implement a library of homegrown scripts to automate the scheduling and notification of periodic maintenance jobs, including REORG, BACKUP, RUNSTATS, LOAD,RECOVERY
- Fixing bind issues that will occur during the application testing in various phases.
- Maintainedcreate/alter/drop/grant/revoke/rebind/free DB2 objectusing IBM DB2 Admin tool.
- Migration of database objects between databases and DB2 subsystems using IBM DB2 Admin tool.
- Automated the DB2 space reports generation for all the databases we support.
- Created and managed production emergency fix process on the mainframes.
- Executing DB2 reports/utilities for performance/database organization & database growth analysis.
TECHNICAL SKILLS:
Big Data: Apache Hadoop, Cloudera, Hortonworks
Hadoop Ecosystem: HDFS,YARN,MapReduce,Hive,Pig,HBase,Zookeeper, Kafka,Sqoop,Oozie,Flume,Spark,Kerberos,Ambari
Devops Tools: Chef, Puppet, Ansible, Jenkins, Git, SVN
Big Data Analytics: Hive
BI Tools: Business Objects, SSRS, Tableau
ETL Tools: Informatica8.6/9.1, Talend
Cloud Technologies: AWS-EC2,S3, Google Cloud
No SQL: Mongo DB, Cassandra
Database: DB2, IMS, Oracle11g,10g, SQL Server 2008R2
Scheduling Tools: Control-M,CA7
Languages: C,C++, Cobol, Python, Bash
Application Servers: Web logic, WebSphere.
Operating system: Linux(RedHat,CentOS,Debian,Ubuntu)/Unix, Windows
PROFESSIONAL EXPERIENCE:
Confidential
Hadoop Administrator
Responsibilities:
- Involved in start to end process of hadoop cluster setup where in installation, configuration and monitoring the Hadoop Cluster.
- Responsible for Cluster maintenance, commissioning and decommissioning Data nodes,Troubleshooting, Manage and review data backups, Manage & review Hadoop log files.
- Monitoring systems and services, architecture design and implementation of hadoop deployment, configuration management, backup, and disaster recovery systems and procedures.
- Importing and exporting data into HDFS using Sqoop.
- Experienced in define the job flows with Oozie.
- Loading log data directly into HDFS using Flume.
- Experienced in managing and reviewing Hadoop log files.
- Installation and configuration of Sqoop and Flume, Hbase,Hive,Hue,Kafka,Spark, Impala.
- Managed and reviewed Hadoop Log files as a part of administration for troubleshooting purposes. Communicate and escalate issues appropriately.
- As a admin followed standard Back up policies to make sure the high availability of cluster.
- Involved in Analyzing system failures, identifying root causes, and recommended course of actions. Documented the systems processes and procedures for future references.
- Worked with systems engineering team to plan and deploy new hadoop environments and expand existing hadoop clusters.
- Monitored multiple hadoop clusters environments using Ganglia and Nagios. Monitored workload, job performance and capacity planning using Cloudera Manager.
- Involved in Installing and configuring Kerberos for the authentication of users and hadoop daemons.
- Experience in setup, configuration and management of security for hadoop clusters using Kerbero
- Configuring HBASE indexer and Solr engines.
- Reporting Nagios and Ganglia metrics for better cluster utilization.
- Perform ongoing capacity management forecasts including timing and budget considerations.
- Coordinate root cause analysis (RCA) efforts to minimize future system issues.
- Installation, configuration, supporting and managing Hadoop Clusters using Apache, Cloudera (CDH4,CDH5), and Yarn acrhitecture.
- Perform data analysis using Hive and Pig.
- Supporting Sandbox Hadoop cluster users with basic Hadoop options
- User/Disk quota Administration.
- Communicated all issues and participated in weekly strategy meetings.
- Implemented system wide monitoring and alerts.
- Migration from traditional data center to AWS using EC2 and S3.
- Upgrading hadoop cluster hbase/zookeeper from CDH3 to CDH4.
Environment: ApacheHadoop, Cloudera, Hive, Pig, Sqoop, Hbase,Zookeeper, Impala,Linux
Confidential
Hadoop Administrator
Responsibilities:
- Work with Client’s internal support teams to Build/Troubleshoot/Fix Hadoop platforms
- Work with the application teams to design and develop an effective Hadoop/Data Science solution.
- Develop best practices for developing and deploying Hadoop applications and assist the team to manage compliance to the standards.
- Manage the backup and disaster recovery for Hadoop data.
- Optimize and tune the Hadoop environments to meet performance requirements.
- Install and configure monitoring tools.
- Work with big data developers and developers designing scalable supportable infrastructure.
- Work with Linux server admin team in administering the server hardware and operating system
- Assist with management of formal system runbooks.
- Create and publish various production metrics including system performance and reliability information to systems owners and management.
- Load log data into HDFS using Flume.
- Monitoring Hadoop cluster using tools like Nagio's, Ganglia and Cloudera Manager.
- Automating the script to monitor HDFS and HBase through cronjobs.
- Administrative support for parallel computation research on a 24-node Fedora/ Linux cluster.
- Integrate Kickstart, Kerberos with open source technologies.
- Hadoop Cluster capacity planning, performance tuning, cluster Monitoring, Troubleshooting.
- Design Big Data solutions for traditional enterprise businesses.
- Used Network Monitoring Daemons like Ganglia and Service monitoring tools like Nagios.
- Adding/removing new nodes to an existing hadoop cluster.
- Backup configuration and Recovery from a NameNode failure.
- Decommissioning and commissioning the Node on running hadoop cluster.
- Installation of various Hadoop Ecosystems tools like Hive, Pig, Sqoop, Hbase, Zookeeper, Impala, SparkManaged and reviewed Hadoop log files.
- Tested raw data and executed performance scripts.
- Shared responsibility for administration of Hadoop, Hive and Pig.
- Report development using splunk on top of the EDW
- Designing and creating data sources for tableau reporting.
- Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.
- File system management and monitoring.
- Performance tuning of Hadoop clusters and Hadoop MapReduce routines.
- Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades when required.
- Point of Contact for Vendor escalation
- Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments.
- Monitor Hadoop cluster connectivity and security
- HDFS support and maintenance.
- Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.
Environment: Apache Haddop, DB2, Tableau, Hadoop, HDFS, Hive, Sqoop, Flume, Zookeeper and HBase, Big Data Cloudera CDH Apache Hadoop, Toad, SQL plus, Redhat/Suse Linux
Confidential
DB2 Administrator
Responsibilities:
- Analyzing and Gathering and documenting Requirements as per client by directly interacting with them and Physical database design and implementation.
- Physical table and index design, referential integrity, check constraints based on application requirement. Test environment set-up, and production deployment
- Defining stored procedures and creation of Joins, Views Triggers,etc.Experience in JCL, application programs performance monitoring and tuning
- Check database status, monitor daily, monthly batch cycle, resolve any job abends in production Environment. Authorization and privileges for DB2 objects.
- Created and altered the database objects tablespace, tables, indexes, views, stored procedures.Created the physical data modeling diagrams using ER-Studio.
- Granted the appropriate privileges to the users on the database objects.
- Prepared the housekeeping utility jobs image copy, reorg and runstat for new database objects and released these housekeeping jobs into control-M.
- Re engineered the database objects into physical data model using ER-Studio.
- Involved in improve the performance of the SQL queries.
- Deployed the database objects from test region to system/production region using BMC change Manager.
- Automation of database maintenance activities in production like finding the image copy gaps and alerts for space threshold.
- Handled and administered the large data volumes in production tables.
- Storage management including space allocation, space enhancement based on the growth, re-distribution of data or re-designs of database objects based on existing growth pattern or anticipated growth.
- Application development support on creation of collection id, package, plan creation, bind/rebind, dclgen.
- Technical support to the programmers and users.
- Debugging application programs, resolution of database errors, support for application errors and bad application performance.
- Supported performance issues of the database system.
- Worked closely with the development teams and data architects.
- Installed and scheduled backups, reorganization, recovery, and check point process.
- Coordinated database changes in production to prevent outage, avoid application impacts, and smoothly execute Modifications.
- Provided 24-hour 7-day production system support to ensure each service level agreement is achieved.
- Granting and Revoking Utilities access on Table Spaces.
- Creation, instance upgrade and version migration. Implement a library of homegrown scripts to automate the scheduling and notification of periodic maintenance jobs, including REORG, BACKUP, RUNSTATS, LOAD,RECOVERY.
- Performance tuning of the Test and development databases using explain tools.
- Schedule and execute reorg/runstat on the database tables.
- Identifying and Streamlining the weak processes within the system
Confidential
DB2 Database Administrator
Responsibilities:
- Created and managed production emergency fix process on the mainframes.
- Expertise in Database design, Installation, Upgradations, Configuration, Backup, Recovery, Database Security, and Query optimization
- Experience in Data Modeling (logical modeling and physical modeling) using Erwin
- Executed DB2 reports/utilities for performance/database organization & database growth analysis.
- Generate the Image copy gaps reports and took the backup copy of an table spaces and alert the application DBA teams to fill the IC gaps
- Involved in re-engineering the application to identify the highly CPU consuming jobs and improved the performance of those applications.
- Facilitated execution of load/unload/DSN1COPY of data for application development teams and to migrate the data between different regions.
- Altered the database objects like table spaces, tables, indexes, views, stored procedures.
- Executed the database utility jobs like Image Copy, Unload, Load, Rebuild IMS database.
- Runstats,Check Data, Rebuild and Recovery Utilities.
- Rotational On-call production DBA support.
- Expertise in Backup and restore of the databases.
- Expertise in Performance tuning & optimization.
- Responsible for database integrity & database backup and recovery.DB2 performance and problem determination,fixing production job abends in a 24 x 7 environment. Authorization and privileges for DB2objects.
- Responsible for the evaluation and creation of appropriate backup and recovery strategies for new database requirements.
- Bind plans & packages. Creation of Package, reviewing & providing DBA approval for production packagesUsing CHANGEMAN tool.
- Removing the restriction of pending status against DB2 object. Analyze data & develop and execute appropriate DDL to support projects.
- Responsible for database integrity & database backup and recovery.DB2 performance and problem determination, fixing production job abends in a 24 x 7 environment. Authorization and privileges for DB2objects.
- Maintained create/alter/drop/grant/revoke/rebind/free DB2 Objects using IBM DB2 Admin Tool
- Migration of database objects between databases and DB2 subsystems using IBM DB2 Admin tool.
- Automated the DB2 space reports generation for all the databases we support.
- Provided relations between tables Using Erwin Tool.
- SQL Server 2008 R2/2008/2005 Database Administration.
- Experienced in Performance tuning, Optimization, capacity planning.
- Migrated databases from SQL Server 2000 to SQL Server 2005 and 2008 R2
Confidential
Application Developer and DBA
Responsibilities:
- Created the new database objects like databases, table spaces, tables, index, views, and stored procedures. Altered the attributes of existing database objects.
- Application development support on creation of collection id, package, plan creation, bind/rebind, dclgen. Loaded and unloaded the data using IBM load and unload utility.
- Worked on writing and modifying native stored procedures as per business requirements.
- Performed the unit testing, system testing on stored procedures.
- Deployed the database objects from test region to system region/production region.
- Prepared and executed the DB2 Utility jobs runstat, rebuild index, rebind, and check data utilities.
- Fixed bind issues that will occur during the application testing in various phases.
- Converting from plan bind to package bind.
- Converting from simple Table Space to Segmented Table Space.
- Converting from Partitioned by Range to Partition by Growth Table Space.
- Granting and Revoking Utilities access on Table Spaces.
- Experience in writing complex queries using SQL
- Providing 24 hour on call support.
- Analyzing Client requirements and preparation of low level design document.
- Developing new COBOL, DB2, application programs as per the customer requirements and tested versions are delivered to the customer.
- Creating batch jobs and procs to extract programs and moving the code to production.
- Coding programs which writes data to XML file by reading DB2 tables.
- Validating XML files According to schema. Provided technical guidance to the team.
- Preparation of test plans, test cases, code reviews and delivery documents.