Hadoop Administrator Resume
Atlanta, GA
SUMMARY
- Having 12 + years of IT Experience in Database and Dataware housing domain.
- Having 3 + years of working experience on Hadoop Admin.
- Strong technical expertise in Hadoop Administration (BIG DATA Administration) with Hadoop 2.0 Multi - Node cluster Set up, Name node (HDFS) High Availability, HDFS Federation, MRv2 and YARN Framework, Apache Oozie workflow scheduler, Hadoop cluster Implementation, Cloudera Manager and Hortonworks Ambari.
- Having 8 years of working experience on IBM DB2 DBA.
- IBM Mainframe technologies with Data warehouse tools & concepts - DB2 Admin tool, Omegamon II for DB2, DB2 Log Analysis Tool, IBM Tivoli performance Analyzer tool/Performance Tuning tool for IMS/DB, Z/OS 390, MVS, TSO, ISPF, VS COBOL for OS/390, PL/1, CICS, COBOL, JCL, DB2, VSAM, IMS/DB, IMS DB/DC, SQL, Embedded SQL, CICS, JCL, VSAM, File-aid, IDCAMS, SORT, IEBGENER, IEFBR14, QMF, SPUFI, FTP, ISPF, TSO, Change Man, EXPEDITER (Batch & Online), Db2 tools & Utilities, Control-M, Performance tuning & optimization, Connect Direct, Lotus notes & Web requester.
- 24x7 production support as a primary response for both batch & online abends.
- Experience in all phases of Software Development Life cycle (SDLC) -- Requirements Gathering, Analysis, Construction, Unit Testing, and Support to various Testing techniques - Integration, Quality Assurance (QA) Test, User Acceptance (UAT) Test, and Regression Test.
- Strong experience in Upgradation/Migration/Conversion projects.
- Roles played: Hadoop Administrator (Big Data Admin), DB2 DBA, IMS/DB DBA, Technical Lead and Application Developer.
- Interested/Willing to learn new technologies based on the business requirement (rapid adopting ability for the new technology).
- Worked in both Highly-Structured and Less-Structured IT Development/Maintenance environments.
- Served on on-call rotation, handling overnight problem calls.
- Ability to prepare technical design, project plan, unit test case development, test result capturing, status report documentation.
- Good analytical, programming and communication skills with accent on constructive teamwork.
- Technical knowledge on Dataware Housing tools (Data stage and Cognos).
TECHNICAL SKILLS
Hadoop / Big Data: HDP 2.6.1, HWX Ambari, HDFS, MapReduce, Yarn, Hive, Pig, HBase, Sqoop, Flume, Oozie, Kerberos, Spark, Scala, Zookeeper, Splunk,YUM, Solar, Kafka, Apache Ranger, Bigtop, Big Insights, Cloudera Manager, Nagios, Ganglia, Graphite, Strom, Unix, Shell Scripting, Mongo DB.
Programming languages: Pl/1, Cobol, Jcl, Sql, Rexx Asm, Pls, Clist
ETL: Informatica, Datastage, Cognos.
Database: HBase, DB2, IMS/DB
Operating Systems: Windows, Ubuntu, Centos, Red Hat Linux, UNIX, MVS
PROFESSIONAL EXPERIENCE
Confidential, Atlanta, GA
Hadoop Administrator
Responsibilities:
- Experience using Hortonworks platform and their eco systems, installing, configuring and using ecosystem components like MapReduce, HDFS, Hive, Sqoop and Flume.
- Upgrade HDP / Ambari with minimum down time and work with HWX to resolve bugs/fixes for the HDP 2.5.3 / HDP 2.6.1.
- Involved in Hadoop cluster task like Adding and Removing Nodes without any effect to running jobs and data.
- Installed and configured Hadoop ecosystem like HBase, Flume.
- Load log data into HDFS using Flume. Worked extensively in creating MapReduce jobs to power data for search and aggregation
- Kerberos/LDAP skills Falcon, Ranger, Knox, Ambary - Deep knowledge w/practical exp.
- Experience on Apache Knox Gateway security for Hadoop Clusters.
- Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Hive.
- Involved in installing Hadoop Ecosystem components.
- Experienced in providing security to Hadoop cluster with Kerberos and integration with LDAP/AD Confidential Enterprise level.
- Responsible to manage data coming from different sources.
- Involved in HDFS maintenance and loading of structured and unstructured data.
- Building massively scalable multi-threaded applications for bulk data processing primarily with Apache Spark and PIG on Hadoop.
- Worked along with the Hadoop Operations team in Hadoop cluster planning, installation, maintenance, monitoring and upgrades.
Environment: Environment: HDP, Ambari, Hbase, Hive, Pig, Kerberos, Kafka, Strom, Sqoop, Knox, Unix, Shell Script, Linux, Apache Ranger, Yarn, Apache Oozie workflow, Flume, Zookeeper
Confidential, Morrisville, NC.
Hadoop Administrator
Responsibilities:
- Deployed multi-node development, testing and production Hadoop clusters with different Hadoop components (HIVE, PIG, SQOOP, OOZIE, FLUME, HCATALOG, ZOOKEEPER) using Hortonworks Ambari.
- Configured Capacity Scheduler on the Resource Manager to provide a way to share large cluster resources.
- Deployed Name Node high availability for major production cluster and designed automatic failover control using zookeeper and quorum journal nodes.
- Configured Oozie for workflow automation and coordination.
- Good experience in troubleshoot production level issues in the cluster and its functionality.
- Backed up data on regular basis to a remote cluster using distcp.
- Regular Ad-Hoc execution of Hive and Pig queries depending upon the use cases.
- Regular Commissioning and Decommissioning of nodes depending upon the amount of data.
- Experience in Disaster Recovery and High Availability of Hadoop clusters/components.
- Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.
- Diagnose and resolve performance issues and scheduling of jobs using
- Configured Fair scheduler to share the resources of the cluster.
- Experience designing data queries against data in the HDFS environment using tools such as Apache Hive.
- Imported data from MySQL server to HDFS using Sqoop.
- Manage the day-to-day operations of the cluster for backup and support.
- Used the RegEx, JSON and Avro SerDe’s for serialization and de-serialization packaged with Hive to parse the contents of streamed log data.
- Writing shell scripts to automate the administrative tasks.
- Implemented Hive custom UDF’s to integrate the Weather and geographical data with business data to achieve comprehensive data analysis.
- Used Oozie workflow engine to manage interdependent Hadoop jobs and to automate several types of Hadoop jobs such as Java map-reduce, Hive and Sqoop as well as system specific jobs.
- Worked along with the Hadoop Operations team in Hadoop cluster planning, installation, maintenance, monitoring and upgrades.
Environment: HDFS, Hbase, Hive, Pig, Kerberos, Kafka, Bigtop, Strom, Sqoop, Knox, Unix, Shell Script, Linux, Apache Ranger, Mongo DB, Splunk, Yarn, Apache Oozie workflow, Flume, Zookeeper, RegEx, JSON.
Confidential, NYC, NY
Hadoop Administrator
Responsibilities:
- Setup a Multi Node Cluster. Plan and Deploy a Hadoop Cluster using with Cloudera Manager.
- Start, Stop and restart clusters and performed all administrative functions by Cloudera Manager
- Performed in developing purge/archive criteria and procedures for historical.
- Performance tuning of Hadoop clusters and Hadoop MapReduce routines.
- Screen Hadoop cluster job performances and capacity planning
- Monitor Hadoop cluster connectivity and security. Manage and review Hadoop log files.
- File system management and monitoring. HDFS support and maintenance.
- Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.
- Collaborating with application teams to install operating system and Hadoop updates, patches, version upgrades when required. Point of Contact for Vendor escalation.
- Created reports for the BI team using Sqoop to export data into HDFS and Hive.
- Used sqoop for bringing in the raw data, populate staging tables and store the refined data in partitioned tables.
- Created Hive queries that helped market analysts spot emerging trends by comparing fresh data.
- Developed Hive queries to process the data for analysis by imposing read only structure on the stream data.
- Performed minor and major upgrades, commissioning and decommissioning of data nodes on Hadoop cluster.
- The Hive tables created as per requirement were internal or external tables defined with proper static and dynamic partitions, intended for efficiency.
- Used Oozie workflow engine to manage interdependent Hadoop jobs and to automate several types of Hadoop jobs such as Java map-reduce, Hive and Sqoop as well as system specific jobs.
- Worked along with the Hadoop Operations team in Hadoop cluster planning, installation, maintenance, monitoring and upgrades.
- Adding, removing, or updating user account information, resetting passwords, etc.
- Installing and updating packages using YUM.
Environment: Cloudera Manager, Hbase, HAWQ, Hive, Pig, Kerberos, Kafka, Sqoop, Nagios, Ganglia, Graphite Apache Ranger, Splunk, Yarn, Apache Oozie workflow scheduler, Flume, Zookeeper, RegEx, JSON, Spark
DB2 Administrator
Confidential
Responsibilities:
- Performed performance, maintenance and utilities associated with each structure (REORG, LOAD, UNLOAD).
- Performed in DB Extent addition in production and test databases as well.
- Performed U1 refresh / create a new test environment 1st Sunday of every month
- Performed storage pool balancing on required weekends
- Performed Stored procedures, Functions and Triggers.
- Develop and execute appropriate Data Definition Language (DDL) to support project.
- Performed in developing purge/archive criteria and procedures for historical.
- Created JCL to take a BACKUP for the daily balancing dataset.
- Worked on REORG utility to reorg the database, DBRM, Package and Plans.
- Created JCL for IMAGE COPY, UNLOAD DATABASE Recovery and BACKUP.
Environment: DB2, VM, QMF, SPUFI, IMS/DB, REXX, RTC, RPM, MQ series and TOD.
DB2 Administrator
Confidential
Responsibilities:
- Performed performance, maintenance and utilities associated with each structure (REORG, LOAD, UNLOAD).
- Performed Stored procedures, Functions and Triggers.
- Develop and execute appropriate Data Definition Language (DDL) to support project.
- Created JCL to take a BACKUP for the daily balancing dataset.
- Making changes in the base code to in corporate additional functionalities.
- Reviewing code changes and testing updates.
- Analysis of defects encountered in the project on monthly basis & sharing it with the team.
- Responsibility As a team member I was involved in analyzing the Business requirements, preparing the Technical specification document and coding.
- Taking up tasks independently from onsite, analyze them, clarify queries with onsite and deliver the completed task.
Environment: PL/1, CICS, JCL, COBOL, DB2, IMS/DB, REXX, ASM, PLS, RMDSSCLM, FM, RTC, RPM, MQ series.
DB2/COBOL Developer
Confidential, Southbury, Connecticut.
Responsibilities:
- Gather requirements via discussions/meetings with users, IT Management.
- 24x7 support to production, provide resolution to night-time production job abends, attend conference calls with business operations, system managers if any issues in batch stream.
- Develop solutions to the business under the direction of IT Management.
- Serve on on-call rotation, handling overnight problem calls.
- Develop database changes when applicable under the direction of I/T Management.
- Analyze problems and solve on a day to day basis with appropriate 360-degree user communication.
- Created regressive test plans and conduct unit testing
- Work with Client I/T Management & staff for co-ordination of System and end user testing.
- Participate in Client/Team meetings send regular status updates and Task assignments to the Team members.
- Implementation of application changes as required, in accordance with SDLC policy.
- Assess user inquiries and work with I/T management to determine priorities.
- Create weekly status report and participation in team meetings.
- Providing class room training related to technology & application to the team and new joiners.
Environment: PL/1, CICS, JCL, COBOL, DB2, REXX, ASM, PLS, DEM, OSPCADP, IBM FA, FM.