Hadoop Administrator Resume
Warren, NJ
SUMMARY
- Hands on experience in installing, configuring and using eco - System components like Hadoop Map Reduce, HDFS, Hbase, Pig, Flume, Hive and Sqoop.
- Hands-on experience with "Productionalizing" Hadoop applications (such as administration, configuration management, debugging, and performance tuning).
- Worked on Multi Clustered environment and setting up Cloudera Hadoop eco-System.
- Implemented in setting up standards and processes forHadoopbased application design and implementation.
- Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.
- Familiar with data architecture including data ingestion pipeline design, Hadoop information architecture, data modeling and data mining, machine learning and advanced data processing.
- Support and utilize multipleOraclebased applications including: SQL and PL/SQL, TOAD, Oracleviews, stored procedures, triggers, and the Microsoft Office suite of tools.
- Design, create/modify, and implement documented solutions as agreed to by all business partners responsible to design and integrate a computing system, from start to finish.
- Excellent communication skills, interpersonal skills, problem solving skills a very good team player along with a can do attitude and ability to effectively communicate with all levels of the organization such as technical management and customers.
- Interpersonal, communication, Planning, Organizational and Computer Skills
- Ability to prioritize work and manage daily workload
- Work on multiple projects with efficiency.
- Extensive experience in Software Development Life Cycle (SDLC) which includes Requirement Analysis, System Design, Development, Testing, and Implementation.
- Created numerous simple to complex queries involving self joins, correlated sub queries for diverse business requirements. Tuned and optimized queries by altering database design, analyzing different query options, and indexing strategies.
- Hands on experience in Unit Testing & System Testing.
- Expert in RDBMS concepts and constructs to design relational database model and implementation as per business needs.
- In-depth knowledge of T-SQL (DDL, DML)
- Professional Expertise in MS-Office i.e.MS-Excel, MS-Word, MS-Power point, MS Outlook
- Expertise in Database design, Entity Normalization and Database Creation, Maintenance, Monitoring, Performance Tuning and Troubleshooting.
- Extensive experience in Microsoft SQL 2005 Server upgrade from SQL Server 2000.
- In depth knowledge of SQL in DB2 environment.
- Ability to prepare documents including Technical Design, Testing strategy, and supporting documents.
- Outstanding analytical and technical problem-solving skills.
TECHNICAL SKILLS
Hadoop eco system: Hadoop, Hive, sqoop, Hdfs, Map-reduce, Yarn, hcatalog, Pig-latin
Hadoop Distributions: Apache hadoop, cloudera 3, cloudera 4.3.0, Pivotal hd 1.1.1(hadoop2.0.5 version), horton works hadoop.
Analytics tools: Datameer, tableau, GPtext
Monitoring tools: Nagios, Ganglia, Cloud era manager, Pivotal command center
RDBMS versions: Oracle 9i, 10g, 11g Mysql, Greenplum, Postgresql
RDBMS Tools: Toad, Tkprof, SQLTRACE
Operating Systems: REDHAT Enterprise Linux 4.0, 5.0, OEL 4.x. 5.x, 6.x
Windows: 2KServer, Windows XP
Scripting: UNIX Shell Scripting, Hadoop fs, shell, sed, Perl
Languages: C, C++, Java, SQL, PLSQL
PROFESSIONAL EXPERIENCE
Confidential, Warren, NJ
Hadoop Administrator
Responsibilities:
- Expertise in recommending hardware configuration for Hadoop cluster
- Installing, Upgrading and Managing Hadoop Cluster on Cloudera distribution.
- Trouble shooting many cloud related issues such as Data Node down, Network failure and data block missing.
- Managing and reviewing Hadoop and HBase log files.
- Experience with Unix or Linux, including shell scripting
- Strong problem solving skills
- Sharp analytical abilities and proven design skills
- Proven results-oriented person with a focus on delivery
- Built and configured log data loading into HDFS using Flume.
- Performed Importing and exporting data into HDFS and Hive using Sqoop.
- Managed cluster coordination services through Zoo Keeper.
- Mystifying and Demystifying nodes from the Cluster environment.
- Provisioning, installing, configuring, monitoring, and maintaining HDFS, Yarn, HBase, Flume, Sqoop, Oozie, Pig, Hive
- Patching and upgrading Cloudera and Hortonworks clusters
- Recovering from node failures and troubleshooting common Hadoop cluster issues
- Scripting Hadoop package installation and configuration to support fully-automated deployments
- Supporting Hadoop developers and assisting in optimization of map reduce jobs, Pig Latin scripts, Hive Scripts, and HBase ingest Required
Confidential, Bloomington, IL
Hadoop Administrator
Responsibilities:
- Built automated set up for cluster monitoring and issue escalation process.
- Administration, installing, upgrading and managing distributions of Hadoop (CDH3, CDH4, Cloudera manager), Hive, Hbase.
- Advanced knowledge in performance troubleshooting and tuning Hadoop clusters.
- Experience with Hadoop stack Map Reduce, Sqoop, Pig, Hive, Hbase
- Plan and execute on system upgrades for existing Hadoop clusters.
- Ability to work with incomplete or imperfect data
- Experience with real-time transactional data
- Strong collaborator and team player
- Installs, manages and configures the Hadoop clusters;
- Monitors the Hadoop jobs and performance
- Participate in development/implementation of Cloudera hadoop environment.
- Work with developers to advice on standards and best practices.
- Create and maintain technical documentation
Confidential, San Diego, CA
Hadoop Administrator
Responsibilities:
- Started with POC on ClouderaHadoopconverting one small, medium, complex legacy system intoHadoop.
- Installation of various Hadoop Ecosystems and Hadoop Daemons.
- Experienced in loading data from UNIX file system to HDFS.
- Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
- Integrate data from various sources intoHadoopand Move data fromHadoopto other databases using Sqoop import and Export.
- Use Cloudera manager to pull metrics on various cluster features like JVM, Running Map and reduce tasks etc.
- Backup configuration and Recovery from a Namenode failure.
- Decommissioning and commissioning the Node on running Hadoop cluster.
- Wrote SQL queries to load the required data to HDFS.
- Experienced in managing and reviewingHadooplog files.
- Created user accounts and given users the access to the Hadoop cluster.
- Worked with application teams to install operating system andHadoopupdates, patches, version upgrades as required.
- Advised file system team on optimizing IO forHadoop/ analytics workloads.
Confidential, Chicago
Environment: COBOL, DB2, JCL, VSAM, SPUFI, IBM FILEAID and ENDEVOR
Responsibilities:
- Analyze the new business requirements and convert business requirements into high level and detail designs. Prepare the design documents.
- Analysis, design and coding the actual changes utilizing COBOL, JCL and DB2.
- Preparation of Test plans and Test cases with QA Department.
- Build complex SQL Queries during some Business Requirements.
- Implemented different types of Constraints on tables designed in the data staging area for the migration.
- Created complex Stored Procedures, Triggers, and Functions for validation process & for loading data from the staging are to the product database in SQL 2005.
- Implemented Views (Partition View, Parameterized Views) and other T-SQL code for making joins between the tables.
- Developed, Maintained, and Monitored all import/export and Data Transformation into the Staging, Production and Data Warehouse environments.
- Generated Reports using Global Variables, Expressions and Functions for the reports.
- Maintained scheduling the subscription reports with the subscription report wizard.
- Serve as technical expert guiding choices to implement analytical and reporting solutions for client.
- Worked closely with DBA’s deploying packages for staging/production server with shell scripts.
- Documented all database objects, which were used in the project and for further upgrades and maintenance.