Lead Hadoop Developer Resume
Miami, FL
PROFESSIONAL SUMMARY:
- 10+ years of professional IT experience which includes around 4 years of proven experience in Hadoop Administration using Cloudera (CDH), Hortonworks (HDP) and extensive years of experience in Linux and System Administration.
- Experienced in installation, configuration, supporting and monitoring of Hadoop clusters using Cloudera and Hortonworks distributions.
- Worked with clusters with size of data over 7 PB and 150+ nodes
- Experienced in understanding the client's Big Data business requirements and transform it intoHadoopcentric technologies.
- Analyzing the clients existingHadoopinfrastructure and understand the performance bottlenecks and provide the performance tuning accordingly.
- Working experience on designing and implementing complete end to end Hadoop Infrastructure.
- Experience on Commissioning, Decommissioning, Balancing, setting up high availability on master services, managing servers and tuning them for optimal performance for Hadoop work loads
- Experience in Cloudera Hadoop Upgrades, Patches and Installation of Ecosystem Products through Cloudera manager along with Cloudera Manager Upgrade.
- Hadoop cluster capacity planning, performance tuning, Optimization of Cluster to meet the SLA’s.
- Job and user management using Capacity Scheduler and Fair Scheduler.
- Setting up Kerberos authentication for Hadoop.
- Manage & review of Hadoop log files.
- Extensive experience on Linux and System administration.
- Excellent troubleshooting skills in Hardware, Software, Application and Network using iDRAC (Integrated Dell Remote Controller), Dell Open Manage, Iperf and other tools.
- Worked directly with vendors, partners and internal clients on gathering and refining technical requirements and designs in order to develop a working solution that addressed needs.
- Created and maintained user accounts, profiles, security, rights, disk space and process monitoring.
- Experience in deployingHadoopcluster on Public and Private Cloud Environment like Amazon AWS and OpenStack.
- Provisioning, building and support of Linux servers both Physical and Virtual using VMware for Production, QA and Developers environment.
- Strengths include good team player, excellent communication interpersonal and analytical skills and ability to work effectively in a fast - paced, high volume, deadline-driven environment.
- Ability to adapt to evolving technology, strong sense of responsibility and .
TECHNICAL SKILLS:
Technologies/Tools: Hadoop, HDFS, YARN, MapReduce, Cloudera, Cloudera Manager, HBase, Hive, Pig, Oozie, Sqoop, Flume, Spark, Kafka, TEZ, Storm, ZooKeeper, AWS, RackSpace, VMWare, HortonWorks, CDH 4, CDH 5, Shell Scripting
Databases: SQL Server, Oracle 10g, SQL Profiler
Servers: Tomcat
Operating Systems: Linux RedHat, CentOs, Windows Server 2003/2008, Win 7/8
WORK EXPERIENCE:
Lead Hadoop Developer
Confidential, Miami, FL
Responsibilities:
- AnalyzingHadoopcluster and different Big Data analytic toolsincluding Pig, HiveandSqoop.
- Developed variousJAVA UDF functionsto use in bothHive and Impalafor ease of usage in various requirements.
- Transformed theImpala queriesintohive scriptswhich can be run using theshell commandsdirectly for higher performance rate.
- Created the shell scripts which can be scheduled using Oozie workflows and even the Oozie Coordinators.
- Developed theOozie workflowsto generate monthly report files automatically.
- Creating multiple MapReduce jobs in Pig and Hive for data cleaning and preprocessing.
- Successfully loading files to Hive and HDFS fromOracle,SQL Server using SQOOP.
- Writing Hive jobs toparsethe logs and structure them in tabular format to facilitate effective querying on the log data.
- Creating Hive tables, loading with data and writing Hive queries.
- Using Hive to analyze the partitioned data and compute various metrics for reporting.
- Managing and reviewing theHadooplog files.
- Using Pig to do Transformations, even joins and some pre-aggregations.
- Exporting data from HDFS environment into RDBMS using Sqoop for report generation and visualization purpose.
Environment: Hadoop, MapReduce, Sqoop, HDFS, Hive, Pig, Oozie, Java, Oracle 10g, MySQL, Impala.
Hadoop Administrator
Confidential, Plano, TX
Responsibilities:
- Installing, configuring, and administrating Hadoop cluster using CDH
- Responsible for implementation and ongoing administration ofHadoopinfrastructure.
- Deploy new hardware and software environments required forHadoopand to expand memory and disks on nodes in existing environments.
- Handle the data exchange between HDFS and different Web Applications and databases using Flume and Sqoop.
- Monitor the data streaming between web sources and HDFS.
- Configuring YARN and optimizing Memory related settings.
- Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.
- Collaborating with application teams to install operating system andHadoopupdates, patches, version upgrades when required.
- ScreenHadoopcluster job performances and capacity planning.
- Configuration change management where changing the configuration properties of the cluster based on volume of the data being processed.
- MonitorHadoopcluster connectivity and security.
- Importing and exporting data using SQOOP from RDBMS to Hadoop.
- Working with data delivery teams to setup newHadoopusers which also includes setting up.
- Linux users and testing HDFS, Hive, Pig and Map Reduce access for the new users.
- Performing Linux systems administration on production and development servers (RedHat Linux, CentOS and other UNIX utilities).
- Commissioning and Decommissioning the Data nodes in Hadoop Cluster.
- Configured a 20-30 node (Amazon EC2 spot Instance) Hadoop cluster to transfer the data from Amazon S3 to HDFS and HDFS to Amazon S3 and also to direct input and output to the Hadoop MapReduce framework.
- Job and user management using Capacity Scheduler.
- Installing Patches and packages on Unix/Linux Servers.
- Install and Configure VMware vSphere client, Virtual Server creation and resource allocation.
- Performance Tuning, Client/Server Connectivity and Database Consistency Checks using different Utilities.
- Inputs to development regarding the efficient utilization of resources like memory and CPU utilization based on the running statistics of Map and Reduce tasks.
Environment: Hadoop, Cloudera, Redhat Linux/Centos 4, 5, 6, Logical Volume Manager,HDFS, Hive, Pig, Sqoop, Flume, VMware ESX 5.1/5.5, Apache and Tomcat Web Server, Oracle 11,12
System Administrator
Confidential
Responsibilities:
- Performing remote installations/upgrades of Confidential applications.
- Actively monitor and act on alerts for production environments.
- Installation, configuration and Administration of Red hat Linux Servers, HP-UX and Solaris Servers.
- Worked on Server migration/Data Center move projects in various stages for all Linux Servers.
- Acquired comprehensive requirements from Project Managers, Team Leads pertaining to the servers to be migrated.
- Setup and configure failover load balancing using IP bonding for network cards.
- Configured volume groups and logical volumes, extended logical volumes for file system growth needs using Logical Volume Manager (LVM) commands.
- Updating YUM Repository and RedHat Package Manager (RPM)
- Applied appropriate support packages/patches to maintain system integrity.
- Setting up cron schedules for backups and monitoring processes.
- Server on various business critical database & application production servers.
- Monitored Linux server for CPU Utilization, Memory Utilization, and Disk Utilization for performance monitoring.
- Data migration from existing data stores to Hadoop.
- Setup NFS file systems and shared them to clients.
- Performed capacity analysis, monitored and controlled disk space usage on systems.
- Monitored system activities and fine-tuned system parameters and configurations to optimize performance and ensure security of systems.
- Work to resolve production issues and meet the laid down SLA’s.
- Administration of RHEL 3, 4, 5 which includes installation, configuration, testing on both physical and virtual machines.
- Installation and configuration on System imager server on RHEL 5
Environment: Red Hat Enterprise Server (RHEL 3, 4 and RHEL 5), HPUX-11i, 11.x, 10.x, Solaris 8, 9 and 10, EMC Power path, Oracle RAC, Perl scripting, VMware ESX Server, Emulex, DELL Power Edge 2850, 2950, 6850, 6950, HP DL 585, 380, 385, SE 1210, HP Blade servers, EMC and Clarion Storage
System Administrator
Confidential
Responsibilities:
- Create technology specific forums as requested by MSFT employees.
- Granting moderator permissions to the forums users, banning users from the forum.
- Adding and removing owners from the blogs.
- Removing spam from the forums and newsgroups.
- Administrating the forums.
- Granting the publishing rights to the user for different locales
- Unlocking the content from the data base.
- Creating the newsgroups on the NNTP server.
- Reporting to the client on daily basis.
- Escalate the issues to Tier3 as per the need.
- Required login credentials to be created and ensured are active for all team members.
- Create Quarterly performance reports and SLA documents.
System Administrator
Confidential
Responsibilities:
- Installing, Maintaining, Administering, Troubleshooting Hardware, Peripherals, LAN Networks (including wireless).
- Patch Management process for securing Desktops (Windows XP & Vista). Installation of patches and service packs.
- Installing and configuring Win XP and Vista.
- Installing, Configuring and Maintaining Confidential Outlook and Outlook Express.
- Troubleshooting the day-to-day problem in LAN.
- Troubleshooting Printer related issues
- Good Experience in solving the problems remotely, by capturing the desktop with the help of Remote Desktop & Net meeting.
- Installation, Implementation and support of LAN in Win XP & Win 2000 server.
- Third party vendor interaction.