Hadoop Developer Resume
Charlotte, NC
PROFESSIONAL SUMMARY:
- Overall 9+ years Experience in the industry, worked in BigData Hadoop, Middleware and Java/J2EE technologies with Multiple domains including Banking, Healthcare, Insurance and Public sectors.
- Having 4+ years hands on Experience on Hadoop ecosystem components like Hadoop MapReduce, HDFS, HBase, Oozie, Hive, Sqoop, Pig, Zookeeper, Flume and Avro including their installation and configuration.
- In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node, and MapReduce concepts and experience in working with MapReduce programs using Apache Hadoop for working with Big Data to analyze large data sets efficiently.
- Experience in developing applications using Map Reduce for analyzing Big Data with different file formats.
- Experience in importing and exporting terra bytes of data using Sqoop from HDFS to Relational Database Systems and vice - versa.
- Knowledge of job workflow scheduling and monitoring tools like oozie and Zookeeper, administrative tasks such as installing Hadoop, Commissioning and decommissioning, and its ecosystem components such as Flume, Oozie, Hive and Pig.
- Knowledge of NoSql databases such as MongoDb and Cassandra.
- Experience in developing custom UDF's for Pig and Hive.
- Expertise in developing Pig Latin Scripts and using Hive Query Language.
- Installation, Configuring, troubleshooting and maintaining the WebLogic 7.0/8.1 and Apache Webserver.
- Deployed applications on Multiple WebLogic Servers and maintained Load balancing, high availability and Fail over functionality.Using Databases for configuring with the support of Weblogic Server.
- Experience in Distributed, Client-server, Middelware and Database application using Weblogic, SOA application servers and databases like Oracle and mysql servers.
- Having excellent domain knowledge and techno-functional skills, provided support for identifying gaps and designing solutions.
TECHNICAL SKILLS:
Operating Systems: Windows, Linux, UNIX.
Languages: Java, SQL,Python,NOSQL
Big Data Technologies/Tools: HDFS, MapReduce, PIG, Hive, HBase, Zookeeper, Sqoop.
Databases: Oracle 8i/9i/10g, MySQL,MS SQL Server.
NoSQL: Cassandra, MongoDB, CouchDB.
Web Technologies: Javascript, Servlets, EJB, JSP, HTML, XML, jQuery
Web/App Servers: Apache Tomcat, BEA Weblogic, JBoss
Version Control: SVN, GIT,CVS.
IDEs: Eclipse for Java.
Other: TOAD, MS Excel, MS Visio
PROFESSIONAL EXPERIENCE:
Hadoop Developer
Confidential, Charlotte, NC
Responsibilities:
- Supported Map Reduce programs those are running on the cluster.
- Developed the Pig UDF'S to pre-process the data for analysis.
- Implemented multiple Map Reduce Jobs in java for data cleansing and pre-processing.
- Experienced in loading data from Linux file system to HDFS.
- Developed job workflow in Oozieto automate the tasks of loading the data into HDFS.
- Responsible for creating Hive tables, loading data and writing Hive queries.
- Effectively involved in creating the partitioned tables in Hive.
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and extracted data from Teradata into HDFS using Sqoop.
- Worked extensively with Sqoop for importing metadata from Oracle.
- Configured Sqoop and developed scripts to extract data from SQL Server into HDFS.
- Expertise in exporting analyzed data to relational databases using Sqoop.
- Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
- Cluster co-ordination services through ZooKeeper.
- Responsible for running Hadoop streaming jobs to process terabytes of xmldata.
- Gained experience in managing and reviewing Hadoop log files.
Environment: HadoopHDFS, Map Reduce, Hive, Pig, Sqoop, Linux, Java, XML,SQL Server.
Hadoop Developer / Administrator
Confidential, Atlanta, GA
Responsibilities:
- Gathered the business requirements from the Business Partners and Subject Matter experts.
- Involved in installing Hadoop ecosystem components.
- Responsible to manage data coming from different sources.
- Monitoring the jobs to analyze performance statistics.
- Performing Unit Testing of completed jobs.
- Involved in database connection by using SQOOP.
- Applying optimization techniques at both Hadoop and Database level.
- Involved in creating Hive tables, and loading and analyzing data using hive queries.
- Developed Simple to complex MapReduce Jobs using Hive and Pig.
- Involved in running Hadoop jobs for processing millions of records of text data.
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required .
- Experienced in defining job flows.
- Experienced in managing and reviewing Hadoop log files.
- Extracted files from MongoDB through Sqoop and placed in HDFS and processed.
- Experienced in running Hadoop streaming jobs to process terabytes of XML format data.
- Load and transform large sets of structured, semi structured and unstructured data.
- Automation script to monitor HDFS and HBase through cronjobs.
- Plan, design, and implement processing massive amounts of marketing information, complete with information enrichment, text analytics, and natural language processing.
- Prepare multi-cluster test harness to exercise the system for performance and failover.
Environment: HadoopHDFS, Map Reduce, Hive, Pig, Sqoop, Oozie, HBase, Linux, Java, Xml, MongoDB.
Oracle WebLogic Administrator
Confidential, Oklahoma, OK
Responsibilities:
- Installed and Configured WebLogic 8.1 and 9.2 Application Servers.
- WebLogic Integration, Domains, Clusters on Solaris and wrote shell scripts for automation of WebLogic startup, and for log maintenance.
- Configured Clusters and troubleshooting issues raised with it.
- Application deployment (WAR, JAR, EAR) in clustered environment.
- Configured Database Support for WebLogic Application Server.
- Configured and administered Connection Pools, multi-connection pools and Data source objects for JDBC connections.
- Configured clustering EJB objects, JDBC connections and JMS connection factories.
- Configured JVM tuning for WebLogic application server.
- Monitored JVM Heap size using samurai.
- Troubleshooting, Load balancing, Testing and Code Profiling to measure the performance of the various applications deployed in WebLogic Server.
- Developed many shell scripts to automate the maintenance process of the WebLogic and recover the backed up WebLogic configuration.
- Configured and handled Jmeter and monitored the performance of Production J2EE components such as EJBs, JSPs and Servlets.
- Supported on call 24/7 for Production support.
- Configured SSL (Secure Sockets Layer) and obtained digital s, private key for WebLogic server to provide secure connections.
Environment: WebLogic Server 8.1/9.1/10.x, Apacheweb server,WebLogic portal, JDK1.5, J2EE, JSP, Servlets, EJB, JDBC, XML, JNDI Oracle9i, Ant1.5.1, Linux, FTP.
Oracle WebLogic Administrator
Confidential
Responsibilities:
- Troubleshooting, Load Balancing, Clustering, Deploying Applications, Performance Tuning and Maintenance of WebLogic servers.
- Create, configure and tune WebLogic domains/applications for performance.
- JDBC Connection Pool and Multi pool configuration with Oracle, MySql, DB2 and SqlServer, etc.,
- Deployment and troubleshooting of JAR, WAR, and EAR files in domain and clustered environments.
- Code deployments/promotions, performance monitoring and performance tuning.
- Upgradation of WebLogic servers in development, testing and production environment and applying patch and service packs.
- WebLogic Administration, Monitoring and Troubleshooting using Admin Console and JMX.
- Cluster Configuration - Single Cluster, Multiple Clusters and Deployment over clusters.
- Web Logic’s HttpProxyServlet and HttpClusterServlet configurations.
- Server Performance Tuning - Thread Dump Analysis, Core Dump Analysis when Server crashed unevenly.
- Creating security roles, group policy, user authentication, auditing and authorization.
- Work on Web Security tools- LDAP Server, SSL s and client authentication.
- Node Manager Configuration - remote managed server start and stop. SSL communication between admin server and Node Manager.
- Tuning JVM heap size to maximize the number of clients accessing the applications.
- Configuring digital s, using java key tool for secured applications.
- Automation of domain creation using WLST, ANT, and Domain Templates.
- Analyzing the Log Files, Thread Dumps, JVM Dumps, Exception Stack Traces.
- Patch updates using BSU, Service Pack Upgrades and Version Upgrades.
- Security Support with SSI, SSL, DES, RSA, TDEA etc.,
- Performed sanity testing on QA, Stage and Production environments after deployment.
- Configure and maintain the scheduled tasks like Thread dumps, log files backup etc.,
- Maintain the applications among Web Logic run-time processes in a cluster of application servers.
- Performance tuning of Web Logic server with respect to heap, threads and connection pools.
Environment: s: Solaris 2.8/5.8, Linux 4.x, Weblogic Server 8.1 sp3/sp5, I Planet 6.0, Apache 2.0., JDK 1.4, Jrockit Mission Control, IIS 6.0.
Oracle WebLogic Administrator
Confidential
Responsibilities:
- Provided support to Production, Development, Testing and Staging environments.
- Installation, Configuring, troubleshooting and maintaining the WebLogic 7.0/8.1 and Apache web server.
- Deployed applications on multiple WebLogic Servers and maintained Load balancing, high availability and Fail over functionality.
- Using Databases for configuring with the support of Weblogic Server.
- Created datasources depends upon the databases.
- Configured JDBC resources, data sources and bounded to the J2EE applications, configured the connection pools for the various data sources.
- Created the connection pools, various data sources, JMS connection Factories, Queues and topics.
- Configuring Execute Threads, Connection Pools and JMS Bridges etc. for different components of the system.
- Configured Application Servers BEA Weblogic8.1 on a single Server and a cluster of Servers.
- Involved in 24/7 support for production environments.
- Configure and maintain the scheduled tasks like Thread dumps, log files backup etc.,
- Maintain the applications among Web Logic run-time processes in a cluster of application servers.
- Performance tuning of Web Logic server with respect to heap, threads and connection pools.
Environment: s: SUN Solaris 2.x,Weblogic8.1 Application Server, iPlanet4.1, Sun One6.1, IIS 6.0, Windows (95, 98, NT, 2000, XP), Linux, Shell Scripting, Oracle 9i.