We provide IT Staff Augmentation Services!

Hadoop Analyst Resume

0/5 (Submit Your Rating)

Franklin Lakes, NJ

SUMMARY

  • 9+ years of comprehensive IT experience in Design, Development, Implementation, Testing and System Analysis using various technologies in Financial and Healthcare business domain.
  • 2 years of experience as Hadoop Developer and Data Analyst in Big Data Infrastructure, Distributed file system HDFS and MapReduce Framework.
  • Working knowledge in Hadoop ecosystem tools like Hive, Pig, Scoop, Oozie, Flume, Zookeeper
  • Hands on experience in developing MapReduce jobs using JAVA. Experience in installing, configuring and administrating the Hadoop Cluster of Major Hadoop Distributions.
  • Experience in writing Complex Pig and Hive Scripts including UDFs based on analytical requirements
  • Experience in working with NOSQL Database HBase for creating, inserting, querying and programming MapReduce operations
  • Extensive design and development experience in RDBMS - DB2, TERADATA and MySQL
  • In-depth understanding of Data Structure and Algorithms
  • Analysis and Development experience in Data Warehousing tools(ETL) AB INITIO (GDE 3.0.2) and Web EME
  • Extensive design and development experience in Mainframe Technology using COBOL, JCL, DB2, CICS, VSAM, IDMS, EAZYTRIEVE, MQ Series
  • Experience in FICO product Debt Manager V6.2.1
  • Experience in Business analysis, data analysis, data mapping, and gap analysis, creating reports, business process related documents like System Requirement, Flow Diagram, and System Understanding Documents.
  • Expertise in Financial service business domain
  • Extensive domain knowledge of Credit Card applications and processes
  • Experience in Pharmacy Benefit Management (PBM) processes in Healthcare business domain
  • Excellent analytical, problem solving, communication and interpersonal skills with ability to interact with individuals at all levels and can work as a part of a team as well as independently.

TECHNICAL SKILLS

Big Data: Hadoop, MapReduce, HDFS, Hive, Pig, Sqoop, Oozie, Flume, HBase, Zookeeper

Operating Systems: Red hat Linux, UNIX, Ubuntu 12.0.4, MVS/ESA, OS/390, Z/OS, Windows

Languages: Core JAVA, COBOL, CICS, MQ Series, C, Linux shell scripts

Distribution: Cloudera(CDH 3 and CDH 4), Apache Hadoop

Relational Database: DB2, TERADATA, MySQL, VSAM, IDMS

Tools: and Utilities: Eclipse, AB INITIO, Web EME, Teradata SQL Assistant(Queryman), QMF, SPUFI, DB2 Admin tool, Fileaid, Changeman, Endevor, Panvalet, Expeditor, Intertest, Infoman, Syncsort, DB2 Stored Procedure, REXX, CLIST, ICETOOL, CA-7, Control-M, Ezytrieve, Omegamon, SAS, FTP, NDM, FileZilla, Winscp, Connect Direct, Debt Manager V6.2.1(FICO Tool), Quality Center 10.0, HPSM

PROFESSIONAL EXPERIENCE

Confidential, Franklin Lakes, NJ

Hadoop Analyst

Responsibilities:

  • Extensive data analysis using HIVE and HIVE UDF’s for predictive analysis and create reports for business users.
  • Develop Pig Latin scripts to extract the data from the web server output files to load into HDFS
  • Develop Pig UDF’S to pre-process the data for analysis
  • Develop workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig
  • Managing and scheduling Jobs on a Hadoop cluster
  • Manage Hadoop clusters and involvement with setup, install, monitor and maintenance
  • Coordination with Data Scientists and involvement in research and predictive analysis
  • Provide quick response to ad hoc internal and external requests for data and creation of ad hoc reports.
  • Solve complex business problems with due considerations for hardware/software capacity and limitations, operating times and desired results.
  • Cluster co-ordination services through ZooKeeper
  • Support and maintain Map Reduce Programs (in Java) running on the cluster
  • Collected the logs data from web servers and integrated into HDFS using Flume
  • Data extraction from Teradata into HDFS using Sqoop.

Environment: Hadoop, MapReduce, HDFS, Hive, Pig, Zookeeper, Oozie, Sqoop, Flume, Java (jdk1.6),Eclipse, Hadoopdistribution of Cloudera, Teradata, MySQL, Flat files, Windows, UNIX Shell Scripting, “R” programming language

Confidential, Richmond, Virginia

Hadoop Consultant

Responsibilities:

  • Install and configure Hadoop MapReduce, HDFS and eco-system tools like Hive, Pig, Zookeeper, Flume and Sqoop
  • Develop multiple MapReduce jobs in Java for data cleaning and preprocessing.
  • Develop big data analytic models for customer fraud transaction pattern detection usingHivefrom customer transaction data.
  • Extensive data validation usingHIVE and Pig.
  • Importing and exporting data into HDFS using Sqoop.
  • Develop workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig
  • Design and implement Map Reduce jobs to support distributed data processing.
  • Designing NoSQL schemas in Hbase.
  • Experienced in managing and reviewingHadooplog files coming from different sources.
  • Experienced in runningHadoopstreaming jobs to process terabytes of xml format data.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Data extraction from IBM DB2 into HDFS using Sqoop.
  • Support Map Reduce Programs those are running on the cluster.

Environment: Hadoop, MapReduce, HDFS, HBase, Hive, Pig, Zookeeper, Oozie, Sqoop, Flume, Java (jdk1.6),Eclipse, Hadoopdistribution of Cloudera, Flat files, Windows, MySQL, IBM DB2, UNIX Shell Scripting

Confidential

Lead Data Warehouse (ETL)/Mainframe Developer

Responsibilities:

  • As a of part of Card-Recoveries IT team, primarily responsible for delivery of technical solutions and development to meet business intent by leading and managing the team interactions both internally and with external teams/business, and ensuring compliance with Confidential ’s policies, standards and best practices.
  • Develop new ETL process/projects using ABINITIO(GDE 3.0.2)
  • Data Analysis and reporting using TERADATA and QUERYMAN
  • Mainframe Development using COBOL, DB2, DB2 Store Procedure and VSAM
  • Review/Create Logical data model, Physical data model, High level and Detail design.
  • Create and review functional and non-functional system requirements and data mapping.
  • Perform analysis by looking at Mainframe code, AB INITIO graphs, Debt Manager Router code and running SQL queries in production.
  • Review Code, Test Plan, Test Scenarios, Test cases.
  • Performance analysis and testing to make sure all the SQL queries are tuned to use minimum database (DB2) resources and optimum access path.

Environment: AB INITIO(GDE 3.0.2), TERADATA, WEB EME, Unix, QUERYMAN, Mainframe, COBOL, JCL, CICS, EZYTRIEVE, DB2, VSAM, QMF, SPUFI, DB2 ADMIN tool, FILEAID, CHANGEMAN, SYNCSORT, DB2 Stored Procedure, REXX, CLIST, ICETOOL, Control-M, OMEGAMON, SAS, FTP, NDM, Connect Direct, Debt Manager V6.2.1(FICO Tool), Quality Center 10.0, HPSM Confidential, Newark, Delaware, USA

Confidential

Mainframe Developer

Responsibilities:

  • Design different functional gaps needed to be fulfilled for introduction of different types of accounts, automation logic for Invoicing and GL processing, new interfaces with three different sub systems SAP, CIIS and SPS.
  • Responsible for developing Mainframe modules of at least 10 major functional gap of ACE application, automated the GL and Invoice processing, debugging and providing resolutions of several technical issues.
  • UAT Support, coordination with business users and solve UAT issues.
  • Onsite-offshore coordination, preparing technical specification documents, knowledge transitions, Status reporting and leading Mainframe Development Team.

Environment: Mainframe, COBOL, JCL, CICS, EZYTRIEVE, MQ Series, DB2, VSAM, QMF, SPUFI, DB2 ADMIN tool, FILEAID, ENDEVOR, EXPEDITOR, INTERTEST, INFOMAN, SYNCSORT IBM Batch Utilities, DB2 Stored Procedure, REXX, CLIST, ICETOOL, Control-M

Confidential

Mainframe Developer

Responsibilities:

  • Mainframe development using COBOL, DB2, CICS, JCL and various tools as part of Chargeback and Interchange team
  • Analyzing Business Requirement Document (BRD) and interact with business users.
  • Perform system, integration and user acceptance testing.
  • Create process improvement automation tools
  • Handle production batch job abends and support as a production on-call person
  • Create technical documentation and flow diagrams for different processes

Environment: Mainframe, COBOL, JCL, CICS, EZYTRIEVE, MQ Series, DB2, VSAM, IDMS, QMF, SPUFI, DB2 ADMIN tool, FILEAID, ENDEVOR, PANVALET, EXPEDITOR, INTERTEST, INFOMAN, SYNCSORT IBM Batch Utilities, DB2 Stored Procedure, REXX, CLIST, ICETOOL, CA-7

We'd love your feedback!