We provide IT Staff Augmentation Services!

Bigdata Senior Analyst/architect Resume

0/5 (Submit Your Rating)

Los, AngeleS

SUMMARY

  • Professional IT Experience of 12 years, including 4 years of work experience in BigData, Hadoop solution architect, design, development and Ecosystem Analytics in Banking and Financial sectors
  • Excellent understanding and knowledge of Big Data and Hadoop architecture
  • Hands on experience on major components in Hadoop Ecosystem like Spark, MapReduce, HDFS, HIVE, PIG, HBase, Sqoop, Oozie, Flume, Kafka, Tableau
  • Strong knowledge in using Map reduce programming model for analyzing the data, implementing custom business logic and perform join optimization, secondary, custom sorting
  • Hands on experience writing complex Hive, SQL queries, PIG scripts for data analysis to meet business requirements
  • Excellent understanding of NoSQL databases like HBase, Cassandra and MongoDB
  • Experience in importing and exporting data using Sqoop from HDFS (Hive & HBase) to Relational Database Systems and vice - versa
  • Worked on Spark SQL, Streaming and involved in Performance optimizations in Spark
  • Extensive hands-on experience using Informatica Power Center 9.x/8.x/7.x (Repository Manager, Mapping Designer, Workflow Manager, Workflow Monitor,) and SAS EG
  • Experience in reporting tools SAP Business Objects (IDT, UDT, Interactive Analysis and BI Launch Pad/InfoView, CMC, CCM, Dashboards, Lumira)
  • Certified and Strong experience with MPP database Netezza, its architecture and underlying data distribution
  • Experience in using different file formats like AVRO, Parquet, RC, Sequence and CSV
  • Worked on Performance Tuning, identifying and resolving performance bottlenecks in various levels
  • Able to work collaboratively, quickly understand system architecture and business needs while committed to achieving corporate goals
  • Experienced in Agile/Lean and PMI project management methodologies

TECHNICAL SKILLS

Big Data/Hadoop: HDFS, Hive, Pig, Sqoop, HBase, Oozie, Spark

ETL Tools: Informatica, BODS, SSIS, SAS EG

BI Tools: SAP BusinessObjects, QlikView, Tableau

Database: Netezza, Oracle, DB2, SQL Server, MySQL, MS Access

Operating Systems: UNIX (AIX/Solaris), Linux, Windows 2008, AS400

Language: NZSQL, PL\SQL, T-SQL, Shell Script, Java, Scala

Database Tools: SQL Developer, SQL Data Modeler, SSMS, TOAD

Testing Tools: HP Quality Center, Jira, Remedy

Methodologies: PMP, Agile

PROFESSIONAL EXPERIENCE

Confidential, Los Angeles

BigData Senior Analyst/Architect

Responsibilities:

  • Gathered business requirements in meetings for successful implementation and moving it to Production
  • Analyse the business requirements and converting into technical documentation for easy understanding for the development team
  • Job duties involved the design, development of various modules in Hadoop Big Data Platform and processing data using Spark, HBase, Sqoop, Hive
  • Implemented POC to migrate MapReduce jobs into Spark RDD transformations using Scala
  • Ingested huge amount of ClickStream data into Hadoop using Parquet storage format
  • Performed advanced procedures like text analytics and processing, using the in-memory computing capabilities of Spark
  • Developed Spark scripts by using Scala shell commands
  • Successfully loaded files to Hive and HDFS from Oracle SQL Server using SQOOP
  • Worked with Oozie Workflow manager to schedule Hadoop jobs and high intensive jobs
  • Extensively used Hive/HQL or Hive queries to query data in Hive Tables and Loaded data into HIVE tables
  • Worked closely in setting up of different environment and updating configurations
  • Introduced Tableau Visualization from hadoop to produce reports for Business and BI team
  • Daily Scrum Meetings with the team for the Updated status and the action plan of the day

ENVIRONMENT: Hadoop, HDFS, HBase, Spark, Spark Streaming, SparkSQL, Pig, Hive, Sqoop, Tableau, Linux, Oracle, Shell Script, Scala

Confidential, Los Angeles

BigData Senior Analyst/Architect

Responsibilities:

  • Involved in design, development of various modules in Hadoop Big Data Platform and processing data using MapReduce, Hive, Pig, Sqoop, Oozie
  • Designed & implemented Java MapReduce programs to support distributed data processing
  • Importing and exporting the data using Sqoop from HDFS to Oracle and vice-versa
  • Developed Hive scripts to handle Avro and JSON data in Hive using Hive SerDe's
  • Automated all the jobs for pulling data from FTP Server to load data into Hive tables, using Oozie workflows
  • Used Oozie to automate the data loading into Hadoop Distributed File System
  • Involved in implementation of Hadoop Cluster for Development and Test Environment
  • Used different file formats like Text files, Sequence Files, Avro
  • Assisted in creating and maintaining Technical documentation to launching HADOOP Clusters and even for executing Hive queries and Pig Scripts.

ENVIRONMENT: Hadoop, MapReduce, Pig, Hive, Sqoop, Linux, Oracle, Shell Script

Confidential

Senior Project Associate

Responsibilities:

  • Participated in planning sessions with offshore technical team for resource allocation and capacity planning
  • The project requirement is to build a scalable and reliable Big Data platform that will be core component for forthcoming business needs
  • Designed & implemented MapReduce programs to support distributed data processing
  • Importing and exporting data using Sqoop from HDFS to Netezza and vice-versa
  • Developed UDF functions for Hive and wrote complex queries in Hive for data analysis
  • Involved in loading and transforming large sets of Structured, Semi-Structured and Unstructured data and analyzed them by running Hive queries
  • Loaded and transformed large sets of structured, semi structured data using Pig Scripts

ENVIRONMENT: Hadoop, MapReduce, HBase, Hive, Pig, Sqoop, Netezza

Confidential

Senior Project Associate

Responsibilities:

  • Interacting with various internal functional teams and team members to analyze business requirements and to outline the initial setup architecture and reporting solution
  • Involved in Dimensional modeling (Star Schema) of the Data warehouse and created conceptual, logical, physical data models
  • Involved in building the ETL architecture and Source to Target mapping to load data into Data warehouse
  • Created technical ETL mapping documents to outline data flow from sources to targets
  • Extracted the data from the flat files and SQL Server databases into staging area and populated onto Data warehouse
  • Review existing code, lead efforts to tweak and tune the performance of existing Informatica processes
  • Extensively used SQL* loader to load data from flat files to the database tables in Oracle
  • Build the new universes as per the user requirements by identifying the required tables from Data Warehouse and by defining the universe connections
  • Used Derived tables to create the universe for best performance, and use context and alias tables to solve the loops in Universe
  • Created complex reports stating Trading Statistics per month and per year using cascading and user objects like measure objects by using @aggregate aware function to create the summarized reports
  • Daily/Weekly/Monthly MAS Regulatory reporting with consolidated exchange statistics for Market Operations Dept and Clearing & Settlement Dept

ENVIRONMENT: Informatica 8.6, SAP BusinessObjects XIR3.1, Netezza, Oracle 11g, PL\SQL, SQL Server 2008, Shell/Batch/VB Script, Linux, Win 2008

Confidential

Senior Consultant

Responsibilities:

  • Interacted with business analysts to identify information needs and business requirements for reports Involved in the design of Conceptual, Logical and physical data modeling of the data warehouse based on star schema methodology, Created Data Models using Erwin
  • Scheduling and monitoring Informatica workflows and batches
  • Maintenance of mappings and load process, Enhancements of existing mapping and developing new mappings to facilitate change requests
  • Responsible for supporting, troubleshooting, Data Quality and Enhancement Issues.
  • Debugging, performance tuning of mappings, session recovery and maintenance of workflow jobs in production
  • Build the new universes as per the user requirements by identifying the required tables from Data mart and by defining the universe connections.
  • Developed critical reports like drill down, Slice and Dice, master/detail for analysis.
  • Created complex reports using the multiple data providers and synchronized the data providers by linking the common objects to create the common reports.
  • Developed universe user documentation guide for the end users reference.

ENVIRONMENT: Informatica 7.1, Business Objects, SAS Base, SAS EG, Oracle 10g, PL/SQL, DB2, Shell Scripting, Clear Case, Clear Quest, HP Quality Center, Control-M, UNIX, AS400 and Win 2000

Confidential

Senior Consultant

Responsibilities:

  • Developed complex mapping using Expression, Aggregate, Lookup, Sequence Generator, Update Strategy, Stored Procedure, Router and Dynamic parameters.
  • Implemented slowly changing dimensions to maintain historical data using Type I Type II
  • Scheduling and monitoring Informatica workflows and batches
  • Responsible for supporting, troubleshooting, Data Quality and Enhancement Issues.

ENVIRONMENT: Informatica 7.1, Windows 2000, Oracle 9i, Erwin, Cognos Series 7 and Cognos Reportnet 1.1 Mr1

Confidential

Developer

Responsibilities:

  • Responsible for complete life cycle of the system.
  • Responsible client side requirement study, preparation of system specification, requirement specification.
  • Responsible for creating application setup, installation, testing and troubleshooting

ENVIRONMENT: HP, MySQL, SQL, Java script, Linux, Apache

Confidential

Developer

Responsibilities:

  • Cleaned up and rewrote much of the front end code to meet validation requirements
  • Provide detailed project requirements, technical specifications as well as a step by step map for project execution
  • Actively participated in all database related activity

ENVIRONMENT: PHP, MySQL, SQL, Java script, Linux, Apache

We'd love your feedback!