We provide IT Staff Augmentation Services!

Bigdata Lead Developer Resume

4.00/5 (Submit Your Rating)

SUMMARY:

  • 14+ years of total work experience in IT Industry in Analysis, Design, Development, Maintenance and Migration of various software applications
  • Worked primarily in the domain of Insurance, Banking, Finance & Telecom
  • Extensively worked in designing & building ETL systems, automated batch jobs
  • 5+ Years of Big data experience
  • 2+ years of Oracle DBA experience
  • 5+ years of Team Lead experience
  • Technological Fortes: Installation & Configuration of Big Data clients (HDP 2.6.5, CDP 7.1.7), hadoop securities - Kerberos/Knox/Ranger, Hive, Sqoop, Spark, Kafka, Scala programming, Oozie, Presto, Solr, Google Cloud (GCP) - IAM, Virtual Instances, Docker, Containerization (GCR, Artifact Registry), CLoud storage Buckets, Tableau, Talend 7.1, Oracle 19c/12c/11g/10g SQL & PL/SQL programming, Oracle GoldenGate, Informatica 9.5, Datastage, Autosys/Control-M scheduler, Unix, Oracle Forms & reports 10g, BIRT Reporting, Gitlab, Snowflake
  • Performing Big data Architect/Lead developer roles in designing and developing high performance, robust and secure solutions based on Horton Works ecosystems (HDP 2.6.5) and Cloudera Data Platform (CDP 7.1.7)
  • Provides Data Migration plans, optimum data load mechanisms using Hive, Sqoop, Spark & Unix
  • Design and Builds ETL loads and other Data Pipelines for the ETL loads into Target Hive.
  • Builds Secure Kerberized/Knox ODBC/JDBC connectivities between various third party tools (SQL developer, Toad Data Point, Teradata SQL assistant, Java Application) with Hive Database.
  • Builds Scala program for data loading processes.
  • Builds & Automates monitoring scripts for tracking various services, data load statuses, file archival processes etc.
  • Designed and Built a Replication system from Oracle GoldenGate to Hive, Oracle Staging area into Hive.
  • Working on Google cloud platform (GCP) in creating IAM roles, VM, Managed Instance Groups (MIG), Dockerization, Artifact Registry, Big Query. Cloud Shell.
  • Expertise in Installation & Configuration of Telecom application “Subex” in Hortonworks and Cloudera platform along with installation of Oracle, Presto & Solr.
  • Performed installation & configuration of - Oracle SQL connector for Hadoop (OSCH), Oracle copy to Hadoop, use of oraoop jar in sqoop import for direct path loading from Oracle to Hive.
  • Performed integration of SQL server, Teradata, Talend, Tableau, TOAD data point with kerberized Hadoop cluster and establishing connection with Hive database.
  • Scheduling Oozie jobs and supports error in data load if any.
  • Staying current with emerging tools and technologies and recommend adoption that will provide competitive advantage and development/delivery efficiencies
  • Lead a team of designers/developers and guide them throughout the system implementation life cycle
  • Expertise in working with Kerberos security system, Knox connectivity, Ranger Policy
  • Designed and Developed Data Ingestion Framework to automate source systems data into Hadoop in pulling Full & Incremental data sets on a daily, weekly & monthly basis.
  • Performed Data migration from Oracle 12c to Hadoop hive, for the entire Risk Assurance application data for Verizon wireline
  • Experience with real-time processing and streaming technologies such as Spark & Kafka
  • Build ETL system using spark scala for multiple sources.
  • Performed Production support for the Application’s data load tasks, Biweekly Release deployments.
  • Expertise in working with Oozie scheduler using Scala.
  • Expert in Unix scripting.
  • Developed dashboard reports using Tableau.
  • Knowledge of Data Processing platforms like Talend 7.1
  • Very good in-depth knowledge of Subex Product and its Technical Architecture
  • Built spring boot Java application for an REST API service to read/write HDFS files using WebHDFS/curl method.
  • Extensive experience as Oracle Developer in building various ETL systems using PL/SQL Packages, Stored Procedures, Functions, Cursors, Triggers, Views, Materialized Views, Global temporary tables, Oracle partitions (Exchange partition) & SQL Loader
  • Performed Oracle DBA activities - Oracle expdp/impdp, generating AWR report, managing database for tablespace, indexes, disk space, grants/synonyms and User/Roles/Privileges, rebuilding environment prior to new code releases
  • Involved in Performance Tuning - Analyzing explain plan, rebuilding complex sqls, adding Hints, gathering stats, generating TKPROF and trace files, Partitioning highly growth tables, Exchange partitioning
  • Very Good ETL experience using Informatica PowerCenter 9.x/8.x and familiarity with Data stage 9x
  • Troubleshooting nightly batch job issues in production involving technologies such as: informatica, Oracle, Unix, Control-M/Autosys
  • Very much comfortable in building and modifying UNIX shell scripts for data migration, batch job processing, file manipulation and file archival policies.
  • Extensive experience in building various Autosys/Control-M batch jobs to perform data load, Archival of files and supporting production jobs

PROFESSIONAL EXPERIENCE:

Confidential, Telecom, TX

Big data Architect

Responsibilities:

  • Performing Requirement gathering, Analysis, Design & preparing proof of concepts (POC) for the new challenges and enhancements on the project.
  • Developed Ingestion framework using Sqoop, Hive & UNIX scripts to load source system data into Hadoop hive database.
  • Built Replication Data Pipeline along with Initial data migration from Oracle database to Hive database.
  • Built Replication system between Oracle GoldenGate to Hive using Oracle GoldenGate HDFS handler
  • Performing proof of concepts (POC) on Google cloud using Big Query, IAM, VM, Cloud Shell, Managed Instance group, Docker
  • Performing proof of concepts (POC) using SnowFlake connecting AWS for bucketing/cloud storage.
  • Migrating data & application from HDP 2.6.5 to CDP 7.1.7
  • Migrating Spark/scala program from HDP 2.6.5 to CDP 7.1.7
  • Setting up HDP & CDP edge node with service id keytab, user access, installation Java, sqoop jar, teradata jars, oozie job setup, Installation of Vendor product - Subex 6.x
  • Performing Installation & Configuration of - Hadoop client on remote machines, Oozie, Presto, Solr, Kafka
  • Established kerberized security connection between Remote servers/Tools with Hive DB
  • Implemented high performing data load features in sqoop 1.4.6 and loaded tables of multi-terabyte sizes efficiently into Hive.
  • Performance tuning of Hive loads using various recommended configuration parameters.
  • Installed, Configured & Migrated the risk assurance product from vendor product Subex 5.0 to Subex 6.3 ( Product Components of Subex are - Hadoop 2.7, Hive, Kafka messaging, Spark streaming and batch jobs, Presto, Solr & Oracle)
  • Performing patch deployment of Vendor product Subex on Non prod & Prod environment.
  • Installed/Configured Oracle SQL Connector for Hadoop (OSCH), Big Data SQL 3.2 to read HDFS files & Hive tables by querying Oracle database external tables.
  • Used Yarn RM browser for monitoring jobs, queue usage.
  • Built spring boot REST API to read JSON files from kerberized HDFS on the chrome browser (using curl/WebHDFS method) to perform CRUD operations on the JSON files.
  • Performed operations (using Scala) such as - Kafka reading and writing into console, creating Kafka topics, start/stop of Kafka services, Spark SQL for creating RDD, Data frame on structured, semi structured and unstructured data formats.
  • Monitoring/maintaining Data Ingestion processes which uses Oracle Golden Gate replications
  • Participates in meetings and discussions with Business users, Cloudera/Hortonworks Support team, Infrastructure team, Unix SA team, DBAs and other stakeholders to understand the current and upcoming road map, prioritizing requirements, setting up the dependencies among deliverables.
  • Setting up Gitlab/Git Bash & Selenium automation testing suite for the project.
  • Helping team members in clarifying the business need and technical changes for the changes
  • Doing code review and other process related initiatives for the team members.

Environment: Hortonworks 2.6.5, Cloudera (CDP) 7.1.7, Hive 1.4 & 3.0, Sqoop, Ambari, Kafka, Spark, Oracle golden gate, Oracle 12c/19c DB, Eclipse IDE, Scala IDE, SQL Developer for Oracle & Hive, TOAD, Teradata SQL Assistant, Talend, Tableau, MS SQL Server management Studio 7.1, Oozie, IntelliJ, Gitlab, OneJIRA, SnowFlake

Confidential, Boston, MA

Bigdata Lead Developer

Responsibilities:

  • Performing Analysis and design of Oracle DB changes.
  • Building PL/SQL package, procedure/Function, Views to enhance LMS application to support new business rules while introducing new source systems or any changes to existing ecosystem.
  • Performing SQL tuning and Data modeling using ERwin 9.5
  • Build conceptual/logical/physical data modeling using ERwin 9.5 for the new stage area and target application tables, considering partitioning management options.
  • Implementing Oracle VPD rules & making analytical queries/reports for BIRT reports development
  • Enhancing Moody’s Risk Origin Application and BIRT report change in XML.
  • Participating in rewriting current LMS applications in Hadoop ecosystems by - creating Hive partitioned external tables (Avro and Parquet), Sqoop queries to import data from various RDBMS systems.
  • Writing spark (Scala) job to process Hive data.
  • Built Spark SQL to process incoming files and push results into Hive and log tables.
  • Building and enhancing Oozie jobs using workflows & coordinator XMLs and their properties files.
  • Building impala views and monitoring Oozie jobs via HUE.
  • Understanding distributed data streaming using Kafka
  • Experience with NoSQL database MongoDB

Environment: Oracle 11g, TOAD, SQL developer, ERwin 9.5, Putty, RTC, Clear case, Moody’s Risk Foundation product, Eclipse, MongoDB, Big data ecosystems - HDFS, hive, sqoop, impala, pig, Spark, Oozie, HUE, Kafka, Python, Data Governance Framework - DGF

Confidential, Malvern - PA

Sr. PL/SQL & ETL Developer

Responsibilities:

  • Performed Analysis and design of Oracle & Informatica requirements
  • Enhanced ADS Loader workflows for various vendors by Making changes to Informatica mapping, session and Oracle Package/procedure/functions
  • Providing support to ETL jobs in system/integration/UAT environments
  • Carried out design, construction, review and defect analysis activities to Oracle Loader and data movement packages.
  • Performed data modeling to set up new vendor load
  • Involved in the creation of database objects like Tables, Views, Stored Procedures, Functions, Packages and Indexes.
  • Performing data migration activities by applying various DMLs to sync up current data to support new business logic on any key attribute.
  • Completing Tasks with proper documentation, thorough Unit testing and code review.
  • Following Agile Scrum methodology to manage & deliver assigned tasks

Environment: Oracle 11g, TOAD, SQL developer, Putty, Informatica 9.5, WinScp, Control-M, SourceTree, UNIX, Rally, Confluence, TOAD Data Modeler

Confidential, Wilmington - DE

Sr. PL/SQL & ETL Developer

Responsibilities:

  • Playing important role in Requirement gathering, Data Analysis and Designing ETL systems for setting up new partner or vendor feeds
  • Building ETL systems by creating Oracle packages, procedures, Functions, Table partitions, Datastage & Control-M jobs.
  • Moving from Oracle tables to Big data Hive tables, creating partitioned external tables, data exp/imp using sqoop
  • Setting up HDFS and Hive related config files,
  • Loading JSON files into MongoDB
  • Performing SQL query tuning & Partitioning table to the long running sqls.
  • Involved in various discussions with Business, Data Modeler and Client in setting the expectations, priority and dependencies for any new requirement.
  • Performing code review, deployment planning, helping team members in clarifying doubts, fixing production issues.
  • Adopted Agile Scrum methodology to deliver assigned tasks

Environment: Oracle 11g, TOAD, Putty, Quality Center, Datastage, WinScp, Control-M, WinCVS, Unix, Rally - Task management tool, Big data - Hive, Sqoop, MongoDB

Confidential, Newark, DE

Sr. Oracle PL/SQL developer

Responsibilities:

  • Performed database and ETL process design to integrate new source systems data into current system.
  • Built PL/SQL packages, procedures, triggers, functions, Indexes and Collections to implement business logic to perform various calculations needed by the netting engine.
  • Built UNIX script and Autosys scheduler jobs to perform PNE batch jobs, File and Data Archival jobs
  • Provided L3 production support for Oracle, Informatica, Unix and Autosys issues for 3 PNE products - Securities Netting and Derivative netting and Fixed Income netting
  • Build Forward and reverse Data models using ERwin r8 and make sync with each release specific changes.
  • Developed Informatica mappings and workflows to load data xml files, regular csv files

Environment: Oracle 11g, Erwin r8, TOAD, Putty, Quality Center, Informatica 9.1.0, WinScp, Autosys, Subversion, GreenHopper for JIRA

Confidential, CT

Sr. Oracle PL/SQL developer

Responsibilities:

  • Developed automated Oracle testing tool -“ACT” which could accept testing sqls, from various projects and primarily used by QA & Production support team for enhanced regression testing.
  • Performed requirement gathering, analysis and design for Oracle, Informatica and Autosys scheduler
  • Developed Data mart to store input test sqls, metadata information of each sql, resulted data after execution of sqls, summary of each test sql
  • Developed Oracle package, procedure, functions using oracle DBMS SCHEDULER to create jobs, running test sqls in parallel thread by assigning run id using ORA HASH function.
  • Designed and setup Autosys batch jobs by -Setting up autosys profile, built JIL script, UNIX script, dependency set up with other existing jobs
  • Worked on SQL*Loader to load data from flat files obtained from various facilities every day. Used standard packages like UTL FILE, DMBS SQL, and PL/SQL Collections and used BULK Binding involved in writing database procedures, functions and packages for Front End Module.

Environment: Oracle Exadata, Informatica 9.0.1, TOAD, PVCS, Putty, Quality Center, WinScp, Autosys, Ultraedit, Subversion, ERwin

Confidential, NY

Sr. Oracle PL/SQL Developer

Responsibilities:

  • Built a common repository system in Oracle for the “Life and Health” division of Swiss Re to store financial, summary of Policies, and treaty information to facilitate decision making process for other business domains.
  • Oracle replication mechanism - Streams has been implemented to refresh staging OLTP data with near real time data.
  • Involved in the continuous enhancements and fixing of production problems. Designed, implemented and tuned interfaces and batch jobs using PL/SQL. Involved in data replication and high availability design scenarios with Oracle Streams.
  • Flashback views with required transformation logics are built on OLTP tables.
  • Informatica mappings are developed to use above created source views to load into target tables. Additionally, the load is supported by other oracle procedures and packages.
  • Implemented various validations during the load process-Preload checking (Replica checking, Mapping validation check), Load completion checking and then Post load checking (Audit and Reconciliation checks)
  • Prepared test cases to validate each transformation logic and business logic and the result data.
  • Performed testing and code release of developed components in various environments.

Environment: Oracle 10g, Informatica 8.6, TOAD, CVS (version control), Putty, Quality Center, eCAB

Confidential

System Analyst

Responsibilities:

  • Designed a common ETL system to support multiple LOB’s load process.
  • Developed PL/SQL package, Procedure, Function, Materialized View to load various LOB’s (Investment Banking, Treasury Security System) data into target tables
  • Used Sql loader to perform ad hoc data loads as needed by business.
  • Prepared UNIX scripts to support data load
  • Developed materialized views as the end result of the ETL system
  • Data management activities like upload user supplied files into the system, Other DML activities to update key business tables, preparing ad hoc sql reports
  • Designed and configured a new Oracle database to make this system a new source of data load.
  • Enhanced the ETL script to support the above newly configured database, to support application front end change.
  • Performed production releases in every 6 weeks for every new set of requirements to support new GUI change as well as ETL code change.
  • Documented each set up changes to the environments, adoption of new process or rules, change in deployment methodology and inclusion of any new services
  • Worked as a single point of contact for JPMC business users, offshore team members and other JPMC stakeholders for - Gathering new requirements, publishing weekly as well as monthly status reports, client meetings, proposing estimation, code reviews.

Environment: Oracle 10g, UNIX, WinSCP, Subversion (version control)JIRA (status tracking), SQL developer, Gvim (Unix editor), Autosys (Job scheduling)

Confidential

System Analyst

Responsibilities:

  • Build Oracle packages, procedure/Functions, Triggers, Synonyms, Views
  • Migrated Oracle DB 8i to 10g, performed DB design/modeling
  • Build new forms/reports as well as enhanced during migration from 6i to 10g
  • Enabled 10g forms features such as WebUtil

Environment: Oracle 8i/10g, Forms/reports 6i/10g, PVCS version control, Oracle Enterprise, UNIX, Solaris application server.

We'd love your feedback!