Bigdata Solution Architect/sr Manager Resume
Mountain View, CA
SUMMARY
- Big Data expert with over Seventeen years’ experience, who has discovered and solved problems in a variety of business and industrial applications, saving millions of dollars.
- Seasoned IT Technologist. With 3+ years management experience leading a highly talented software development team while delivering high quality/Bigdata projects. 5+ years Hands - on Hadoop/Big data Solution Architect working on Kafka//Spark/Storm/NoSQL/Yarn/Hadoop/NoSQL on AWS platform. Expert in building, design and managing multi petabyte scale cluster in AWS/hybrid cloud. Certified Hadoop Administrator.
- Expert in Data integration moving large amounts of data & ETL ingestion in/out of HDFS/ RDBMS/Data Warehousing/ NoSQL(Cassandra/Hbase) designing data pipelines with transformation topologies and using 3rd party tools like Snaplogic tools. Advance replication between heterogeneous databases and applications using Goldengate.
- Expert in setting up and develop a complex data pipeline dealing with large data sets (Real-time streaming and batch data )using Kafka/Spark/Storm to HDFS/Cassandra and trained data using ML for advance analytics from common de-normalized data lake.
- Very strong with RDBMS -Oracle, MS SQL Server, MySQL, Sybase with Active GoldenGate Replication setup. Data design and data integration from OLTP to data warehouse.
- Expert in data governance, Enterprise wide data-driven security Perimeter and Fine grained access - Knox, Ranger, Kerberos, LDAP, Data Encryption, archiving and tokenization in Hadoop, implement HIPAA, PCI & SOX compliance
- Managed teams, Release planning, merging cross functional teams for better and faster results as a Technical program Manager. Drove high-stake technical projects using Agile development framework.
- Excellent Analytical, Problem Solving, Presentation and Communication skills. Goal oriented team player and highly motivated with ability to learn new technologies very quickly.
TECHNICAL SKILLS
HDFS Platforms: Hortonworks HDP2.3, Cloudera CDH4, Pivotal 2.1, Apache Hadoop, AWS
RDBMS/No-SQL: Cassandra, HBase, RDS, Redshift, Oracle 12c, 11gR2, Exadata, GreenplumTools/ Utilities/ Decision Support Spark, Kafka, Storm, ETL tools like SQOOP, Talend, Apache Ranger, Ambari, Ganglia, Nagios, Hadoop ecosystem (Pig, Hive, Hbase, Oozie, Zookeeper, Flume). Solr, Elastic Search, Splunk, Protegrity HDFS Tokenization and encryption. Data masking, sub-setting & archiving. Oracle Advanced security options 11gR2 Grid, Data torrent streaming, Tableau, OEM, RAT, SPA, Golden Gate HA, UNIX (shell scripts, Python, ftp, grep, awk, vi)
Hardware/ OS: Linux RHEL, VMware, HDFS, Oracle Enterprise, Unix
PROFESSIONAL EXPERIENCE
Confidential, Mountain View, CA
Bigdata Solution Architect/Sr Manager
Responsibilities:
- Working on AWS integration and migration. Building 100 PB authoritative data lake(ADL) under unified security. Design and develop a complex ETL ingestion data pipelinefor both streaming and batch data setsusing Kafka/Storm to HDFS/NoSQL and trained data. Setup ETL migration, backup and archival using Kafka, Storm, Spark, Cassandra, Hbase, Druid & Hadoop infrastructure.
- Deployed data governance and security - Knox, Ranger, Kerberos, LDAP, data Encryption, archiving and tokenization in Hadoop & legacy data platforms.
- Managed Analytics Engineering teams for sprints/releases. Worked with Product management for Analytics, worked with internal/external customers
- On-boarded new customers on Cloud Analytics Platform, understand resource, size and data requirement and designed the analytics cluster for users . Worked with Engineering to troubleshoot issues related to cluster provisioning and JVM on Hortonworks HDP2.4 platform in Openstack and AWS IaaS.
Confidential, San Ramon, CA
Sr. Technical Program Manager
Responsibilities:
- Managed Analytics administration team, Long Range Planning. Scoping, Defining and delivering programs using Agile. Managed large end-to-end enterprise Industrial Datalake projects team and Predix (predictive analytics platform) products & applications.
- Rolled out 3 Analytics project to production, resulted in cost savings of 20M per year - Major one was strategic direction to consolidate and bring data on one big data platform as Industry standard eliminating 20 data center out of 200+. Others related to Transportation & Sourcing, Power and Water.
- Managed BI - ETL data migration from RDBMS/MPP to DW, HDFS using ODI, Talend, Historian-HD & Tableau
- Managing mission critical HDFS migration project from Pivotal to Horton works platform - working with 3rd party vendors on infrastructure planning, installation, POC and implementation resulted in scalability and Reliability.
- Collaborate with internal and external customers of Industrial Data lake platform for big data product life cycle
- Data security expert; HDFS tokenization, encryption using Protegrity, Kerberization & rule based security Authorization. Oracle Datavault and data TDE encryption implementations
- Led Big-Data team technologies: HWX HDP2.3 & Pivotal. Hadoop & Cassandra performance tuning for 1.5PB, 100+ node Cluster setup, Install, POC & implementation of all ecosystem tools use cases - Hive, MR patterns/API, Oozie, Flume, SQOOP, Talend, Zookeeper, Storm, Hawq with MR-2 Yarn, Tableau, Knox, Ranger, Solr, Kafka, Ambari, Gangila, Nagios & Data torrent Instream analytics. HDFS cluster performance benchmark and testing
- Define a Data lake product roadmap, review requirements for each release, Agile- Scrum methods.
Confidential, Foster City, CA
Hadoop, Data Migration Specialist, DevOps
Responsibilities:
- BigData Hadoop administration, cluster setup using Cloudera 4.2 HDFS and Scripting in Hive. Data transfer using Flume, SQOOP. Setup & testing Oozie workflow, Managing and Monitoring with Cloudera Manager 4/5. Data migration, Ab Initio ETL from In-memory DBs/RDBMS/ERP to No-SQL & HDFS
- No-SQL administration -production migration to Cassandra integrated with Hadoop, Worked and understand the architecture of other No-SQL DBs like Hbase
- Designed and implemented Oracle DB automation to provide hands free deployment of Linux servers that involves automatic deployment of OS, Oracle database 11gR2 (Standalone and RAC) and application following ‘Development as a service’ concept. Deployed applications in Cloud, PaaS and SaaS model.
- Oracle DB Administration, Upgrade and platform migration specialist: Managed DB Upgrades & migration on high stake project implementations to 12c (POC), 11gR2 & 11gR1 with RAC, ASM, Solaris & Linux
- DevOps Technical project manager for multiple mission critical projects involving Cyber source Confidential application integration and data-center migration projects. Projects involved IDC data center application and DB migration, Sybase to Oracle 11gR2 Production migration and upgrade. 700TB Data Migration to in-memory Exadata X3-2 Half Rack with 10x compression for data warehouse performance.
- Designed HA solution using Golden Gate bi-directional replication for production databases. Designed complex architecture having 16 nodes bi-directional replication. Sybase to Oracle and Oracle (11gR2 on Solaris) to Oracle (Linux) on remote data center with 5 Nines availability and seamless failover enabled.
- Completed SAPM (Advanced Project Management) as part of Confidential leadership program affiliated to Stanford (Graduated in Apr ’14)
Confidential, Mountain View, CA
Solution Architect
Responsibilities:
- Led enterprise wide efforts to design, standardize & implement HA-DR solution involving active-active replication for data integration across data center using Golden Gate 11.1.1.x involving heterogeneous environments - OS (Solaris & Linux) & Databases from Oracle to SQL server for Financial applications.
- Developed a standard DR solution & monitoring framework for 12 mission critical application environments that needed to be ready for HA-DR using Golden Gate replication in two phases. Scope also involves data security features like ASO & data encryption.
- Actively engaged & evaluated different products for HDFS Implementation Led HDFS POC for the data center
Confidential, Pleasanton, CA
DBA Lead
Responsibilities:
- Managed mission critical Oracle HR & Finance IT databases infrastructure engineering projects, new releases, upgrades and application deployments and hand over to production support with proper documentation. Tasks include evaluating, conduct POCs and finalizing new technologies. Prepared functional and technical design documents, create comprehensive test plan, and follow it through SDLC deployment.
- PeopleSoft critical HR & Finance Database upgrade: Upgraded the DB from 8.1.7 to 10gR2 along with platform upgrade to Solaris 10. Provided complete solution involved upgrade to 10gR2, platform & data migration and integration with new technologies & features ASM, RAC, new DB security modules like DB vault & ASO - Transparent data encryption, Data masking and Data sub-setting in a seamless manner.