Big Data Architect Resume
Norfolk, VA
PROFESSIONAL SUMMARY:
- Having 18+ years of IT experience predominantly wif large Companies, in architect, design and implementing large - scale systems, 5+ Years in Big Data Consulting wif result-oriented driving & delivering software solutions through continual innovation in techniques, tools & processes
- Designed and implemented scalable Big Data solutions for various applications needs and worked wif rapidly evolving technologies to Analyze and define unique solutions
- Has strong hands-on experience implementing big data solutions using technologies including HDFS, Map Reduce, Sqoop, Attunity, Flume, Spark SQL, Scala, Hive, Kafka, Oozie, Python, HBase, Informatica Big Data Edition, Snap logic, spark 2.2.0, Beeline, Hue, Impala, Mongo DB 3.4.9
- Worked in multiple capacities as Solution Architect, Technical Lead and Project Management
- An individual wif excellent interpersonal & communication skills, strong business acumen, creative problem-solving skills, technical competency, team-player spirit and leadership skills
- Technically strong hands-on wif very positive attitude & passion for excellence
- Domain: Telecom, Healthcare, e-Commerce, Retail & Banking
- Architect, Design,Development,Big Data Implementation
- Integration Analysis, Process improvements, Strategic& Tactical Planning
TECHNICAL SKILLS:
Big Data Technologies: Hadoop, HDFS, MapReduce, Hive, spark, HBase, Kafka, Sqoop,Attunity Flume, Oozie, Hue, Splunk,AWS-Kinesis, Zeppelin, Zookeeper, HBase, MongoDB, Neo4j
Languages: HiveQL, SQL, PL/SQL, UNIX Shell Scripting, Python, C, C++, CSS, XML, COBOL, Visual Basic, Java, Clipper5.2, SQL, Perl Script, HTML
Databases: HBase, GreenPlum, DB2, Oracle 9i, Oracle 10g, Teradata 14.3, Oracle, MySQL, MS SQL Server, MySQL, MS Access
Tools and IDEs: Eclipse, TOAD, SQL Developer, Tidal Job Scheduler, Autosys, AccuRev, Remedy, Control-M
Source Control: Git/Bitbucket, Subversion, CVS, Clear Case, TFS
Data Warehousing Tools: Informatica, Informatic BDE, Talend,Pentaho, SSIS
PROFESSIONAL EXPERIENCE:
Confidential, Norfolk, VA
Big Data Architect
Responsibilities:
- Architectdesignabigdata application solution,build a Pharmacy data resiliency repositoryto has consolidated pharmacy member demographic & eligibility data to be sent via eligibility service APIs exposed to external system for prior authorization. This is fail-over system when primary SOA is unable to respond wifin 5sec or during downtime
- Lead teh implementation of teh defined tactic approach through teh design and architecture
- Understand and translate project requirements into technical design & solution document physical, technical architecturefor engineering team to execute
- Deploy guidelines, standards, and processes to ensure teh highest data quality and integrity deployment process and version control
- Collaborate and partner wif business stakeholders, data SMEs to determine usage of analytic solutions to drive business value for member Prior Authorisation for commercial stream
- Drive innovations through developing concept’s and prototypes to help illustrate approaches to technology and business problems
- Design a core engine using SparkSQL andScalato handle high volume of data and load into Mongo DB Collections and integrate wif External system using Web Services.
- Developed driver program in Scala,DataFrame for Member detail and extended & group tables, created aggregate DataFrame for Medical Source system to list teh subscriber wif more than 1 Plan type and join wif teh member detail and populate to MongoDB
- Developed supporting programs in Scala utils for reusable business logic for teh patient Info, Plan info, patient Communication, member detail and aggregate plan for members
- Lead teh team of Onsite & Offshore developer resources from multivendor team
Environment: Cloudera Enterprise Data Hub - Major Version: CDH 5.12.2, Mongo DB 3.4.9 Enterprise, Mongo Compass,Spark 2.2.0, Scala, Beeline, Hive, Impala,Bamboo Git/Bitbucket,ConfluenceControl-M, JIRA
Confidential, Emids Technologies, TN
Lead Architect
Responsibilities:
- Responsible for Architect, Design Application Solution for solving teh data integration challenges of Magellan Med Management systems.
- Design Solution for Aetna Coventry Specialty business build an ingestion and Outgestion Framework wif integration tool Snap logic.
- Create solution map, physical, technical architecture for Analytic Platform that provides users wif reporting and ad-hoc querying capabilities
- Responsible for building teh solution which entails creating a comprehensive system wif loosely coupled modules which can be configured easily per business need of Magellan Medical Management Teh system shall be flexible enough for teh business user to set up a secured and efficient data integration process from different data sources in Magellan environment as well as raw data received through different external data sources and store in big data repository.
Environment: ClouderaCDH 5.12.2, Zaloni, Hive 1.2, Snap logic, Spark, SQL Server
Confidential, ATLANTA,GA
Big Data Architect
Responsibilities:
- Responsible for Architect, Design Application Solutions &deployment for Contour2 IP-Playback Data for Various Reporting Requirement
- Create solution map, physical, technical architecture for Analytic Platform that provides users wif reporting and ad-hoc querying capabilities
- Interact wif Business to Finalise\prioritize requirement, create an interactive data analysis using Zeppelin for business discussion and understanding.
- Optimizing process, implement naming standards and best practices. Identify and Document bad\Incomplete Data Issue, Analyse & Validate Data from Kinesis Stream using Splunk
- Apply Business Rules using Hive Queries to Enrich data wif various data from EDW, ODS and build hive Aggregate table data for Reporting connecting to Tableau
- Automate teh Process of data Load wif Oozie Workflows and Scheduling wif different option to deliver Report Email, SFTP and Hive Aggregate Data Table
- Build Vendor Payment Analytics for TV Everywhere Mobile App usage for IP-Playback Data Provide extensible solution to generate Report monthly, using hive Scripts to perform ETL, Extract enrich data wif Customer and Site info from different source
- Build Kafka topics for teh consumption of teh IP playback Streams like Reconnect, Pageview, Keypress, Error streams form teh IP playback devices for Video on demand
- Build VOD-Analytics by Unique House hold and App Usage wif different dimension TVOD, SVOD in comparison to Linear, VOD, DVR, ROVI, DTA & C1 users
- App Usage Analytics to Determine True C2 App usage, Reduce Calls generated by App Users, Tracking and Trending Plant Issues, Behavioural Analytics for In-home and Out of Home usage which Provides insights to Marketing, Advanced Advertising, Incentivize Users
Environment: Horton Works HDP 2.7.3, Kerberos, Hive1.2, Splunk6.3.3, Kinesis Stream, Zeppelin0.7.0, YARN, Ambari, Tez Views 0.7.2.6, Oozie Workflow 1.0.0, Toad for Apache Hadoop 1.5.3, Nifi, Spark, HBase, Linux, Kafka, Oracle 11i, SQL Developer 4.1.5, SQL Server, Tableau.
Confidential, Tampa, FL
ArchitectSr. Big Data Consultant
Responsibilities:
- Responsible for leading efforts of modernizing and migrating legacy Analytics System to big data-based Analytics System, Responsible for Design EDW Application Solutions & deployment, optimizing processes, definition and implementation of best practices
- Created solution maps, physical, technical and data architectures for Analytics Platform
- Designed and implemented data acquisition and ETL pipeline using Attunity tool
- Successfully architecture for Data Ingestion process, ingested more than 2500 Tables ingest to Landing, Raw zone and transform teh data wif business logic to refined zone and to Green plum data marts for reporting layer for consumption thro Tableau
- Engage wif business analysts to understand business requirements and translate them to functional specs and technical design. Ensure full requirement traceability
- PEGA: Business Process Management Application: Build and deploy end to end data integration on Hadoop for PEGA for their Care, Appeals & Grievances (CAG) and Medical Management Platform (MMP) AUTH application, Members, Claims & Pharmacy Analytics Design and develop Informatica BDE Application and Hive Queries
- Implemented critical solution components using technologies including Spark Streaming, Spark SQL, Python, Hadoop, MapReduce, Hive, HDFS, Sqoop, Oozie, Shell scripting and other big data technologies
- Implemented Spark by leveraging Interactive SQL queries for processing large volumes of data. Semi Automate Chart Review and Doctor notes
- Working wif Data science Analytics team for teh POCs on newer tech Jupyter Notebook, NumPy, SciPy and Pandas, R, Tableau, H2O, Spark MLIB
Environment: Horton WorksHDP 2.3, Spark, Hadoop, Map-Reduce, HBase, Hive, Linux, Agile-Scrum, Storm, Kafka, Oozie, Informatica 9.6.1 BDE, Oracle 11i, PLSQLDevelopers, SQL Server, Green plum, Attunity, Autosys, Hue, YARN, Python, Spark, Tableau.
Confidential, Dayton, OH
Architect
Responsibilities:
- Architect future technology goals & strategy initiative to implement Hadoop
- Technical assessment of current architecture & identify gaps
- Design technical footprint and implementation of Hadoop Data Lake
- Data Integration Tool Assessment of Pentaho, Talend wif Informatica BDE
- Designed Tactical architectural road map for teh enterprise Hadoop Platform.
- Envisionedintegration framework as teh overall Hadoop Implementation solution
- Signoff deliverables from Assessment Phase to teh implementation phase
- Designed Business use cases & documented processes wif data flow & diagrams
Environment: Horton Works HDP 2.2, Informatica 9.6.1, Teradata 14.11, HDFS, Flume, Sqoop, YARN, Hive, Ambari, Zookeeper & Oozie.
Confidential, Atlanta, GA
Architect, Sr. Consultant
Responsibilities:
- Analyse, Source and Parse teh web log activity data of my.t-mobile.com, identify search activity patterns of customers before churn
- Ingest data sets like Customer usage data, event logs, and web logs to HDFS.
- Responsible for Design EDW Application Solutions & deployment, optimizing processes, definition and implementation of best practices
- Designed and implemented Hadoop to store high volume data such as billing P and Contact history, Data delivered as batch and stream as well
- Designed and implemented system wide data retention and archive strategy in legacy EDW system to avert a multimillion-dollar expansion in 2013, by saving existing Teradata system
- Decreased teh incremental EDW load time by removing and or reorganize EDW load schedule by Tuning ETL jobs by employing enhanced TD functionality and techniques
- Collaborated wif departments and wif cross-organizational teams to implement improvements in processes, Business Support System/Operations Support System support. Designing, executing and updating of Implementation Plan
Environment: Horton Works HDP 2.0,Informatica 9.1, Teradata 14, Oracle 11i, SQL Server, Perl script, AccuRev, Remedy,TIBCO CEP, Hue, Flume, Sqoop, Hive, Spark, Ambari, Tableau.
Confidential, Raleigh, NC
Tech Lead
Responsibilities:
- Developed Informatica Mappings Re-design - for teh Enterprise Data warehouse
- Architected teh ETL flow to capture teh record count of different layers staging, EDW and Semantic layer to provide a data lineage
- Standardized process & procedures. Enforced processes to enable proactive involvement in value-added activities for teh organization, developed innovative testing strategies & streamlined process to reduce testing redundancy
Environment: Informatica 9.5.1, Teradata 14, Oracle 11i
Confidential, Chicago, IL
Architect
Responsibilities:
- Architect teh Enterprise EDW to Integrate wif SAP
- High-level design document wif teh new architectural set up
- Requirements into technical solutions. Get signoff and finalize ETL technical design
- Designed & Developed 30+ mappings that extract data for Sales, Returns, Inventory, Forecast, Contract, and Chargeback from different feeds Internal and external IMS,Chargeback, Cardinal, McKesson, AHA. Apply business logic & load data to SAP PSA
- ed Customer Excellence Project Star in Recognition for managing End to End implementation of teh project wif Successful client requirements gathering onsite and wif development to implementation
Environment: Informatica 9.1, SAP BW, Xcelsius, SAP Business Objects 4.0, Bex Queries
Confidential
Off Shore Lead
Responsibilities:
- Lead teh delivery of teh entire offshore development activity
- Re-Design Framework for Profitability, Sales, Order, and Inventory& Purchasing as per teh business rules
- Lead a team of 4 resources, Allocation, Track and update status to Delivery Manager Onsite.
- Configure pre-packaged 40+ ETL, SDE and SIL mappings, sessions using Oracle-DAC
- Configured teh DAC execution plans for full and incremental loads
- ed Spot - Recognition for managing End to End implementation of project 1 month ahead of timeline
Environment: Informatica 8.6, Oracle-DAC 10.1.3.4.1, Oracle 11i and OBIEE 10.1
Confidential
Sr Developer
Responsibilities:
- Implementation of KANA Phase-3 Changes
- Redesigned teh Informatica logic to handle teh new data flow as per client requirement to include Metrics stored in teh subject areas Feedback, Rating, Content, Freshness and Search.
- Implement complicated business to maintain history of changes to teh subject areas
- Developed Mapping, Optimized teh mappings written by others and guided team members
Environment: Informatica 8.6.1 & Oracle 11
Confidential
Offshore Lead
Responsibilities:
- GDR is teh Central Repository; objective is to lift teh data from Smith-Berney & shift to Morgan Stanley Central Repository
- Used Informatica for data extraction/conversion from DB2 to Teradata
- Designed & developed various Informatica mappings to populate teh data in teh GDR
- Developed and implemented one-time migration of Accounts, Contacts and Assets into GDR
Environment: HP UNIX 11i, Teradata, DB2, Informatica8.1, Synergy7.1
Confidential
Consultant
Responsibilities:
- Content Management services for major Nationalised Banks and Schools in India, State Bank of India, State Bank of Mysore. Content management are required to keep teh content on teh screen fresh and updated. Scheduling and purging to keep teh content relevant
- Responsive Web Design Services, Interactive Devices for Teaching
- Provides accessibility of website across all devices on teh same set of URLs