Hadoop Architect Resume
New, JerseY
SUMMARY:
Financial Services Analytics & Hadoop Architect with hands - on experience in developing Predictive Analytic models using Machine Learning, Hadoop (Spark, Kafka, HBase, Hive), Data Governance, developing transactional applications, platform infrastructure build, application security and DMZ design. Strong communications skills with all levels of business users and technical community/management. Led global teams, workforce planning, evaluation of vendors and products in a Fortune 500 company. Reduced Confidential of the platform by $5 million.
TECHNICAL SKILLS:
DATA SCIENCE/ENGINEERING, ANALYTICS, BI:
- Machine Learning Models Random Forrest, Gradient Boosting Machine, Recommendation Engines
- Python, H20.ai, Scikit-learn
- IBM WebSphere, Tomcat Application Server
- Java (Sun Certified), J2EE, SQL
- SOAP and REST Web Services
- Architecture and Infrastructure build
- DMZ design
- Workforce planning, Client relationship
- Incident management & root cause analysis
- Hadoop Spark, Kafka, HBase, Flume, Hive, Sqoop, Cloudera Manager - CDH 5.3, 5.5
- Data Governance, Metadata Management
- Actuate, UNIX, DB2 analytical platform
- DB2, Oracle Design and Development
- Single Sign-On using SAML Assertion
- Java batch, DB2 Optimization and capacity planning
- Inter-tier communications security
- Evaluating & selecting vendors and products
- Monitoring/profiling/benchmarking processes
PROFESSIONAL EXPERIENCE:
Confidential, New Jersey
HADOOP ARCHITECT
Responsibilities:
- Contributing on multiple Big Data solution proposals for Confidential .
- Define solution architecture, requirements, solution phases, roadmap.
Confidential, New York, NY
VICE PRESIDENT
Responsibilities:
- Implemented Random Forest and Gradient Boosting Machine Learning models to predict whether a corporate equity client participant is a reinvestment opportunity for the firm or not. In this proof of concept (POC), the model was implemented using H20.ai (an Open Source Machine Learning platform) and Python. The data was gathered from data sources of various applications, merged, cleaned, and stored in Hadoop (Hive) for further analysis. The Financial Advisors, users of this predictive analytic engine, will be able to do more focused marketing campaigns.
- Implemented Spark Streaming on Hadoop to investigate Denial of Service attacks by extracting IP addresses and cumulative count of each one of them (using State operations) to determine sites sending maximum number of requests. This information is used by infrastructure security groups within the firm for further action. In another effort, used this technology to investigate Trade execution issues reported by corporate client participants suspecting fraud. Extracted login and all features used/transactions of a given user from application logs and shared with the Fraud detection team.
- Led the effort to migrate Actuate reporting platform to Business Objects and selected a vendor after multiple presentations from four vendors. The migration to Business Objects will result in Confidential reduction of $5 million over a period of few years.
- Enterprise Metadata Management and Data Governance: Made significant contributions towards Data Strategy for the CES group of applications. The requirement is to have one Data Hub for all applications instead of each application using their own database. The objective is to improve data quality, data standardization, reduce future development costs and provide new data analytics platform.
- As team Architect, led SAPPHIRE JV integration of Citigroup Smith Barney with Confidential . Responsible for all technical discussions/decisions for the SAPPHIRE application that included Actuate reporting plant, DB2 Data Migration, remediation of VA Scan findings, use of NAS Share vis-à-vis SAN for data storage, and DMZ design review with Enterprise Infrastructure and Security Architect.
- Led SAPPHIRE migration effort from IBM Data Center to Confidential Data Center under Enterprise Stack Adoption (ESA) program.
- Led DB2 access and Java processes optimization to handle high data volumes resulting from on-boarding of big clients like Google and Amazon.
- Led the effort to automate Client On-boarding/migration by providing the implementation team a Workflow Engine that guides the user to verify Client and Plan setup, schedule files in the correct order, run data reconciliation reports, and complete the Confidential checklist.
- Managed delivery of major SAPPHIRE releases, with the biggest release of approximately 9 Confidential .
- Instrumental in formation of first SAPPHIRE Level1 production support team in terms of setting processes, application knowledge transfer, team formation etc.
- Managed Desktop call center application for a period of one year.
- Technology stack: Cloudera Hadoop CDH 5.3 and 5.5, Spark, Kafka, HBase, Flume, Hive, Sqoop, H2O.ai, Python, Java, IBM WebSphere Application Server 7.1, DB2 v11, IBM MQ 6.0, Actuate iServer v11, AutoSys for Job scheduling, J2EE, SOAP and REST Web Services, JIRA, Git, TeamCity
Confidential, New York, NY
PROJECT LEAD
Responsibilities:
- Designed and developed New Accounts Web Approval portal that supports approval of New Retail Accounts, Managed Accounts and Margin Accounts.
- Technology stack: Apache Axis 1.1 for Web Services, IBM WebSphere Application Server 4.0, Apache Tomcat, DB2, CICS Transaction Gateway (CTG)
Confidential
SENIOR SOFTWARE ENGINEER
Responsibilities:
- Delivered projects.