We provide IT Staff Augmentation Services!

Quantitative Analyst/data Scientist Resume

3.50/5 (Submit Your Rating)

CT

SUMMARY:

  • Over 9+ years of professional experience as a Quantitative Analyst /Data Scientist workingacross Finance, Telecom and Retail industries with amaster’s degree in Science and Specialization in Mathematics and Quantitative Finance.
  • Proven ability to translate high - level objective into practical analysis and deliver actionable recommendations. Record of managing complex projects and creating solutions that work. Self-directed innovator searching for challenges.
  • Proficient in Machine Learning Techniques,R, Python,SAS,Tableau, SQL & Advanced Excel.
  • Developed predictive models to identify the most significant behavioral patterns that lead to a member conversion which eventually increased the ROI of 0.96 million$ per quarter
  • Re-built the existing model and increased its accuracy from 68% to 89% using Advanced statistical Algorithms
  • Proficient in managing entire data science project life cycle and actively involved in all the phases of project life cycle including data acquisition, data cleaning, data engineering, features scaling, features engineering, statistical modeling (Decision trees, regression models, neural networks, SVM, KMeans Clustering), dimensionality reduction using Principal Component Analysis and Factor Analysis, testing and validation using ROC plot, K- fold cross validation and data visualization.
  • Good practical knowledge in performing Data Analysis process using Python like Importing datasets, Data wrangling, Exploratory Data Analysis, Model development and Model Evaluation.
  • Hands on Expertise on Classification,Regression,Time Series Data,Churn Prediction, Home Price valuation, Exit Strategies using various packages in R and Python
  • Adept and deep understanding of Statistical modeling, Multivariate Analysis, model testing, problem analysis, model comparison and validation.
  • Skilled in performing data parsing, data manipulation and data preparation with methods including describe data contents, compute descriptive statistics of data, regex, split and combine, Remap, merge, subset, reindex, melt and reshape.
  • Experience in using various packages in R and libraries in Python.
  • Working knowledge in Hadoop, Hive and NOSQL databases like Cassandra and HBase.
  • Ability to handle multiple tasks simultaneously.
  • Proven leader with outstanding relationship building skills and strong communication abilities
  • Highly motivated team player with ability to work cross-organizationally and manage strict deadline.
  • Extensive working experience in developing mappings in Informatica, tuning them to achieve optimal performance and migrate objects in all environments including DEV, QA testing and PROD.

TECHNICAL SKILLS:

Expertise: Caret, Tidyverse, MASS, GGPLOT2, Scikit-Learn, NumPy, SciPy, Deep learning, RNN, CNN, Tensor flow, Keras, matplotlib, Microsoft Visual Studio, Microsoft Office

MachineLearning Algorithms: Multinomial Regression, Logistic Regression, Decision Trees, Random Forest, K Means Clustering, Support Vector Machines, Gradient Boost Machines & XGBoost, Neural Networks.

RDBMS: SQL Server 2005/2008/2012 , MySQL,Teradata

NoSQL DB: Cassandra

Frameworks: Hadoop Ecosystem, Apache Spark

Programming Languages: R,Python

Tools: /Platforms: RStudio, Tableau, Informatica, MicroStrategy, Toad, SAS, Eclipse, Windows,SQLdeveloper,Toad for Oracle,Microsoft SQL, Teradata, Hadoop

PROFESSIONAL EXPERIENCE:

Confidential, CT

Quantitative Analyst/Data Scientist

Responsibilities:

  • Created a dataset with data in 7-8 different repositories using SQL Teradata involving complex querying
  • Analyzed dataset of about 1.34M records and identified trends and effective factors for data modelling
  • Preprocessed the data and developed various visualizations using packages ggplot/choroplethr/caret
  • Rebuilt the propensity model with increased accuracy from 68%-89% using Advanced ML Algorithms
  • The model estimates to increase the ROI and least estimated standard error 2.05%
  • Developed dashboards using Tableau and automated the process using SSIS for daily update
  • Documented the phase2 Analysis for the project and also designed roadmap and conducted KT sessions
  • Built multiple models using various Machine Learning Algorithms like Multinomial Regression,Decision Tree, SVM, Random Forest,Neural Networks etc., and finalized with the better model using ROC Curve
  • Used various Parameter Tuning Techniques to get better results from the model
  • Used different methods like Univariate approach,Boxplots,Cook’s distance to find the outliers
  • Developed time series forecasting model(ARIMA) & provided strategic analysis for the business demands and supply, in which the model predicted actual demand with 92% accuracy
  • Created Exploratory Data Analysis to identify trend, seasonality and outliers etc.,
  • Managed team processes and deliverables for Ramp-up and Ramp down demand forecasts
  • Responsible for providing reports, analysis and insightful recommendations to business leaders on key performance metrics pertaining to employee performance
  • Built predictive models to identify the most significant behavioral patterns that lead to employee churn
  • Created Propensity model to identify the most influential attributes contributing to Indent/Demand cancellation

Confidential

D ATA S CIENTIST

Responsibilities:

  • Understanding business context and strategic plans and develop a data-driven business plan to support the attainment of business goals
  • Data manipulation/treatment based on nature of data (for example missing value imputation, Information Value (IV), Weight of Evidence (WOE), Data profiling, correlation matrix, relative importance between predictors, variable clustering, univariate and bivariate plots, etc.)
  • Building predictive models from start-to-finish (i.e. extract data, manipulate data, Data Profiling, build and validate model) and then deploy model on real data and track model performance/model accuracy
  • Preparation of final project presentation documents for overall significance of the project in a well-defined manner
  • Identify the most significant behavioral patterns that lead to customer churn and build an attrition model to understand the probability of a subscriber to stay or attrite after subscribing
  • Data exploration to make the data useable and ready for analysis.
  • Analyze the employee behavior and characteristics of terminated employees.
  • Build a model to check the factors affecting the termination of employees and their reasons for termination.
  • Predict the employees who are high risk to maintain the talent pool and to effectively retain talent at every level(High Potential and High performer)
  • Responsible for working with stakeholders to troubleshoot issues, communicate to team members, leadership and stakeholders on findings to ensure models are well understood and optimized.

Confidential

DATA ANALYST

  • Creating output to explain data analysis, data visualization, and statistical modeling results to managers.
  • Modeling survey data responses with ordinal logistic regression in R.
  • Experience with working on clickstream activities, Customer Journey activities, Fraud Detection, Sales and managing Store items.
  • Analyzing and visualizing user behavior migration.
  • Created mappings to load data from source and target to staging, staging to reporting tables by applying business requirements using Informatica Power Center.
  • Applying machine learning concepts to capture insights.
  • Handled importing data from various data sources, performed transformations using Hive, MapReduce, and loaded data into HDFS.
  • Providing timely, relevant, accurate reports and analysis of the organization’s performance to facilitate decision-making towards achievement of the budget and strategic plan.
  • Documented all phases of project implementation for future reference and conducting KT sessions

Confidential

Emerging Market Analyst

  • Gathered requirements from onsite coordinators, performed Requirement Gap analysis and finalized design documents using Erwin Data Modeler and Microsoft Visio.
  • Designed complex SQL queries to input at the beginning of mappings to filter the data as per requirements.
  • Created reporting tables for comparing source and target data and report data discrepancies (mismatch, missing scenarios) found in the data.
  • Performed validations not received in the requirement document from the customer end and learnt the SQL queries which helped to attend defect triage calls.
  • Results obtained from report mappings were displayed using MicroStrategy which is a better User Interface tool.
  • Extensive hands on experience of HP Quality Center tool used for performing production support activities.
  • Implemented Microsoft Visio and Rational Rose for designing the Use Case Diagrams, Class model, Sequence diagrams, and Activity diagrams for SDLC process of the application.
  • Performed debugging of the code as per inputs given by IST (Integrated System Testing) team and deployed code into PROD environment after receiving approval from IST team.
  • Performed performance improvement of the existing Data warehouse applications to increase efficiency of the existing system.
  • Finalized the factors required for predictive model development by gathering inputs from MCC (Customer service) team, performed feature engineering to identify factors best suited for model development and gathered data from various repositories as per required factors using SQL complex querying n Teradata.

Confidential

FSM/Data Analyst

  • Gathered requirements from onsite coordinators, performed Requirement Gap analysis and finalized design documents using Erwin Data Modeler and Microsoft Visio.
  • Understanding the requirements and develop various packages in SSIS.
  • Gathered requirements from JAD/JAR sections with developers and business clients.
  • Designed the business requirement collection approach based on the project scope and SDLC methodology.
  • Designs and develops the logical and physical data models to support the Data Marts and the Data Warehouse
  • Create SQL queries for product components to update FACETS backend tables and create product prefixes.
  • Involved in formatting data stores and generate UML diagrams of logical and physical data.
  • Developed project plans and managed project scope.
  • Performed user acceptance and parallel testing for coding, pricing, and benefit builds in Facets.
  • Prepared a handbook of standards and Documented standards for Informatica code development.

We'd love your feedback!