Data Scientist Resume
Richmond, VA
SUMMARY
- 4 years of experience with Data Analysis, Machine Learning using R language and Python.
- Strong with ETL, Data warehousing, Data Store concepts and Data mining.
- Experienced in SQL programming and creation of relational database models.
- Experienced in creating cutting edge data processing algorithms to meet project demands.
- Experienced in SAS includes detailed knowledge of statistical analysis of financial.
- Good knowledge in ETL tools, ORACLE & SQL Server.
- Experience in creating different visualizations using bars, lines and pies, maps, scatter plots, Gantts, bubbles,histograms, bullets, heat maps and highlight tables.
- Involved in troubleshooting, performance tuning of reports and resolving issues within Tableau server and Tableaudesktop report.
- Good understanding of relationaldatamodeling,dataprofiling methods to analyzedatasources.
- Involved in creating database objects like tables, views, procedures, triggers, and functions to provide definition,structure and to maintaindataefficiently.
- Excellent experience in creating indexes, indexed views in observing business rules and creating effective functionsand appropriate triggers to assist efficientdatamanipulation anddataconsistency.
- Well experience indataextraction, transform and loading (ETL) using various tools such as SQL Server IntegrationServices (SSIS) and SQL Server Reporting Services (SSRS).
- Extracted, transformed and loadeddatafrom various heterogeneousdatasources and destinations like Access, Excel,flat files using connectors, tasks and transformations provided by SSIS.
- Excellent interpersonal, analytical, problem - solving and communication skills.
- Ability to effectively manage time and prioritize tasks in order to consistently complete projects according tospecifications and on time.
- Worked with applications like R and Python to develop regression analysis, classification analysis, cluster analysis, neural network algorithms.
- Worked on Tableau to create dashboards and visualizations.
- Strong experience working with SQL Server, RStudio, MATLAB, Oracle10i.
- Expertise in analyzing data using big data technologies (Hadoop, Hive, Pig, Mahout, Spark).
- Identified problems and provided solutions to business problems using data processing, data visualization and graphical data analysis.
- Worked with clients to identify analytical needs and documented them for further use.
- Solid knowledge of mathematics and experience in applying it to technical and research fields.
TECHNICAL SKILLS
Programming Skills: SQL, Python, R, SAS, UNIX, Spark, Pig, Hive, JAVA, C#
MS-Office Package: Microsoft Office (Windows, Word, Excel, PowerPoint, Visio, Project)
Database: SQL Server, Oracle 9i, PostgreSQL, Teradata
ETL tools: SSIS, SSAS, SSRS
Visualization Technologies: Tableau, Power BI, Matplotlib, Seaborn, ggplot2
Machine Learning Algorithms: Logistic Regression, Linear Regression, Decision Tree, Random Forest, Nearest Neighbor classifier, K-Means Clustering, Principle Component Analysis, Support Vector Machine, Naïve Bayes, Gradient Boosting
PROFESSIONAL EXPERIENCE
Confidential, Richmond, VA
Data Scientist
Responsibilities:
- As a member of the Analytics Team, support advanced analytics projects to meet the needs of business partners.
- Wrote a python-based SQL generator that helped speed up a weekly reporting from several days effort to just running the automation and getting the report done in a few hours.
- Investigate existing and emerging technologies to explore analytics solutions for business partners, and make recommendations for enterprise-wide implementation.
- Maintained SQL scripts to create and populate tables indatawarehouse for daily reporting across departments.
- Worked on adataset that predicts the length of stay in dentist office based on claim records
- Created, coded, tested, modified and installed programs for a host of application
- Use SQL, cluster analysis, and text analytics methodologies to investigate account data, including client interaction, to anticipate client dissatisfaction and determine statistical patterns that could be precursors closing or diminishing accounts, and to discover a relationship between client interaction and fraud.
- Perform data extraction, sampling, advance data mining and statistical analysis using linear and logistic regression, time series analysis and multivariate analysis within R and Python.
- Discover and visualize patterns in Tableau.
- Examine customer feedback and activity for use in detecting or confirming fraud, using a combination of text analytics, statistical modeling, and classification.
- Used SQL language to write queries inside the SQL server database.
- Developed predictive models using Decision Tree, Random Forest and Naïve Bayes.
Confidential, Newbury Park, CA
Data Analyst
Responsibilities:
- Worked on an application that took care of online account services which included Bill paying, checking the bill history, changing customer profile, ordering auto insurance ID cards Worked on Commercial lines Property and Casualty Insurance including both policy and claim processing and reinsurance.
- Provide customer service and support for Property & Casualty risk management needs
- Query optimization using SQL Profiler and performance monitors to enhance the performance of database servers.
- Involved in creating automated Test Scripts representing various Transactions, Documenting the Load Testing Process and Methodology. Created meaningful reports for analysis and integrated the Performance Testing in the SDLC.
- Converted various SQL statements into stored procedures thereby reducing the Number of database accesses (since Stored Procedures passes the whole block at one time.)
- Used the DataStage Designer to develop processes for extracting, cleansing, transforms, integrating and loading data into data warehouse database.
- Used DataStage Manager for importing metadata from repository, new job categories and creating new data elements.
- Designed and developed Use Cases, Activity Diagrams, Sequence Diagrams, OOAD using UML and Business Process Modeling.
- Created UML (Class, Sequence and Component Diagrams) diagrams using Visio.
- Used the DataStage Director and its run-time engine to schedule running the solution, testing and debugging its components, and monitoring the resulting executable versions (on an ad hoc or scheduled basis).
- Implemented local containers to simplify the job design.
- Implemented shared containers to use in multiple jobs, which have same business logic.
- Involved in writing Job Automation process i.e. writing shell scripts using Korn shell, as scheduled daily, weekly, monthly, and sending these jobs to client’s email-ids.
- Involved in DBA tasks such as table creation and maintenance, create indexes, optimize queries, and grant privileges.
- Involved in creation of Database Tables, Global Temporary Tables, Cluster Tables, Partition Tables, And Index-by Tables using storage parameters.
- Wrote many stored procedures, stored functions, Packages and used in many Forms and Reports.
- Created indexes on tables and Optimizing Stored procedure queries.
- Involved in interaction with the End Users and implemented reviews of the application.
Confidential, Houston, TX
Data Analyst
Responsibilities:
- Involved in defining the business/transformation rules applied for sales and servicedata.
- Document, clarify, and communicate requests for change requests with the requestor and coordinate with the development and testing team.
- Worked on metadata clean-up and processing at regular intervals for better quality ofdata.
- Involved indataload/export utilities like Fast Load, Multi Load, Fast Export and UNIX/Mainframes environments.
- Responsible to build and maintain well-manageddatasolutions and deliver capabilities to tackle business problems.
- Partner with the business to provide consultancy and translate the business needs to design and develop tools, techniques, metrics, and dashboards for insights anddatavisualization.
- Drive an understanding and adherence to the principles ofdataquality management including metadata, lineage, and business definitions.
- Work collaboratively with appropriate Tech teams to manage security mechanisms anddataaccess governance.
- Build and execute tools to monitor and report ondataquality.
- Answering the technical queries, driving the product initiatives and metric collection and analysis.
- Generate daily, weekly and monthly reports as per the specifications/requirements by business users.
- Discuss business solutions with client business team, resolving existing problems and improving the report quality.
- Performed Source to Targetdataanalysis anddatamapping.
- Created SQL queries to validateDatatransformation and ETL Loading.
- Building, publishing customized interactive reports and dashboards, report scheduling using Tableau server.
Confidential
DataAnalyst
Responsibilities:
- Created new reports based on requirements. Responsible in Generating Weekly ad-hoc Reports
- Planned, coordinated, and monitored project levels of performance and activities to ensure project completion in time.
- Automated and scheduled recurring reporting processes using UNIX shell scripting and Teradata utilities such as MLOAD, BTEQ and Fast Load
- Experience with Perl
- Worked in a Scrum Agile process & Writing Stories with two weeks iterations delivering product for each iteration
- Worked on transferring the data files to vendor through sftp &Ftp process
- Involved in defining and Constructing the customer to customer relationships based on Association to an account & customer
- Created action filters, parameters and calculated sets for preparing dashboards and worksheets in Tableau.
- Experience in performing Tableau administering by using tableau admin commands.
- Worked with architects and, assisting in the development of current and target state enterprise level data architectures
- Worked with project team representatives to ensure that logical and physical data models were developed in line with corporate standards and guidelines.
- Involved in defining the source to target data mappings, business rules and data definitions.
- Responsible for defining the key identifiers for each mapping/interface.
- Performed data analysis and data profiling using complex SQL on various sources systems including Oracle and Teradata.
- Migrated three critical reporting systems to Business Objects and Web Intelligence on a Teradata platform
- Created Excel charts and pivot tables for the Adhoc data pull