We provide IT Staff Augmentation Services!

Python Data Analyst Resume

4.00/5 (Submit Your Rating)

Mt Laurel New, JerseY

SUMMARY

  • Data Scientist/Python Developer with extensive experience in Machine Learning, Predictive Modeling, Statistical Analysis, Data Mining, Data cleaning, Data Acquisition, Data validation, and Data Visualization with large data sets of Structured and Unstructured data.
  • Expertise in transforming business requirements into analytical models, designing algorithms, building models, developing data mining and reporting solutions in various domains like Networking, Banking, Healthcare, Insurance and E - Commerce.
  • Extensive programming skills in analytical and statistical programming language, Python 2.x/3.x
  • Experienced in Statistical Modelling and Machine Learning techniques like Linear and Logistic Regression, Decision Tree, Random Forest, Naïve Bayes, Neural Networks, Support Vector Machines, KNN classifier and Cluster analysis and other machine learning algorithms.
  • Adept in using Python libraries such as Pandas, Numpy, Seaborn, Matplotlib, Scikit-learn to analyze the insights of data and perform data cleaning, data visualization and build models.
  • Experience in using Jupyter notebook in Anaconda an open source distribution for Python programming.
  • Working experience in RDBMS such as MySQL, MemSQL, Oracle, SQL server.
  • Experience in NoSQL databases like MongoDB (RoboMongo) and exposure in Cassandra.
  • Exposure in Big Data Topics like Hadoop, HDFS architecture, MapReduce, PySpark.
  • Experience in messaging systems such as RabbitMQ and Kafka.
  • Expert in designing visualizations using Tableau and Power BI and publishing and presenting Dashboards, Storylines, Text analysis on web and desktop platforms.
  • Strong experience in Software Development Life Cycle (SDLC) including Requirements Analysis, Design Specification in Agile methodologies.
  • Experience in using Hypothesis testing and ANOVA testing to evaluate models.
  • Well versed in Git repository for version control.
  • Ability to independently and effectively organize and manage multiple assignments with excellent analytical and problem-solving skills and ability to work independently and as a part of the team.

TECHNICAL SKILLS

Python Packages\Languages\: Numpy, Pandas, Matplotlib, Seaborn, Scikit-learn\Python 2.x/3.x\

Hadoop Ecosystem\IDE’s: Hadoop, MapReduce, HDFS, PySpark\Sublime Text, Jupyter Notebook 4.1\

Reporting Services\Microsoft Tools\: Tableau, Power BI, Pivot Tables.\ Office 2016, Visio 2013, PowerPoint \

Databases\: MySQL, MemSQL, Oracle, SQL Server\

MongoDB (Robo-Mongo), Cassandra\: Machine Learning Algorithms\

PROFESSIONAL EXPERIENCE

Confidential, Mt Laurel, New Jersey

Python Data Analyst

Responsibilities:

  • Developing Python scripts to login and obtain card and port data for all of Confidential 's network CMTS, CRAN, VOD, Edge QAM's and other network routers.
  • Ranking the CMTS, VOD and CRAN devices at division, region, market, site levels based on the devices capacity and usage.
  • Aggregate the base level data for each type of device to various levels such as Divisions, Regions, Markets and Sites.
  • Migrating data from MySQL to Mongo DB (No SQL).
  • Gathering feature specific requirements and making infrastructure setup to handle feature data.
  • Communicating the feature specific requirements with UI/UX development teams.
  • Provide excel reports with daily router inventory information to various internal teams.
  • Writing various scripts to query several networking devices, databases to gather information and store that data into MySQL, MemSQL and NoSQL databases
  • Developed service group mapping information for highly valuable internal reports.
  • Involved in infrastructure tasks such as installing and maintaining all the required packages for scripts to run on several servers including implemented Rabbit MQ for queuing tasks, Nagios and ELK stack.
  • Designed My SQL database schema, Mongo Documents to store the required data.
  • Working on Data Analytics using PyData Stack.
  • Assisted with development of PHP5, ReactJS, Redux.

Environment: Linux, Python2.6/3.5, MySQL, MemSQL, Mongo DB, Rabbit MQ, ELK Stack, Nagios, Celery, Cent OS servers (v6 and v7),Git-hub, Sublime text editor, Jupyter Notebook

Confidential, Memphis, Tennessee

Data Scientist

Responsibilities:

  • Used Predictive Modeling, Statistics, Machine Learning, Data Mining, and other aspects of data analytics techniques to collect, explore, and extract insights from structured and unstructured data.
  • Involved in gathering requirements while uncovering and defining multiple dimension. Extracted data from Oracle database using SQL and performed data analysis.
  • Performed Data Cleaning, Feature Scaling, and Feature engineering using Python packages such as Pandas, Numpy, Matplotlib, Sci-kit Learn.
  • Developed various machine learning algorithms and statistical modeling like Support Vector Machines (SVM), Logistic Regression to build predictive models.
  • Summarized the performance of models using Confusion matrix, Recall rate and Precision rate.
  • Used Hypothesis testing to compare and identify the effectiveness of creative models.
  • Visualized, interpreted, report findings and developed strategic uses of data using Tableau 9.3, and created interactive Dashboards.
  • Used Agile methodology to build the different phases of software development life cycle and used Git for version control.
  • Documented all programs and procedures to ensure an accurate historical record of work completed on assigned project as well as to improve quality and efficiency.

Environment: Python, Pandas, Numpy, Sci-kit Learn, Oracle 11g, Support Vector Machine, Logistic regression, Sublime Text, Tableau 9.3, Agile, Git

Confidential, Nashville, TN

Python Data Analyst

Responsibilities:

  • Worked in all phases of research like Data Cleaning, Data Mining, Feature Engineering, Developing tools, Validation, Visualizations and performance monitoring.
  • Handling data and performing creating, reading, updating and deleting(CRUD) operations on MongoDB.
  • Used Python 2.7 programming for managing and transforming and integrating with datasets in preparation for analytics.
  • Performed exploratory data analysis like statistical calculation, data cleaning and data visualizations using Numpy, Pandas and Matplotlib.
  • Developed predictive models using Logistic regression, Random Forest and KNN algorithms in Scikit-learn library.
  • Used Cross Validation for checking overfitting in suggested model.
  • To check the performance of model Confusion matrix, Recall Rate and Precision Rate are used.
  • Performed Anova testing to find the effectiveness of the model created using machine learning algorithms.
  • Created interactive Dashboards on desktop platform to visualize the data by using Tableau 9.0 desktop.
  • Used Anaconda Navigator an open-source tool for running Python in Jupyter notebook and Spyder.
  • Build various life cycles of project using Agile Methodology.
  • Interacted with the other departments to understand and identify data needs and requirements and work with other members to deliver and address those needs.

Environment: Python 2.7, MongoDB, Numpy, Pandas, Matplotlib, Sci-kit learn, Logistic Regression, Random Forest, KNN classifier, Anaconda Navigator, Jupyter notebook, Spyder, Tableau 9.0

Confidential

Python Data Analyst

Responsibilities:

  • Developed business process models using MS Visio 2013 to create case diagrams and flow diagrams to show flow of steps that are required.
  • Designed, Implemented and automated modelling and analysis procedures on existing and experimentally created data using Python packages like Pandas, Numpy, Matplotlib, Scikit Learn.
  • Created PL/SQL packages and Database Triggers and developed user procedures and prepared user manuals for the new programs.
  • Created dynamic linear models to perform trend analysis using Python
  • Used MS Excel, MS Access and SQL to write and run various queries.
  • Used traceability matrix to trace the requirements of the organization.
  • Analyze the data and create dashboards using Tableau 9.0.
  • Reviewed the logical model with application developer, ETL teams, DBA’s and testing team to provide information about the data model and business requirements.
  • Involved in the daily maintenance of the database that involved monitoring the daily run of the scripts as well as troubleshooting in the event of any errors in the entire process

Environment: Python, SQL, Oracle 11g, MS Office, MS Visio, Tableau 9.0

Confidential

PL/SQL developer

Responsibilities:

  • Requirement gathering from the Business Team.
  • Analyze the Existing code and do the impact analysis.
  • Performed DML and DDL operations in Oracle 9i as per business requirement
  • Developed complex database objects like Stored Procedures, Functions, Packages and Triggers using Oracle Database, SQL and PL/SQL.
  • Used SQL Loader to upload the information into the database and using UTL FILE packages write data to files.
  • Involved in peer to peer code reviews.
  • Code validation as per client requirement.

Environment: Oracle 10g, SQL, PL/SQL

Confidential

PL/SQL developer

Responsibilities:

  • Worked on requirements gathering, analysis, design, change management and deployment.
  • Automated the process of rebuilding indexes at regular interval for better performances.
  • Generated reports that can handle both dynamic grouping and sorting.
  • Created database objects like Tables, Views, Sequences, Synonyms, Stored Procedures, Indexes, Table Spaces and integrity constraints.
  • Wrote SQL queries using joins, Sub queries and correlated sub queries to retrieve data from database.
  • Debugged many PL/SQL packages procedures, function, cursors and types for application.
  • Query optimization, execution plan and Performance tuning of queries for better performance.

Environment: Oracle 10g, SQL, PL/SQL

We'd love your feedback!