Sr. Python Developer Resume
Thomasville, GA
SUMMARY:
- I have 10 years of experience Analyzing, Developing, Managing and implementing various stand - alone, client-server enterprise applications using Python, Django and mapping the requirements to the systems. I am Well versed with Agile with SCRUM, Waterfall Model and Test-driven Development (TDD) methodologies.
- I do have Experience in developing web applications by using Python, Django, C++, XML, CSS, HTML, JavaScript and jQuery also analyzing data using Python, R, SQL, Microsoft Excel, Hive, PySpark, Spark SQL for Data Mining, Data Cleansing, Data Munging and Machine Learning.
- Experience in developing machine learning models like Classification, Regression, Clustering, Decision Tree also Sound knowledge in Data Quality & Data Governance practices & processes. I have Good experience in developing web applications implementing Model View Control (MVC) architecture using Django, Flask, Pyramid and Python web application frameworks.
- Experienced in developing web-based applications using SaaS, Python, Django, Kafka, RPC, CSS, Artificial Intelligence, AI/ML, HTML, JavaScript and jQuery based in Ansible automation script, Angular.JS, Boto 3, React.
- Experience in working with number of public and private cloud platforms like Confidential Web Services ( Confidential ), Microsoft Azure and Extensive experience in Confidential Web Services ( Confidential EC2, Confidential S3, Confidential Simple DB, Confidential RDS, Confidential Elastic Load Balancing, Elastic Search, Confidential MQ, Confidential Lambdas, Confidential SQS, Confidential Identity and access management, Confidential Cloud Watch, Confidential EBS and Confidential CloudFormation) also Proficient in SQLite, MySQL and SQL databases with Python.
- Experience in Python development, design, methodologies, tools and processes.
- Good experience in working with various Python Integrated Development Environments like PyCharm, Spyder, Jupyter Notebook and code editors like Sublime Text, Visual Studio Code.
- Experience in Software Development (Python) and have python framework knowledge.
- Experienced in working with various Python IDE’s using PyCharm, PyScripter, Spyder, PyStudio, PyDev, IDLE, NetBeans and Sublime Text also Experience with Requests, Report Lab, NumPy, SciPy, Pytables, cv2, imageio, Python-Twitter, Matplotlib, HTTPLib2, Urllib2, Beautiful Soup, Data Frame and Pandas python libraries during development lifecycle.
TECHNICAL SKILLS:
Operating Systems: Windows 98/2000/XP/7,8, Mac OS and Linux CentOS, Debian, Ubuntu
Programming Languages: Python, R, C, C++
Web Technologies: HTML/HTML5, CSS/CSS3, XML, jQuery, JSON, Bootstrap, Angular JS
Python Libraries/Packages: NumPy, SciPy,Boto, Pickle, PySide, PyTables, Data Frames, Pandas, Matplotlib, SQLAlchemy, HTTPLib2, Urllib2, Beautiful Soup, Py Query
Statistical Analysis Skills: A/B Testing, Time Series Analysis, Marko
Machine Learning and Analytical Tools: Supervised Learning (Linear Regression, Logistic Regression, Decision Tree, Random Forest, SVM, Classification), Unsupervised Learning (Clustering, KNN, Factor Analysis, PCA), Natural Language Processing, Google Analytics Fiddler, Tableau.
Cloud Computing: Confidential, Azure, Rackspace, OpenStack
Confidential Services: Confidential EC2, Confidential S3, Confidential Simple DB, Confidential MQ, Confidential ECS, Confidential Lambdas, Confidential Sagemaker, Confidential RDS, Confidential Elastic Load Balancing, Elastic Search, Confidential SQS, Confidential Identity and access management, Confidential Cloud Watch, Confidential EBS and Confidential CloudFormation
Databases/Servers: MySQL, SQLite3, Cassandra, Redis, PostgreSQL, CouchDB, MongoDB,TerraData, Apache Web Server 2.0, NginX, Tomcat, JBoss, WebLogic
ETL: Informatica 9.6, Data Stage, SSIS.
Web Services/ Protocols: TCP/IP, UDP, FTP, HTTP/HTTPS, SOAP, Rest, Restful
Miscellaneous: Git, GitHub, SVN, CVS
Build and CI tools: Docker, Kubernetes, Maven, Gradle, Jenkins, Hudson, Bamboo
SDLC/Testing Methodologies: Agile, Waterfall, Scrum, TDD
IDE: PyCharm, PyScripter, Spyder, PyStudio, PyDev, IDLE, NetBeans, Sublime Text, Visual Code
PROFESSIONAL EXPERIENCE:
Sr. Python Developer
Confidential, Thomasville, GA
Responsibilities:
- Develop a data platform from scratch and took part in requirement gathering and analysis phase of the project in documenting the business requirements.
- Developing web-based applications using SaaS, Python, Django, Kafka, RPC, CSS, Artificial Intelligence, AI/ML, HTML, JavaScript and JQuery based in Ansible automation script, Angular.JS, Boto 3, React.
- Worked in designing tables in Hive, MYSQL using SQOOP and processing data like importing and exporting of databases to the HDFS, involved in processing large datasets of different forms including structured, semi-structured and unstructured data.
- Developed rest API's using python with flask and Django framework and done the integration of various data sources including Java, JDBC, RDBMS, Shell Scripting, Spreadsheets, and Text files.
- Worked with Hadoop architecture and the daemons of Hadoop including Name-Node, Data Node, Job Tracker, Task Tracker, and Resource Manager.
- Used Confidential data pipeline for Data Extraction, Transformation and Loading from homogeneous or heterogeneous data sources and built various graphs for business decision-making using Python matplot library
- Developed scripts to load data to hive from HDFS and involved in ingesting data into Data Warehouse using various data loading techniques and Scheduled Jobs using crontab, run deck and control-m.
- Build Cassandra queries for performing various CRUD operations like create, update, read and delete, also used Bootstrap as a mechanism to manage and organize the html page layout
- Developed entire frontend and backend modules using Python on Django Web Framework and created User Interface (UI) using JavaScript, bootstrap, Cassandra with MySQL and HTML5/CSS3
- Importing and exporting data jobs, to perform operations like copying data from HDFS and to HDFS using Sqoop and developed Spark code and Spark-SQL/Streaming for faster testing and processing of data.
- Analyzed SQL scripts and designed the solutions to implement using PySpark.
Environment: Hadoop, Hive, Sqoop, Pig, java, Django, Flask, XML, MySQL, MS SQL Server, Linux, Shell Scripting, Mongo dB, SQL, Python 3.3, Django, HTML5/CSS, Cassandra, JavaScript, PyCharm, GIT, Linux, Shell Scripting, RESTful, Docker, Jenkins, JIRA, jQuery, MySQL, Bootstrap, HTML5, CSS, Confidential, EC2, S3.
Python developer
Confidential, LA
Responsibilities:
- Developed Data pipelines using python for medical image pre-processing, and Testing.
- Developed Artificial Intelligence Platform which helps Data Scientist’s to Train, Test and develop A.I. models on Confidential Sage maker.
- Used Pandas, OpenCV, NumPy, Seaborn, TensorFlow, Kera’s, Matplotlib, Sci-kit-learn, NLTK in Python for developing data pipelines and various machine learning algorithms.
- Design and engineer REST APIs and/or packages that abstract feature extraction and complex prediction/forecasting algorithms on time series data.
- Developed Python application for Google Analytics aggregation and reporting and used Django configuration to manage URLs and application parameters.
- Developed pre-processing pipelines for DICOM and NONDICOM Images.
- Developed and presented analytical insights on medical data, image data.
- Implement Confidential Lambdas to drive real-time monitoring dashboards from system logs.
- Cleansing the data for normal distribution by applying various techniques like missing value treatment, outlier treatment, and hypothesis testing.
- Perform Data Cleaning, features scaling, features engineering using pandas and numpy packages in python.
- Create several types of data visualizations using Python and Tableau.
- Collected data needs and requirements by Interacting with the other departments.
- Worked on different data formats such as JSON, XML.
- Performed preliminary data analysis using descriptive statistics and handled anomalies such as removing duplicates and imputing missing values.
- Developed various graph methods to visualize and understand the data like Scatter plot, Pi-plot, bar charts, box-plot, and histograms Also Involved in development of Web Services using REST API’s for sending and getting data from the external interface in the JSON format.
- Configured EC2 instances and configured IAM users and roles and created S3 data pipe using Boto API to load data from internal data sources.
- Developed rest API's using python with flask framework and done the integration of various data sources including Java, JDBC, RDBMS, Shell Scripting, Spreadsheets, and Text files.
Environment: Python 2.7, R, SAS 9.4, HDFS, MongoDB 3.2, Elastic Search, Hadoop, Hive, Linux, Spark, Scala, Kafka, Tableau 9.4, D3.js, SQL Server 2012, Spark SQL, PL/SQL.
Python Developer
Confidential, NY
Responsibilities:
- Worked on Python Open stack API's and used Python scripts to update content in the database and manipulate files.
- Involved in using Confidential for the Tableau server scaling and secured Tableau server on Confidential to protect the Tableau environment using Confidential VPC, security group, Confidential IAM and Confidential Direct Connect.
- Configured EC2 instances and configured IAM users and roles and created S3 data pipe using Boto API to load data from internal data sources.
- Built a mechanism for automatically moving the existing proprietary binary format data files to HDFS using a service called Ingestion service.
- Involved and worked on Python Open stack API's and used several python libraries such as wxPython, NumPy and matplotlib
- Performed Data transformations in HIVE and used partitions, buckets for performance improvements.
- Ingestion of data into Hadoop using Sqoop and apply data transformations and using Pig and HIVE.
- Used Python and Django creating graphics, XML processing, data exchange and business logic implementation
- Used Git, GitHub, and Confidential EC2 and deployment using Heroku and Used extracted data for analysis and carried out various mathematical operations for calculation purpose using python library - NumPy, SciPy.
- Developed server-based web traffic using RESTful API's statistical analysis tool using Flask, Pandas.
- Used Pandas API to put the data as time series and tabular format for east timestamp data manipulation and retrieval.
- Added support for Confidential Confidential S3 and RDS to host static/media files and the database into Confidential Cloud.
- Participate in the design, build and deployment of NoSQL implementations like MongoDB.
- Wrote and executed various MYSQL database queries from Python using Python-MySQL connector and MySQL dB package
Environment: Python, Hive, Oozie, Confidential Confidential S3, MySQL, HTML, Python 2.7, Django, HTML5, CSS, XML, MySQL, MS SQL Server, GIT, Jenkins, JIRA, MySQL, Cassandra, Pig, Hadoop, Confidential Cloud Watch, Confidential Redshift, SQL, SOAP, Rest APIs, Confidential EC2, XML, JavaScript, Confidential, Linux, Shell Scripting, AJAX, Mongo dB
Python developer
Confidential, Oaks, PA
Responsibilities:
- Developed entire frontend and backend modules using Python on Django Web Framework.
- Used Django framework for application development and Worked on CSS Bootstrap to develop web applications.
- Designed and developed the UI of the website using HTML, AJAX, CSS and JavaScript
- Used update strategy to effectively migrate data from source to target.
- Moved the mappings from development environment to test environment.
- Designed ETL Process using Informatica to load data from Flat Files, and Excel Files to target Oracle Data Warehouse database also Interacted with the business community and database administrators to identify the Business requirements and data realties.
- Created various transformations according to the business logic like Source Qualifier, Normalizer, Lookup, Stored Procedure, Sequence Generator, Router, Filter, Aggregator, Joiner, Expression and Update Strategy.
- Created Informatica mappings using various Transformations like Joiner, Aggregate, Expression, Filter and Update Strategy.
- Improving workflow performance by shifting filters as close as possible to the source and selecting tables with fewer rows as the master during joins.
- Used connected and unconnected lookups whenever appropriate, along with the use of appropriate caches.
- Created tasks and workflows in the Workflow Manager and monitored the sessions in the Workflow Monitor.
Environment: Python, Django, MySQL, Linux, Informatica Power Centre 9.5.1, PL/SQL, HTML, XHTML, CSS, AJAX, JavaScript, Apache Web Server, NO SQL, XML, jQuery.