Cloud Engineer Resume
0/5 (Submit Your Rating)
Elkridge, MD
SUMMARY
- Highly experienced Cloud Engineer with specialty in DevOps, Infrastructure, Systems, and Configuration.
- Experienced with AWS & AZURE, Splunk, Hadoop, Aruba and Cisco Installs.
- Successfully worked with both medium to large size clients.
- Excellent communication and written skills.
TECHNICAL SKILLS
Operating Systems: Windows 20080/2012/2016 / Win7, AIX, Linux CentOS Redhat, Ubuntu, Fedora OSX and Solaris
IBM ECM/BPM/ICM: ECM 4.x/4.5.x/5.1x/5.2x, BPM 4.x/4.5.x/5.1x/5.2x, IBM Case Manager (ICM), IBM Content Navigator (ICN), Business Process Framework (BPF), Infosphere Content Collector Records Manager(ICC)
RDBMS: Oracle 10/11g/12c, MS - SQL Server 2005/2008/2012 , 2016 DB2 9.7,10.1
VMWARE: ESX 3.5, Vsphere 4.0, Vcenter 5.0, 6.0 Workstation 6.0/6.5/7.0
Cloud: AWS, AZURE, Splunk
PROFESSIONAL EXPERIENCE
Confidential, Elkridge MD
Cloud Engineer
Responsibilities:
- Automate and manage our AWS infrastructure and deployment processes, including production, test and development environments.
- Automate of infrastructure using Terraform and Ansible
- Develop, Maintain and support Continuous Integration framework based on Jenkins
- Work with Jenkins Pipeline develop Pipeline Development, build configure with suite of Jenkins features, which is installing plugins, then enable implementation of continuous delivery pipelines, which is to automated the customer processes for getting software from source control through deployment to end users.
- Lead the development of innovative service solutions for Azure cloud service offerings
- Used Ansible and Ansible Tower as Configuration management tool, to automate repetitive tasks, quickly deploys critical applications, and proactively manages change.
- Wrote Python Code using Ansible Python API to Automate Cloud Deployment Process.
- Setup complete CI/CD Pipelines
- Automate instance schedule using Lambda Cloud Watch S3 and RDS services in AWS
- Edit and repurpose WordPress plugins under customers’ needs in AWS
- Write and extend WordPress plugins in AWS
- Developed procedures to unify streamline and automate applications development and deployment procedures with Linux container technology using Docker swarm.
- Worked in all areas of Jenkins setting up CI for new branches, build automation, plugin management and securing Jenkins and setting up master/slave configurations.
- Involved in deploying systems on Amazon Web Services Infrastructure services EC2, S3, RDS, SQS, Cloud Formation.
- Manage the Azure environments Network Design and Infrastructure Setup using Azure Services for both Development and Production systems.
- Build AWS-based services supporting production SaaS platform including web applications and data analytic services
- Provided leadership in developing innovative service capabilities for Azure Cloud and in managing Azure capability development project. plan, configure, optimization and deploy Microsoft Azure solutions (IaaS, PaaS, VMs, AD, Automation, Monitor, etc
- Migrate existing on-premises services to an AWS cloud infrastructure.
- Build/Maintain Docker container clusters managed by Kubernetes, Linux, Bash, GIT, Docker, on GCP. Utilized Kubernetes and Docker for the runtime environment of the CI/CD system to build, test deploy.
- Responsible for design and implementation of the Codex is Network and server infrastructure.
- Provide following duties as Sr Engineer include Firewall, Switch and Router configuration and maintenance
- Secured configured locked down Hadoop multi-tenant data sets to users and grant access to resources based on each user’s unique needs.
- Work with OS and application teams to ensure client service success.
- Worked on large enterprise supporting, HIPAA, FISMA, DOD and DCI which required data encrypted while it is in-flight being transferred over the network. Supported when it is at-rest whiles it was being stored durably on disk
- Performed Vulnerability Assessment & Penetration Testing on the infrastructure on AWSfor security.
- Installed configured maintained Key Trustee Server with Apache Sentry on the current AWS cloud.
- Responsible for auditing and tracking usage across multiple tenants and multiple clusters.
- Build a technical and security architecture in Azure for the selected apps/workloads
- Lead compliance assessments and application portfolio assessment with the customer on designed Azure architecture
- Select a migration approach to lift and shift the workloads to Azure or architecting a greenfield development and/or production platform for new applications
- Configured supported monitored Key Trustee Server with Apache Sentry within customers datacenter environments located offsite.
- Configured, data read from and written to HDFS directories while its transparently encrypted and decrypted without requiring any changes to user application code.
- Supported encryption end-to-end data that is protected both in-flight and at-rest, and can only be encrypted and decrypted by the customers and clients within DOD DHS and commercial sector.
- Configured encryption layers in traditional data management software/hardware stack.
- Supported and deployed encryption at a given layers in a traditional data management software/hardware stack with different advantages and disadvantages. Application-level encryption, Database-level encryption, Filesystem-level encryption, and Disk-level encryption
- Integrated various Version control tools, build tools, nexus and deployment methodologies (scripting) into Jenkins to create an end to end orchestration build cycles.
- Troubleshoot build issues in Jenkins, performance and generating metrics on master's performance along with jobs usage.
- Implemented enterprise-grade authorization mechanisms based on user directories and authentication technologies such as Kerberos.
- Installed configured Kerberos to allow Master/Slave replication cluster with consist of any number of hosts which stores all information, both account and policy data, in application databases.
- Ensure plan execution and Azure consumption targets are met
- Implemented Kerberos software distribution which includes software replication, such as copying data to other servers.
- Installed configured design Kerberos which gives client applications ability to attempt authentication against secondary servers if the primary master is down.
- Create data level security rules for IDH Hive users leveraging Apache Sentry
- Create new infrastructure Load Balancing, Packet Routing and SSH protocol designs to Maximize Network routing efficiency. Daily network monitoring and troubleshooting of network operation deficiencies
- Administering & designing LANs, WANs internet/intranet, and voice networks.
- Work with Tanium Foundation, Operations, & IR Deep Dive tools in customer enterprise AWS space
- Standardize Splunk forwarder deployment, configuration and maintenance across a variety of platforms
- Deploying and using enterprise EDR products such as Tanium
- Define, manage, and promote various development activities for DevOps practices, including continuous integration, continuous delivery, continuous testing, and continuous monitoring
- Support AWS Cloud infrastructure automation with multiple tools including Gradle, Chef, Nexus, Knife, Docker and monitoring tools such as Splunk, New Relic and Cloudwatch
- Responsible for designing, scaling and deploying various cloud services, modernizing processes and workflows along with building a consolidated and collaborative integration of IaaS, SaaS, and PaaS cloud services
- Manage all components of the DevOps Configuration Management platform (Jenkins, Nexus, GitLab, Sonar, etc.)
- Perform security log analysis during Information Security related events, identifying and reporting possible security breaches, incidents, and violations of security policies.
- Responsible for designing, developing, testing, troubleshooting, deploying and maintaining Splunk solutions, reporting, alerting and dashboards
- Implemented and supported Cloud Networks. Collaborate with security and network team to ensure all cloud platforms adhere to security models and compliance requirements for the cloud infrastructure for either on-premises or Cloud network. Assist in the support and troubleshooting of cloud network infrastructure along with the network support team to resolve complex operational issues
- Manage, configure and install VMware vSphere environment: vCenter, hypervisor on new hosts, virtual machines, datastore creation and maintenance
- Perform daily system monitoring of Virtual Infrastructure which includes VMware and Amazon Cloud Service
- Work with various teams to design, implement, integrate and operate AWS cloud solutions for high availability and scalable service delivery.
- Conduct and remediate Windows Security Content Automation Protocol (SCAP) and NESSUS system scans
- Configure ACAS (Security Center) Webinspect, Appdective and NESSUS to manage Windows server patches
- Work with Automating configuration management, infrastructure, and application deployments in a toolset such as Puppet, Chef, Ansible or Salt
- Implemented distributed data storage system using Accumulo and Hadoop Distributed File System (HDFS) for storing and running analytics on large volumes of data.
- Install, configure, and manage VMware vSphere environment: vCenter, hypervisor on new hosts, virtual machines, datastore creation and maintenance.
- Responsible for system administration, engineering, provisioning, operation, maintenance of vCenter, vRealize Operations, VMware Configuration Manager and support..
- Assist in the proper operation and performance of Splunk, loggers and connectors
- Worked configured responsible for Installation and configuration of Hadoop, YARN, Cloudera manager, Cloudera BDR, Hive, HUE and MySQL applications
- Reviewed performance stats and query execution/explain plans, and recommends changes for tuning Hive/Impala queries
- Enforce best practices in while maintaining customers environment as well as Service request management, Change request management and Incident management by using the standard tools of preference
- Review security management best practices which includes ongoing promotion of awareness on current threats, auditing of server logs and other security management processes, as well as following established security standards.
- Work with Cloudera maintenance, monitoring, and configuration tools to accomplish task goals and build reports for the management review.
- Responsible to build and maintain the Cloudera distribution of Hadoop.
- Perform cluster maintenance as well as creation and removal of nodes using tools like Ganglia, Nagios, Cloudera Manager Enterprise, Dell Open Manage and other tools
- Integrate data feeds (logs) into Splunk administering Splunk and Splunk App for Enterprise Security (ES) log management
- Standardize Splunk agent deployment, configuration and maintenance across a variety Of UNIX and Windows platform
- Work on System Center and Tanium design and deployment initiatives
Confidential, Hanover, MD
Cloud Engineer
Responsibilities:
- Participate in the upgrading of operating systems and design of systems enhancements.
- Provided consistent environment using Kubernetes for deployment scaling and load balancing to the application from development through production, easing the code development and deployment pipeline by implementing Docker containerization
- Developed Docker images to support Development and Testing Teams and their pipelines; distributed Jenkins, Selenium and JMeter images, and ElasticSearch, Kibana and Logstash (ELK & EFK) etc
- Infrastructure buildout, maintenance, & automation: Collaborated with infrastructure and product engineers to maintain ~1300 servers using Terraform for provisioning, Puppet for platform config & Ansible for deployment. Servers were spread across 14 datacenters/regions, from 3 cloud providers and 1 non-cloud provider
- Developed Python Modules for Ansible Customizations.
- Used Ansible Playbooks to setup Continuous Delivery Pipeline. Deployed micro services, including provisioning AWS environments using Ansible Playbooks.
- Used Ansible to document all infrastructures into version control.
- Work with the partner to identify, architect and design new cloud based solutions based on Azure technologies that the partner will sell to their customers.
- Identify, build and drive programs that establish new technical practices within your partner. These practices will be partner architects and consultants who are able to deliver consulting services to their customers using Azure cloud services
- Used Kubernetes, I have controlled and automated application deployments and updates and orchestrated deployment.
- Performed Vulnerability Assessment & Penetration Testing on the infrastructure on AWSfor security.
- Setup AWS VPC's for dev, staging and Prod environments.
- Used Amazon S3 to store and retrieve media files such as images and Amazon Cloud Watch is used to monitor the application and to store the logging information.
- Involved in writing JavaAPI for Amazon Lambda to manage some of the AWSservices.
- Configured and managed site counter-intelligence systems using Tripwire and Cisco Firewalls to protect servers and collect audit logs for the network packet filtering.
- Design and implement container orchestration systems with Docker
- Implement and managing private register and containers orchestration with tools such as Artifactory, Nexus, Docker and Docker Register
- Support the implementation of VMware hardware and operating systems solutions to provide hosting services to multiple data centers.
- Provision Virtual Machines and patches to the software and hardware hosting infrastructure
- Research, design and develop end to end technology stack (front end/backend) in support of Api’s to help support high volume web transactions
- Develop technical roadmaps for future AWS cloud implementations.
- Automate/configure management using Docker, Puppet, and Chef
- Design and develop web applications, RESTful API’s, prototypes, or proofs of concepts (POC’s)
- Architect and deploy Splunk Enterprise implementations in small to medium sized customers.
- Administer Splunk and Splunk App for Enterprise Security (ES) log management.
- Integrate Splunk with a wide variety of legacy data sources that use various protocols.
- Consulting with customers to customize and configure Splunk in order to meet their requirements.
- Perform implementation of security and compliance-based use cases based on the NIST Rev4 security controls.
- Technical writing/creation of formal documentation such as reports, training material, slide decks, and architecture diagrams.
- Work closely with middleware (e.g., WebLogic, Tomcat), database, UNIX, network and storage administrators for routine operations such as performance tuning, upgrades and backup.
- Deploy applications on multiple Weblogic Server and maintain Load balancing, high Availability and Fail over functionality
- Deploying, managing, and operating scalable, highly available, and fault tolerant systems on AWS
- Migrating an existing on-premises application to AWS
- Design and help to lead the implementations of cloud security solutions such as Web Application Firewalls, SIEM integrations, monitoring and auditing tools, and more.
- Implementing and controlling the flow of data to and from AWS
- Assist AWS Security Assurance team in determining the strategic direction of the various AWS compliance programs based on customer interaction and demonstrative metrics.
- Selecting the appropriate AWS service based on compute, data, or security requirements
- Proven ability to consultatively engage with Enterprise Clients to evaluate and translate functional requirements to a technology solution on Azure / AWS. Help design and implement hosting stack using AWS and Docker.
- Work with developers on understanding identified vulnerabilities and their underlying causes to develop plans of mitigating actions and comprehensive corrections.
- Install and configure AppDetective, WebInspect and Nessus out the box
- Uses a variety of tools (Nessus, HP WebInspect, AppDetective, Fluke Network Tester) to provide full range of system security testing.
- Configure applications in the C2S AWS environment on Chef configuration management tool
- Work with engineers on Docker and debugging bad builds using docker-machin, docker-compose, etc
- Conduct formal tests on web-based applications, networks, and other types of computer systems.
- Work on physical security assessments of servers, computer systems, and networks.
- Work with a team in charge of the management, maintenance, and operation of the customers HPC systems.
- Plan design, engineer, and project support for HPC hardware and software;
- Design and managing petabyte-scale data storage, with uses ranging from collaborative software development environments to multi-terabyte scientific datasets;
- Establish strategic relationships with vendors; collaborating with peers across the DOD .
- Collaborate with customers to address security and compliance challenges, and implementing and migrating customer solutions and workloads onto AWS
- Experience in design, build, test cloud apps on Cloud; should have a multi-year technical consulting and solutioning experience on AWS / Azure infrastructure and at least one PaaS platforms like Open Shift or PCG; a good understanding of pricing models will be an added advantage.
- Conduct regular security audits from both a logical/theoretical standpoint and a technical/hands-on standpoint.
Confidential, Hanover MD
Engineer
Responsibilities:
- Lead and contributing to the development, maintenance, and usage of deployment and task automation (OS, database services, virtual networks, or other platform services)
- Design enterprise collaborative cloud computing and hybrid cloud solutions with a focus on Microsoft Office 365 and Azure
- Implemented the UltraDNS & NS1 providers for Terraform, and a Nagios-compatible monitoring plugin to check for divergence.
- Provisioned load balancer, auto-scaling group and launch configuration for micro services using Ansible.
- Implemented Ansible to manage all existing servers and automate the build/configuration of new servers.
- Provided consistent environment using Kubernetes for deployment scaling and load balancing to the application from development through production, easing the code development and deployment pipeline by implementing Docker containerization.
- Crated images stored in the Docker container and uploaded to Docker hub.
- Created Linux containers in CoreOS and Docker and Automated system using Chef.
- Write, automation code using Chef, Puppet, or Ansible
- Integrated Kubernetes with network, storage, and security to provide comprehensive infrastructure and orchestrated container across multiple hosts.
- Provisioning of new services, mostly on new Juniper network, sometimes on former network before future migration.
- Reverse-engineering and troubleshooting of a complex system and hosting infrastructure.
- Responsible for the analysis, design and planning of infrastructure and architecture of solutions in Azure, and other related technologies
- Plan, create and manage cloud infrastructure in a Microsoft Azure environment
- Install and tune Hadoop clusters which includes benchmarking Hadoop cluster, supporting HA Namenodes, balancing HDFS block data and Datanode adding and decommissioning in secure enterprise environment.
- Monitor and troubleshoot Zookeeper and Yarn (configure different schedule options, manage and monitor workloads, maintain a multi-tenant environment, implement security controls, and manage high availability features of Hadoop.
- Troubleshoot Apache Sentry, Kerberos (both Kerberos RPC and HTTP SPNEGO), SSSD and HDFS interdependencies.
- Manage Hadoop I/O (including Data Integrity, Data compression and serialization), encrypting HDFS data at rest including encryption zones, manage HDFS snapshots and HDFS Backup DR.
- Manage Hadoop and Spark cluster environments, on bare-metal and container infrastructure, including service allocation and configuration for the cluster, capacity planning, performance tuning, and ongoing monitoring.
- Work with data engineering groups in the support of deployment of Hadoop and Spark jobs.
- Responsible for monitoring Linux, Hadoop, and Spark communities and vendors and report on important defects, feature changes, and or enhancements to the team.
- Create design and implement hosting stack using AWS and Docker.
- Secure and mobility-ready management of the whole network and system infrastructure (IPv4 and IPv6);
- Isolation of private addresses (RFC 1918) from the global routing table;
- Perform as a data scientist leveraging expertise with distributed scalable Big Data store, including Apache Accumulo, Apache Hadoop, MapReduce programming and technologies, and Real-time data processing with Apache Spark. to maintain day to day applications development implantation with tools such as Cloud Foundry, Chef, Puppet, Kubernetes, Docker, Heroku buildpacks and BOSH.
- Perform administration of VMware environment by managing the following VM components to include but not limited to the VMware Virtual Center, Site Recovery Manager, Operations Manager, Cloud Director, and other VMware products.
- Administer VMware 5.1/5.5 environment of approximately 300 hosts and 1500 virtual servers
- Maintain and manage VM resources to include (CPU, Memory and Disk) usage.
- Build new VM hosts and instances to support customer requirements.
- Transfer all production data to the newest VM infrastructure/platform.
- Resolve VM related incidents in compliance with organizational incident management process.
- Work with the Storage team to attach and manage Fibre Channel storage on VMware clusters.
- Work with Network Transport team to configure or acquire network resources and configure either standard switches on ESXi hosts or dynamic switches on VMware clusters
- Design, build, support and maintain Splunk infrastructure in a highly available configuration
- Standardize Splunk forwarder deployment, configuration and maintenance in Linux and Windows platforms
- Collaborate with internal teams to integrate data feeds to a centralized Splunk platform
- Develop and maintain production quality dashboards, custom views, saved searches and alerts for Splunk Operations and for other clients as per their requirements
- Assist internal users of Splunk in designing and maintaining production quality dashboards
- Monitor Splunk infrastructure for capacity planning and optimization
- Supported and responsible for Installation and configuration of Hadoop, YARN, Cloudera manager, Cloudera BDR, Hive, HUE and MySQL applications
- Reviewed performance stats and query execution/explain plans, and recommends changes for tuning Hive/Impala queries
- Enforce best practices in while maintaining customers environment as well as Service request management, Change request management and Incident management by using the standard tools of preference
- Reviewed security management best practices which includes ongoing promotion of awareness on current threats, auditing of server logs and other security management processes, as well as following established security standards.
- Work with Cloudera maintenance, monitoring, and configuration tools to accomplish task goals and build reports for the management review.
- Responsible to build and maintain the Cloudera distribution of Hadoop.
- Perform cluster maintenance as well as creation and removal of nodes using tools like Ganglia, Nagios, Cloudera Manager Enterprise, Dell Open Manage and other tools
- Installed, modify, support and maintain scripts, policies, procedures and documentation for automation and configuration management.
- Install configure application servers administration, including WebLogic, Tomcat, IIS and JBoss.
- Setup high availability for application servers e.g., Tomcat, Weblogic, ESB, JBoss
- Build, configure, install, maintain, diagnose, troubleshoot, repair, and debug EMC and NetApp products.
- Manage & customize Cloud Foundry Buildpacks & Services
- Build, manage and operate highly available systems utilizing Docker, Linux, Ubuntu, CoreOS, HAproxy, nginx, uWSGI, Couchbase, Zookeeper, Mesos, Marathon, Rabbitmq and Percona clusters