Job Seekers, Please send resumes to resumes@hireitpeople.com
Must Have Skills:
Must have at least 5 - 6 years of experience in development, design, administration, configuration, design and architecture of the following:
- Big Data and Distributed Processing Technologies and framework ( Spark, Python, SQL-Snowflake)
- AWS Cloud Services (S3, EC2, Lambda, SQS)
- Experience with Data preparation tools such as Trifacta & Alteryx & Understanding of Tableau for reporting
Nice to Have Skills:
- Linux Administration (Ability to troubleshoot OS issues)
- Data Science and Machine Learning ( Familiarity with Data preparation and Feature Engineering techniques, Understanding of Machine learning techniques)
- Understanding of SAS & R programming languages.
Detailed Job Description:
- Onboard large volumes of data (Terabyte scale) in an agile fashion (ingest Data from source to S3, Use Spark for Data Validation and Preparation, use spark/ Bulk load for Data loading to snowflake)
- Provide application, tool and Infrastructure support to Data Stewards and Data Scientists (Resolve issues related to tools, resolve data connectivity issues, Provision Infrastructure for Data processing needs)
- Provide solutions around Big Data processing, Data Management & Data Science use cases.
Minimum years of experience: 5+
Certifications Needed: No
Top 3 responsibilities you would expect the Subcon to shoulder and execute:
- Onboard large volumes of data Terabyte scale in an agile fashion
- Provide application, tool and Infrastructure support to Data Stewards and Data Scientists
- Provide solutions around Big Data processing, Data Management Data Science use cases.
Interview Process (Is face to face required?) Yes
Does this position require Visa independent candidates only? No