Job Seekers, Please send resumes to resumes@hireitpeople.com
Required Skills:
- 3 - 6 years experience in Hadoop stack and storage technologies, HDFS, MapReduce, Yarn, HIVE, sqoop, Impala, spark, flume, kafka and oozie
- Extensive Knowledge on Bigdata Enterprise architecture (Cloudera preferred)
- Excellent analytical capabilities - Strong interest in algorithms
- Experienced in HBase, RDBMS, SQL, ETL and data analysis
- Experience in No SQL Technologies (ex., Cassandra/ MongoDB, etc )
- Experienced in scripting(Unix/Linux) and scheduling (Autosys)
- Experience with team delivery/release processes and cadence pertaining to code deployment and release
- Research oriented, motivated, pro-active, self-starter with strong technical, analytical and interpersonal skills.
- A team player with good verbal and written skills, capable of working with a team of Architects, Developers, Business/Data Analysts, QA and client stakeholders
- Versatile resource with balanced development skills and business acumen to operate at a fast and accurate speed
- Proficient understanding of distributed computing principles. Continuously evaluate new technologies, innovate and deliver solution for business critical applications
Desired Skills:
- Object-oriented programming and design experience.
- Degree in Computer Science or equivalent.
- Experience with automated testing methodologies and frameworks, including JUnit, is a plus
- Python IDEs(Django, Flask), data wrangling and analytics in a python based environment
- Fundamentals of Python - Data Structures, Collections, Pandas for file and other type of data handling, visualizations etc.
- Visual Analytics Tools knowledge ( Tableau )
- Experience with Big Data Analytics & Business Intelligence and Industry standard tools integrated with Hadoop ecosystem. ( R, Python )
- Data Integration, Data Security on Hadoop ecosystem. ( Kerberos )
- Any Big Data certification(ex. Clouderas CCP, CCA) is a plus