RESPONSIBILITIES:
- Architect, develop and deploy data science solutions on scalable data science platforms
- Evaluate and choose the right technologies & tools to solve a given problem
- Integrate various open source tools used at for various purposes into a cohesive platform
- Ensure systems meet business requirements and industry practices
- Stay updated on the evolving technologies and trends
REQUIREMENTS:
- Utilize Big Data technologies for producing technical designs
- Created architecture components with cloud and visualization methodologies
- Experience with Big Data tools like Hadoop, Hive, Spark etc..
- Experience to any of the following tool sets is a big plus, but not a must
- Data flow tools like Luigi/Azkaban etc.,
- Scrapping tools like Scrappy/Nutch etc.,
- Data blending/ wrangling using tools (such as Talend, Alteryx etc.)
- Scripting languages (Python, etc.)
- AWS/AZURE/IBM Big Data Platforms
- Ability to use a wide variety of databases (both SQL and NoSQL) and tools
- Scripting with Python, Ruby or Perl
- Ability to quickly learn and evaluate new technologies