
Data Scientist
- Pune, Maharashtra
- Permanent
- Full-time
- Execute. This means being constantly biased towards action, knowing the data, being in the details, setting a high bar and building strong collaborative relationships with other teams, thinking boldly, disambiguating, always thinking of the customer first, and rapidly iterating towards something demonstrably awesome.
- Learn. Always exploring new techniques and technologies. Collaborating with team members and helping to push the practice forward.
- Design and Deploy. Ability to design and deploy end-to-end AI systems and pipelines
- Communicate. Ability to communicate clearly with clients and peers.
- Requirements Gathering. Ability to work with business to define requirements to solve business problems and translate those into solutions
- Management. Ability to manage your time and client expectations, as well as complete quality work on time and within budget
- A Master's Degree in Computer Science, Data Science, or any other related fields; or equivalent related professional experience
- Robust experience of working with cloud technologies such as Azure Databricks and MLflow and similar tools on the Azure stack.
- Excellent interpersonal and communication skills enabling you to translate technical concepts for business groups, present to executive audiences or drive effective discussion with your team
- Ability to work independently and participate in multiple diverse collaborating teams, prioritize effectively, and thrive in a fast-paced, dynamic environment
- In-depth knowledge of various modeling algorithms e.g., regression, classification, tree-based models, deep learning, clustering, reinforcement learning, dimensionality reductions, and time series models.
- Strong feature engineering and feature selection capabilities
- Strong model evaluation and selection capabilities
- Strong EDA (Exploratory Data Analysis) and visualization skills and experience/knowledge
- Ability to understand complex and ambiguous business needs and apply the right tools and approaches.
- Proficiency in Python/PySpark and SQL
- Understanding of distributed systems (compute and storage)
- Ability to build models for streaming, batch, and real-time serving