
Data Engineer
- Gurgaon, Haryana
- Permanent
- Full-time
Location: India (Hybrid)
Experience: 3–6 Years
Employment Type: Full-timeCompany Profile:NPS Prism is a market-leading, cloud-based CX benchmarking and operational improvement platform owned by Bain & Company. NPS Prism provides its customers with actionable insights and analysis that guide the creation of game-changing customer experiences. Based on rock-solid sampling, research, and analytic methodology, it lets customers see how they compare to their competitors on overall NPS®, and on every step of the customer journey.With NPS Prism you can see where you’re strong, where you lag, and how customers feel about doing business with you and your competitors, in their own words. The result: Prioritize the customer interactions that matter most. NPS Prism customers use our customer experience benchmarks and insights to propel their growth and outpace the competition.Launched in 2019, NPS Prism has rapidly grown to a team of over 200, serving dozens of clients around the world. NPS Prism is 100% owned by Bain & Company, one of the top management consulting firms in the world and a company as one of the world’s best places to work. We believe that diversity, inclusion, and collaboration is key to building extraordinary teams. We hire people with exceptional talents, abilities and potential, then create an environment where you can become the best version of yourself and thrive both professionally and personally.Position Summary:
We are seeking a highly skilled and experienced Data Engineer to join our team. The ideal candidate will have strong expertise in Python, SQL and PySpark, with proven experience working on Databricks and cloud platforms such as Azure and AWS. A solid understanding of ETL tools like Python as well as basic knowledge of DevOps practices and CI/CD pipelines, will be advantageous. This is a unique opportunity to work in a dynamic and fast-paced environment to design and implement robust data solutions for scalable business needs. Working with Git and versioning.Key Responsibilities:Data Pipeline Development:
- Design, build, and optimize ETL/ELT workflows using tools like Databricks, SQL, Python/pyspark & Alteryx (Good to have).
- Develop and maintain robust, scalable, and efficient data pipelines for processing large datasets. from source to emerging data.
- Use Databricks for big data processing, analytics, and real-time data processing.
- Leverage Apache Spark for distributed computing and handling complex data transformations.
- Create and manage SQL-based data solutions, ensuring high availability, scalability, and performance.
- Develop and enforce data quality checks and validation mechanisms.
- Collaborate with cross-functional teams, including data scientists, analysts, and business stakeholders, to deliver impactful data solutions.
- Understand business requirements and translate them into technical solutions.
- Proficiency in Python, SQL, and PySpark for data processing and manipulation.
- Proven experience in Databricks and Apache Spark.
- Expertise in working with cloud platforms like Azure, AWS.
- Sound knowledge of ETL processes and tools like Alteryx. (Good to have)