
Data Engineer - AWS/PySpark/Hadoop - 5+ Years - Bengaluru
- Bangalore, Karnataka
- Permanent
- Full-time
- Collaborate with product, analytics, and engineering teams to improve models feeding BI tools and enable data-driven decision-making.
- Implement data quality monitoring, reliability SLAs, and documentation (engineering wiki); write unit/integration tests.
- Troubleshoot complex data issues via deep analysis; define & optimize Spark SQL/Hive SQL jobs to populate models.
- Design robust data integrations and a data quality framework aligned to compliance and reporting needs.
- 5+ years with Hive, Hadoop, Spark and data warehouse design; strong Python, SQL, PySpark.
- Hands-on with AWS (S3, EMR) or Azure; familiarity with CI/CD, Docker, and automation.
- Experience supporting enterprise reporting and data compliance use cases; strong documentation habits.
- Excellent problem-solving, communication, and collaboration skills; thrive in agile, fast-paced environments.
- Build at scale for global clients across US, Europe, APAC & India.
- Work in an automation-first culture with modern data stacks.
- Cross-functional exposure (product, engineering, analytics).