
Senior Analyst - Data Platform
- Hyderabad, Telangana
- Permanent
- Full-time
- Design, develop, and maintain data pipelines using Databricks, Spark, and other Azure cloud technologies.
- Optimize data pipelines for performance, scalability, and reliability, ensuring high speed and availability of data warehouse performance.
- Develop and maintain ETL processes using Databricks and Azure Data Factory for real-time or trigger-based data replication.
- Ensure data quality and integrity throughout the data lifecycle, implementing new data validation methods and analysis tools.
- Collaborate with data scientists, analysts, and stakeholders to understand and meet their data needs.
- Troubleshoot and resolve data-related issues, providing root cause analysis and recommendations.
- Manage a centralized data warehouse in Azure SQL to create a single source of truth for organizational data, ensuring compliance with data governance and security policies.
- Document data pipeline specifications, requirements, and enhancements, effectively communicating with the team and management.
- Leverage AI/ML capabilities to create innovative data science products.
- Champion and maintain testing suites, code reviews, and CI/CD processes.
- Strong knowledge of Databricks architecture and tools.
- Proficient in SQL, Python, and PySpark for querying databases and data processing.
- Experience with Azure Data Lake Storage (ADLS), Blob Storage, and Azure SQL.
- Deep understanding of distributed computing and Spark for data processing.
- Experience with data integration and ETL tools, including Azure Data Factory.
- Data warehouse and data lake concepts and architectures.
- Optimizing performance of databases and servers.
- Managing infrastructure for storage and compute resources.
- Writing unit tests and scripts.
- Git, GitHub, and CI/CD practices.
- Experience with big data technologies, such as Kafka, Hadoop, and Hive.
- Familiarity with Azure Databricks Medallion Architecture with DLT and Iceberg.
- Experience with semantic layers and reporting tools like Power BI.
- 5+ years of experience as a Data Engineer, ETL Developer, or similar role, with a focus on Databricks and Spark.
- Experience working on internal, business-facing teams.
- Familiarity with agile development environments.
- Bachelor's degree in Computer Science, Engineering, or a related field, or equivalent work experience.