Databricks_Bangalore

Diverse Lynx

  • Bangalore, Karnataka
  • Permanent
  • Full-time
  • 1 day ago
HDatabricksJob SummaryWe are seeking an experienced Databricks Developer with 57 years of expertise in big data engineering and analytics platforms. The ideal candidate will have deep experience in Apache Spark, Databricks, PySpark, and cloud platforms such as Azure or AWS, and a strong background in data pipeline development, performance optimization, and data governance.Key ResponsibilitiesDesign, develop, and maintain scalable data pipelines using Apache Spark on Databricks.Write efficient and production-ready PySpark or Scala code for data transformation and ETL processes.Integrate data from various structured and unstructured sources into a unified platform.Implement Delta Lake and manage data versioning, updates, and schema evolution.Optimize data processing workflows for performance, scalability, and cost efficiency.Collaborate with data scientists, analysts, and business stakeholders to deliver high-quality datasets.Implement data quality checks, validation routines, and logging mechanisms.Monitor and debug production jobs using Databricks jobs, notebooks, and clusters.Ensure security, privacy, and compliance standards are met throughout the data lifecycle.Provide guidance and mentorship to junior team members.Required Skills Qualifications57 years of experience in Big Data development.Hands-on experience with Databricks (including Workflows, Notebooks, Delta Live Tables, Unity Catalog).Strong programming skills in PySpark andor Scala.Solid understanding of Delta Lake architecture.Proficient in SQL for data analysis and transformation.Experience with cloud platforms such as Azure (Azure Data Lake, Data Factory, Synapse) or AWS (S3, Glue, Redshift).Familiarity with CICD for Databricks deployments (e.g., using GitHub Actions, Azure DevOps).Knowledge of data governance, cataloguing, and security best practices.Experience working in an AgileScrum environment.Preferred SkillsExperience with Databricks Unity Catalog and Delta Live Tables.Exposure to machine learning workflows in Databricks.Experience with Apache Airflow, Kafka, or other orchestrationmessaging tools.Certifications Databricks Certified Data Engineer AssociateProfessional, AzureAWS certification

Diverse Lynx