
Data Engineering Engineer II
- Chennai, Tamil Nadu
- Permanent
- Full-time
- Employees in this job function are responsible for designing, building, and maintaining data solutions including data infrastructure, pipelines, etc. for collecting, storing, processing and analyzing large volumes of data efficiently and accurately
- Collaborate with business and technology stakeholders to understand current and future data requirements
- Design, build and maintain reliable, efficient and scalable data infrastructure for data collection, storage, transformation, and analysis
- Plan, design, build and maintain scalable data solutions including data pipelines, data models, and applications for efficient and reliable data workflow
- Design, implement and maintain existing and future data platforms like data warehouses, data lakes, data lakehouse etc. for structured and unstructured data
- Design and develop analytical tools, algorithms, and programs to support data engineering activities like writing scripts and automating tasks
- Ensure optimum performance and identify improvement opportunities
- Google Cloud Platform - Biq Query,
- Data Flow, Dataproc, Data Fusion,
- TERRAFORM, Tekton,Cloud SQL, AIRFLOW,
- POSTGRES, Airflow PySpark, Python, API
- GenAI
- 4+ years Data Engineering work experience
- Strong proficiency and hands-on experience in both Python(Must-have) and Java(Nice to have).
- Experience building and maintaining data pipelines (batch or streaming) preferably on Cloud platforms(especially GCP).
- Experience with at least one major distributed data processing framework (e.g., DBT, DataForm, Apache Spark, Apache Flink, or similar).
- Experience with workflow orchestration tools (e.g., Apache Airflow, Qlik replicate etc).
- Experience working with relational databases (SQL) and understanding of data modeling principles.
- Experience with cloud platforms (Preferably GCP. AWS or Azure will also do) and relevant data services (e.g., BigQuery, GCS, Data Factory, Dataproc, Dataflow, S3, EMR, Glue etc.).
- Experience with data warehousing concepts and platforms (BigQuery, Snowflake, Redshift etc.).
- Understanding of concepts related to integrating or deploying machine learning models into production systems.
- Experience working in an Agile development environment & hands-on in any Agile work management tool(Rally, JIRA etc.).
- Experience with version control systems, particularly Git.
- Solid problem-solving, debugging, and analytical skills.
- Excellent communication and collaboration skills.
- Experience working in a production support team (L2/L3) for operational support.
- Familiarity with data quality and data governance concepts.
- Experience building and consuming APIs (REST, gRPC) related to data or model serving.
- Bachelor's or Master's degree in Computer Science, Engineering, Data Science, or a related field.
- Bachelor's Degree
- Bachelor's Degree