
Associate III - Data Engineering
- Bangalore, Karnataka
- Permanent
- Full-time
- Implement ETL (Extract Transform Load) processes to facilitate efficient data movement and transformation.
- Integrate data from multiple sources including databases APIs cloud services and third-party data providers.
- Establish data quality checks and validation procedures to ensure data accuracy completeness and consistency.
- Develop and manage data storage solutions including relational databases NoSQL databases and data lakes.
- Stay updated on the latest trends and best practices in data engineering cloud technologies and big data tools.
- Adherence to schedule / timelines
- Adhere to SLAs where applicable
- # of defects post delivery
- # of non-compliance issues
- Reduction of reoccurrence of known defects
- Quickly turnaround production bugs
- Completion of applicable technical/domain certifications
- Completion of all mandatory training requirementst
- Efficiency improvements in data pipelines (e.g. reduced resource consumption faster run times).
- Average time to detect respond to and resolve pipeline failures or data issues.
test cases
and results.Configuration: * Follow configuration processes diligently.Testing: * Create and conduct unit tests for data pipelines and transformations to ensure data quality and correctness.
- Validate the accuracy and performance of data processes.
- Understand data schemas in relation to domain-specific contexts
and retest defects in accordance with project standards.Estimation: * Estimate timeeffort
and resource dependencies for personal work.Knowledge Management: * Consume and contribute to project-related documentsSharePoint
libraries
and client universities.Design Understanding: * Understand design and low-level design (LLD) and link it to requirements and user stories.Certifications: * Obtain relevant technology certifications to enhance skills and knowledge.Skill Examples: * Proficiency in SQL Python or other programming languages utilized for data manipulation.
- Experience with ETL tools such as Apache Airflow Talend Informatica AWS Glue Dataproc and Azure ADF.
- Hands-on experience with cloud platforms like AWS Azure or Google Cloud particularly with data-related services (e.g. AWS Glue BigQuery).
- Conduct tests on data pipelines and evaluate results against data quality and performance specifications.
- Experience in performance tuning data processes.
- Proficiency in querying data warehouses.
- Understanding of data warehousing principles and practices.
- Proficiency in SQL for analytics including windowing functions.
- Familiarity with data schemas and models.
- Understanding of domain-related data and its implications.