
Data Engineering Consultant
- Gurgaon, Haryana
- Permanent
- Full-time
- Automation and Data Modeling
- Design, build, and maintain automated data pipelines for extracting, transforming, and loading data from diverse sources (enterprise platforms, SharePoint, NAS drives, etc.)
- Develop robust and scalable data models to support risk surveillance analytics and reporting needs
- Implement and maintain workflows for scheduling and monitoring ETL/ELT jobs to ensure data freshness and reliability
- Utilize scripting and workflow automation tools to reduce manual intervention in data movement and processing
- Integrate new data sources and automate ingestion processes to expand surveillance coverage
- Data Management and Governance
- Ensure data quality, completeness, and consistency across all risk surveillance datasets
- Develop and enforce data validation, cleansing, and transformation procedures to support accurate analysis
- Implement data security and access controls in compliance with regulatory and organizational standards
- Maintain detailed metadata, data dictionaries, and lineage documentation for all data assets
- Support data governance initiatives, including data cataloguing, retention policies, and audit readiness
- Collaboration and Communication
- Partner with Risk Surveillance partners, data analysts, and audit teams to understand requirements and deliver analytical-ready datasets
- Collaborate with IT, data stewards, and business partners to resolve data issues and facilitate access to new data sources
- Communicate data pipeline status, issues, and solution approaches clearly to both technical and non-technical stakeholders
- Provide training and support for users on data tools, repositories, and best practices
- Document data workflows, processes, and solutions for knowledge sharing and operational continuity
- Comply with the terms and conditions of the employment contract, company policies and procedures, and any and all directives (such as, but not limited to, transfer and/or re-assignment to different work locations, change in teams and/or work shifts, policies in regard to flexibility of work benefits and/or work environment, alternative work arrangements, and other decisions that may arise due to the changing business environment). The Company may adopt, vary or rescind these policies and directives in its absolute discretion and without any limitation (implied or otherwise) on its ability to do so
- Overall 8+ years of program experience in Computer Science, Information Technology, Mathematics, Engineering, Data Analytics or related field
- 4+ years of SQL programming
- 4+ years programming in Python and/or R
- 2+ years of data modeling and scaled automation experience
- 2+ years of data visualization experience (Tableau and/or PowerBI)
- Solid interpersonal and analytical skills while working effectively with a matrixed team
- 2+ years experience in developing scalable solutions with SSIS, Data Factory, Python, or R
- Extensive program experience in Computer Science, Information Technology, Mathematics, Engineering, or related field
- Internal Audit / Control experience
- Cloud computing experience including Azure, AWS, Databricks, and/or Spark computing
- Experience working in a Healthcare Industry and or a complex IT environment
- Experience with conducting automation surrounding API calls
- Working knowledge of Big Data tools, Cloud platforms, SQL Server database engineering
- Data Science experience including regression analysis and machine learning techniques
- Change management tool experience (e.g., Github, Jenkins, or similar)