Lead Data Engineer
FM India
- Bangalore, Karnataka
- Permanent
- Full-time
- Define and implement the strategy for bringing structured and unstructured data sources into Data Analytics. Perform as the technical lead with the Product teams (Underwriting and Risk; Client Service, Sales and Marketing; Claims; Account and Location Engineering) as well as emerging data sources (purchased data sets; external data; etc.)
- Partner with product owners, developers, solution architects, business analysts, data engineers, data analysts, data scientists and others to understand data and reporting needs
- Assist in or lead the development and implementation of data solutions using data modeling techniques and technologies such as Fabric, Synapse Analytics, SQL Server, Azure services, and others as required
- Validate solutions are accurate through detailed and disciplined testing methodologies and taking opportunities to use AI in this process.
- Responsible for ensuring the developed solutions are accurate, good quality, and delivered according to schedule
- Ensure tables and views are designed for data integrity, efficiency and performance, and are easy to comprehend
- Design and implement the data flow, infrastructure pipelines, ETL/ELT, structured and unstructured data movement and storage solutions
- Design data models and data flows into and out of Data Analytics databases
- Understand and design data relationships between business and data subject areas
- Follow standards for naming conventions, code documentation and code review
- Support data exploration and transformation needs
- Lead the strategy and implementation of data cleansing and mentor other team members on data quality and data cleansing tasks as needed
- Define and implement approach to data profiling to identify data anomalies and resolve issues
- Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
- Lead the build-out of the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and Azure technologies
- Lead the definition and creation of prototypes and proof of concepts for the selected solutions, and implement complex data engineering projects
- Project planning and facilitation
- Create the development project implementation plans for strategic data analytics projects
- Prioritize work within assigned projects and assign work to development team
- Create estimates, timelines and resourcing needs for development team
- Lead the team in the development and implementation of the project
- Support production applications and reduce maintenance
- Identify repeat support issues and implement solutions to reduce those issues
- Manage and address operational data issues by determining and fixing root cause
- Mentor developers, data analysts and data scientists who need to interact with data
- Liaison with dba(s) and Infrastructure Teams on the infrastructure and configuration needed to support Data Analytics solutions.
- Monitor system performance and execute opportunities for optimization
- Monitor storage capacity and reliability
- Remediate production issues quickly, with appropriate validation and deployment steps
- Supply ad-hoc data extracts and analysis to respond to tactical business needs
- Provide clear and concise communication to users, teammates and leadership
- Identify work tasks and capture them in team backlog
- Participate in backlog and technical debt planning and prioritization
- Resolve colliding priorities and escalate as needed
- Network with product teams and infrastructure teams to keep abreast of databases, business processes and technology changes which may impact data delivery.
- 6-7 years of experiencerequired to perform essential job functions.
- Ability to read and create data models Relational (3rd Normal Form) and non-relational (Inmon/Kimball) database theory
- Strong SQL skills
- Build scalable, fault-tolerant data pipelines
- Design and support ETL/ELT processes
- Design, build, maintain data lakes and data warehouses
- Programming languages SQL, Python, Spark, Kafka, Azure, JSON
- Implement data governance and data security measures
- SQL
- Spark / Pyspark, Data lakes, Data warehouses
- Fabric Platform,
- Ability to read and create data models
- ETL/ELT Pipeline development experience
- Ability to keep pace with rapidly evolving cloud services and AI technologies
- Demonstrate importance of data governance, quality and observability
- Preferred: Orchestration and automation knowledge for managing complex data workflows
- Communication and collaboration skills
- Python,
- Azure Data Factory, Kafka,
- Synapse Analytics Platform,
- Relational (3NF) and non-relational (Inmon/Kimball) database theory
- AI/LLM Integration experience