
Data Engineer - Business Insights & Analytics
- Bangalore, Karnataka
- Permanent
- Full-time
- Create and maintain optimal data pipeline architecture ETL/ ELT into Structured data
- Assemble large, complex data sets that meet functional / non-functional business requirements and create and maintain Multi-Dimensional modelling like Star Schema and Snowflake Schema, normalization, de-normalization, joining of datasets.
- Expert level experience creating Fact tables, Dimensional tables and ingest datasets into Cloud based tools. Job Scheduling, automation experience is must.
- Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
- Setup and maintain data ingestion, streaming, scheduling and job monitoring automation. Connectivity between Lambda, Glue, S3, Redshift, Power BI needs to be maintained for uninterrupted automation.
- Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and “big data” technologies like AWS and Google
- Build analytics tools that utilize the data pipeline to provide actionable insight into customer acquisition, operational efficiency and other key business performance metrics
- Work with stakeholders to assist with data-related technical issues and support their data infrastructure needs
- Create data tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader
- 5-7 years of in-depth hands-on experience in data warehousing Redshift or any OLAP to support business/data analytics, business intelligence (BI)
- Advanced working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases and Cloud Data warehouse like Redshift
- Data Model development, additional Dims and Facts creation and creating views and procedures, enable programmability to facilitate Automation
- Prior Data Modelling, OLAP cube modelling in SQL Server, SSAS and Power BI experience
- Experience with Redshift and OLAP systems is must. GLUE pipeline skill is must
- Data compression into PARQUET to improve processing and finetuning SQL programming skills required
- Experience building and optimizing “big data” data pipelines, architectures and data sets
- Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement
- Strong analytic skills related to working with structured and unstructured datasets
- Experience with manipulating, processing and extracting value from large disconnected unrelated datasets
- Working knowledge of message queuing, stream processing, and highly scalable “big data” stores
- Experience supporting and working with cross-functional teams and Global IT
- Familiarity of working in an agile based working models
- Experience with relational SQL and NoSQL databases, including AWS Redshift, Postgres and Cassandra
- Experience with AWS cloud services Preferable: S3, EC2, EMR, RDS, Sage Maker
- Experience with stream-processing systems: Storm, Spark-Streaming, etc.
- Experience with object-oriented/object function scripting languages: Python, Java, R, C++ Scala, etc.
- Working knowledge of Databricks Platform
- Bachelor’s or master’s degree on Technology and Computer Science background