Job Title: Data Engineer
Location: Sacramento, CA (Onsite)
Description
- Excellent written and verbal communication
- Must have worked on several (5+) projects building solutions for Data Pipelines, Data warehousing, Data Modelling on Azure Data Platform (AWS, Google Cloud Platform - good to have)
- Expertise in design and development of Data pipelines and ETL using ADF, Databricks to move data from relational/structured/
unstructured data from source to data lake to data warehouse, on Azure - Creating Pipelines and integrations of multiple data sources
- Data processing and transformation using Databricks (Scala/Python)
- Expertise in optimizing cost and performance of data pipelines
- Experience working with Data Lake, Synapse on Azure
- Designing Data Lake storage layer (Landing/staging, raw, trusted/curated zones)
- Designing data storage DB schema, data models and processing on Azure Synapse DW
- Experience with Synapse Spark Pool
- Experience in Data Modelling
- Building Data Models from structured and semi-structured data (Data Model, Common Data Model etc.)
- Performance optimization of data model for high performance analytical and reporting workloads
- Expertise with SQL Server and extensive experience with SQL programming
- Programming and optimizing DB objects Views, Stored Procs, Functions
- Good experience in defining data access policy and security for azure data services
- Experience with Oracle EDW
- Knowledge/Experience migrating from Oracle EDW to any cloud DW
- Experience integrating AI skills/ML models with Data and reporting solutions
- Experience in building API layer for downstream consumption