Key Responsibilities:
Data Pipeline Engineering
- Design and build batch and near-real-time pipelines using:
- Azure Fabric Data Factory
- Snowflake ingestion patterns (COPY, Snowpipe, Streams, Tasks)
- Implement lakehouse-style data layers:
- Raw → Curated → Consumption
- Manage data structures using:
- Azure Fabric OneLake
- Snowflake databases, schemas, and secure objects
Data Consolidation & Modeling
- Consolidate data from multiple enterprise systems into governed datasets
- Apply enterprise data modeling patterns:
- Conformed dimensions
- Canonical fact tables
- Shared reference data
- Manage schema design, evolution, and versioning for long-term stability
Data Quality & Trust
- Implement data validation for completeness, consistency, and freshness
- Detect and surface data quality issues early
- Partner with business teams to define clear data contracts
Platform Optimization & Operations
- Optimize performance and cost across Snowflake and Fabric through:
- Partitioning, clustering, and pruning
- Warehouse sizing and query tuning
- Support monitoring, alerting, and incident response for data pipelines
Required Skills & Experience:
- 4+ years of hands-on data engineering experience
- Strong SQL skills with Snowflake
- Experience building pipelines in Microsoft Azure Fabric
- Solid understanding of data modeling (dimensional, canonical, wide tables)
- Experience integrating data from SaaS, ERP, operational, and internal systems
- Familiarity with CI/CD and source-controlled data pipelines
- Ability to work independently in remote, client-embedded environments
Nice to Have:
- Experience with Microsoft Purview, Fabric governance, or Snowflake Data Sharing
- Exposure to AI/ML-ready or feature datasets
- Knowledge of data quality frameworks or automated testing tools
- Experience in regulated or finance-heavy environments
Location: Onsite for KHI & ISB | Remote for other cities.
Job Type: Full-time
Work Location: In person