Qureos

FIND_THE_RIGHTJOB.

DE&A - Core - DataBricks Data Architect

JOB_REQUIREMENTS

Hires in

Not specified

Employment Type

Not specified

Company Location

Not specified

Salary

Not specified

About the Role:

We are seeking an experienced Data Architect with deep expertise in Databricks and modern data lakehouse architectures across cloud platforms such as Azure, AWS, or GCP . The ideal candidate will be responsible for designing, governing, and leading large-scale data modernization and analytics initiatives. This role involves architecting data platforms that enable scalable data engineering, analytics, and AI/ML workloads using Databricks as the core platform.

Key Responsibilities:

Define and implement end-to-end data architecture using Databricks (Delta Lake, Unity Catalog, Spark, MLflow).

Design data ingestion, transformation, and storage frameworks supporting structured, semi-structured, and streaming data.

Architect data lakehouse solutions for enterprise-scale analytics and AI workloads.

Collaborate with stakeholders to translate business, analytics, and ML requirements into scalable data platform designs .

Develop standards and best practices for Databricks clusters, job orchestration, and optimization.

Work with data engineers, data scientists, and platform teams to ensure seamless integration and scalability.

Establish governance, security, and data lineage using Databricks Unity Catalog or third-party tools.

Define and implement data modeling , metadata management , and data quality frameworks .

Support multi-cloud deployment strategies (Azure, AWS, or GCP) and hybrid data environments.

Lead efforts on performance tuning, cost optimization, and DevOps integration for Databricks workloads.

Provide technical leadership and mentorship to offshore engineering teams.

Required Skills & Experience:

12+ years of overall experience in data architecture, engineering, and analytics platforms .

Strong expertise with Databricks , Delta Lake , Apache Spark , PySpark , and MLflow .

Hands-on experience with at least one major cloud platform ( Azure , AWS , or GCP ) and its native data services:

Azure: Data Lake Storage, Synapse, ADF

AWS: S3, Glue, Redshift, EMR

GCP: BigQuery, Dataflow, Dataproc

Proven experience designing lakehouse and ELT frameworks for large-scale enterprise data.

Strong understanding of data modeling, schema design, partitioning , and data lifecycle management .

Expertise in Python , SQL , or Scala for data processing and orchestration.

Experience implementing CI/CD pipelines , GitOps , and infrastructure automation (Terraform, CloudFormation).

Knowledge of data governance, cataloging, and access control frameworks (Unity Catalog, Purview, Glue Data Catalog).

Strong communication and architecture documentation skills.

Preferred / Nice-to-Have Skills:

Experience with real-time streaming (Kafka, Event Hubs, Kinesis).

Exposure to MLOps or AI/ML lifecycle management in Databricks.

Familiarity with Snowflake , dbt , or modern ELT tools .

Cloud certifications such as:

Databricks Certified Data Engineer / Architect Professional

AWS Solutions Architect / Azure Architect Expert / GCP Data Engineer

Education:

Bachelor’s or Master’s degree in Computer Science, Information Systems, or Data Engineering .

Advanced cloud or Databricks certifications preferred.

© 2025 Qureos. All rights reserved.