Position Summary:
We are seeking a Professional Data Engineer to join our dynamic team, where you will play
a crucial role in developing and maintaining robust data solutions.
As a Professional Data Engineer, you will collaborate with data science, business analytics,
and product development teams to deploy cutting-edge techniques and utilise best-in-class
third-party products. The Data team operates with engineering precision, prioritising security, privacy, and regulatory compliance in every initiative. As a Professional Data Engineer, you will contribute to the team's commitment to utilising the latest tools and methodologies, ensuring that our data solutions align with industry best practices.
Our Tech Stack:
-
Languages: SQL and Python
-
Pipeline orchestration tool: Dagster (Legacy: Airflow)
-
Data stores: Redshift, Clickhouse
-
Platforms & Services: Docker, Kubernetes
-
PaaS: AWS (ECS/EKS, DMS, Kinesis, Glue, Athena, S3 and others.)
-
ETL: FiveTran & DBT for transformation
-
IaC: Terraform (with Terragrunt)
-
GenAI: Bedrock, LangChain, LLMs
Key Responsibilities:
-
Develop and maintain ETL pipelines using SQL and/or Python.
-
Use tools like Dagster/Airflow for pipeline orchestration
-
Collaborate with cross-functional teams to understand and deliver data requirements.
-
Ensure a consistent flow of high-quality data using stream, batch, and CDC processes.
-
Use data transformation tools like DBT to prepare datasets to enable business users to self-service.
-
Ensure data quality and consistency in all data stores.
-
Monitor and troubleshoot data pipelines for performance and reliability.
Essential Experience:
-
3+ years of experience as a data engineer.
-
Proficiency in SQL is a must
-
Experience with modern cloud data warehousing, data lake solutions like Snowflake, BigQuery, Redshift, Azure Synapse.
-
Experience with ETL/ELT, batch, streaming data processing pipelines.
-
Excellent ability to investigate and troubleshoot data issues, providing fixes and proposing both short and long-term solutions.
-
Knowledge of AWS services (like S3, DMS, Glue, Athena, etc.)
-
Familiar with DBT or other data transformation tools.
-
Familiarity with GenAI, and how to leverage LLMs to resolve engineering challenges
Other Desired Experience:
-
Experience with AWS services and concepts (like EC2, ECS, EKS, VPC, IAM, etc).
-
Familiar with Terraform and Terragrunt.
-
Experience with Python
-
Experience with orchestration tools like Dagster, Airflow, AWS Step functions, etc.
-
Experience with pub-sub, queuing, and streaming frameworks such as AWS Kinesis, Kafka, SQS, SNS.
-
Familiar with CI/CD pipelines and automation