Design, build, and maintain batch and/or streaming data pipelines to ingest, transform, and load data from various internal and external systems into data warehouses, data lakes, or lakehouses.
Develop and optimize data models, schemas, and storage layouts to support analytics, reporting, and machine learning use cases.
Implement data quality checks, monitoring, and alerting to ensure the accuracy, completeness, and timeliness of critical datasets.
Collaborate with data scientists, application developers, analysts, and business stakeholders to understand data requirements and deliver well-documented datasets and tables.
Manage and tune data infrastructure components, including databases, warehouses, orchestration tools, and distributed processing frameworks.
Apply security and governance best practices, including access controls, data lineage, and compliance with privacy regulations.
Contribute to engineering standards, documentation, and reusable tooling to improve the reliability and productivity of the data platform.