Qureos

FIND_THE_RIGHTJOB.

Data Engineer - Databricks/PySpark/AWS

JOB_REQUIREMENTS

Hires in

Not specified

Employment Type

Not specified

Company Location

Not specified

Salary

Not specified

Experience: 2 - 4 years

Location: Chennai

Minimum Qualifications:
  • B.E./ master’s in computer science, Information systems, or Computer engineering, Systems Engineering with 2+ years of relevant experience
  • Minimum of 2+ years of work experience using Databricks, Spark and Databricks workflows, Autoloader.
  • Knowledge of other scripting languages such as Scala, JavaScript, and API development.
  • Knowledge of other Data Orchestration/ETL tools - Alteryx, Snowflake, Azure Data Factory etc.
  • Knowledge of any of BI Tools - Tableau, Power BI, QuickSight.

Job Description:
  • 2-3 years of relevant work experience
  • To work in the capacity of AWS Databricks, Python, PySpark Hands-On developer
  • Work with stakeholders for regular updates, requirement understanding and design discussions. Hands-on experience on designing and developing scripts for custom ETL processes and automation in AWS, data bricks, Python, PySpark etc.
  • Should be skilled in creating architecture and design documentation and have knowledge of data engineering tools & technologies
  • Expertise in Cloud-related Big Data integration and infrastructure Tec stack using Databricks & Apache Spark framework
  • Well-versed with scripting languages – Python
  • Well-versed with Data Orchestration/ETL.
  • Good SQL/PL SQL programming Skills – writing complex procedures and queries for Dataframes and Notebooks
  • Very good data engineering mindset to build and maintain pipelines, and ability to closely monitor critical pipeline batch jobs and resolve failures on an urgent basis.
  • Manage and mentor junior developers in adopting and implementing the data bricks workflows.
  • Excellent communication skills is a must

Similar jobs

No similar jobs found

© 2025 Qureos. All rights reserved.