B.E./ master’s in computer science, Information systems, or Computer engineering, Systems Engineering with 2+ years of relevant experience
Minimum of 2+ years of work experience using Databricks, Spark and Databricks workflows, Autoloader.
Knowledge of other scripting languages such as Scala, JavaScript, and API development.
Knowledge of other Data Orchestration/ETL tools - Alteryx, Snowflake, Azure Data Factory etc.
Knowledge of any of BI Tools - Tableau, Power BI, QuickSight.
Job Description:
2-3 years of relevant work experience
To work in the capacity of AWS Databricks, Python, PySpark Hands-On developer
Work with stakeholders for regular updates, requirement understanding and design discussions. Hands-on experience on designing and developing scripts for custom ETL processes and automation in AWS, data bricks, Python, PySpark etc.
Should be skilled in creating architecture and design documentation and have knowledge of data engineering tools & technologies
Expertise in Cloud-related Big Data integration and infrastructure Tec stack using Databricks & Apache Spark framework
Well-versed with scripting languages – Python
Well-versed with Data Orchestration/ETL.
Good SQL/PL SQL programming Skills – writing complex procedures and queries for Dataframes and Notebooks
Very good data engineering mindset to build and maintain pipelines, and ability to closely monitor critical pipeline batch jobs and resolve failures on an urgent basis.
Manage and mentor junior developers in adopting and implementing the data bricks workflows.