FIND_THE_RIGHTJOB.
JOB_REQUIREMENTS
Hires in
Not specified
Employment Type
Not specified
Company Location
Not specified
Salary
Not specified
Job Summary
As a Product Engineer – Big Data, you will design, build, and optimize large-scale data processing pipelines using modern Big Data technologies. You will collaborate with data scientists, analysts, and product managers to ensure data accessibility, security, and reliability. Your work will focus on delivering scalable, high-quality data solutions while driving continuous improvements across the data lifecycle.
Key Responsibilities
Design and implement complex, end-to-end ETL pipelines for large-scale data ingestion and processing.
Optimize performance, scalability, and resilience of data pipelines.
Develop and optimize real-time and batch data workflows using Apache Spark, Scala/PySpark, and Apache Kafka.
Ensure fault-tolerant, high-performance data processing.
Knowledge of Java and NoSQL is a plus.
Build scalable, cost-efficient cloud-based data infrastructure leveraging AWS services.
Ensure pipelines are resilient to variations in data volume, velocity, and variety.
Work with business teams and data scientists to deliver high-quality datasets aligned with business needs.
Perform data analysis to uncover trends, anomalies, and actionable insights.
Present findings clearly to technical and non-technical stakeholders.
Enable seamless integration of real-time streaming and batch datasets from systems like AWS MSK.
Ensure consistency and reliability across data ingestion sources and formats.
Use Jenkins (or similar tools) to implement CI/CD pipelines.
Automate testing, deployment, and monitoring of data solutions.
Ensure pipelines comply with relevant data governance and regulatory frameworks (e.g., GDPR, HIPAA).
Implement controls for data integrity, security, and traceability.
Partner with engineers, product managers, and data teams in an Agile environment.
Contribute to sprint planning, architectural discussions, and solution design.
Identify and resolve bottlenecks in data pipelines.
Conduct performance tuning and adopt best practices for ingestion, processing, and storage.
Required Experience
4–8 years (or adjust as needed) of hands-on experience in Big Data engineering, cloud data platforms, and large-scale data processing.
Proven experience delivering scalable data solutions in production environments.
Mandatory Skills
AWS Expertise
Hands-on experience with EMR, Managed Apache Airflow, Glue, S3, DMS, MSK, EC2, and cloud-native data architectures.
Big Data Technologies
Proficiency in PySpark/Scala Spark and SQL. Experience with Apache Spark, Kafka, and large-scale data processing.
Data Frameworks
Strong knowledge of Spark DataFrames and Datasets.
Database Modeling & Data Warehousing
Experience designing scalable OLAP/OLTP data models and warehouse solutions.
ETL Pipeline Development
Proven ability to build robust real-time & batch pipelines across various platforms.
Data Analysis & Insights
Strong analytical skills with the ability to extract meaningful insights and support business decisions.
CI/CD & Automation
Practical experience with Jenkins or similar tools for automating deployment and monitoring.
Good-to-Have Skills
Familiarity with data governance frameworks and compliance standards.
Experience with monitoring tools such as AWS CloudWatch, Splunk, or Dynatrace.
Working knowledge of Java or NoSQL databases.
Exposure to cost optimization strategies in cloud environments.
Apache Spark,Scala experience,Aws,Big Data
Similar jobs
UST
Hyderabad, Pakistan
6 days ago
Tata Consultancy Services (TCS)
Hyderabad, Pakistan
6 days ago
Tata Consultancy Services (TCS)
Hyderabad, Pakistan
6 days ago
UST
Hyderabad, Pakistan
6 days ago
Tata Consultancy Services (TCS)
Hyderabad, Pakistan
6 days ago
© 2025 Qureos. All rights reserved.