
5– 8 Years of Total Experience in Hadoop Administration using either Hortonworks or Cloudera distributions. – Top priority
3+ years of experience Big Data, Hadoop( tools like Hive, Zookeeper, Sqoop, Spark, YARN ), Data Lake,
Microsoft SQL Server, Oracle, Unix platforms.
Programming Languages: Proficiency in Spark or Python.
SQL and Database Knowledge: Understanding of SQL (Structured Query Language) and familiarity with relational databases like MySQL, PostgreSQL,Oracle or Microsoft SQL Server. This includes the ability to write and optimize SQL queries.
Linux/Unix Basics: Understanding of basic Linux/Unix commands and shell scripting for data processing and automation
Good to have below skills.
Data Integration and ETL: Familiarity with Extract, Transform, Load (ETL) processes and tools used to integrate data from various sources into data warehouses or data lakes.
Data Cleansing and Preprocessing: Understanding of data cleaning techniques and data preprocessing to ensure data quality and consistency.
Big Data Technologies: Basic knowledge of big data technologies, such as Hadoop, Spark, or Apache NiFi, and their role in processing and managing large-scale data.
Version Control: Familiarity with version control systems like Git for collaborative development and code management.
Data Warehousing Concepts: Understanding the fundamentals of data warehousing, including data aggregation, star schema, and data cube concepts.
Data Formats: Knowledge of various data formats commonly used in data engineering, such as JSON, CSV, Parquet, Avro, etc.
Basic Cloud Computing: Familiarity with cloud computing platforms like AWS, Azure, or Google Cloud, and the ability to work with cloud-based data storage and processing services.
Data Visualization: Basic knowledge of data visualization tools like Tableau or Power BI,Qlik View to create visual representations of data.
Basic Domain Knowledge in Banking/Finance area.
Job Type: Full-time
Education:
- Bachelor's (Required)
Work Location: In person
