Qureos

FIND_THE_RIGHTJOB.

• AI Red Teamer

Date Posted:
13 January, 2026
Industry:
IT Services and IT Consulting
Location:
VAPORVM IT SERVICES DMCC

Job Description:

AI Red Teamer: (1ppl.)

Position Overview:
We are seeking a highly skilled and detail-oriented AI Red Teamer to join our
organization. As an AI Red Teamer, you will play a critical role in assessing,
testing, and improving the security, reliability, and ethical integrity of AI
systems. You will be responsible for identifying vulnerabilities, weaknesses,
and potential risks in AI models and systems, ensuring they meet the highest
standards of performance and security.

Key Responsibilities:

  • AI System
    Vulnerability Assessment:
    • Conduct thorough
      security assessments of AI models, algorithms, and systems to identify
      vulnerabilities, biases, and risks.
    • Simulate adversarial
      attacks to test the robustness and resilience of AI systems.
  • Adversarial Testing:
    • Design and execute
      adversarial scenarios to evaluate how AI systems respond to malicious
      inputs or attempts to manipulate their behavior.
    • Develop adversarial
      examples to test the limits of AI models and identify areas for
      improvement.
  • Risk Analysis and
    Mitigation:
    • Analyze risks
      associated with AI systems, including ethical concerns, data privacy, and
      potential misuse.
    • Provide
      recommendations for mitigating identified risks and improving system
      security.
  • Collaboration with
    Development Teams:
    • Work closely with AI
      developers, data scientists, and engineers to share findings and
      implement improvements.
    • Assist in the
      development of secure and robust AI systems by providing actionable
      feedback.
  • Compliance and Ethical
    Standards:
    • Ensure AI systems
      comply with relevant regulations, standards, and ethical guidelines.
    • Identify and address
      biases in AI models to promote fairness and inclusivity.
  • Documentation and
    Reporting:
    • Prepare detailed
      reports on findings, vulnerabilities, and recommendations for
      stakeholders.
    • Maintain
      documentation of testing methodologies and results for future reference.
  • Continuous Learning
    and Research:
    • Stay updated on the
      latest advancements in AI security, adversarial techniques, and ethical
      AI practices.
    • Research emerging
      threats and develop innovative strategies to counteract them.

Qualifications:

  • Bachelor’s or Master’s
    degree in Computer Science, Artificial Intelligence, Cybersecurity, or a
    related field.
  • Proven experience in
    AI security, adversarial testing, or ethical AI development.
  • Strong understanding
    of machine learning algorithms, neural networks, and AI frameworks.
  • Knowledge of
    cybersecurity principles and practices, including penetration testing and
    vulnerability assessment.
  • Proficiency in
    programming languages such as Python, R, or Java.
  • Familiarity with tools
    and libraries for adversarial testing (e.g., Foolbox, CleverHans, ART).
  • Excellent
    problem-solving skills and attention to detail.
  • Strong communication
    skills to effectively convey findings and recommendations.

Preferred Skills:

  • Experience with large
    language models (LLMs) and generative AI systems.
  • Knowledge of data
    privacy regulations and ethical AI guidelines.
  • Certifications in
    cybersecurity or AI-related fields (e.g., CEH, CISSP, TensorFlow
    Developer).

© 2026 Qureos. All rights reserved.