Master Works is seeking an AI Engineer with strong generative AI prototyping experience to rapidly turn ideas into working AI-powered demos, proofs of concept, and early-stage products. The role sits between research and production engineering — combining hands-on use of large language models, multimodal models, retrieval-augmented generation, and agent frameworks to validate feasibility, explore new capabilities, and de-risk feature delivery. The role is central to accelerating AI innovation across the data, analytics, and intelligence portfolio, with a particular emphasis on bilingual (Arabic/English) AI use cases.
-
Rapidly design and build generative-AI prototypes and proofs of concept — from initial idea to working demo — using modern LLM APIs, open-source models, and prototyping frameworks
-
Design and implement retrieval-augmented generation (RAG) pipelines, including document ingestion, chunking strategies, embedding generation, vector storage, and retrieval orchestration
-
Build agentic systems that use tool/function calling, planning, and multi-step reasoning to automate complex tasks; integrate AI agents with internal APIs, databases, and external services
-
Engineer prompts, system instructions, and structured-output schemas; iterate on prompt design through measurable evaluation rather than guesswork
-
Work with multimodal AI capabilities — text, vision, speech (STT/TTS), and document understanding — and combine them into coherent end-to-end AI pipelines
-
Build and run AI evaluation frameworks (offline benchmarks, LLM-as-judge, golden-set regression testing) to measure quality, accuracy, latency, and cost across model and prompt variants
-
Adapt and fine-tune open-source models (LoRA, QLoRA, full fine-tuning) when proprietary or domain-specific behaviour is required, especially for Arabic-language tasks
-
Apply responsible-AI practices — safety filtering, hallucination mitigation, prompt-injection defence, PII handling, and bias awareness — especially for sensitive or regulated environments
-
Hand off mature prototypes to production engineering teams, providing clear documentation, evaluation results, deployment notes, and migration guidance
-
Demo prototypes and AI capabilities to internal stakeholders and clients; explain technical concepts, limitations, and risks to non-technical audiences
-
Track and evaluate fast-moving developments in the generative-AI ecosystem (new models, tools, techniques, and providers) and recommend adoption where they materially improve outcomes
-
Collaborate cross-functionally with product, design, data, and software engineering teams to shape AI features end-to-end and ensure they deliver real user value
Requirements
-
Bachelor's degree in Computer Science, Artificial Intelligence, Machine Learning, Data Science, Software Engineering, or a related discipline.
-
Postgraduate qualification (MSc/PhD) in Machine Learning, Artificial Intelligence, Natural Language Processing, or a related field is preferred
-
Recognised certifications in machine learning, deep learning, generative AI, or cloud AI platforms (AWS, Azure, GCP) are a strong plus
-
3-5+ years of AI/ML engineering experience, with at least 1-2 years hands-on building generative-AI applications using modern LLMs
-
Demonstrable portfolio of shipped prototypes or production AI features — ideally including RAG systems, AI agents, or multimodal pipelines
-
Hands-on experience with leading LLM APIs (e.g., Anthropic Claude, OpenAI GPT, Google Gemini) and open-source model serving (e.g., Hugging Face, vLLM, Ollama)
-
Strong Python proficiency, including modern AI libraries such as PyTorch, Hugging Face Transformers, LangChain or LlamaIndex, and standard data tooling
-
Working experience with embedding models, vector databases (e.g., Pinecone, Weaviate, Qdrant, FAISS), and retrieval evaluation techniques
-
Familiarity with rapid-prototyping tools (Streamlit, Gradio, Jupyter), API integration, and lightweight web back-ends
-
Experience designing AI evaluation harnesses (offline benchmarks, LLM-as-judge, golden sets) and using observability/tracing tools for AI systems
-
Exposure to fine-tuning techniques (LoRA, QLoRA, instruction tuning) and to Arabic-language NLP is strongly preferred
-
Working knowledge of containerisation (Docker), version control (Git), and cloud AI platforms (AWS Bedrock, Azure OpenAI, GCP Vertex AI) is a strong plus
-
Excellent leadership and team development skills
-
Strong bias for action and a rapid-prototyping mindset; able to ship a working demo in days, not months
-
Comfortable working under uncertainty — rapidly forming hypotheses, running experiments, and iterating based on evidence
-
Strong analytical and problem-solving skills, with the discipline to evaluate AI behaviour quantitatively rather than relying on intuition alone
-
Excellent verbal and written communication, with the ability to demo AI capabilities and explain technical concepts, limitations, and risks to non-technical audiences
-
Curiosity and continuous-learning mindset; actively follows the rapidly evolving generative-AI landscape and brings new ideas back to the team
-
Collaborative team player; comfortable in cross-functional teams of researchers, software engineers, designers, and client business owners
-
Cultural awareness and sensitivity, particularly when designing AI experiences for Arabic-speaking users in regional contexts