Senior AI / LLM Engineer

Senior AI / LLM R&D Engineer

Location: Remote - Europe Time Zones

Level: Senior / Staff

 

Experience required: 5+ years in machine learning / deep learning engineering with direct hands-on work on large language models

 

About the role

We are seeking an exceptional Senior LLM R&D Engineer to join our frontier AI research & development team. You will work on next-generation large language models, post-training techniques, reasoning & agentic systems, efficient scaling, and multimodal capabilities.

Your work will directly contribute to training / post-training of models in the 100B–1T+ parameter range, pushing state-of-the-art performance in reasoning, long-context understanding, tool use, and safety/alignment.

This is a highly technical, research-oriented engineering role — ideal for people who love reading recent papers, implementing them at scale, and shipping improvements to production models.

 

Key Responsibilities

  • Design, implement and scale pre-trainingpost-training (SFT, DPO, PPO, RLVR, online RL, etc.) and inference pipelines for very large language models
  • Conduct cutting-edge experiments in model architecture, mixture-of-experts (MoE), state-space models, test-time compute scaling, long-context training, and multimodal integration
  • Own end-to-end ownership of major R&D initiatives: from data curation → training recipe → evaluation → iteration
  • Optimize training & inference efficiency (memory, throughput, energy) using latest distributed training techniques and hardware-aware optimizations
  • Build robust evaluation harnesses (human & automated) for reasoning, agentic capabilities, safety, and factual accuracy
  • Collaborate closely with researchers, systems engineers, data team, and product teams to turn research ideas into production-grade model improvements
  • Stay at the absolute forefront of the field — implement ideas from recent ArXiv / conferences within days/weeks

 

Required Qualifications

  • 5+ years of professional experience building, training or fine-tuning large-scale deep learning models (LLMs ≥ 13B strongly preferred)
  • Very strong PyTorch proficiency (JAX bonus)
  • Deep hands-on experience with modern LLM training & post-training stacks:
    • Hugging Face TransformersAcceleratePEFT (LoRA/QLoRA/DoRA/LoHa…)
    • DeepSpeedMegatron-LMvLLMSGLangTRLAxolotlUnslothLiger-kernelFSDP2 / torch.compile
    • Alignment toolkits: TRL, Aligners, RLAIF pipelines, reward modeling
  • Proven track record shipping LLM improvements in at least one of: reasoning, long-context, tool use / agents, multimodality, multilingual, or safety
  • Strong software engineering skills: clean, modular, well-tested code; experience with large codebases
  • Comfortable working at scale: multi-node (hundreds of GPUs), massive datasets (trillions of tokens), long-running jobs
  • Solid understanding of latest advances in LLMs (2024–2026 papers): MoE scaling laws, test-time scaling, inference-time search, chain-of-thought distillation, RL for reasoning, etc.

 

Strongly Preferred / Differentiating Skills

  • Experience training or heavily customizing models ≥ 70B parameters
  • Publications (even workshop) or strong open-source contributions in LLM space
  • Hands-on work with frontier open models (Llama-3/4, DeepSeek-R1/R2, GLM-4.x, Qwen-2.5, Mistral Large / Nemo, Gemma-2/3…)
  • Expertise in any of: agent frameworks (LangGraph, CrewAI, AutoGen), RAG at scale, long-context architectures (RingAttention, Infini-Attention, Mamba-2 hybrids), speculative decoding, quantization (GPTQ/AWQ/FP8), multimodal (vision-language, video-language)
  • Experience with very large-scale data pipelines (Spark, Ray Data, web-scale deduplication/filtering)
  • Familiarity with modern evaluation suites (Arena-Hard, GPQA, SWE-bench Verified, FrontierMath, LiveCodeBench…)

 

What we offer

  • Work on frontier-scale models with meaningful compute budget
  • Direct impact on next-generation AI systems
  • Competitive compensation package (base + meaningful equity + performance bonus)
  • Top-tier compute resources and latest hardware
  • Flat structure — your ideas matter regardless of title

If you're passionate about building the next generation of reasoning & agentic foundation models and have shipped serious LLM work in the past 5 years — we want to talk to you.

 

Apply with: CV + GitHub / Hugging Face profile + short note about your most impressive LLM project (training, post-training, inference optimization, agent system…)

Required skills experience

AI/ML API 5 years
LLM 5 years
Big Data 5 years

Required domain experience

Machine Learning / Big Data 5 years

Required languages

English C1 - Advanced
Python, Machine Learning, LLM, LLM/Llama/Mistral/GPT/RAG/FAISS, Multi-agent LLMs, LLM & AI Agents:, LLM Integration, LLM tuning, LLM-tools, OpenAI/LlamaIndex/LLaMa/Moralis/LangChain/LLM/LLMOps
Published 10 February
29 views
·
4 applications
25% read
To apply for this and other jobs on Djinni login or signup.
Loading...