AI Engineer $$$
appflame - Ukrainian product-driven tech company building world-class products: Hily, Taimi, AdConnect, Mailkeeper, and more.
About us:
β’ 7 years in the market, 500+ team members, offices in Kyiv, London, Limassol, and a co-working hub in Warsaw.
β’ In 2025, appflame ranked 5th among the top 50 employers in Ukraine according to Forbes and won the βBest Place for Growthβ award.
β’ Our apps Hily and Taimi are among the top 5 dating apps in the US with over 70 million users. AdConnect and Mailkeeper focus on building proprietary B2B and B2C solutions.
Our mission:
β’ To put Ukrainian-built products on the global map.
Our goals:
β’ Break into the global top 5 product companies.
β’ Become a unicorn.
β’ Make Ukraine a country where unicorns are born.
Weβre looking for bold, driven people who are passionate about building real products and dream of launching and scaling great startups. You bring the ambition β we provide the environment to make it happen.
What youβll do:
β’ Design and develop an AI-powered productivity analytics platform β from data pipeline architecture to the final analytical product that helps teams make data-driven decisions.
β’ Build scalable LLM pipelines (Claude, GPT): develop data chunking strategies, implement MapReduce approaches for parallel processing of large datasets, and synthesize results into structured reports and insights.
β’ Create a meta-workflow system where LLMs generate, test, and deploy automation scripts in an isolated environment β with automatic self-correction loops and production deployment without manual intervention.
β’ Develop system-level prompt engineering: build and maintain a library of prompt templates for various analytical scenarios β from summaries and profiles to deep performance analysis.
β’ Build an evaluation framework for AI output quality control: hallucination detection, consistency scoring, regression tests β ensuring the product delivers reliable and reproducible results.
β’ Scale the platform to new domains and analysis types without linear growth in manual effort β through an architecture that allows adding new modules via configuration, not code.
β’ Document AI architecture, define automation specs, and present product insights to stakeholders and clients.
Itβs a match if you have:
β’ 2+ years of experience working with LLMs in production: prompt engineering, pipeline development, API integration β with at least 1 year of hands-on experience with advanced features (tool use, structured outputs, agents).
β’ Strong Python skills (async, dataclasses, type hints, API integrations) and a commitment to writing clean, testable, and maintainable code.
β’ Understanding of MapReduce patterns for LLM processing: ability to choose chunking strategies, organize parallel processing, and aggregate results into a cohesive analytical product.
β’ Experience building agentic systems: tool use, self-correcting loops, multi-agent workflows β and the judgment to know when an agent works better than a rigid pipeline.
β’ Proficiency in SQL and experience working with analytical databases.
β’ English at C1 level β comfortable reading documentation, writing technical specs, and communicating asynchronously.
β’ Ownership mentality: you take tasks end-to-end, debug production issues independently, and iterate to deliver results without micromanagement.
Nice to have:
β’ Experience with orchestration/automation platforms (Windmill, Dagster, Prefect) β understanding how to build reliable automated workflows.
β’ Knowledge of RAG architectures, vector databases, and embedding pipelines β ability to build systems that work with large volumes of unstructured data.
β’ Experience building evaluation systems for LLMs (LangSmith, PromptFoo, or custom solutions) β understanding how to measure and improve AI product quality.
β’ Familiarity with Databricks / Delta Lake or Snowflake β experience working with modern data platforms.
β’ Experience working at product-driven tech companies or AI startups where you needed to build and iterate quickly.
β’ Understanding of product team metrics (DAU, retention, unit economics) and the ability to connect technical decisions to product impact.
Preferred tech stack:
Core:
β’ Languages: Python (primary), SQL
β’ LLM APIs: Claude API (Anthropic), OpenAI API
β’ Databases: PostgreSQL, ClickHouse
β’ Infrastructure: Docker, Git, FastAPI, Pydantic, pytest, asyncio
Nice to have:
β’ Languages: TypeScript/JavaScript, Bash
β’ LLM APIs: Google Gemini API
β’ Orchestration & Automation: Windmill, Dagster, Prefect, Airflow, Temporal
β’ RAG & Embeddings: LangChain, LlamaIndex, ChromaDB, Pinecone, Weaviate, pgvector, FAISS
β’ Eval & Observability: LangSmith, PromptFoo, Weights & Biases, Arize AI
β’ Databases: Redis, MongoDB
β’ Data Platforms: Databricks, Delta Lake, Snowflake, BigQuery
β’ Infrastructure: Kubernetes, AWS (Lambda, SQS, S3), GCP
β’ CI/CD & DevOps: GitLab CI, GitHub Actions, Terraform
Hiring process: recruiter outreach > interview > test task > final interview > job offer.
Required languages
| English | B2 - Upper Intermediate |
| Ukrainian | Native |