Senior Python / AI Engineer

Senior Python / AI Engineer

Role Overview

We are looking for a Senior Python/AI Engineer with strong experience in machine learning, data engineering, cloud-based architectures, and AI agent pipelines.

You will work directly with the system architect to design and build:

  • AI-powered data pipelines,
  • prediction and scoring models,
  • market-resolution oracles,
  • agentic processing pipelines (AWS Bedrock / LangChain / custom),
  • high-performance AWS-backed microservices and event-driven workflows.

This role is suited for a self-driven engineer who can operate in a fast-moving architecture.

 

Key Responsibilities

AI / ML

  • Develop and train ML models for market signals, behavioral patterns, user risk segmentation, anomaly detection.
  • Implement embedding pipelines, vector search and semantic analysis using:
    AWS Bedrock (Titan, Claude), SageMaker, LangChain, FAISS, OpenSearch, or local pipelines.
  • Build LLM-based agents using LangGraph, LangChain, AWS Bedrock Agents, or custom orchestration.
  • Work with HuggingFace, PyTorch, scikit-learn, Transformers, Nomic embeddings, etc.

 

 

Python Engineering

  • Design clean, modular services for data collection, processing, analytics and agentic workflows.
  • Build real-time pipelines using:
    asyncio, WebSockets, FastAPI, Redis Streams, Kafka, Celery, Apache Beam (optional).
  • Implement microservices interacting with internal APIs, AWS services and data layers.
  • Write production-quality Python (3.10+) with Pydantic, SQLAlchemy, Poetry/pipenv, type checking (mypy), and tests (pytest).

 

Data Engineering

  • Create ETL/ELT pipelines aggregating both on-chain and off-chain datasets using:
    AWS Glue, AWS Lambda, Step Functions, Athena, S3, DynamoDB Streams, Kinesis.
  • Optimize storage and data access: PostgreSQL, DynamoDB, Redis, S3, OpenSearch.
  • Implement observability and monitoring: CloudWatch Logs, Metrics, X-Ray, OpenTelemetry.

 

DevOps / Cloud (nice to have)

  • Experience with AWS:
    • Lambda (Python runtime)
    • ECS Fargate
    • Bedrock (LLMs, embeddings, agents)
    • SageMaker (model training & deployment)
    • SQS, SNS, EventBridge
    • API Gateway
    • OpenSearch
    • Neptune (graph DB)
    • KMS, IAM best practices
  • Build and monitor ML services in production using:
    SageMaker endpoints, CI/CD, Docker, Terraform, GitLab CI.

 

Requirements

Must-Have

  • 5+ years of Python engineering experience.
  • Strong background in AI/ML, especially NLP and agent-based architectures.
  • Experience with LLMs, embeddings, RAG, and vector DBs (FAISS, OpenSearch, Pinecone).
  • Strong understanding of async Python and distributed systems.
  • Experience with data pipelines (ETL/ELT), real-time event-driven processing.
  • Ability to work independently and architect solutions end-to-end.
  • Familiarity with AWS cloud services (at least S3, Lambda, API Gateway, CloudWatch).

 

Nice-to-Have

  • Experience with blockchain (EVM, Polygon, oracles).
  • Experience with AWS SageMaker training pipelines.
  • Understanding of smart-contract-driven workflows.
  • Experience with graph analytics: Neo4j, AWS Neptune, RDF/Gremlin.
  • Basic Solidity understanding.
  • Experience with agent frameworks such as LangGraph.

 

Required skills experience

Python 5 years
AWS Sagemaker 3 years
LangChain 3 years
Open Search 3 years
FastAPI 4 years

Required languages

English C1 - Advanced
AWS Bedrock, FAISS, WebSockets, Kafka, Celery
Published 4 May 2020 · Updated 9 December
Statistics:
27 views
·
3 applications
67% read
·
34% responded
Last responded yesterday
To apply for this and other jobs on Djinni login or signup.
Loading...