We are looking for a Senior DevOps Engineer to own the infrastructure for our new Medallion Lakehouse and AI-powered applications. You will not be maintaining legacy on-prem servers; you will be building the cloud-native foundation that drives our growth.
Tech Stack
• Cloud: Azure PaaS (AKS, Container Apps, Functions, Key Vault, ADLS Gen2, Azure SQL).
• IaC: Terraform (Module design, state management).
• CI/CD: GitHub Actions (Automated pipelines, quality gates).
• Orchestration: Kubernetes (AKS), Docker.
• Observability: Prometheus, Grafana, Azure Monitor.
• Scripting: Python, Bash.
Responsibilities
• Infrastructure as Code: Design and maintain scalable Azure infrastructure using Terraform. Implement module-based architectures for our Data Lakehouse (Bronze/Silver/Gold layers) and AI microservices.
• CI/CD Mastery: Build bulletproof GitHub Actions pipelines for Python applications and data workflows. Implement automated testing, security scanning, and seamless deployment strategies.
• Reliability & Observability: Implement monitoring dashboards (Grafana/Azure Monitor) and alerting strategies. Define and track SLIs/SLOs to distinguish between application and infrastructure noise.
• Security & Compliance: Enforce "Zero Trust" and "Least Privilege" principles. Manage identity (Azure AD/Entra ID), secrets (Key Vault), and network security (VNet, Private Endpoints).
• Disaster Recovery: Design and test backup/recovery strategies for databases and data lakes to ensure business continuity.
• Collaboration: Act as a bridge between Data Engineering and Backend teams, helping optimize Docker containers and troubleshoot production issues.
Requirements
• 5+ years of DevOps/Platform Engineering experience.
• Deep expertise in Azure ecosystems (proven track record of delivering Azure-based systems).
• Strong proficiency with Terraform in a production environment.
• Experience managing Kubernetes (AKS) and containerized workloads.
• Solid understanding of networking (VNets, DNS, Firewalls) and security (RBAC, IAM).
• English B2 – We communicate daily with the US team.
Nice to Have
• Experience supporting Data Engineering teams (Data Factory, Databricks, Spark).
• Background in Python development or complex scripting.
• Experience with GitOps workflows (ArgoCD).
What We Offer
• Greenfield Project: No legacy code or technical debt. You choose the right tools for the job.
• High Autonomy: You are the infrastructure owner. Your architecture decisions matter.
• Remote-First: Flexible schedule
• Stability: Long-term role with a stable US company
• Direct Impact: Your work directly powers AI models predicting roof damage and optimizing multimillion-dollar construction projects.
About the Project
TeamCraft is a large U.S. commercial roofing company undergoing an ambitious AI transformation. We’re building a centralized data platform from scratch - a unified Azure Lakehouse that integrates multiple operational systems into a single source of truth.
This is greenfield development with real business outcomes - not legacy maintenance.