Data Engineer Offline
We are Boosta — an international IT company with a portfolio of successful products, performance marketing projects, and our investment fund, Burner. Boosta was founded in 2014, and since then, the number of Boosters has grown to 600+.
We’re looking for a Data Engineer to join our team in the iGaming industry, where real-time insights, affiliate performance, and marketing analytics are at the center of decision-making. In this role, you’ll own and scale our data infrastructure, working across affiliate integrations, product analytics, and experimentation workflows.
Your primary responsibilities will include building and maintaining data pipelines, implementing automated data validation, integrating external data sources via APIs, and creating dashboards to monitor data quality, consistency, and reliability. You’ll collaborate daily with the Affiliate Management team, Product Analysts, and Data Scientists to ensure the data powering our reports and models is clean, consistent, and trustworthy.
WHAT YOU’LL DO
- Design, develop, and maintain ETL/ELT pipelines to transform raw, multi-source data into clean, analytics-ready tables in Google BigQuery, using tools such as dbt for modular SQL transformations, testing, and documentation.
- Integrate and automate affiliate data workflows, replacing manual processes in collaboration with the related stakeholders.
- Proactively monitor and manage data pipelines using tools such as Airflow, Prefect, or Dagster, with proper alerting and retry mechanisms in place.
- Emphasize data quality, consistency, and reliability by implementing robust validation checks, including schema drift detection, null/missing value tracking, and duplicate detection using tools like Great Expectations or
- Build a Data Consistency Dashboard (in Looker Studio, Power BI, Tableau or Grafana) to track schema mismatches, partner anomalies, and source freshness, with built-in alerts and escalation logic.
- Ensure timely availability and freshness of all critical datasets, resolving latency and reliability issues quickly and sustainably.
- Control access to cloud resources, implement data governance policies, and ensure secure, structured access across internal teams.
- Monitor and optimize data infrastructure costs, particularly related to BigQuery usage, storage, and API-based ingestion.
- Document all pipelines, dataset structures, transformation logic, and data contracts clearly to support internal alignment and knowledge sharing.
- Build and maintain postback-based ingestion pipelines to support event-level tracking and attribution across the affiliate ecosystem.
- Collaborate closely with Data Scientists and Product Analysts to deliver high-quality, structured datasets for modeling, experimentation, and KPI reporting.
Act as a go-to resource across the organization for troubleshooting data discrepancies, supporting analytics workflows, and enabling self-service data access.
WHAT WE EXPECT FROM YOU
- Strong proficiency in SQL and Python.
- Experience with Google BigQuery and other GCP tools (e.g., Cloud Storage, Cloud Functions, Composer).
- Proven ability to design, deploy, and scale ETL/ELT pipelines.
- Hands-on experience integrating and automating data from various platforms.
- Familiarity with postback tracking, attribution logic, and affiliate data reconciliation.
- Skilled in orchestration tools like Airflow, Prefect, or Dagster.
- Experience with Looker Studio, Power BI, Tableau, or Grafana for building dashboards for data quality monitoring.
- Use of Git for version control and experience managing CI/CD pipelines (e.g., GitHub Actions).
- Experience with Docker to build isolated and reproducible environments for data workflows.
- Exposure to iGaming data structures and KPIs is a strong advantage.
Strong sense of data ownership, documentation, and operational excellence.
HOW IT WORKS
- Stage 1: pre-screen with a recruiter.
- Stage 2: test task.
- Stage 3: interview.
- Stage 4: bar-raising.
- Stage 5: reference check.
- Stage 6: job offer!
A trial period for this position is 3 months, during which we will get used to working together.
WHAT WE OFFER
- 28 business days of paid off
- Flexible hours and the possibility to work remotely
- Medical insurance and mental health care
- Compensation for courses, trainings
- English classes and speaking clubs
- Internal library, educational events
- Outstanding corporate parties, teambuildings
The job ad is no longer active
Look at the current jobs Data Engineer →