Data Engineer โ Azure DWH / Databricks
Role Summary
We are seeking a Data Engineer to join our team and assist in building a modern Data Warehouse (DWH) solution on Azure.
You will play a key role in designing and developing data pipelines, building and maintaining the DWH layer, and supporting colleagues in creating dashboards, AI-based clients, and visualizations that drive business insights.
About the Projects
We help build and evolve a modern Azure data platform (lakehouse) for several product and service initiatives in healthcare and B2B SaaS.
Core stack: ADLS, Databricks (PySpark), Delta Lake, Azure Synapse (SQL), Azure Data Factory, Power BI.
Primary sources include Microsoft Dynamics 365, Salesforce, and operational MS SQL systems. Work happens in short delivery cycles (1โ3 months), often across 1โ2 concurrent projects with clear SLAs and strong collaboration between Data Engineering, BI, and Product teams.
Key Responsibilities
- Design, build, and maintain scalable data pipelines in Azure using PySpark, Pandas, and T-SQL.
- Develop and optimize ETL/ELT processes to integrate multiple data sources (Microsoft D365, Salesforce, MS SQL databases, etc.).
- Implement and maintain an Azure-based DWH architecture to ensure reliability, performance, and scalability.
- Ensure data quality, consistency, and integrity across all stages of the data lifecycle.
- Collaborate with business analysts and BI developers to support reporting, dashboards, and visualization needs (e.g., Power BI).
- Support the team in building AI-based applications/clients that consume DWH data.
- Monitor and troubleshoot data processes to ensure timely and accurate data delivery.
Contribute to best practices, coding standards, and documentation for the data platform.
Role Requirements
- 3+ years of experience in data engineering or related roles.
- Strong proficiency in Python and SQL (including T-SQL).
- Hands-on experience with Pandas and PySpark for large-scale data transformations.
- Proven experience with Azure cloud services (Data Lake, Synapse, Data Factory, Databricks) or other major cloud providers (AWS, GCP).
- Experience working with relational databases (especially MS SQL) and integrating CRM/ERP systems (e.g., Dynamics 365, Salesforce).
- Solid understanding of data warehousing concepts, including dimensional modeling and performance optimization.
- Knowledge or understanding of columnstore and MPP systems (e.g., Azure Synapse, Snowflake, BigQuery, Redshift).
Good communication skills and the ability to work collaboratively in a team environment.
Nice to Have
- Experience with data visualization tools (Power BI, Tableau, or similar).
- Familiarity with CI/CD pipelines and DevOps practices for data solutions.
- Knowledge of data governance, security, and compliance in cloud environments.
- Experience with API integrations and handling semi-structured/structured data (JSON, XML, Parquet).
Background in Agile/Scrum development methodologies.
Selection Process (transparent & practical)
Stage 1 โ Intro + Talent Acquisition (TA) + Tech Screen (40โ60 min, Zoom):
A call with TA to confirm project expectations (multi-project mode, 1โ3-month cycles), discuss your experience with Azure/ELT, and walk through a brief SQL/Python scenario.
Goal: align on must-haves and context-switching readiness.
Stage 2 โ Deep-Dive Technical Interview (75โ90 min, 2 engineers):
Live SQL challenge (CTE, window functions, SCD2, incremental load plan) + PySpark mini-exercises (joins, aggregations, skew/broadcast, UDF vs built-ins) + architecture discussion (ADLS/ADF/Databricks/Synapse, Delta/Medallion, CDC from D365/Salesforce). No take-home task โ we simulate day-to-day work in the session.
Why this order matters: we quickly validate the core skills that drive success in the role (SQL, PySpark, Azure) and your ability to deliver in a multi-stream setup; two interviewers reduce bias and provide a balanced assessment; live exercises reflect real tasks and collaboration style.
What We Offer
โ Competitive salary based on experience and skills.
โ Ability to work remotely or from abroad (gig contract; over time, if necessary, conversion to payroll employment is possible).
โ All the needed equipment & help with home-office setup (if you prefer working from home).
โ 20 days of vacation, respecting UA holidays.
โ Unlimited sick leave (no need to provide a certificate).
โ Lawyer, accountant, negotiation expert for both work and personal occasions.
โ Public speaking coach and designer at your disposal if you want to present at events or publish.
โ Paid sabbatical after 5 years in the company.
The team is made up of colleagues and friends who share the same values and passion for what they are doing.
We invite new teammates who have a similar mindset and professional views.
Required skills experience
DWH | 3 years |
ETL/ELT | 3 years |
Python | 3.5 years |
SQL | 3 years |
Azure DevOps | 1 year |
Databricks | 3 years |
PySpark | 3 years |
GenAI | 2 years |
Required domain experience
Machine Learning / Big Data | 2.5 years |
SaaS | 1.5 years |
GovTech | 1 year |
Required languages
English | B2 - Upper Intermediate |