Jobs Data Engineer

147
  • Β· 80 views Β· 1 application Β· 15d

    Data Engineer

    Countries of Europe or Ukraine Β· Product Β· 5 years of experience Β· English - B2
    We’re looking for a highly skilled Data Expert! Product | Remote ​​We’re looking for a data expert who bridges technical depth with curiosity. You’ll help Redocly turn data into insight β€” driving smarter product, growth, and business decisions. ...

     

    πŸ”₯ We’re looking for a highly skilled Data Expert!πŸ”₯

     

    Product | Remote

     

    ​​We’re looking for a data expert who bridges technical depth with curiosity. You’ll help Redocly turn data into insight β€” driving smarter product, growth, and business decisions.

     

    This role combines data governance and development. You’ll build reliable data pipelines, improve observability, and uncover meaningful patterns that guide how we grow and evolve.

     

    You’ll work closely with product and technical teams to support data collection, processing, and consistency across systems.

     

    What you’ll do 

    • Analyze product and user behavior to uncover trends, bottlenecks, and opportunities.
    • Build and maintain data pipelines and ETL processes.
    • Design and optimize data models for new features and analytics (e.g., using dbt).
    • Work with event-driven architectures and standards like AsyncAPI and CloudEvents.
    • Collaborate with engineers to improve data quality, consistency, and governance across systems.
    • Use observability and tracing tools (e.g., OpenTelemetry) to monitor and improve performance.
    • Support existing frontend and backend systems related to analytics and data processing.
    • Build and maintain datasets for analytics and reporting.

     

    You’re a great fit if you have 

    • 5+ years of software engineering experience, with 3+ years focused on data engineering.
    • Strong SQL skills and experience with data modeling (dbt preferred).
    • Strong proficiency with Node.js, React, JavaScript, and TypeScript.
    • Proven experience in data governance and backend systems.
    • Familiarity with columnar databases or analytics engines (ClickHouse, Postgres, etc.).
    • Strong analytical mindset, attention to detail, and clear communication.
    • Passionate about clarity, simplicity, and quality in both data and code.
    • English proficiency: Upper-Intermediate or higher.

     

    How you’ll know you’re doing a great job

    • Data pipelines are trusted, observable, and performant.
    • Metrics and dashboards are used across teams β€” not just built once.
    • Teams make better product decisions, faster, because of your insights.
    • Data pipelines are trusted, observable, and performant.
    • You’re the go-to person for clarity when questions arise about β€œwhat the data says.”

     

    About Redocly

    Redocly builds tools that accelerate API ubiquity. Our platform helps teams create world-class developer experiences β€” from API documentation and catalogs to internal developer hubs and public showcases. We're a globally distributed team that values clarity, autonomy, and craftsmanship. You'll work alongside people who love developer experience, storytelling, and building tools that make technical work simpler and more joyful.

    Headquarter – Austin, Texas, US. There is also an office in Lviv, Ukraine.

     

    Redocly is trusted by leading tech, fintech, telecom, and enterprise teams to power API documentation and developer portals. Redocly’s clients range from startups to Fortune 500 enterprises.

    https://redocly.com/

     

    Working with Redocly

    • Team: 4-6 people (middle-seniors)
    • Team’s location: Ukraine&Europe
    • There are functional, product, and platform teams and each has its own ownership, and line structure, and teams themselves decide when to have weekly meetings.
    • Cross-functional teams are formed for each two-month cycle, giving team members the opportunity to work across all parts of the product.
    • Methodology: Shape Up

     

    Perks

    • Competitive salary based on your expertise 
    • Full remote, though you’re welcome to come to the office occasionally if you wish.
    • Cooperation on a B2B basis with a US-based company (for EU citizens) or under a gig contract (for Ukraine).
    • After a year of working with the company, you can buy a certain number of company’s shares
    • Around 30 days of vacation (unlimited,  but let’s keep it reasonable)
    • 10 working days of sick leave per year
    • Public holidays according to the standards
    • No trackers and screen recorders
    • Working hours – EU/UA timezone. Working day – 8 hours. Mostly they start working from 10-11 am
    • Equipment provided – MacBooks (M1 – M4)
    • Regular performance reviews

     

    Hiring Stages

    • Prescreening (30-45 min)
    • HR Call (45 min)
    • Initial Interview (30 min)
    • Trial Day (paid)
    • Offer

     

    If you are an experienced Data Scientist, and you want to work on impactful data-driven projects, we’d love to hear from you! 


    Apply now to join our team!

    More
  • Β· 59 views Β· 6 applications Β· 15d

    Lead Data Engineer

    Full Remote Β· Countries of Europe or Ukraine Β· 7 years of experience Β· English - B2
    We are seeking a highly skilled Lead Data Engineer to design, develop, and optimize our Data Warehouse solutions. The ideal candidate will have extensive experience in ETL/ELT development, data modeling, and big data technologies, ensuring efficient data...

    We are seeking a highly skilled Lead Data Engineer to design, develop, and optimize our Data Warehouse solutions. The ideal candidate will have extensive experience in ETL/ELT development, data modeling, and big data technologies, ensuring efficient data processing and analytics. This role requires strong collaboration with Data Analysts, Data Scientists, and Business Stakeholders to drive data-driven decision-making.

     

    Does this relate to you?

    • 7+ years of experience in Data Engineering field
    • At least 1+ year of experience as Lead\Architect
    • Strong expertise in SQL and data modeling concepts.
    • Hands-on experience with Airflow.
    • Experience working with Redshift.
    • Proficiency in Python for data processing.
    • Strong understanding of data governance, security, and compliance.
    • Experience in implementing CI/CD pipelines for data workflows.
    • Ability to work independently and collaboratively in an agile environment.
    • Excellent problem-solving and analytical skills.

       

    A new team member will be in charge of:

    • Design, develop, and maintain scalable data warehouse solutions.
    • Build and optimize ETL/ELT pipelines for efficient data integration.
    • Design and implement data models to support analytical and reporting needs.
    • Ensure data integrity, quality, and security across all pipelines.
    • Optimize data performance and scalability using best practices.
    • Work with big data technologies such as Redshift.
    • Collaborate with cross-functional teams to understand business requirements and translate them into data solutions.
    • Implement CI/CD pipelines for data workflows.
    • Monitor, troubleshoot, and improve data processes and system performance.
    • Stay updated with industry trends and emerging technologies in data engineering.

       

    Already looks interesting? Awesome! Check out the benefits prepared for you:

    • Regular performance reviews, including remuneration
    • Up to 25 paid days off per year for well-being
    • Flexible cooperation hours with work-from-home
    • Fully paid English classes with an in-house teacher
    • Perks on special occasions such as birthdays, marriage, childbirth
    • Referral program implying attractive bonuses
    • External & internal training and IT certifications
    More
  • Β· 26 views Β· 4 applications Β· 15d

    Senior Data Engineer

    Full Remote Β· EU Β· 3 years of experience Β· English - B2
    We are looking for an experienced Data Engineer to join a long-term B2C project. The main focus is on building Zero ETL pipelines, as well as maintaining and improving existing ones. Responsibilities: - Build and maintain scalable Zero ETL pipelines. -...

    We are looking for an experienced Data Engineer to join a long-term B2C project. The main focus is on building Zero ETL pipelines, as well as maintaining and improving existing ones.

    Responsibilities:
    - Build and maintain scalable Zero ETL pipelines.
    - Design and optimize data warehouses and data lakes on AWS (Glue, Firehose, Lambda, SageMaker).
    - Work with structured and unstructured data, ensuring quality and accuracy.
    - Optimize query performance and data processing workflows (Spark, SQL, Python).
    - Collaborate with engineers, analysts, and business stakeholders to deliver data-driven solutions.

    Requirements:
    - 5+ years of experience in Data Engineering.
    - Advanced proficiency in Spark, Python, SQL.
    - Expertise with AWS Glue, Firehose, Lambda, SageMaker.
    - Experience with ETL tools (dbt, Airflow etc.).
    - Background in B2C companies is preferred.
    - JavaScript and Data Science knowledge are a plus.
    - Degree in Computer Science (preferred, not mandatory).

    We offer:
    - remote time job, B2B contract
    - 12 sick leaves and 18 paid vacation business days per year
    - Comfortable work conditions (including MacBook Pro and Dell monitor on each workplace)
    - Smart environment
    - Interesting projects from renowned clients
    - Flexible work schedule
    - Competitive salary according to the qualifications
    - Guaranteed full workload during the term of the contract
     

    More
  • Β· 42 views Β· 4 applications Β· 15d

    Senior Data Engineer (Data Competency Center)

    Full Remote Β· Countries of Europe or Ukraine Β· 5 years of experience Β· English - B2
    Are you a Senior Data Engineer passionate about building scalable, secure, and high-performance data solutions? Join our Data Engineering Center of Excellence at Sigma Software and work on diverse projects that challenge your skills and inspire...

    Are you a Senior Data Engineer passionate about building scalable, secure, and high-performance data solutions? Join our Data Engineering Center of Excellence at Sigma Software and work on diverse projects that challenge your skills and inspire innovation.

     

    At Sigma Software, we value expertise, continuous learning, and a supportive environment where your career path is shaped around your strengths. You’ll be part of a collaborative team, gain exposure to cutting-edge technologies, and work in an inclusive culture that fosters growth and innovation.

    Project

    Our Data Engineering Center of Excellence (CoE) is a specialized unit focused on designing, building, and optimizing data platforms, pipelines, and architectures. We work across diverse industries, leveraging modern data stacks to deliver scalable, secure, and cost-efficient solutions.

    Job Description

    • Collaborate with clients and internal teams to clarify technical requirements and expectations
    • Implement architectures using Azure or AWS cloud platforms
    • Design, develop, optimize, and maintain squad-specific data architectures and pipelines
    • Discover, analyze, and organize disparate data sources into clean, understandable data models
    • Evaluate new tools for analytical data engineering or data science
    • Suggest and contribute to training and improvement plans for analytical data engineering skills, standards, and processes

    Qualifications

    • 5+ years of experience with Python and SQL
    • Hands-on experience with AWS services (API Gateway, Kinesis, Athena, RDS, Aurora)
    • Proven experience building ETL pipelines for analytics/internal operations
    • Experience developing and integrating APIs
    • Solid understanding of Linux OS
    • Familiarity with distributed applications and DevOps tools
    • Strong troubleshooting/debugging skills
    • English level: Upper-Intermediate
    • WILL BE A PLUS:
    • 2+ years with Hadoop, Spark, or Airflow
    • Experience with DAGs/orchestration tools
    • Experience with Snowflake-based data warehouses
    • Experience developing event-driven data pipelines
    • Personal Profile

    PERSONAL PROFILE:

    • Passion for data processing and continuous learning
    • Strong problem-solving skills and analytical thinking
    • Ability to mentor and guide team members
    • Effective communication and collaboration skills
    More
  • Β· 40 views Β· 1 application Β· 16d

    Data Engineer (with Azure)

    Full Remote Β· EU Β· 3 years of experience Β· English - B1
    Main Responsibilities: Data Engineer is responsible for helping select, deploy, and manage the systems and infrastructure required of a data processing pipeline to support customer requirements. You will work on cutting-edge cloud technologies,...

    Main Responsibilities:

    Data Engineer is responsible for helping select, deploy, and manage the systems and infrastructure required of a data processing pipeline to support customer requirements.

     

    You will work on cutting-edge cloud technologies, including Microsoft Fabric, Azure Synapse Analytics, Apache Spark, Data Lake, Data Bricks, Data Factory, Cosmos DB, HD Insights, Stream Analytics, Event Grid in the implementation projects for corporate clients all over EU, CIS, United Kingdom, Middle East.

    Our ideal candidate is a professional passionated with technologies, a curious and self-motivated person.

     

    Responsibilities revolve around DevOps and include implementing ETL pipelines, monitoring/maintaining data pipeline performance, model optimization

     

    Mandatory Requirements:

    – 3+ years of experience, ideally within a Data Engineer role.

    – understanding of data modeling, data warehousing concepts, and ETL processes

    – 2+ years of experience with Azure Cloud technologies

    – experience in distributed computing principles and familiarity with key architectures, broad experience across a set of data stores (Azure Data Lake Store, Azure Synapse Analytics, Apache Spark, Azure Data Factory)

    – Understanding of landing, staging area, data cleansing, data profiling, data security and data architecture concepts (DWH, Data Lake, Delta Lake/Lakehouse, Datamart)

    – SQL-skills

    – communication and interpersonal skills

    – English β€”Π’2

     

    Will be beneficial if a candidate has experience in SQL migration from on-premises to cloud, data modernization and migration, advanced analytics projects, and/or professional certification in data&analytics.

     

    We offer:

    – professional growth and international certification

    – free of charge technical and business trainings and the best bootcamps (worldwide, including HQ Microsoft- Redmond courses)

    – innovative data & analytics projects, practical experience with cutting-edge Azure data&analytics technologies at various customers’ projects

    – great compensation and individual bonus remuneration

    – medical insurance

    – long-term employment

    – ondividual development plan

    More
  • Β· 34 views Β· 0 applications Β· 16d

    Senior Data Engineer

    Ukraine Β· 4 years of experience Β· English - B2
    We are a global audience and location intelligence company that helps marketers connect the digital and physical world. We provide data-driven solutions to enhance marketing campaigns by leveraging location and audience data to reveal consumer behavior...

    We are a global audience and location intelligence company that helps marketers connect the digital and physical world. We provide data-driven solutions to enhance marketing campaigns by leveraging location and audience data to reveal consumer behavior and enable more precise targeting and measurement. We work on high-end / high-performance / high-throughput systems for in-time analysis of data for autonomous driving and other big data applications e.g. for E-commerce.


    Job Description

    You have 4+ years of experience on similar position.

    You have significant experience with Python. Familiarity with Java or Scala is a plus.

    Hands-on experience building scalable solutions in AWS.

    Proficiency in NoSQL and SQL databases and in high-throughput data-related architecture and technologies (e.g. Kafka, Spark, Hadoop, MongoDB, AWS Batch, AWS Glue, Athena, Airflow, dbt).

    Excellent SQL and data transformation skills.

    Excellent written and verbal communication skills with an ability to simplify complex technical information.

    Experience guiding and mentoring junior team members in a collaborative environment.


     

    Job Responsibilities

    Work in a self-organised agile team with a high level of autonomy, and you will actively shape your team's culture.

    Design, build, and standardise privacy-first big data architectures, large-scale data pipelines, and advanced analytics solutions in AWS.

    Develop complex integrations with third-party partners, transferring terabytes of data.

    Align with other Data experts on data (analytics) engineering best practices and standards, and introduce those standards and data engineering expertise to the team in order to enhance existing data pipelines and build new ones.

    Successfully partner up with the Product team to constantly develop further and improve our platform features.

    More
  • Β· 46 views Β· 7 applications Β· 16d

    Senior Data Engineer

    Full Remote Β· Countries of Europe or Ukraine Β· 4 years of experience Β· English - B1
    GlobalLogic is searching for a motivated, results-driven, and innovative software engineer to join our project team at a dynamic startup specializing in pet insurance. Our client is a leading global holding company that is dedicated to developing an...

    GlobalLogic is searching for a motivated, results-driven, and innovative software engineer to join our project team at a dynamic startup specializing in pet insurance. Our client is a leading global holding company that is dedicated to developing an advanced pet insurance claims clearing solution designed to expedite and simplify the veterinary invoice reimbursement process for pet owners.
    You will be working on a cutting-edge system built from scratch, leveraging Azure cloud services and adopting a low-code paradigm. The project adheres to industry best practices in quality assurance and project management, aiming to deliver exceptional results.
    We are looking for an engineer who thrives in collaborative, supportive environments and is passionate about making a meaningful impact on people’s lives. If you are enthusiastic about building innovative solutions and contributing to a cause that matters, this role could be an excellent fit for you.



    Requirements

    • Strong hands-on experience with Azure Databricks (DLT Pipelines, Lakeflow Connect, Delta Live Tables, Unity Catalog, Time Travel, Delta Share) for large-scale data processing and analytics
    • Proficiency in data engineering with Apache Spark, using PySpark, Scala, or Java for data ingestion, transformation, and processing
    • Proven expertise in the Azure data ecosystem: Databricks, ADLS Gen2, Azure SQL, Azure Blob Storage, Azure Key Vault, Azure Service Bus/Event Hub, Azure Functions, Azure Data Factory, and Azure CosmosDB
    • Solid understanding of Lakehouse architecture, Modern Data Warehousing, and Delta Lake concepts
    • Experience designing and maintaining config-driven ETL/ELT pipelines with support for Change Data Capture (CDC) and event/stream-based processing
    • Proficiency with RDBMS (MS SQL, MySQL, PostgreSQL) and NoSQL databases
    • Strong understanding of data modeling, schema design, and database performance optimization
    • Practical experience working with various file formats, including JSON, Parquet, and ORC
    • Familiarity with machine learning and AI integration within the data platform context
    • Hands-on experience building and maintaining CI/CD pipelines (Azure DevOps, GitLab) and automating data workflow deployments
    • Solid understanding of data governance, lineage, and cloud security (Unity Catalog, encryption, access control)
    • Strong analytical and problem-solving skills with attention to detail
    • Excellent teamwork and communication skills
    • Upper-Intermediate English (spoken and written)

    Job responsibilities

    • Design, implement, and optimize scalable and reliable data pipelines using Databricks, Spark, and Azure data services
    • Develop and maintain config-driven ETL/ELT solutions for both batch and streaming data
    • Ensure data governance, lineage, and compliance using Unity Catalog and Azure Key Vault
    • Work with Delta tables, Delta Lake, and Lakehouse architecture to ensure efficient, reliable, and performant data processing
    • Collaborate with developers, analysts, and data scientists to deliver trusted datasets for reporting, analytics, and machine learning use cases
    • Integrate data pipelines with event-based and microservice architectures leveraging Service Bus, Event Hub, and Functions
    • Design and maintain data models and schemas optimized for analytical and operational workloads
    • Identify and resolve performance bottlenecks, ensuring cost efficiency and maintainability of data workflows
    • Participate in architecture discussions, backlog refinement, estimation, and sprint planning
    • Contribute to defining and maintaining best practices, coding standards, and quality guidelines for data engineering
    • Perform code reviews, provide technical mentorship, and foster knowledge sharing within the team
    • Continuously evaluate and enhance data engineering tools, frameworks, and processes in the Azure environment
    More
  • Β· 44 views Β· 3 applications Β· 16d

    Data Engineer ( with Snowflake and insurance companies experience)

    Full Remote Β· Countries of Europe or Ukraine Β· 5 years of experience Β· English - B2
    Senior Data Engineer with experience in insurance & Snowflake. On behalf of our Client from the USA, Mobilunity is looking for a Senior Data Engineer. Our client is a well-established US-based financial services organization with a long history in the...

    Senior Data Engineer with experience in insurance & Snowflake.
    On behalf of our Client from the USA, Mobilunity is looking for a Senior Data Engineer.
    Our client is a well-established US-based financial services organization with a long history in
    the insurance and long-term financial security space. The company operates as a member-oriented, non-profit institution, focusing on life insurance, retirement programs, and community-driven initiatives. The product ecosystem is mature, data-heavy, and highly
    regulated, with a strong emphasis on reliability, accuracy, and compliance.
    We are looking for a Senior Data Engineer with deep Snowflake expertise and hands-on experience in insurance companies. 

    This domain background is a must-have requirement.
     

    Requirements:
    - 5+ years of experience as a Data Engineer.
    - Strong hands-on experience with Snowflake (data modeling, optimization, ELT/ETL pipelines).
    - Mandatory experience in the insurance domain (policies, claims, underwriting, actuarial, or related datasets).
    - Solid experience with data warehousing and analytical platforms.
    - Proven ability to build and maintain scalable, reliable data pipelines.
    - Advanced SQL skills.
    - Experience working with cloud platforms (AWS / GCP / Azure).
    - Upper-Intermediate+ / Advanced English – direct communication with US stakeholders.
     

    Nice to have:
    - Experience with dbt, Airflow, or similar orchestration tools.
    - Background in regulated or compliance-heavy environments.
    - Previous experience working with US clients or distributed teams.
     

    In return we offer
    The friendliest community of like-minded IT-people.
    Open knowledge-sharing environment – exclusive access to a rich pool of colleagues willing to share their endless insights into the broadest variety of modern technologies.


    Languages
    English - B2-Π‘1

    More
  • Β· 64 views Β· 6 applications Β· 17d

    Senior Data Engineer (for Ukrainians in EU)

    Full Remote Β· Countries of Europe or Ukraine Β· Product Β· 6 years of experience Β· English - B2
    About our Customer It's a European company turning bold ideas into reality. We build innovative products for startups and guide established companies on their journey to data-driven innovation and AI-powered solutions. Our expertise spans EnergyTech,...

    About our Customer
    It's a European company turning bold ideas into reality.  We build innovative products for startups and guide established companies on their journey to data-driven innovation and AI-powered solutions. Our expertise spans EnergyTech, FinTech, ClimateTech, SocialTech, PropTech , and more.
     

    Founded in Ukraine with a Scandinavian-inspired culture.
     

    We value skills, passion, excellence, equality, openness, mutual respect, and trust. You’ll join a growing company, work with creative, inspiring colleagues, explore cutting-edge technologies, and build AI-driven solutions that make a real impact.
     

    Project
    Our client is an Icelandic energy company  providing electricity, geothermal water, cold water, carbon storage, and optic networks.
     

    We are looking for a Senior Data Engineer ready to dive deep into data, solve challenging problems, and create maximum value for internal stakeholders. You’ll handle complex issues, design long-term improvements, and develop new data pipelines as part of an enthusiastic and collaborative Data Engineering team.
     

    Tech Stack:
    πŸ—„οΈ MS SQL Server | Azure/Databricks | Power BI, Tableau | Microsoft BI stack (SSRS, SSIS, SSAS) | TimeXtender | exMon
     

    Responsibilities:

    • Develop & maintain enterprise data warehouse, data marts, staging layers, and transformation logic
    • Design, implement & optimize ETL/ELT pipelines (SQL Server, Azure, Databricks)
    • Build & maintain robust data models (dimensional/star-schema, semantic layers, analytical datasets)
    • Improve BI environment and ensure data is reliable and actionable
    • Implement controlled data delivery processes to analysts & BI specialists
    • Support data quality frameworks, testing & validation procedures
    • Investigate 3rd-line operational issues & guide 2nd-line support
    • Run stakeholder workshops to translate business needs into elegant technical solutions
    • Identify opportunities to improve data usability, value, and automation
    • Document all processes, models, and pipelines in Confluence
    • Collaborate with on-site Team Lead for sprint planning, backlog refinement, and prioritization
       

    Requirements

    • Bachelor’s or Master’s in Computer Science or related field
    • 6+ years of experience with DWH solutions & data pipelines
    • Strong SQL development skills (MS SQL Server preferred)
    • ETL/ELT workflow experience using:
      • Databricks
      • Azure Data Factory / cloud orchestration tools
      • Azure data platform services (storage, compute, data lake)
    • Solid understanding of data warehouse architectures & dimensional modeling
    • Experience with data quality checks, validation, and monitoring
    • Understanding of BI concepts & ability to prepare user-friendly datasets
    • Strong communication, able to explain data concepts to stakeholders
    • Willingness to document solutions and share knowledge
    • Experience in distributed, cross-cultural Agile environments
    • English: upper-intermediate / advanced


    πŸ”Ή Bonus / Nice to Have

    • Python or similar for data processing
    • Performance tuning for SQL or data pipelines
    • Interest in visual clarity & usability of data models
    More
  • Β· 43 views Β· 1 application Β· 17d

    Senior Data Engineer

    Full Remote Β· EU Β· 6 years of experience Β· English - B2
    OUR COMPANY HBM is a European company building exciting new products from scratch for startups and helping mature companies in their journey towards data-driven innovation and AI based solutions. Our expertise refers to EnergyTech, FinTech, ClimateTech,...

    OUR COMPANY  

    HBM is a European company building exciting new products from scratch for startups and helping mature companies in their journey towards data-driven innovation and AI based solutions. Our expertise refers to EnergyTech, FinTech, ClimateTech, SocialTech, PropTech, etc. 

    Founded in Ukraine and developed based on Scandinavian culture, HBM is hiring both in Ukraine and the EU for our customers located in Europe and USA.  

      

    Our values include skills, passion, excellence, equality, openness, mutual respect, and trust. 

      

    At HBM, you can become a part of growing company, work with creative colleagues, and enjoy modern technologies and creating AI-based solutions. You’ll be part of a strong corporate culture combined with the agility and flexibility of a start-up backed by proven outsourcing and development practices, a human-oriented leadership team, an entrepreneurial mindset, and an approach to work-life balance. 

      

    PROJECT 

    Our customer is an Icelandic energy company providing electricity, geothermal water, cold water, carbon storage and optic network.  

    We are looking for a Senior Data Engineer who will be responsible for developing, enhancing, and maintaining enterprise data warehouse, data platform, and analytical data flows. The role supports all company’s subsidiaries and contributes to creating maximum value from data for internal stakeholders. 

    The qualified candidate will work as part of the Data Engineering team and will handle complex 3rd-line issues, long-term improvements, and new data development. The work will be aligned with the team’s structured 3-week planning cycles, and tight collaboration with the on-site Team Lead is expected. 

    Tech stack: MS SQL Server, Azure/Databricks, Power BI, Tableau, Microsoft BI stack (SSRS, SSIS,SSAS [Olap and Tabular]) , TimeXtender, exMon. 

     

    WE PROVIDE YOU WITH THE FOLLOWING EXCITING CHALLENGES 

    • Develop and maintain the enterprise data warehouse, data marts, staging layers, and transformation logic 
    • Design, implement, and optimize ETL/ELT pipelines (SQL Server, Azure data components, Databricks, etc.) 
    • Build and maintain robust data models (dimensional/star-schema, semantic layers, analytical datasets) 
    • Develop and improve the BI environment and the underlying data processes used by analysts across the company 
    • Implement processes for controlled, reliable data delivery to BI specialists, analysts, and modelling teams (e.g., forecasting, scenario modelling) 
    • Support data quality frameworks and implement testing/validation procedures 
    • Investigate and resolve escalated 3rd-line operational issues and guide 2nd-line support in root cause analysis 
    • Conduct stakeholder workshops to understand business requirements and translate them into technical data solutions 
    • Identify opportunities to improve data usability, analytical value, and process automation 
    • Document data processes, models, pipelines, and architectural decisions in Confluence 
    • Collaborate with the on-site Team Lead during sprint planning, backlog refinement, and prioritization. 

     

      

    WE EXPECT FROM YOU 

    • Degree (bachelor or master) in computer science or a comparable course of study 
    • 6+ years of experience working with DWH solutions and data pipelines 
    • Strong SQL development skills, preferably in MS SQL Server 
    • Experience building and maintaining ETL/ELT workflows using: 
    • Databricks 
    • Azure Data Factory or similar cloud-based data orchestration tools 
    • Azure data platform services (e.g., storage, compute, data lake formats) 
    • Solid understanding of data warehouse architectures and dimensional modelling 
    • Experience with data quality checks, validation frameworks, and monitoring 
    • Understanding of BI concepts and ability to prepare user-friendly analytical datasets 
    • Experience collaborating with business stakeholders and capturing analytical or operational data requirements 
    • Strong communication skills and the ability to explain data concepts clearly 
    • Willingness to document solutions and share knowledge within the team 
    • Excellent communication skills, ability to communicate to stakeholders on multiple levels 
    • Action and quality-oriented 
    • Experience of work the distributed, cross-culture Agile environment 
    • English: upper-intermediate / advanced 

     

    WOULD BE A PLUS 

    • Experience with Python or similar languages for data processing 
    • Experience with performance tuning for SQL or data pipelines 
    • Interest in visual clarity, usability of data models, and BI-driven design 

     

     

     WE OFFER YOU 

      

    • Modern technologies, new products development, different business domains. 
    • Start-up agility combined with mature delivery practices and management team. 
    • Strong focus on your technical and personal growth. 
    • Transparent career development and individual development plan. 
    • Flexible working mode (remote/work from office), full remote possibility. 
    • Competitive compensation and social package 
    • Focus on the well-being and human touch. 
    • Flat organization where everyone is heard and is invited to contribute. 
    • Work-life balance approach to work. 
    • Passion and Fun in everything we do. 
    More
  • Β· 44 views Β· 5 applications Β· 17d

    Senior Data Platform Engineer

    Full Remote Β· Countries of Europe or Ukraine Β· Product Β· 7 years of experience Β· English - B2
    What You’ll Actually Do Architect and run high-load, production-grade data pipelines where correctness and latency matter. Design systems that survive schema changes, reprocessing, and partial failures. Own data availability, freshness, and trust - not...

    🎯 What You’ll Actually Do

    • Architect and run high-load, production-grade data pipelines where correctness and latency matter.
    • Design systems that survive schema changes, reprocessing, and partial failures.
    • Own data availability, freshness, and trust - not just pipeline success.
    • Make hard calls: accuracy vs cost, speed vs consistency, rebuild vs patch.
    • Build guardrails so downstream consumers (Analysts, Product, Ops) don’t break.
    • Improve observability: monitoring, alerts, data quality checks, SLAs.
    • Partner closely with backend engineers, data analysts, and Product - no handoffs, shared ownership.
    • Debug incidents, own RCA, and make sure the same class of failure doesn’t return.

    This is a hands-on IC role with platform-level responsibility.

     

    🧠 What You Bring

    • 5+ years in data or backend engineering on real production systems.
    • Strong experience with columnar analytical databases (ClickHouse, Snowflake, BigQuery, similar).
    • Experience with event-driven / streaming systems (Kafka, pub/sub, CDC, etc.).
    • Strong SQL + at least one general-purpose language (Python, Java, Scala).
    • You think in failure modes, not happy paths.
    • You explain why something works - and when it shouldn’t be used.

    Bonus: You’ve rebuilt or fixed a data system that failed in production.

     

    πŸ”§ How We Work

    • Reliability > elegance. Correct data beats clever data.
    • Ownership > tickets. You run what you build.
    • Trade-offs > dogma. Context matters.
    • Direct > polite. We fix problems, not dance around them.
    • One team, one system. No silos.
    •  

    πŸ”₯ What We Offer

    • Fully remote.
    • Unlimited vacation + paid sick leave.
    • Quarterly performance bonuses.
    • Medical insurance for you and your partner.
    • Learning budget (courses, conferences, certifications).
    • High trust, high autonomy.
    • Zero bureaucracy. Real engineering problems.

       

    πŸ‘‰ Apply if you see data platforms as systems to be engineered - not pipelines to babysit.

    More
  • Β· 43 views Β· 10 applications Β· 17d

    Senior Data Engineer

    Full Remote Β· Countries of Europe or Ukraine Β· Product Β· 7 years of experience Β· English - B2
    What You’ll Actually Do Design and run high-throughput, production-grade data pipelines. Own data correctness, latency, and availability end to end. Make hard trade-offs: accuracy vs speed, cost vs freshness, rebuild vs patch. Design for change - schema...

    🎯 What You’ll Actually Do

    • Design and run high-throughput, production-grade data pipelines.
    • Own data correctness, latency, and availability end to end.
    • Make hard trade-offs: accuracy vs speed, cost vs freshness, rebuild vs patch.
    • Design for change - schema evolution, reprocessing, and new consumers.
    • Protect BI, Product, and Ops from breaking changes and silent data issues.
    • Build monitoring, alerts, and data quality checks that catch problems early.
    • Work side-by-side with Product, BI, and Engineering β€” no handoffs, shared ownership.
    • Step into incidents, own RCA, and make sure the same class of failure never repeats.

    This is a hands-on senior IC role with real accountability.

     

     

    🧠 What You Bring (Non-Negotiable)

    • 5+ years in data or backend engineering on real production systems.
    • Strong experience with analytical databases
      (ClickHouse, Snowflake, BigQuery, or similar).
    • Experience with event-driven or streaming systems
      (Kafka, CDC, pub/sub).
    • Solid understanding of:
      • at-least-once vs exactly-once semantics
      • schema evolution & backfills
      • mutation and reprocessing costs
    • Strong SQL and at least one programming language
      (Python, Java, Scala, etc.).
    • You don’t just ship - you own what happens after.

       

    πŸ”§ How We Work

    • Reliability > cleverness.
    • Ownership > process.
    • Impact > output.
    • Direct > polite.
    • One team, one system.

       

    πŸ”₯ What We Offer

    • Fully remote (Europe).
    • Unlimited vacation + paid sick leave.
    • Quarterly performance bonuses.
    • Medical insurance for you and your partner.
    • Learning budget (courses, conferences, certifications).
    • High trust, high autonomy.
    • No bureaucracy. Real data problems.

       

    πŸ‘‰ Apply if you treat data like production software - and feel uncomfortable when numbers can’t be trusted.

    More
  • Β· 51 views Β· 9 applications Β· 17d

    Senior Data Engineer

    Full Remote Β· Countries of Europe or Ukraine Β· Product Β· 5 years of experience Β· English - None
    About us: Data Science UA is a service company with strong data science and AI expertise. Our journey began in 2016 with the organization of the first Data Science UA conference, setting the foundation for our growth. Over the past 9 years, we have...

    About us:
    Data Science UA is a service company with strong data science and AI expertise. Our journey began in 2016 with the organization of the first Data Science UA conference, setting the foundation for our growth. Over the past 9 years, we have diligently fostered the largest Data Science Community in Eastern Europe, boasting a network of over 30,000 AI top engineers.

    About the client:
    We are working with a new generation of data service provider, specializing in data consulting and data-driven digital marketing, dedicated to transforming data into business impact across the entire value chain of organizations. The company’s data-driven services are built upon the deep AI expertise the company’s acquired with a 1000+ client base around the globe. The company has 1000 employees across 20 offices who are focused on accelerating digital transformation.

    About the role:
    We are seeking a Senior Data Engineer (Azure) to design and maintain data pipelines and systems for analytics and AI-driven applications. You will work on building reliable ETL/ELT workflows and ensuring data integrity across the organization.

    Required skills:
    - 6+ years of experience as a Data Engineer, preferably in Azure environments.
    - Proficiency in Python, SQL, NoSQL, and Cypher for data manipulation and querying.
    - Hands-on experience with Airflow and Azure Data Services for pipeline orchestration.
    - Strong understanding of data modeling, ETL/ELT workflows, and data warehousing concepts.
    - Experience in implementing DataOps practices for pipeline automation and monitoring.
    - Knowledge of data governance, data security, and metadata management principles.
    - Ability to work collaboratively with data science and analytics teams.
    - Excellent problem-solving and communication skills.

    Responsibilities:
    - Transform data into formats suitable for analysis by developing and maintaining processes for data transformation;
    - Structuring, metadata management, and workload management.
    - Design, implement, and maintain scalable data pipelines on Azure.
    - Develop and optimize ETL/ELT processes for various data sources.
    - Collaborate with data scientists and analysts to ensure data readiness.
    - Monitor and improve data quality, performance, and governance.

    More
  • Β· 71 views Β· 4 applications Β· 17d

    Data Engineer

    Full Remote Β· Ukraine Β· Product Β· 3 years of experience Β· English - None
    About us: Data Science UA is a service company with strong data science and AI expertise. Our journey began in 2016 with uniting top AI talents and organizing the first Data Science tech conference in Kyiv. Over the past 9 years, we have diligently...

    About us:
    Data Science UA is a service company with strong data science and AI expertise. Our journey began in 2016 with uniting top AI talents and organizing the first Data Science tech conference in Kyiv. Over the past 9 years, we have diligently fostered one of the largest Data Science & AI communities in Europe.

    About the client:
    Our client is an IT company that develops technological solutions and products to help companies reach their full potential and meet the needs of their users. The team comprises over 600 specialists in IT and Digital, with solid expertise in various technology stacks necessary for creating complex solutions.

    About the role:
    We are looking for a Data Engineer (NLP-Focused) to build and optimize the data pipelines that fuel the Ukrainian LLM and NLP initiatives. In this role, you will design robust ETL/ELT processes to collect, process, and manage large-scale text and metadata, enabling the Data Scientists and ML Engineers to develop cutting-edge language models.

    You will work at the intersection of data engineering and machine learning, ensuring that the datasets and infrastructure are reliable, scalable, and tailored to the needs of training and evaluating NLP models in a Ukrainian language context.

    Requirements:
    - Education & Experience: 3+ years of experience as a Data Engineer or in a similar role, building data-intensive pipelines or platforms. A Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field is preferred. Experience supporting machine learning or analytics teams with data pipelines is a strong advantage.
    - NLP Domain Experience: Prior experience handling linguistic data or supporting NLP projects (e.g., text normalization, handling different encodings, tokenization strategies). Knowledge of Ukrainian text sources and data sets, or experience with multilingual data processing, can be an advantage given the project’s focus.
    Understanding of FineWeb2 or a similar processing pipeline approach.
    - Data Pipeline Expertise: Hands-on experience designing ETL/ELT processes, including extracting data from various sources, using transformation tools, and loading into storage systems. Proficiency with orchestration frameworks like Apache Airflow for scheduling workflows. Familiarity with building pipelines for unstructured data (text, logs) as well as structured data.
    - Programming & Scripting: Strong programming skills in Python for data manipulation and pipeline development. Experience with NLP packages (spaCy, NLTK, langdetect, fasttext, etc.). Experience with SQL for querying and transforming data in relational databases. Knowledge of Bash or other scripting for automation tasks. Writing clean, maintainable code and using version control (Git) for collaborative development.
    - Databases & Storage: Experience working with relational databases (e.g., PostgreSQL, MySQL), including schema design and query optimization. Familiarity with NoSQL or document stores (e.g., MongoDB) and big data technologies (HDFS, Hive, Spark) for large-scale data is a plus. Understanding of or experience with vector databases (e.g., Pinecone, FAISS) is beneficial, as the NLP applications may require embedding storage and fast similarity search.
    - Cloud Infrastructure: Practical experience with cloud platforms (AWS, GCP, or Azure) for data storage and processing. Ability to set up services such as S3/Cloud Storage, data warehouses (e.g., BigQuery, Redshift), and use cloud-based ETL tools or serverless functions. Understanding of infrastructure-as-code (Terraform, CloudFormation) to manage resources is a plus.
    - Data Quality & Monitoring: Knowledge of data quality assurance practices. Experience implementing monitoring for data pipelines (logs, alerts) and using CI/CD tools to automate pipeline deployment and testing. An analytical mindset to troubleshoot data discrepancies and optimize performance bottlenecks.
    - Collaboration & Domain Knowledge: Ability to work closely with data scientists and understand the requirements of machine learning projects. Basic understanding of NLP concepts and the data needs for training language models, so you can anticipate and accommodate the specific forms of text data and preprocessing they require. Good communication skills to document data workflows and to coordinate with team members across different functions.

    Nice to have:
    - Advanced Tools & Frameworks: Experience with distributed data processing frameworks (such as Apache Spark or Databricks) for large-scale data transformation, and with message streaming systems (Kafka, Pub/Sub) for real-time data pipelines. Familiarity with data serialization formats (JSON, Parquet) and handling of large text corpora.
    - Web Scraping Expertise: Deep experience in web scraping, using tools like Scrapy, Selenium, or Beautiful Soup, and handling anti-scraping challenges (rotating proxies, rate limiting). Ability to parse and clean raw text data from HTML, PDFs, or scanned documents.
    - CI/CD & DevOps: Knowledge of setting up CI/CD pipelines for data engineering (using GitHub Actions, Jenkins, or GitLab CI) to test and deploy changes to data workflows. Experience with containerization (Docker) to package data jobs and with Kubernetes for scaling them is a plus.
    - Big Data & Analytics: Experience with analytics platforms and BI tools (e.g., Tableau, Looker) used to examine the data prepared by the pipelines. Understanding of how to create and manage data warehouses or data marts for analytical consumption.
    - Problem-Solving: Demonstrated ability to work independently in solving complex data engineering problems, optimizing existing pipelines, and implementing new ones under time constraints. A proactive attitude to explore new data tools or techniques that could improve the workflows.

    Responsibilities:
    - Design, develop, and maintain ETL/ELT pipelines for gathering, transforming, and storing large volumes of text data and related information.
    - Ensure pipelines are efficient and can handle data from diverse sources (e.g., web crawls, public datasets, internal databases) while maintaining data integrity.
    - Implement web scraping and data collection services to automate the ingestion of text and linguistic data from the web and other external sources. This includes writing crawlers or using APIs to continuously collect data relevant to the language modeling efforts.
    - Implementation of NLP/LLM-specific data processing: cleaning and normalization of text, like filtering of toxic content, de-duplication, de-noising, detection, and deletion of personal data.
    - Formation of specific SFT/RLHF datasets from existing data, including data augmentation/labeling with LLM as teacher.
    - Set up and manage cloud-based data infrastructure for the project. Configure and maintain data storage solutions (data lakes, warehouses) and processing frameworks (e.g., distributed compute on AWS/GCP/Azure) that can scale with growing data needs.
    - Automate data processing workflows and ensure their scalability and reliability.
    - Use workflow orchestration tools like Apache Airflow to schedule and monitor data pipelines, enabling continuous and repeatable model training and evaluation cycles.
    - Maintain and optimize analytical databases and data access layers for both ad-hoc analysis and model training needs.
    - Work with relational databases (e.g., PostgreSQL) and other storage systems to ensure fast query performance and well-structured data schemas.
    - Collaborate with Data Scientists and NLP Engineers to build data features and datasets for machine learning models.
    - Provide data subsets, aggregations, or preprocessing as needed for tasks such as language model training, embedding generation, and evaluation.
    - Implement data quality checks, monitoring, and alerting. Develop scripts or use tools to validate data completeness and correctness (e.g., ensuring no critical data gaps or anomalies in the text corpora), and promptly address any pipeline failures or data issues. Implement data version control.
    - Manage data security, access, and compliance.
    - Control permissions to datasets and ensure adherence to data privacy policies and security standards, especially when dealing with user data or proprietary text sources.

    The company offers:
    - Competitive salary.
    - Equity options in a fast-growing AI company.
    - Remote-friendly work culture.
    - Opportunity to shape a product at the intersection of AI and human productivity.
    - Work with a passionate, senior team building cutting-edge tech for real-world business use.

    More
  • Β· 18 views Β· 1 application Β· 17d

    Senior Snowflake Data Engineer

    Full Remote Β· Ukraine Β· 5 years of experience Β· English - B2
    The project is for one of the world's famous science and technology companies in pharmaceutical industry, supporting initiatives in AWS, AI and data engineering, with plans to launch over 20 additional initiatives in the future. Modernizing the data...

    The project is for one of the world's famous science and technology companies in pharmaceutical industry, supporting initiatives in AWS, AI and data engineering, with plans to launch over 20 additional initiatives in the future. Modernizing the data infrastructure through the transition to Snowflake as a priority, as it will enhance capabilities for implementing advanced AI solutions and unlock numerous opportunities for innovation and growth.

    We are seeking a highly skilled Snowflake Data Engineer to design, build, and optimize scalable data pipelines and cloud-based solutions across AWS, Azure, and GCP. The ideal candidate will have strong expertise in Snowflake, ETL Tools like DBT, Python, visualization tools like Tableau and modern CI/CD practices, with a deep understanding of data governance, security, and role-based access control (RBAC). Knowledge of data modeling methodologies (OLTP, OLAP, Data Vault 2.0), data quality frameworks, Stream lit application development and SAP integration and infrastructure-as-code with Terraform is essential. Experience working with different file formats such as JSON, Parquet, CSV, and XML is highly valued.

    • Responsibilities:

      β€’ In-depth knowledge of Snowflake's data warehousing capabilities.
      β€’ Understanding of Snowflake's virtual warehouse architecture and how to optimize performance
      and cost.
      β€’ Proficiency in using Snowflake's data sharing and integration features for seamless collaboration.
      β€’ Develop and optimize complex SQL scripts, stored procedures, and data transformations.
      β€’ Work closely with data analysts, architects, and business teams to understand requirements and
      deliver reliable data solutions.
      β€’ Implement and maintain data models, dimensional modeling for data warehousing, data marts,
      and star/snowflake schemas to support reporting and analytics.
      β€’ Integrate data from various sources including APIs, flat files, relational databases, and cloud
      services.
      β€’ Ensure data quality, data governance, and compliance standards are met.
      β€’ Monitor and troubleshoot performance issues, errors, and pipeline failures in Snowflake and
      associated tools.
      β€’ Participate in code reviews, testing, and deployment of data solutions in development and production environments.

    • Mandatory Skills Description:

      β€’ 5+ years of experience
      β€’ Strong proficiency in Snowflake (Snowpipe, RBAC, performance tuning).
      β€’ Ability to write complex SQL queries, stored procedures, and user-defined functions.
      β€’ Skills in optimizing SQL queries for performance and efficiency.
      β€’ Experience with ETL/ELT tools and techniques, including Snowpipe, AWS Glue, openflow, fivetran
      or similar tools for real-time and periodic data processing.
      β€’ Proficiency in transforming data within Snowflake using SQL, with Python being a plus.
      β€’ Strong understanding of data security, compliance and governance.
      β€’ Experience with DBT for database object modeling and provisioning.
      β€’ Experience in version control tools, particularly Azure DevOps.
      β€’ Good documentation and coaching practice.

    More
Log In or Sign Up to see all posted jobs