Data Engineer (Python/Databricks)

📲Project:

Team's goal is to deliver Hotel Business Management solution in a few phases starting MVP.

This tool will enable users to focus on Revenue Management, cover Guest Check-In experience, Marketing & Retention, Analytics & Reporting, Service Desk and other features that market requires.

Type of cooperation: Full-time

 

⚡️Your responsibilities on the project will be:

  • Develop and maintain scalable data pipelines using Databricks (Apache Spark) for batch and streaming use cases.
  • Work with data scientists and analysts to provide reliable, performant, and well-modeled data sets for analytics and machine learning.
  • Optimize and manage data workflows using Databricks Workflows and orchestrate jobs for complex data transformation tasks.
  • Design and implement data ingestion frameworks to bring data from various sources (files, APIs, databases) into Delta Lake.
  • Ensure data quality, lineage, and governance using tools such as Unity Catalog, Delta Live Tables, and built-in monitoring features.
  • Collaborate with cross-functional teams to understand data needs and support production-grade machine learning workflows.
  • Apply data engineering best practices: versioning, testing (e.g., with pytest or dbx), documentation, and CI/CD pipelines.


You can be part of the following team: 

Client`s side: Business Owner, Senior AI/ML Engineer, Data Engineers, Subject Matter Experts.

On Softermii`s side: Software Architect, 2-3 React.js/Node.js Engineers, 2-3 AI/ML Engineers, 3 Data Engineers, DevOps, Business Analyst/Proxy PO, UI/UX Designer

2-3 Quality Assurance Engineers, Project Manager.

 

🕹Tools we use: Jira, Confluence, Git, Figma

 

🗞Our requirements to you:

  • 3+ years of experience in data engineering or big data development, with production-level work on the Databricks platform.
  • Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field.
  • Proficiency in Python (or Scala) and SQL for data processing and transformation on Spark.
  • Hands-on experience with Apache Spark, Delta Lake, and Databricks Notebooks.
  • Familiarity with Databricks Workflows, Unity Catalog, and Delta Live Tables for managing and securing data pipelines.
  • Solid understanding of data lakehouse architecture, distributed computing, and performance tuning on Spark.
  • Experience integrating data from cloud sources like AWS S3.
  • Knowledge of orchestration tools (e.g., Airflow, Databricks Jobs), and CI/CD practices for data pipelines.
  • Strong communication and collaboration skills, with a team-oriented mindset.
  • Upper-Intermediate English level.

 

 

👨‍💻Who will you have the opportunity to meet during the hiring process (stages):
Call, HR, Tech interview, PM interview.

 

🥯What we can offer you:

  • We have stable and highly-functioning processes – everyone has their own role and clear responsibilities, so decisions are made quickly and without unnecessary approvals. 
  • You will have enough independence to make decisions that can affect not only the project but also the work of the company.
  • We are a team of like-minded experts who create interesting products during working hours and enjoy spending free time together.
  • Do you like to learn something new in your profession or do you want to improve your English? We will be happy to pay 50% of the cost of courses/conferences/speaking clubs.
  • Do you want an individual development plan? We will form one especially for you + you can count on mentoring from our seniors and leaders.
  • Do you have a friend who is currently looking for new job opportunities? Recommend them to us and get a bonus.
  • And what if you want to relax? Then we have 21 working days off.
  • What if you are feeling bad? You can take 5 sick leaves a year.
  • Do you want to volunteer? We will add you to a chat, where we can get a bulletproof vest, buy a pickup truck or send children's drawings to the front.
  • And we have the most empathetic HRs (who also volunteers!). So we are ready to support your well-being in various ways.

 

👨‍🏫A little more information that you may find useful:

- our adaptation period lasts 3 months, this period of time is enough for us to understand  each other better;

- there is a performance review after each year of our collaboration where we use a skills map to track your growth;

- we really have no boundaries in the truest sense of the word – we have flexible working day is up to you.

Published 18 April
62 views
·
11 applications
60% read
·
60% responded
Last responded 5 days ago
To apply for this and other jobs on Djinni login or signup.