Consultant Data Engineer (Python/Databricks)

Softermii is looking for a part-time Data Engineering Consultant / Tech Lead who will do technical interviews, assist with upcoming projects, and occasionally be hands-on with complex development tasks — including data pipeline design and solution optimization on Databricks.

 


Type of cooperation: Part-time

 

⚡️Your responsibilities on the project will be:

  • Interview and hire Data Engineers
  • Supervise work of other Engineers and have hands on for the most complicated tasks from backlog, focus on unblocking other data Engineers in case of technical difficulties
  • Develop and maintain scalable data pipelines using Databricks (Apache Spark) for batch and streaming use cases.
  • Work with data scientists and analysts to provide reliable, performant, and well-modeled data sets for analytics and machine learning.
  • Optimize and manage data workflows using Databricks Workflows and orchestrate jobs for complex data transformation tasks.
  • Design and implement data ingestion frameworks to bring data from various sources (files, APIs, databases) into Delta Lake.
  • Ensure data quality, lineage, and governance using tools such as Unity Catalog, Delta Live Tables, and built-in monitoring features.
  • Collaborate with cross-functional teams to understand data needs and support production-grade machine learning workflows.
  • Apply data engineering best practices: versioning, testing (e.g., with pytest or dbx), documentation, and CI/CD pipelines



 

🕹Tools we use: Jira, Confluence, Git, Figma

 

🗞Our requirements to you:

  • 5+ years of experience in data engineering or big data development, with production-level work.
  • Architect and develop scalable data solutions on the Databricks platform, leveraging Apache Spark, Delta Lake, and the lakehouse architecture to support advanced analytics and machine learning initiatives.
  • Design, build, and maintain production-grade data pipelines using Python (or Scala) and SQL, ensuring efficient data ingestion, transformation, and delivery across distributed systems.
  • Lead the implementation of Databricks features such as Delta Live Tables, Unity Catalog, and Workflows to ensure secure, reliable, and automated data operations.
  • Optimize Spark performance and resource utilization, applying best practices in distributed computing, caching, and tuning for large-scale data processing.
  • Integrate data from cloud-based sources (e.g., AWS S3), ensuring data quality, lineage, and consistency throughout the pipeline lifecycle.
  • Manage orchestration and automation of data workflows using tools like Airflow or Databricks Jobs, while implementing robust CI/CD pipelines for code deployment and testing.
  • Collaborate cross-functionally with data scientists, analysts, and business stakeholders to understand data needs and deliver actionable insights through robust data infrastructure.
  • Mentor and guide junior engineers, promoting engineering best practices, code quality, and continuous learning within the team.
  • Ensure adherence to data governance and security policies, utilizing tools such as Unity Catalog for access control and compliance.
  • Continuously evaluate new technologies and practices, driving innovation and improvements in data engineering strategy and execution.
  • Experience in designing, building, and maintaining data pipelines using Apache Airflow, including DAG creation, task orchestration, and workflow optimization for scalable data processing.
  • Upper-Intermediate English level.

 

 

👨‍💻Who will you have the opportunity to meet during the hiring process (stages):
Call, HR, Tech interview, PM interview.

 

🥯What we can offer you:

  • We have stable and highly-functioning processes – everyone has their own role and clear responsibilities, so decisions are made quickly and without unnecessary approvals. 
  • You will have enough independence to make decisions that can affect not only the project but also the work of the company.
  • We are a team of like-minded experts who create interesting products during working hours and enjoy spending free time together.
  • Do you like to learn something new in your profession or do you want to improve your English? We will be happy to pay 50% of the cost of courses/conferences/speaking clubs.
  • Do you want an individual development plan? We will form one especially for you + you can count on mentoring from our seniors and leaders.
  • Do you have a friend who is currently looking for new job opportunities? Recommend them to us and get a bonus.
  • And what if you want to relax? Then we have 21 working days off.
  • What if you are feeling bad? You can take 5 sick leaves a year.
  • Do you want to volunteer? We will add you to a chat, where we can get a bulletproof vest, buy a pickup truck or send children's drawings to the front.
  • And we have the most empathetic HRs (who also volunteers!). So we are ready to support your well-being in various ways.

 

👨‍🏫A little more information that you may find useful:

- our adaptation period lasts 3 months, this period of time is enough for us to understand  each other better;

- there is a performance review after each year of our collaboration where we use a skills map to track your growth;

- we really have no boundaries in the truest sense of the word – we have flexible working day is up to you.

 

Of course, we have a referral bonus syst

Published 15 May
21 views
·
4 applications
To apply for this and other jobs on Djinni login or signup.