Senior Data Engineer (offline)

About the project:
Our client is the leading Autonomous Decision Science company providing data-enabled prescriptive and anticipatory analytics and insights for companies across a broad range of industries. They automate a new way of problem-solving and decision-making, going beyond advanced analytics to understand data, perform analytics, generate insights, answer queries, and make decisions at the speed of the market. This powerful capability uniquely enables “End-to-End Value Chain Optimization and Decision-Making” at scale and allows clients to uncover and understand the inter-relationships that lead to innovative new product development and innovation, heightened consumer understanding and targeted marketing, revenue growth tactics, and intelligent demand and supply-chain planning. We help transform market-leading companies into true data-driven digital enterprises.

Stage: Ongoing 1 year +.
Timezone requirements: Europe working time is ok, the client is in the EST, and our team is in Europe.
Location requirements: Ukraine, EU; the ability to travel abroad is essential.
English: Upper-Intermediate.

Requirements:
— Minimum of 5 years hands-on experience as a data engineer or similar position. 
— Minimum of 3 years of commercial experience with Python 
— Minimum of 3 years SQL and experience working with relational databases (Postgres preferred, similar OLTP DB is acceptable). 
— Demonstratable knowledge and experience developing data pipelines to automate data processing workflows. 
— Demonstratable experience in data modeling. 
— Demonstratable knowledge of data warehousing, business intelligence, and application data integration solutions.  
— Demonstratable experience in developing applications and services that run on a cloud infrastructure (Azure preferred, AWS or GCP). 
— Experience with container management and deployment, e.g., Docker and Kubernetes. 
— Experience with Python 3, Airflow, Snowflake. 
— Strong skills with pandas and pyspark/spark. 
— Excellent problem-solving and communication skills.
— Ability to thrive in a fast-paced, remote environment. 
— Comfortable with ambiguity with the ability to build structure and take a proactive approach to drive results. 
— Attention to detail – quality and accuracy in deliverables. 

The following additional skills would be beneficial:  
— Knowledge of one or more of the following technologies: Data Science, Machine Learning, Natural Language Processing, Business Intelligence, and Data Visualization. 
— Knowledge of statistics and experience using statistical or BI packages for analyzing large datasets (Excel, R, Python, Power BI, Tableau etc.). 
— Experience with at least one of the following – Databricks, Spark, Hadoop or Kafka. 
— FastAPI / Flask or similar web backend framework experience. 
Experience with some of the following - Celery, RabbitMQ, ZeroMQ. 
— Other SQL or NoSQL experience - Redis, ElasticSearch, DuckDB, ClickHouse.

What you will do:  
The successful candidate will join a diverse team to:  
— Build unique high-impact business solutions utilizing advanced technologies for use by world class clients. 
— Create and maintain the underlying data pipeline architecture for the solution offerings, from raw client data to final solution output.  
— Create, populate, and maintain data structures for machine learning and other analytics.  
— Use quantitative and statistical methods to derive insights from data. 
— Combine machine learning, artificial intelligence (ontologies, inference engines and rules) and natural language processing under a holistic vision to scale and transform businesses — across multiple functions and processes. 

Responsibilities Include: 
— Create and maintain optimal data pipeline architecture, incorporating data wrangling and Extract-Transform-Load (ETL) flows.  
— Assemble large, complex data sets to meet analytical requirements – analytics tables, feature-engineering etc. 
— Build the infrastructure required for optimal, automated extraction, transformation, and loading of data from a wide variety of data sources using SQL and other ‘big data’ technologies. 
— Build automated analytics tools that utilize the data pipeline to derive actionable insights. 
Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc. 
— Design and develop data integrations and data quality framework. 
— Develop appropriate testing strategies and reports for the solution as well as data from external sources.  
— Evaluate new technology for use within the projects. 
— Work with the team and client personnel to administer and operate client-specific instances of the offerings. 
— Configure the data pipelines to accommodate client-specific requirements to onboard new clients. 
— Perform regular operations tasks to ingest new and changing data – implement automation where possible. 
Implement processes and tools to monitor data quality - investigate and remedy any data-related issues in daily solution operations. 

Tech stack:
Backend
— Programming Language: Python 3.11.2 (latest)
— Package Manager: Poetry
— Web Framework: FastAPI
— Object-Relational Mapping (ORM): SQLAlchemy
— Linting: ruff
— Code Formatter: isort
— Testing Framework: pytest
— Property-based Testing: Hypothesis
— Database: PostgreSQL
Frontend
— Programming Language: TypeScript
— Package Manager: PNPM
— Framework: React
— State Management: Zustand
— Routing: React Router
— Data Fetching: TanStack Query (successor to react-query)
— Build Tool: Vite
— Unit Testing: Vitest
— Component Testing: React Testing Library
— End-to-End Testing: Playwright
— UI Kit (to be determined)
— Linting: ESLint + typescript-eslint
— Code Formatting: Prettier
And:
— Docker and Docker Compose
— Kubernetes
— Airflow
— Azure
— Snowflake
— PostgreSQL

Why choose us?
— “Family and Friends”. We are no longer a start-up, but we still, have a family atmosphere in our supportive and spirited team, who are all working together on the same goal.
— “Just break down all barriers and find a better way”. Every day you’ll meet with interesting and challenging (international) projects that cover industries from commercial aviation to fintech (different technologies, different products).
— “Hungry for learning”. You will get a lot of chances for career advancement and the development of new skills, mentorship opportunities, or learning from more experienced colleagues.

Benefits from 8allocate:
— You’ll work with a supportive and spirited team of professionals.
— Corporate events, holidays, and team buildings for your joy.
— Training and development: we have a huge library (about 500 books!) and a budget for your professional development.
— Sport or massage compensation according to the general policy.
— Compensation for psychological support or medical insurance compensation according to the general policy.
— People-oriented management without bureaucracy.
— Paid vacation and sick leaves.

About 8allocate

8allocate is a provider of software development solutions to companies all over the globe, from North America to the EU. We deliver market-ready solutions for FinTech/InsurTech, eCommerce, AI, Web, Mobile, and other domains.

Company website:
https://8allocate.com/

DOU company page:
https://jobs.dou.ua/companies/8allocate/

The job ad is no longer active

Look at the current jobs Data Engineer Kyiv→