Senior Data Engineer
Project Description
New long-term project for an Energy client, where we will create an application with AI integrated for the comprehensive data analysis. You will be working closely with the customer stakeholders as part of the Scrum team.
Technical Requirements (Must Have):
Python โ 5+ years, production code (not just notebooks)
SQL / PostgreSQL โ 5+ years, complex queries, optimization
Apache Kafka โ event streaming, consumers, producers
pandas / numpy โ expert level, large datasets (1M+ rows)
scikit-learn โ clustering algorithms, metrics, hyperparameter tuning
ETL Pipelines โ 4+ years building production data pipelines
Text Processing โ tokenization, cleaning, encoding handling
Git โ branching, PRs, code reviews
English โ B2+ written and verbal
Would Be a Plus
Sentence-BERT / Transformers (HuggingFace ecosystem)
MLflow or similar ML experiment tracking
Topic Modeling (LDA, NMF)
DBSCAN / Hierarchical Clustering
FastAPI / Flask
Azure DevOps
Kafka Streams / ksqlDB
BI & Visualization tools (Power BI, Tableau, Grafana, Apache Superset, Plotly/Dash, or similar)
Nice to Have
Energy / Utility / SCADA domain experience
Time-series analysis
Prometheus / Grafana monitoring
On-premise ML infrastructure (no cloud APIs)
Data modeling / dimensional modeling
dbt (data build tool)
Job Responsibilities
Strong problem-solving and follow-up skills; must be proactive and take initiative
Professionalism and ability to maintain the highest level of confidentiality
Create robust code and translate business logic into project requirements
Develop code using development best practices, and an emphasis on security best practices
Leverage technologies to support business needs to attain high reusability and maintainability of current and newly developed systems
Provide system design recommendations based on technical requirements
Work independently on development tasks with a minimal amount of supervision
Required languages
| English | B2 - Upper Intermediate |