Candidates 423
$6000 / mo
≈ $72000 / year net
Senior Data Engineer
Poland · 5 years of experience · Advanced/Fluent ·Published today
Poland · 5 years of experience · Advanced/Fluent ·Published today
- Years of experience: 5
- Domain: banking, fintech, healthcare.
- Technologies stack: Python (Spark, Trino, SQLalchemy, Airflow, FastAPI, MinIO, Boto3), Postgres, Kubernetes, Docker, AWS (S3, Athena, Lambda, Glue, QuickSight, LakeFormation), Dask.
- BI Tools: SAP, Alteryx, PowerBI, Tableau
- Version control: Git, BitBucket
- Documentation: GitLab, Confluence
- Pipeline: ArgoCD
- Monitoring: Grafana/Loki
- Education: Carnegie Mellon University (2019)
- Major: Data Analytics/Data Science
- Rated TOP 35% in McKinsey x Alibaba hackathon
- Certified Alteryx designer (ETL tool)
$6000 / mo
≈ $72000 / year net
Data Engineer
Uzbekistan · 4 years of experience · Advanced/Fluent ·Published today · In passive search
Uzbekistan · 4 years of experience · Advanced/Fluent ·Published today · In passive search
Modeling relational databases (ER modeling, Relational schema, Normal forms), data modeling 3NF/Dimensional (Star,Snowflake,DataVault)
Python, Django, FastApi, Pandas
Looker/LookML
Metabase,Tableau
Docker, Git (GitHub, Actions), Airflow,
Nginx, ElasticSearch, Terraform, K8s(Helm), ArgoCD
AWS(EKS, Glue, S3, EC2, Lambda, ElasticBeanstalk, IAM, VPC…)
GCP(Beam, PubSub, Instance, GCS, BigQuery, …)
AZURE(Synapse, data factory, datarbicks, adls…)
PySpark(Datarbicks),
OpenTableFormats (Iceberg, Delta, Hudi), FileFormat(Avro,Parquet,ORC)
Storage(S3,ADLS,GCS)
HadoopStreaming(MapReduce), Kafka
==================================
----September 2023 - now ApexBank
ApexBank.uz
Data Engineer
Financial Sector
• Banking
Building DWH with:
1. Apache Spark(PySpark) on k8s for compute
2. Apache Hudi on S3(MinIO) as datalake
3. Greenplum as DWH
4. Clickhouse(or Starrocks.io) as serving layer for BI (still in progress)
5. Airflow for orchestration
6. Airbyte for CDC
---- September 2022 till 2023 _ Itransition
---- Data Engineer role
1. Converting BigQuery queries to Snowflake after two companies merged
2. Doing ELT(dbt) with Snowflake-tasks, and for some SQL using Airflow. It depends on which team has
created a ticket.
3. Preparing semantic layer for Looker, so that Bi engineers(including me) can build dashboards
and do some aggregations.
4. Building Looker reports and attaching them to business-related boards.
Python, Apache Airflow, GCP/AWS, Snowflake/BigQuery, Looker(LookML), SQL, DBT
---- September 2021 — September 2022 _ EPAM Systems
---- Python Engineer role
Extracting and data validation for RSS feeds.
Extending functionalities of existing lambdas.
Python, AWS(Lambda), S3, Elastic Beanstalk, VPC, EC2
---- August 2020 — September 2021 _ AlifTech
---- Data Engineer role
- Improved initially working with MySQL Tableau performance by replacing it with MongoDB.
- Wrote data pipelines using Apache Airflow (including data retrieving from several MySQL
instances, processing them on Pandas, and storing them into MongoDB as data warehouse).
- Automated daily manual jobs with Airflow including sending daily reports to other departments.
- Automated ML model builds by refactoring code into Airflow DAGs and Tasks.
Airflow, python, pandas, FastAPI, Tableau
-- Modeling relational databases (ER modeling, Relational schema, Normal forms), data modeling 3NF/Dimensional (Star/Snowflake)
-- Python, Django, FastApi, Pandas
-- k8s(Helm), Docker, Git (GitHub, Actions), Airflow, Metabase, Nginx, AWS/GCP
-- Hadoop(MapReduce), PySpark(Datarbicks), TableFormats (Iceberg, Delta table), FileFormat(Avro,Parquet,ORC), Storage(S3,ADLS,GCS,HDFS)
$3000 / mo
≈ $36000 / year net
SQL Developer/Data Engineer
Ukraine · 7 years of experience · Intermediate ·Published today
Ukraine · 7 years of experience · Intermediate ·Published today
Stack:
- MS SQL, PostgreSQL, Oracle, Teradata;
- C#/.Net;
- ADO.Net, ASP.Net Core, MudBlazor
- Microstrategy BI, Power Bi;
In my previous roles, my responsibilities encompassed:
Created, maintained, and automated reports using Excel and BI tools;
Created queries, tables, views, procedures;
Prepered datasets for analitics;
Developed C# services to synchronize data between databases, services, and APIs;
Created data factory pipelines;
Engaged with MudBlazor: crafting pages, components, and dialogs within the administrative panel.
$7000 / mo
≈ $84000 / year net
Team/Tech Lead/Architect/Engineering Manager
Poland · 6 years of experience · Upper-Intermediate ·Published today
Poland · 6 years of experience · Upper-Intermediate ·Published today
Lead Engineer (Oct 2022 – now)
• Optimization of existing processes and procedures
• Creating new pipelines,modeles,views and drivers (Django + AWS)
• Creating unit-tests for our drivers(PyTest+python)
• Technical support storage
• Writing and optimization drivers, functions
• Kept project leaders regularly updated with progress, maintaining open, productive communication.
• Managed multiple projects with differing technologies, including delegating tasks, assessing quality and sign off.
Lead Engineer (June 2022 – now)
• Optimization of existing processes and procedures
• Creating new pipelines and drivers (Java + AWS)
• Creating unit-tests for our drivers(java spring+springBoot)
• Technical support storage
• Writing and optimization drivers, functions
• Kept project leaders regularly updated with progress, maintaining open, productive communication.
Software Engineer (April 2021 – May 2022)
• Optimization of existing processes and procedures
• Creating new pipelines and drivers (python+Airflow+spark+AWS)
• Creating unit-tests for our data-drivers(scala)
• Creating a new data-notebooks(spark)
• Technical support storage
• Big Data migration process development
• Writing and optimization views, stored procedures, and functions
Intetics
Software Engineer (September 2020 – April 2021)
• ETL process development
• Preparation of data for building reports
• Data migration process development
• Writing and optimization views, stored procedures, and functions, construction and support of storage
EPAM Systems Inc
Software Engineer (August 2019 – September 2020)
• data collection from various sources: Source DB (Oracle, Postgres); different api (Beamery api,GoogleAnalytics api and other); flat files (exel, csv, json)
• pre-processing of data before loading into the DWH (the storage is built in MS
SQL server)
• creation and configuration of loading processes in the DWH (using SSIS packages and
services, python-scripts, PowerShell-scripts and .NET Application)
• the creation of procedures, functions and scripts for data transformation for their preparation for
reports
• development, support of storage performance
• optimization of existing processes and procedures
GlowByte Consulting
Software Engineer (May 2018 – July 2019)
• ETL process development
• Preparation of data for building reports
• Data migration process development
$6000 / mo
≈ $72000 / year net
Data engineer
Greece · 10 years of experience · Upper-Intermediate ·Published today
Greece · 10 years of experience · Upper-Intermediate ·Published today
• Building Data Lake House;
• Building Data Warehouse;
• Implementation of data quality process;
• Creating ETL processes from different sources;.
• Implementation BI solutions
• Migration data to cloud
• Developed database object: stored procedures, UDF, calculation views, complicated queries.
• Developed Data Bricks notebooks and workflows
• Implemented and supported Spark structured streaming process (Event Hub & Data Bricks)
• Designed and implemented custom log structure (Dynatrace & Azure Log Analytics)
• Deployed and maintained snapshots in Snowflake
• Designed and implemented of ADF pipelines
• Designed Azure Data Lake structure
• Performance tuning
• Troubleshooting
Data migration, validation, modeling, transformation.
Work with streaming data.
$6000 / mo
≈ $72000 / year net
Senior Data Engineer
Poland · 6 years of experience · Advanced/Fluent ·Published today
Poland · 6 years of experience · Advanced/Fluent ·Published today
- Building Data Warehouses in BigQuery: from raw data ingestion to materialized tables and views
- Advanced SQL: Built custom CTEs daily for 3 years
- Python & R as a working language, occasionally Scala Spark
- Data Modelling & Analysis in big EU eCommerce using AWS and BigQuery stack (Athena, Sagemaker with EMR, dbt, dataform, advanced Tableau)
Education:
- Bachelor degree in IT/Cybersecurity
- Coursera: Data Science, Machine Learning
Working 6 years as a Data Analyst and Data Engineer, combined with 7 years as a Marketing professional, endowed me with a broad and deep perspective on leveraging data insights for strategic decision-making and marketing effectiveness.
I enjoy having both deep and wide expertise in Data. Understanding the needs of data consumers made me a better analyst, and feeling data analysts pain points made me a more efficient engineer.
- Data ingestion with Airbyte and GCP Cloud Functions from various APIs into BigQuery
- Created conceptual data model for marketing needs
- Built materialized tables with dbt for use in Looker Studio
2. "Trust your data" project (data QA):
Initiated a "trust your data" sprint in a company after finding deep problems in data quality from a data modelling perspective. As a small group of couple data engineers and analysts, we defined new core concepts definition which showed a 30% increase in the core KPI of my department, both in current and historical data.
3. Learning Scala
Just for fun, I learned Scala to use it with Spark. After a month I wrote a tool that got an SQL query as an input, downloaded data from Presto by chunks, and converted it into CSV or Hyper (for Tableau) files. Before, Data Analysts constantly had problems because of Presto limits. My tool eliminated this problem.
$2500 / mo
≈ $30000 / year net
Middle QA Engineer
Ukraine · 3 years of experience · Upper-Intermediate ·Published today · In passive search
Ukraine · 3 years of experience · Upper-Intermediate ·Published today · In passive search
$7000 / mo
≈ $84000 / year net
Senior Data Engineer
Azerbaijan · 7 years of experience · Upper-Intermediate ·Published today
Azerbaijan · 7 years of experience · Upper-Intermediate ·Published today
Good knowledge Google Cloud Platform (Dataflow, Bigquery, Dataproc, Bigtable etc..).
Good knowledge PySpark
• Exceptional background in analysis, design, development, customization, and implementation and testing of software applications and products.
• Demonstrated expertise utilizing ETL tools, including SQL Server Integration Services (SSIS), SQL Server Annalise Services (SSAS), SQL Server Reporting Services (SSRS), Data Transformation Services (DTS), and DataStage and ETL package design, and RDBM systems like SQL Servers, Oracle.
• Strong understanding of data warehousing concepts, OLTP, OLAP, Normalization, Star and Snow Flake data models.
• Create OLAP Cubs.
• Strong leader with experience training developers and advising technical groups on ETL best practices.
• Excellent technical and analytical skills with clear understanding of desig
$4500 / mo
≈ $54000 / year net
Data Engineer
Ukraine · Kyiv · 5 years of experience · Upper-Intermediate ·Published today
Ukraine · Kyiv · 5 years of experience · Upper-Intermediate ·Published today
Core bank system implementation and migration in a few large banks in Ukraine and Azerbaijan as a technical business analyst.
Development and implementation of a new B2C application for an insurance company in OAE as a system analyst.
$2000 / mo
≈ $24000 / year net
Python Developer
Uzbekistan · 3 years of experience · Upper-Intermediate ·Published today
Uzbekistan · 3 years of experience · Upper-Intermediate ·Published today
Software Engineer at ITC in Ministry of Finance:
- Deployed 2 web applications to server(nginx/gunicorn/postgresql/ubuntu)
- Changed the production server(CentOS 7) and database (MariaDB) to Ubuntu 20.04 and PostgreSQL respectively
- Built API services for HR project front-end
- Developed Admin Dashboard for Info-kiosk project
- Integrated external API services in SOAP
- Web application development
Python Developer at Asia Finance Group:
- Optimized report systems readability by removing 60% of lines in source code
- Integrated 11 external API services to analyze the data in main risk-management project
- Planned Database architecture for saving data from side API services
Data Engineer at Special Texno Soft:
- 100% automated image uploading process with annotations into Label Studio using API services
- Developed reports about the state of projects of Label Studio(Image annotation), CVAT(Video annotation), Pose Estimation(MongoDB) in Apache Airflow
- Control data flow/quality
- Gather datasets
- Manage annotators
- Built monitoring system of Server state with Grafana+Prometheus
- Action detection tool
Tech: Python3, Django+DRF, PostgreSQL/MySQL, Nginx, Apache Airflow, Label Studio, Docker, OpenCVAT, MongoDB, Grafana, Prometheus, Ansible, Ubuntu, ELK
Accelerated file upload process to Label Studio by 38% using API.
Built monitoring system of Machine Learning/Image/Video Processing server.
Used: Grafana, Prometheus, Node Exporter, cAdvisor, Docker Exporter, shell/python scripts for custom exporters
Developed Action Detection tool web version to enlarge availability for large number of annotators.
Used: Django, DRF, Postgres
Friendly team.
High salary rates.