Middle+ Data Engineer
Dataforest is looking for a Data Engineer to join an interesting software development project in the field of Cybersecurity. Our client’s platform offers Identity Insights drives real-time, multi-contextual digital forensics to protect new account opening workflows and expose fake accounts already in a customer database.
Requirements:
• 3+ years of hands-on experience in data engineering;
• 3+ years of commercial experience with Python;
• Advanced experience with SQL DBs (optimisations, monitoring, etc.);
• Advanced experience with PySpark;
• Solid understanding of ETL principles (architecture/ monitoring/ alerting/search and resolve bottlenecks);
• Familiar with AWS infrastructure (boto3, S3 buckets, etc);
• Experience working with large volumes of data;
• Understanding the principles of medallion architecture.
Responsibilities:
• Maintain data processing architecture using Python and PySpark that interacts with ScyllaDB & PostgreSQL.
• Work with Databricks infrastructure to ensure reliability, scalability, and cost efficiency.
• Proactively identify bottlenecks and suggest technical improvements.
Nice to have:
• Cassandra/Scylla
• PostgreSQL
• Pandas
• Experience with Kafka and Redis
We offer:
• Great networking opportunities with international clients, challenging tasks;
• Building interesting projects from scratch using new technologies;
• Personal and professional development opportunities;
• Competitive salary fixed in USD;
• Paid vacation and sick leaves;
• Flexible work schedule;
• Friendly working environment with minimal hierarchy;
• Team building activities, corporate events.
Required skills experience
| Python | 3 years |
| Data Engineering | 3 years |
| PySpark | 3 years |
| AWS | 3 years |
| ETL | 3 years |
| Databricks | 3 years |
Required languages
| English | B2 - Upper Intermediate |
| Ukrainian | Native |