Data Engineer (Romania)
About the client
Haptiq is an AI-native enterprise solutions company with purpose-built technology for public & private companies, governments, institutions, asset managers, and family offices
With headquarters in New York City and four global offices, Haptiq is supported by more than 300 engineers and delivery professionals across the globe. By centralizing and unifying data, automating workflows, and surfacing predictive insights, Haptiq enables organizations to scale operational excellence and generate alpha across complex enterprise environments
The Opportunity
This position is for a Cloud Data Engineer with a background in Python, DBT, SQL, and data warehousing for enterprise-level systems
Responsibilities and Duties
β Design, develop, and deploy Python scripts and ETL processes with Prefect and Airflow to prepare data for analysis
β Model dimensional and denormalized schemas for optimal performance reporting and discovery
β Design AI-friendly DB schemas and ontologies
β Architect cloud ops solutions for data topologies
β Transform and migrate data with Python, DBT, and Pandas
β Work with event-based/streaming technologies for real-time ETL
β Ingest and transform structured, semi-structured, and unstructured data
β Optimize ETL jobs for performance and scalability to handle big data workloads
β Monitor and troubleshoot ETL jobs to identify and resolve issues or bottlenecks
β Implement best practices for data management, security, and governance with Prefect, DBT, and Pandas
β Write SQL queries and program stored procedures and reverse engineer existing data pipelines
β Perform code reviews to ensure fit to requirements, optimal execution patterns, and adherence to established standards
β Assist with automated release management and CI/CD processes
β Validate and cleanse data and handle error conditions gracefully
Requirements
β 3+ years of Python development experience, including Pandas
β 5+ years writing complex SQL queries with RDBMSes
β 5+ years of Experience with developing and deploying ETL pipelines using Airflow, Prefect, or similar tools
β Experience with cloud-based data warehouses in environments such as RDS, Redshift, or Snowflake
β Experience with data warehouse design: OLTP, OLAP, Dimensions, and Facts
β Experience with cloud-based data architectures, messaging, and analytics
β Bachelorβs degree in Computer Science or equivalent β preferred
Pluses β Experience with
β Docker
β Kubernetes
β CI/CD automation
β AWS lambdas/step functions
β Data partitioning
β Databricks
β Pyspark
β Cloud certifications
Why Join Us?
We value creative problem solvers who learn fast, work well in an open and diverse environment, and enjoy pushing the bar for success ever higher. We do work hard, but we also choose to have fun while doing it
Required languages
| English | B2 - Upper Intermediate |