Senior Data Engineer โ Java
We're looking for a Senior Data Engineer โ DevOps with strong GCP and Java experience to join a stable, long-term project for the biggest insurance tech company in the USA. We're building a completely new, advanced data platform based on the lake house architecture, using bronze/silver/gold layers, and a data model layer. You'll work with advanced technologies and will be able to work with one of the best GCP data architects in the World.
About Company
Our client is a large USA product company, a global leader in insurance technologies, and is seeking an experienced Data Engineer with strong expertise in Google Cloud Platform (GCP). Join us in scaling our Data and Analytics capabilities to drive data-informed decisions across our organization. You will design, build, and maintain efficient data pipelines, optimize data workflows, and integrate data seamlessly from diverse sources.
What You Will Do:
- Build and maintain CI/CD pipelines to enhance productivity, agility, and code quality.
- Optimize data pipelines and workflows for performance and scalability.
- Design efficient processes to minimize data refresh delays, leveraging reusable components and automated quality checks.
- Develop robust, scalable data pipelines supporting business needs.
- Code BigQuery procedures, functions, and SQL database objects.
- Monitor application performance, troubleshoot issues, and implement effective monitoring and alerting.
- Lead design and build-out of production data pipelines using GCP services (BigQuery, DBT, Apache Airflow, Celigo, Python).
- Ensure data quality through rigorous testing and validation.
- Maintain thorough technical documentation and stay current with industry trends.
What You Need to Succeed:
- 8+ years in Data/ETL Engineering, Architecture, and pipeline development, with at least 2 years focused on GCP.
- Proven experience building scalable cloud Data Warehouses (preferably BigQuery).
- 3+ years advanced SQL and strong Python or Java programming experience.
- Extensive experience optimizing ETL/ELT pipelines, data modeling, and schema design.
- Expertise with GCP services: Composer, Compute, GCS, Cloud Functions, BigQuery.
- Proficiency in DevOps tools (Git, GitLab) and CI/CD pipeline integration with GCP.
- Strong automation scripting skills, especially with GCP Composer.
- Solid understanding of Data Lake/Warehouse concepts and data modeling techniques (star schema, snowflake schema, normalization).
- Excellent problem-solving skills; able to work independently and collaboratively.
- Strong communication skills, capable of explaining technical concepts clearly.
- Bachelor's degree in Computer Science, MIS, CIS, or equivalent experience.