You will help internal teams to migrate to a new set of canonical datasets produced by Metadata Distribution squad. Above all, your work will impact the way the world experiences music.
What you’ll do
- Implement canonical datasets for metadata entities that are used to fuel hundreds of experiences on our platform
- Supporting internal teams in migrating their pipelines to the new generation of metadata datasets.
- Getting hands-on experience with Google Cloud Platform and technology/languages such as BigQuery, Scala, Scio, Luigi, Styx and Docker
- Operate large batch data pipelines
- Work closely with our customers and stakeholders to understand, document, troubleshoot and analyze their data requirements
Who you are
- You have Data Engineering experience and you know how to work with high- volume, heterogeneous data, preferably with distributed systems such as Hadoop, BigTable, Cassandra, GCP, AWS or Azure.
- You know Scala language well
- You have experience with one or more higher-level JVM-based data processing frameworks such as Beam, Dataflow, Crunch, Scalding, Storm, Spark, Flink etc.
- You might have worked with Docker as well as Luigi, Airflow, or similar tools.
- You are passionate about crafting clean code and have experience in coding and building data pipelines
- You understand the value of collaboration and partnership within team
Start Date: asap
Workload: 100 %
CV in English
About DataJob
DataJob is a trusted partner for providing staffing, consulting and insourcing services in the Nordics.
We have been involved in a wide range of projects for clients including computer software, information technology, internet, marketing and advertising companies.
Together with selected consultants DataJob provides a complete range of professional services.
Company website:
http://www.datajob.se
DOU company page:
https://jobs.dou.ua/companies/datajob-sweden-ab/
The job ad is no longer active
Job unpublished on
29 September 2020
Look at the current
jobs
Data Science
remote→