Senior AWS Data Engineer (offline)

Ideal Candidates have:

Previous years of experience and a substantial understanding in:
- AWS data ingestion frameworks, such as Kinesis, DMS (in use)
- AWS data processing (CDC,ETL) frameworks, such as EMR (in use),
- AWS Glue/Athena (in use), Sqoop, Spark, Zeppelin, Kinesis tools
- AWS pipeline orchestration technologies, such as Data Pipeline(in use), Step Functions, and Lambda (in use)
- AWS data storage engines, such as S3 (in use), HDFS (in use), Dynamo (in use)
- AWS data access/warehousing tools, such as Athena (in use) and Redshift or any other cloud technology
- Understanding the strengths and weaknesses of each tool/service and what it’s best used for
- Should be comfortable in Linux and using command line
- Proficiency in software development in Python (in use), Scala, Javascript (node in use), or Java
- Deep familiarity with SQL (heavily used)
- Good understanding of distributed systems in general and how they are different from traditional storage and processing systems
- Passionate about “Big Data” and wants to expand their skills in the realm of data science infrastructure and tooling
- Experience with data visualization tools like Tableau (in use) or ElasticSearch will be a big plus

Job Responsibilities:

We are seeking a motivated senior data engineer to architect, build and support AWS data pipelines for our corporate data strategies. As a key player of a dedicated team, you will be instrumental in driving the direction of the pipeline infrastructure from ingestion through to access. As our corporate data strategy evolves from business analytics to data science, there will be numerous opportunities to utilize new technology and learn new skills.

What You Will Do:
- Architect, build and maintain data pipelines as a key player of the company long term data strategies
- Utilizing AWS services, build and support reliable multi-tenant data pipelines combining data multiple diverse data sources
- Providing scalable tools and mechanism for data ingestion, cleansing and ETL
- Make pragmatic iterations of pipelines as the companies needs grow over time
- Providing and support infrastructure for pipelines, data access, and visualization

Department/Project Description:
We develop a supplier collaboration platform for ocherstrating of construction’s heavy work for US-based partner.

This cloud-based platform built on AWS using Javascript and Node.JS, React frameworks. Its purpose is increasing collaboration between contractors and project owners, and their heavy material suppliers, producers and haulers. It facilitates every step in the workflow from initial orders through to paying the final bill.

In addition to this platform has advanced mechanisms for revealing insights collected from a variety of sources, including IoT devices, telematics, jobsite tech, production automatics, etc. Collected analytics help making better business decisions, operate more efficiently, and see actionable info in real time.

Users interact with the platform using web, and mobile applications running Apple iOS and Android devices.

About GlobalLogic


Company website:
https://www.globallogic.com/ua/

The job ad is no longer active
Job unpublished on 15 August 2020

Look at the current jobs Data Science Kyiv→