CLIENT
PROJECT OVERVIEW
TEAM
POSITION OVERVIEW
TECHNOLOGY STACK:
Responsibilities
- Design and implement new version of Data Platform based on GCP Cloud
- Migration of Teradata to BigQuery. Moving from on-premise to the cloud, decomp of legacy systems
- Implement and support near real-time and batch processing data pipelines? Backend integration
- Work collaboratively on designing and implementing modern, scalable data solutions using a range of new and emerging technologies from the Google Cloud Platform
- Manage both real time and batch data pipelines. Our technology stack includes a wide variety of technologies including Spark, Kafka, GCP PubSub, Teradata
Requirements
- Showcase your GCP Data experience when communicating with stakeholders, turning these into technical data solutions
- Expertise in the main components of continuous data delivery: set up, design and delivery of data pipelines (testing, deploying, monitoring and maintaining)
- Expertise in data architecture within web/mobile environments, web/internet-related technologies, architecture across Software-as-a-Service, Platform-as-a-Service, Infrastructure-as-a-Service and cloud productivity suites
- Strong engineering background, with experience with Python, SQL, SparkML or similar frameworks used to ship data processing pipelines at scale
- Demonstrated experience with and solid knowledge of cloud Google Cloud Platform: Cloud Composer, BigQuery, Data Proc, etc.
- Demonstrated experience with Infrastructure-as-code (IaC) – Terraform
- Basic knowledge of Teradata
- Capable of working independently with minimal supervision in remote teams configuration
- Ability to work, communicate effectively and influence stakeholders on internal/external engineering teams, product development teams, sales operations teams and external partners and consumers
- Good spoken English
Nice to have
- Demonstrated experience in Data migration projects. Migration from Teradata to BigQuery