Google Data Engineering Architect ID-3453

About the Role:

We are seeking a highly skilled and experienced Data Engineering Architect to join our growing team. As a Data Engineering Architect, you will play a critical role in designing, building, and scaling Google’s massive data infrastructure and platforms. You will be a technical leader and mentor, driving innovation and ensuring the highest standards of data quality, reliability, and performance.

Responsibilities:

Design and Architecture:

Design and implement scalable, reliable, and efficient data pipelines and architectures for various Google products and services.

Develop and maintain data models, schemas, and ontologies to support diverse data sources and use cases.

Evaluate and recommend new and emerging data technologies and tools to improve Google’s data infrastructure.

Collaborate with product managers, engineers, and researchers to define data requirements and translate them into technical solutions.

Other Jobs You May Be Interested In

Data Processing and Pipelines:

Build and optimize batch and real-time data pipelines using Google Cloud Platform (GCP) services such as Dataflow, Dataproc, Pub/Sub, and Cloud Functions.

Develop and implement data quality checks and validation processes to ensure data accuracy and consistency.

Design and implement data governance policies and procedures to ensure data security and compliance.

Data Storage and Management:

Design and implement scalable data storage solutions using GCP services such as BigQuery, Cloud Storage, and Spanner.

Optimize data storage and retrieval for performance and cost-effectiveness.

Implement data lifecycle management policies and procedures.

Team Leadership and Mentorship:

Provide technical leadership and guidance to data engineers and other team members.

Mentor and coach junior engineers to develop their skills and expertise.

Foster a culture of innovation and collaboration within the team.

Qualifications:

Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field.

8+ years of experience in data engineering or a related field.

Strong understanding of data warehousing, data modeling, and ETL processes.

Expertise in designing and implementing large-scale data pipelines and architectures.

Proficiency in SQL and at least one programming language such as Python or Java.

Experience with Google Cloud Platform (GCP) services such as BigQuery, Dataflow, Dataproc, Pub/Sub, and Cloud Storage.

Experience with open-source data processing frameworks such as Hadoop, Spark, and Kafka.

Excellent communication, interpersonal, and collaboration skills.

Preferred Qualifications:

Experience with data governance and data quality management.

Experience with machine learning and data science.

Experience with containerization and orchestration technologies such as Docker and Kubernetes.

Contributions to open-source projects or communities.

Google Cloud Professional Data Engineer certification.