Google Cloud Platform Data Engineer jobs in United States
info-icon
This job has closed.
company-logo

Jobs via Dice · 4 hours ago

Google Cloud Platform Data Engineer

Dice is the leading career destination for tech experts at every stage of their careers. Our client, Data Capital Inc, is seeking an experienced Senior Data Engineer with strong expertise in Spark, Streaming technologies, and Google Cloud Platform to design and build real-time, scalable data pipelines.

Computer Software

Responsibilities

Design, develop, and maintain scalable ETL/ELT pipelines for batch and real-time data processing
Build and optimize real-time data pipelines using Apache Spark (PySpark/Scala) and streaming technologies such as Kafka, Flink, and Google Cloud Platform Pub/Sub
Develop and manage data architectures including data lakes, BigQuery data warehouses, and streaming platforms
Perform performance tuning of Spark jobs, Spark SQL queries, and data workflows for efficiency and cost optimization
Implement data quality checks, monitoring, and alerting to ensure data reliability and consistency
Work extensively with Google Cloud Platform services such as BigQuery, Dataproc, Pub/Sub, Vertex AI, and Cloud Functions
Write efficient, scalable code using Python, SQL, PySpark, and Spark SQL
Apply strong data modeling techniques to support analytics and ML use cases
Collaborate with analytics, ML, and business teams to deliver high-quality data solutions
Utilize tools like Airflow, Databricks, Docker, and Kubernetes where applicable

Qualification

Google Cloud PlatformApache SparkETL/ELT pipelinesPythonSQLData modelingKafkaFlinkBigQueryAirflowDatabricksDockerKubernetes

Required

Strong expertise in Spark, Streaming technologies, and Google Cloud Platform
Design, develop, and maintain scalable ETL/ELT pipelines for batch and real-time data processing
Build and optimize real-time data pipelines using Apache Spark (PySpark/Scala) and streaming technologies such as Kafka, Flink, and Google Cloud Platform Pub/Sub
Develop and manage data architectures including data lakes, BigQuery data warehouses, and streaming platforms
Perform performance tuning of Spark jobs, Spark SQL queries, and data workflows for efficiency and cost optimization
Implement data quality checks, monitoring, and alerting to ensure data reliability and consistency
Work extensively with Google Cloud Platform services such as BigQuery, Dataproc, Pub/Sub, Vertex AI, and Cloud Functions
Write efficient, scalable code using Python, SQL, PySpark, and Spark SQL
Apply strong data modeling techniques to support analytics and ML use cases
Collaborate with analytics, ML, and business teams to deliver high-quality data solutions
Utilize tools like Airflow, Databricks, Docker, and Kubernetes where applicable

Company

Jobs via Dice

twitter
company-logo
Welcome to Jobs via Dice, the go-to destination for discovering the tech jobs you want.

Funding

Current Stage
Early Stage
Company data provided by crunchbase