Data Architect jobs in United States
cer-icon
Apply on Employer Site
company-logo

Confidential Jobs · 1 day ago

Data Architect

Confidential Jobs is seeking an Azure Data Architect with a strong focus on Databricks and AI/ML enablement. The role involves designing and implementing cloud-native data architectures while integrating AI/ML pipelines and ensuring effective stakeholder engagement.

Computer Software
Hiring Manager
Suresh K
linkedin

Responsibilities

12+ years in data engineering or architecture, with a strong focus on Databricks (at least 4-5 years) and AI/ML enablement
Deep hands-on experience with Apache Spark, Delta Lake
Proficiency in AI/ML pipeline integration using Databricks MLflow or custom model deployment strategies
Strong knowledge of Apache Airflow, Databricks Jobs, and cloud-native orchestration patterns
Experience with structured streaming, Kafka, and real-time analytics frameworks
Proven ability to design and implement cloud-native data architectures
Solid understanding of data modeling, Lakehouse design principles, and lineage/tracking with Unity Catalog
Excellent communication and stakeholder engagement skills
Preferred Qualifications Certification in Databricks Data Engineering Professional is highly desirable
Experience transitioning from in house data platforms to Databricks or cloud-native environments
Hands-on experience with Delta Lake, Unity Catalog, and performance tuning in Databricks
Expertise in Apache Airflow DAG design, dynamic workflows, and production troubleshooting
Experience with CI/CD pipelines, Infrastructure-as-Code (Terraform, ARM templates), and DevOps practices
Exposure to AI/ML model integration within real-time or batch data pipelines
Exposure to MLOps, MLflow, Feature Store, and model monitoring in production environments
Experience with LLM/GenAI enablement, vectorized data, embedding storage, and integration with Databricks is an added advantage

Qualification

DatabricksAI/MLApache SparkCloud-native architectureDelta LakeApache AirflowData modelingCI/CD pipelinesStakeholder engagementCommunication skills

Required

12+ years in data engineering or architecture, with a strong focus on Databricks (at least 4-5 years) and AI/ML enablement
Deep hands-on experience with Apache Spark, Delta Lake
Proficiency in AI/ML pipeline integration using Databricks MLflow or custom model deployment strategies
Strong knowledge of Apache Airflow, Databricks Jobs, and cloud-native orchestration patterns
Experience with structured streaming, Kafka, and real-time analytics frameworks
Proven ability to design and implement cloud-native data architectures
Solid understanding of data modeling, Lakehouse design principles, and lineage/tracking with Unity Catalog
Excellent communication and stakeholder engagement skills

Preferred

Certification in Databricks Data Engineering Professional is highly desirable
Experience transitioning from in house data platforms to Databricks or cloud-native environments
Hands-on experience with Delta Lake, Unity Catalog, and performance tuning in Databricks
Expertise in Apache Airflow DAG design, dynamic workflows, and production troubleshooting
Experience with CI/CD pipelines, Infrastructure-as-Code (Terraform, ARM templates), and DevOps practices
Exposure to AI/ML model integration within real-time or batch data pipelines
Exposure to MLOps, MLflow, Feature Store, and model monitoring in production environments
Experience with LLM/GenAI enablement, vectorized data, embedding storage, and integration with Databricks is an added advantage

Company

Confidential Jobs

twitter
company-logo
Confidential Jobs is operated by ExecThread, Inc. (https://execthread.com).

Funding

Current Stage
Growth Stage

Leadership Team

leader-logo
Amy Nelson
Managing Partner
linkedin
leader-logo
Brandon Nolan
Chief Information Security Officer
linkedin
Company data provided by crunchbase