Galent ยท 21 hours ago
Data Architect
Galent is a digital engineering firm that brings AI-driven innovation to enterprise IT. They are seeking a Data Architect to design and implement end-to-end data architecture for batch and real-time data processing, ensuring data quality and governance across platforms.
Responsibilities
Design and implement end-to-end data architecture for batch and real-time data processing to implement enterprise-grade data solutions using Azure and Microsoft Fabric
Design, build, and optimize ETL/ELT pipelines for ingestion, transformation, and publishing using ADF, Spark, and Databricks
Architect and optimize data pipelines using Azure Data Factory (ADF)
Develop and manage scalable data processing frameworks using Databricks and Apache Spark
Develop efficient data transformation logic using PySpark and Spark SQL
Build reusable, high-performance data models for analytics and reporting
Develop and maintain data ingestion, transformation, and orchestration workflows
Ensure data quality, consistency, security, and governance across platforms
Define and enforce data engineering best practices, including CI/CD, versioning, testing, and monitoring
Mentor and guide data engineers, fostering a culture of innovation and excellence
Collaborate with data engineers, analysts, data scientists, and business stakeholders
Qualification
Required
Azure Data Factory (ADF)
Azure Databricks & PySpark
Azure Synapse
Azure SQL
Python
Spark SQL
12 + years of IT experience in the Development/ Architecture Role
5+ years of Proven experience in Data Engineering and architecture roles.(with Azure Data Factory , Azure Databricks & PySpark, Azure Synapse and Azure SQL)
Bachelor's or master's degree in computer science, Engineering, or a related field
Strong Experience with ETL/ELT tools like ADF, Informatica , Talend etc., and data warehousing technologies like Azure Synapse, Azure SQL, Amazon redshift , Snowflake , Google Big Query etc
Strong hands-on experience with Azure Data Factory (ADF) for data orchestration (for building and managing pipelines), Azure Databricks for big data processing and analytics, and Apache Spark for distributed data processing
Adept in with big data tools(Databricks , Spark etc..)
Experience with Power BI, Tableau/OBIEE etc
Proficiency in PySpark, Python, and Spark SQL
Experience with CI/CD pipelines for data platforms
Solid understanding of data warehouse best practices, development standards and methodologies
Strong expertise in SQL and relational data modeling
Experience designing data lakes, data warehouses, or Lakehouse architectures
Good understanding of data governance, metadata management, and security best practices
Solid understanding of distributed data processing and performance tuning
Experience with ETL/ELT patterns and best practices
Strong analytical, problem-solving skills and ability to work in a fast-paced, dynamic environment
Excellent communication and documentation skills
Company
Galent
Galent is an AI-native digital engineering firm at the forefront of the AI revolution, dedicated to delivering unified, enterprise-ready AI solutions that transform businesses and industries.
Funding
Current Stage
Late StageCompany data provided by crunchbase