Generis Tek Inc · 12 hours ago
Databricks DevOps Engineer/ Data Platform Engineer-Hybrid
Generis Tek Inc is seeking a Databricks DevOps Engineer/Data Platform Engineer to design, develop, and maintain a high-scale, cloud-based data platform. The role involves implementing robust data pipelines and establishing best-in-class DevOps practices to ensure automated deployment and reliability.
Responsibilities
Platform Development: Design, build, and maintain scalable, efficient, and reliable ETL/ELT data pipelines to support data ingestion, transformation, and integration across diverse sources
Big Data Implementation: Serve as the subject matter expert for the Databricks environment, developing high-performance data transformation logic primarily using PySpark and Python. This includes utilizing Delta Live Tables (DLT) for declarative pipeline construction and ensuring governance through Unity Catalog
Cloud Infrastructure Management: Configure, maintain, and secure the underlying AWS cloud infrastructure required to run the Databricks platform, including virtual private clouds (VPCs), network endpoints, storage (S3), and cross-account access mechanisms
DevOps & Automation (CI/CD): Own and enforce Continuous Integration/Continuous Deployment (CI/CD) practices for the data platform. Specifically, design and implement automated deployment workflows using GitHub Actions and modern infrastructure-as-code concepts to deploy Databricks assets (Notebooks, Jobs, DLT Pipelines, and Repos)
Data Quality & Testing: Design and implement automated unit, integration, and performance testing frameworks to ensure data quality, reliability, and compliance with architectural standards
Performance Optimization: Optimize data workflows and cluster configurations for performance, cost efficiency, and scalability across massive datasets
Technical Leadership: Provide technical guidance on data principles, patterns, and best practices (e.g., Medallion Architecture, ACID compliance) to promote team capabilities and maturity. This includes leveraging Databricks SQL for high-performance analytics
Documentation & Review: Draft and review architectural diagrams, design documents, and interface specifications to ensure clear communication of data solutions and technical requirements
Qualification
Required
Experience: 5+ years of professional experience in Data Engineering, focusing on building scalable data platforms and production pipelines
Big Data Expertise: Minimum 3+ years of hands-on experience developing, deploying, and optimizing solutions within the Databricks ecosystem. Deep expertise required in: Delta Lake (ACID transactions, time travel, optimization). Unity Catalog (data governance, access control, metadata management). Delta Live Tables (DLT) (declarative pipeline development). Databricks Workspaces, Repos, and Jobs. Databricks SQL for analytics and warehouse operations
AWS Infrastructure & Security: Proven, hands-on experience (3+ years) with core AWS services and infrastructure components, including: Networking: Configuring and securing VPCs, VPC Endpoints, Subnets, and Route Tables for private connectivity. Security & Access: Defining and managing IAM Roles and Policies for secure cross-account access and least privilege access to data. Storage: Deep knowledge of Amazon S3 for data lake implementation and governance
Programming: Expert proficiency (4+ years) in Python for data manipulation, scripting, and pipeline development
Spark & SQL: Deep understanding of distributed computing and extensive experience (3+ years) with PySpark and advanced SQL for complex data transformation and querying
DevOps & CI/CD: Proven experience (2+ years) designing and implementing CI/CD pipelines, including proficiency with GitHub Actions or similar tools (e.g., GitLab CI, Jenkins) for automated testing and deployment
Data Concepts: Full understanding of ETL/ELT, Data Warehousing, and Data Lake concepts
Methodology: Strong grasp of Agile principles (Scrum)
Version Control: Proficiency with Git for version control
Preferred
AWS Data Ecosystem Experience: Familiarity and experience with AWS cloud-native data services, such as AWS Glue, Amazon Athena, Amazon Redshift, Amazon RDS, and Amazon Dynamo DB
Knowledge of real-time or near-real-time streaming technologies (e.g., Kafka, Spark Structured Streaming)
Experience in developing feature engineering pipelines for machine learning (ML) consumption
Background in performance tuning and capacity planning for large Spark clusters
Company
Generis Tek Inc
Generís Tek is an elite IT staffing firm headquartered in Chicago, IL offering long-term, short-term, temporary-to-permanent and direct placement staffing.
Funding
Current Stage
Growth StageCompany data provided by crunchbase