KAPITAL · 5 months ago
Senior Lead Azure Streaming Data Engineer
KAPITAL is seeking a Senior Lead Azure Streaming Data Engineer to design and build real-time data ingestion and processing pipelines. The role involves collaborating with the Data Architect to develop streaming solutions within Azure, ensuring data is captured, processed, and delivered efficiently to end users.
ConsultingData ManagementRecruitingStaffing Agency
Responsibilities
Design; Develop Real-Time Pipelines: Design and implement scalable, low- latency data pipelines to ingest real-time data from various sources (REST APIs, MQTT feeds, IoT sensors, etc.) using Azure streaming services such as Azure Event Hubs, Azure IoT Hub, Azure Stream Analytics, and Azure Functions
Stream Processing; Delta Lake Integration: Build streaming or micro-batch processing jobs (using Stream Analytics or Synapse Spark Structured Streaming) that cleanse, transform, and write event data into Delta Lake tables on ADLS Gen2 in the Raw and Enriched zones. Ensure that incoming data is captured in the Raw layer and promptly processed into Enriched for use in analytics
Real-Time Data Delivery: Enable real-time and near-real-time analytics by integrating streaming outputs with Power BI (GCC) real-time datasets and dashboards. Ensure that critical events (e.g. facility sensor alerts, airline flight alerts, passenger processing kiosk messages) are pushed to Power BI or other subscribers with minimal latency for situational awareness. Coordinate with the GIS team to incorporate location data (ArcGIS) into streaming insights when applicable (the GIS team will handle ArcGIS specifics, but this role will ensure data feeds include necessary spatial references)
Azure Best Practices; Security: Apply Azure best practices in all solutions – including secure networking (VNet integration, private endpoints for Event Hubs/Storage), authentication/authorization (managed identities, RBAC), encryption, and scalable design (throughput units, partitioning strategies, etc.). Ensure compliance with any relevant security or regulatory standards in this public-sector environment while optimizing for performance
Collaboration; Leadership: Work closely with the Client's Data Architect to review designs and align with the overall lakehouse framework. Provide technical leadership on streaming to more junior engineers if needed, and collaborate with stakeholders in operations and IT to understand real-time data needs. Partake in an agile-lite development process (task refinement, sprint planning, demos) and document your solutions for knowledge sharing
Qualification
Required
7+ years of data engineering experience with at least 2–3 years focused on real-time/streaming data pipeline development
Proven expertise in Azure streaming technologies (Azure Event Hubs, IoT Hub, Azure Stream Analytics) and related patterns (pub/sub, event processing)
Hands-on experience with Azure Synapse Analytics (or Azure Spark environments) to develop data pipelines
Ability to create and manage Synapse pipelines, triggers, and Synapse Spark notebooks for streaming or batch workflows
Familiarity with writing to and reading from Delta Lake storage
Proficiency in Python (PySpark) and/or Scala for Spark, plus experience authoring Azure Functions (in Python, C# or JavaScript) to handle streaming transformations or invoke APIs
Solid understanding of SQL for creating views or analyzing streaming results
Experience designing end-to-end streaming data solutions in Azure
Knowledge of networking, security, and scalability best practices (e.g., Event Hub partitioning, error handling and retry logic in Functions, scaling Stream Analytics jobs)
Ability to optimize pipelines for high throughput and low latency
Experience writing data to data lakes in a structured format (Delta/Parquet) and integrating with BI tools
Understanding of how to output streaming data to Power BI (e.g., through Stream Analytics outputs or Power BI REST endpoints) or similar real-time visualization platforms
Excellent communication skills to work with cross-functional teams
Ability to translate business requirements for real-time data into technical solutions
Experience in a lead or architect capacity on data projects is required, as this is a senior role needing self-direction and mentorship abilities
Preferred
Microsoft Azure certifications such as Azure Data Engineer and/or Azure Solutions Architect
Familiarity with other streaming and IoT technologies (e.g., Kafka, MQTT brokers, Azure Event Grid, Azure Data Explorer)
Experience with big data file formats and tools (Apache Parquet/Delta, Kafka, Spark) and implementing CI/CD for data pipelines (Azure DevOps Pipelines, YAML deployments)
Knowledge of infrastructure-as-code (ARM/Bicep or Terraform) for deploying data services is a plus
Prior experience in secure or regulated environments (government agencies, utilities, healthcare, etc.) dealing with sensitive data
Understanding of compliance requirements and how to design data solutions within those constraints