Senior Data Architect jobs in United States
cer-icon
Apply on Employer Site
company-logo

Divit Technologies, Inc. ยท 14 hours ago

Senior Data Architect

Divit Technologies, Inc. is seeking a Senior Data Architect to develop and evolve data architectures and pipelines. The role involves building event-driven services, mentoring junior team members, and collaborating with various stakeholders to leverage data for business insights.

Information Technology & Services
check
H1B Sponsor Likelynote
Hiring Manager
A Kumar
linkedin

Responsibilities

Develop data pipelines responsible for ingesting large amounts of different kinds of data from various sources
Help evolve data architecture and work on Next Generation real time pipeline algorithms and architecture in addition to supporting and maintaining current pipelines and legacy systems
Write code and develop worker nodes for business logic, ETL and orchestration processes
Develop algorithms for better attribution rules and category classifiers
Work with stakeholders throughout the organization to identify opportunities for leveraging company data to drive search, discovery, and recommendations
Work closely with architects, engineers, data analysts, data scientists, contractors/consultants and project managers in assessing project requirements, design, develop and support data ingestions and API services
Work with Data Scientists in building feature engineering pipelines and integrating machine learning models during the content enrichment process
Able to influence on priorities working with various partners including engineers, project management office and leadership
Mentor junior team members, define architecture, code review, hands-on development and deliver the work in sprint cycle
Participate in design discussions with Architects and other team members for the design of new systems and re-engineering of components of existing systems
Wear an Architect hat when required to bring new ideas to the table, thought leadership and forward thinking
Take a holistic approach to building solutions by thinking of the big picture and overall solution
Work on moving away from legacy systems into next generation architecture
Take complete ownership from requirements, solution design, development, production launch and post launch production support. Participate in code reviews and regular on-call rotations
Desire to apply the best solution in the industry, apply correct design patterns during development and learn best practices and data engineering tools and technologies
Performs any other functions and duties assigned and necessary for the smooth and efficient operation

Qualification

JavaKafkaSparkPythonScalaApache Kafka ConnectorsETLSQLAWSNoSQL DatabasesMachine LearningDebugging SkillsCommunication SkillsProblem-Solving

Required

Must have a minimum of 10-12 years of hands-on development experience implementing batch and events driven applications using Java, Kafka, Spark, Scala, PySpark and Python
Experience with Apache Kafka and Connectors, Java, Springboot in building event driven services, Python in building ML pipelines
Develop data pipelines responsible for ingesting large amounts of different kinds of data from various sources
Help evolve data architecture and work on Next Generation real time pipeline algorithms and architecture in addition to supporting and maintaining current pipelines and legacy systems
Write code and develop worker nodes for business logic, ETL and orchestration processes
Develop algorithms for better attribution rules and category classifiers
Work with stakeholders throughout the organization to identify opportunities for leveraging company data to drive search, discovery, and recommendations
Work closely with architects, engineers, data analysts, data scientists, contractors/consultants and project managers in assessing project requirements, design, develop and support data ingestions and API services
Work with Data Scientists in building feature engineering pipelines and integrating machine learning models during the content enrichment process
Able to influence on priorities working with various partners including engineers, project management office and leadership
Mentor junior team members, define architecture, code review, hands-on development and deliver the work in sprint cycle
Participate in design discussions with Architects and other team members for the design of new systems and re-engineering of components of existing systems
Wear an Architect hat when required to bring new ideas to the table, thought leadership and forward thinking
Take a holistic approach to building solutions by thinking of the big picture and overall solution
Work on moving away from legacy systems into next generation architecture
Take complete ownership from requirements, solution design, development, production launch and post launch production support. Participate in code reviews and regular on-call rotations
Desire to apply the best solution in the industry, apply correct design patterns during development and learn best practices and data engineering tools and technologies
Performs any other functions and duties assigned and necessary for the smooth and efficient operation
BS or MS in Computer Science (or related field) with 12+ years of hands-on software development experience working in large-scale data processing pipelines
Must have skills are Apache Spark, Scala and PySpark with 2-4 years of experience building production grade batch pipelines that handle large volumes of data
Must have at least 8+ years of experience in Java and API / Microservices
Must have at least 5+ years of experience in Python
5+ years of experience in understanding and writing complex SQL and stored procedures for processing raw data, ETL, data validation, using databases such as SQL Server, Redis and other NoSQL DBs
Knowledge of Big data technologies, Hadoop, HDFS
Expertise with building events driven pipelines with Kafka and Java / Spark
Expertise with Amazon AWS stack such as EMR, EC2, S3
Experience working with APIs to collect and ingest data as well build the APIs for business logic
Experience working with setting up, maintaining, and debugging production systems and infrastructure
Experience in building fault-tolerant and resilient systems
Experience in building worker nodes, knowledge of REST principles and data engineering design patterns
In-depth knowledge of Java, SpringBoot, Spark, Scala, PySpark, Python, Orchestration tools, ESB, SQL, Stored procedures, Docker, RESTful web services, Kubernetes, CI/CD, Observability techniques, Kafka, Release processes, caching strategies, versioning, B&D, BitBucket / Git and AWS Cloud Ecosystem, NoSQL Databases, Hazelcast
Strong software development, architecture diagramming, problem-solving and debugging skills
Phenomenal communication and influencing skills

Preferred

Exposure to Machine Learning (ML), LLM models, using AI during coding, build with AI
Knowledge of Elastic APM, ELK stack and search technologies such as Elasticsearch / Solr
Some experience in workflow orchestration tools such as Air Flow or Apache NiFi

Company

Divit Technologies, Inc.

twitter
company-logo
Divit Technologies Inc. specialises in AI & Automation, Data Science, DevOps Security, Quality Engineering, and IT Staffing Solutions.

H1B Sponsorship

Divit Technologies, Inc. has a track record of offering H1B sponsorships. Please note that this does not guarantee sponsorship for this specific role. Below presents additional info for your reference. (Data Powered by US Department of Labor)
Distribution of Different Job Fields Receiving Sponsorship
Represents job field similar to this job
Trends of Total Sponsorships
2025 (6)
2024 (1)

Funding

Current Stage
Growth Stage
Company data provided by crunchbase