Aionics, Inc. · 5 hours ago
Senior Data Engineer
Maximize your interview chances
Advanced Materials
Insider Connection @Aionics, Inc.
Get 3x more responses when you reach out via email instead of LinkedIn.
Responsibilities
Design, build, and maintain scalable and reliable data pipelines for processing structured datasets.
Acquire, transform, manipulate, and analyze data from multiple sources to ensure data is accurate, consistent, and of high quality.
Develop robust ETL processes using Python and SQLAlchemy to automate data transformations and ensure data integrity.
Use Airflow for workflow orchestration and automation of data pipelines.
Collaborate with cross-functional teams to prepare data for model training and analysis.
Optimize database performance to ensure fast, efficient access to datasets.
Implement data governance policies, ensuring data quality, security, and compliance with relevant regulations.
Continuously monitor and improve data workflows, incorporating best practices in data engineering and agile methodologies.
Manage data engineering projects, ensuring timely delivery and clear communication across teams.
Write clean, maintainable code and documentation for all data-related workflows and systems.
Contribute to version control and code reviews using Git, ensuring collaborative, efficient code management across the team.
Qualification
Find out how your skills align with this job's requirements. If anything seems off, you can easily click on the tags to select or unselect skills to reflect your actual expertise.
Required
Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field, or equivalent experience.
5 years or more hands-on experience as a Data Engineer, with expertise in designing and managing complex data architectures.
Proficiency in Python and experience with data processing libraries such as Pandas, Dask, and PySpark.
Hands-on experience with dbt (data build tool) for transforming data and managing data pipelines.
Experience with Airflow for orchestrating workflows and managing data pipelines.
Experience working with relational databases (e.g., PostgreSQL, MySQL) and writing complex SQL queries.
Familiarity with cloud platforms such as AWS, GCP, or Azure.
Strong knowledge of ETL processes, data modeling, and pipeline orchestration tools.
Experience with data quality assurance practices, ensuring consistent and reliable datasets.
Proficiency in Git for version control and collaboration.
Experience working in agile development environments, with familiarity in agile methodologies.
Strong project management skills, with the ability to manage multiple projects and communicate clearly with stakeholders.
Preferred
Experience in machine learning and familiarity with the data needs for AI model training.
Knowledge of computational materials science would be a plus.
Familiarity with containerization tools like Docker.
Benefits
Medical, dental, and vision insurance
Unlimited paid time off and sick leave
401(k) plan and paid parental leave
Two-week full company holiday at year’s end