SageCor Solutions · 14 hours ago
Software Engineer 2 (IDN - 143)
SageCor Solutions is a growing company providing complete engineering services and full lifecycle System Engineering services. They are seeking a Software Engineer 2 with expertise in high performance computing and large data analytics to contribute to their projects.
HardwareInformation TechnologySoftware
Responsibilities
Experience using the Linux CLI and Linux tools
Experience developing Bash scripts to automate manual processes
Recent software development experience using Python
Familiar with Distributed Big Data processing engines including Apache Spark
Experience using Jupyter Notebook
Experience with data wrangling and preprocessing tools such as pandas and NumPy
Experience working with structured, semi-structured, and unstructured data
Familiarity with data quality concepts, data validation, and anomaly detection
Experience with Git Source Control System
Familiar with Apache Airflow (DAG design, scheduling, operators, sensors) to orchestrate, schedule, and monitor complex workflows
Familiar with SQL technologies such as MySQL, MariaDB, and PostgreSQL for querying, joining, and aggregating large datasets
Familiar with HPC Job Scheduling tools including Slurm
Experience using the Atlassian Tool Suite (JIRA, Confluence)
Qualification
Required
Active TS/SCI W/ Polygraph Required
Master's degree in computer science or related discipline from an accredited college or university, plus three (3) years of experience as a SWE, in programs and contracts of similar scope, type, and complexity
OR Bachelor's degree in computer science or related discipline from an accredited college or university, plus five (5) years of experience as a SWE, in programs and contracts of similar scope, type, and complexity
OR Seven (7) years of experience as a SWE, in programs and contracts of similar scope, type, and complexity
Experience using the Linux CLI and Linux tools
Experience developing Bash scripts to automate manual processes
Recent software development experience using Python
Familiar with Distributed Big Data processing engines including Apache Spark
Experience using Jupyter Notebook
Experience with data wrangling and preprocessing tools such as pandas and NumPy
Experience working with structured, semi-structured, and unstructured data
Familiarity with data quality concepts, data validation, and anomaly detection
Experience with Git Source Control System
Preferred
Familiar with Apache Airflow (DAG design, scheduling, operators, sensors) to orchestrate, schedule, and monitor complex workflows
Familiar with SQL technologies such as MySQL, MariaDB, and PostgreSQL for querying, joining, and aggregating large datasets
Familiar with HPC Job Scheduling tools including Slurm
Experience using the Atlassian Tool Suite (JIRA, Confluence)