ALOIS Solutions ยท 23 hours ago
Python and PySpark Developer
Maximize your interview chances
Insider Connection @ALOIS Solutions
Get 3x more responses when you reach out via email instead of LinkedIn.
Responsibilities
Design and implement scalable data pipelines using PySpark and other big data frameworks.
Develop reusable and efficient code for data extraction, transformation, and loading (ETL).
Optimize data workflows for performance and cost efficiency.
Process and analyze structured and unstructured datasets.
Build and maintain data lakes, data warehouses, and other storage solutions.
Collaborate with cross-functional teams to understand business requirements and translate them into technical solutions.
Troubleshoot and resolve performance bottlenecks in big data pipelines.
Write clean, maintainable, and well-documented code.
Ensure compliance with data governance and security policies.
Qualification
Find out how your skills align with this job's requirements. If anything seems off, you can easily click on the tags to select or unselect skills to reflect your actual expertise.
Required
Proficient in Python with experience in data processing libraries like Pandas and NumPy.
Strong experience with PySpark and Apache Spark.
Hands-on experience with big data platforms such as Hadoop, Databricks, or similar.
Familiarity with cloud services like AWS (EMR, S3), Azure (Data Lake, Synapse), or Google Cloud (BigQuery, Dataflow).
Strong knowledge of SQL and NoSQL databases.
Experience working with relational databases like PostgreSQL, MySQL, or Oracle.
Experience with workflow orchestration tools like Apache Airflow or similar.
Ability to solve complex data engineering problems efficiently.
Strong communication skills to work effectively in a collaborative environment.
Benefits
Health/medical/dental benefits
PTO