data pipelines for statistical and machine learning models
. The role involves close collaboration with data scientists and analysts to enable scalable model training, data ingestion, and real-time analytics.
Key Responsibilities:
Design, build, and maintain scalable data pipelines to support ML and statistical modeling.
Integrate structured and unstructured data from multiple sources into analytical environments.
Implement efficient ETL/ELT workflows for data preparation, cleaning, and feature engineering.
Collaborate with data science teams to ensure smooth deployment of ML models to production.
Optimize data workflows for performance, scalability, and reliability.
Manage data quality, lineage, and governance standards.
Requirements:
Bachelor's or Master's degree in Computer Science, Data Engineering, or a related field.
4-15 years of experience as a
Data Engineer
with exposure to
machine learning workflows
.
Strong proficiency in
Python
and SQL; experience with
data manipulation frameworks
(Pandas, PySpark, DuckDB).
Familiarity with
ML model lifecycle
, from data preprocessing to deployment.
Experience with
data pipeline tools
(Airflow, Prefect, dbt) and
cloud platforms
(AWS, Azure, GCP).
Knowledge of
containerization (Docker)
and
CI/CD pipelines
.
Experience working in a
quantitative, trading, or data-driven environment
is advantageous.
When you apply, you voluntarily consent to the disclosure, collection and use of your personal data for employment/recruitment and related purposes in accordance with the Tech Aalto Privacy Policy, a copy of which is published at Tech Aalto's website ( https://www.techaalto.com/privacy/)
Confidentiality is assured, and only shortlisted candidates will be notified for interviews.
Beware of fraud agents! do not pay money to get a job
MNCJobz.com will not be responsible for any payment made to a third-party. All Terms of Use are applicable.