I am a Data Engineer based in Brussels, with a diverse engineering background spanning over a decade. I am currently finalizing my MSc in Applied Computer Science while completing an intensive Data Engineering Masterclass focused on AWS, Big Data, and Modern Data Architectures.
My background in maintaining critical medical systems taught me the value of reliability, compliance (FDA/GDPR), and zero-downtime operations. Today, I apply that same discipline to build robust data pipelines. I specialize in Real-time Streaming (Kafka/Flink), Cloud Infrastructure, and Data Security.
Technologies I use to build reliable data pipelines.
Python (Pandas, PySpark, PyFlink), SQL (PostgreSQL, ClickHouse),Java, Bash.
Apache Airflow, Docker, Kubernetes.
Apache Kafka, Apache Flink, Spark Streaming, Debezium (CDC).
AWS (Glue, Athena, Lambda), GCP (BigQuery).
Grafana (dashboards/alerts), Tableau.
I help teams design, build, and monitor data systems from source to dashboard.
Building reliable Batch and Streaming pipelines to move data from sources to storage using Kafka & Flink.
Designing clean data models to ensure data is organized, fast, and ready for analytical queries.
Setting up Monitoring (Grafana) to catch issues early and ensure pipelines are running smoothly 24/7.
Implementing automated tests to ensure data is accurate, consistent, and trustworthy.
Creating clear Dashboards (Tableau) to help stakeholders visualize trends and make decisions.
Combining engineering discipline with modern data skills.
Intensive, project-based Data Engineering & MLOps bootcamp. Working hands-on with PostgreSQL, ClickHouse, Apache Kafka, Apache Spark, Apache Flink, Delta Lake, Docker, Kubernetes, Apache Airflow, MLflow, Jenkins and AWS.
Capstone projects include:
– Real-time fraud detection pipeline
(Kafka, Spark, PostgreSQL, MySQL, Redis, Grafana, AWS)
– End-to-end ML model deployment with monitoring and security on the cloud
– Building a secure Rust-based message broker (CipherMQ) as a product-focused project
Worked with Google Cloud Platform (GCP) and Power BI to visualize complex datasets. Supported ETL processes and created clear technical reports to help business stakeholders make decisions.
Maintained complex medical imaging systems with 99% uptime. Performed root-cause analysis on system logs and ensured strict compliance with FDA and local healthcare regulations.
Operated high-tech radiology equipment in a fast-paced clinical environment, delivering accurate results while following strict safety protocols.
Selected courses and certificates that support my data engineering path.
Data Engineering School · Issued Mar 2025
CI/CD, PostgreSQL, Git & GitHub, Kubernetes, Apache Airflow, Docker, AWS, Apache Kafka, ETL.
Snowflake · Issued Jun 2025
Data warehousing fundamentals and best practices on the Snowflake platform.
DataCamp · Issued Jun 2025
Analytics engineering with dbt for modular SQL transformations and testing.
Databricks · Issued Apr 2025
Building reliable ingestion pipelines with Delta Lake and the Lakehouse architecture.
Databricks · Issued Apr 2025
Data governance, access control and quality on top of the Databricks Lakehouse.
Maktabkhooneh · Issued Mar 2025
Core cloud concepts, security and cost-awareness for modern cloud platforms.
Coursera · Issued Jun 2021
Classical ML algorithms, logistic regression, neural networks and model evaluation.
Maktabkhooneh · Issued Apr 2021
Advanced Python for data processing, scripting and backend development.
Prefer email? saeidshahriari1@gmail.com