Skip to content

ashkangoleh/ashkangoleh

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

23 Commits
ย 
ย 

Repository files navigation

๐Ÿ‘‹ Hi, I'm Ashkan Golehpour

๐Ÿ’ผ Data / Software Engineer | Data Platform Architect

A dedicated Data Engineer with a strong background in Software Architecture and Microservices, focused on building scalable, reliable, and high-performance data systems.
I specialize in transforming complex datasets into actionable insights using modern open-source technologies such as ClickHouse, Polars, DuckDB, and Kafka โ€” all while ensuring data quality, observability, and performance.


๐Ÿš€ Summary

Over 5 years of hands-on experience designing and implementing end-to-end data platforms, from ingestion and orchestration to serving APIs.
Currently contributing to market-surveillance and real-time analytics systems at Iran FaraBourse (IFB) and leading data architecture design at a BioTech startup in Sweden.


๐Ÿง  Core Competencies

๐Ÿงฉ Data Engineering & Analytics

  • Data ingestion & transformation (ETL/ELT) using Prefect, Airflow, MageAI
  • Real-time & streaming pipelines with Kafka, Redpanda, and ClickHouse
  • DataLakehouse architecture with Iceberg, DuckLake, MinIO (S3)
  • Advanced analytics using Polars, DuckDB, and Apache Arrow

โš™๏ธ Backend & Microservices

  • FastAPI, gRPC, AsyncIO, Celery for high-performance APIs
  • Designed an internal FaaS platform at IFB for dynamic, scalable gRPC endpoints
  • Experience with distributed systems and event-driven design

๐Ÿ“Š Observability & Monitoring

  • Built real-time data quality monitoring with Prometheus, Grafana, and OpenTelemetry
  • Implemented unified logging and metrics systems across multiple data services

๐Ÿงฑ Databases & Query Optimization

  • ClickHouse, PostgreSQL, MongoDB, Neo4j, Qdrant, Elasticsearch
  • Schema design, indexing strategies, and query optimization for OLAP/OLTP workloads

๐Ÿค– AI & Intelligent Systems

  • Integration of LangChain and OpenAI for document embedding and semantic search
  • Built vector search pipelines using Qdrant (768D cosine embeddings)
  • Designed microservices for LLM-powered analytics and biomedical knowledge graphs

๐Ÿข Professional Experience

Stealth Mode BioTech Startup (Sweden) โ€” Data Engineer (Advisory, Remote)

Jan 2025 โ€“ Present

  • Designed full data & analytics workflows using DuckDB, Polars & Apache Arrow
  • Modeled a biomedical knowledge graph in Neo4j and implemented vector search with Qdrant
  • Developed a gRPC microservice exposing vector retrieval and search APIs

Iran FaraBourse (IFB) โ€” Data Engineer (Full-time, On-site)

Oct 2024 โ€“ Present

  • Built in-house FaaS platform with dynamic gRPC endpoints serving 5+ internal systems
  • Refactored 10+ Prefect pipelines, improving reliability from 85% โ†’ 99.9%
  • Migrated analytical schemas to ClickHouse, achieving 40% faster queries

Bit24 (Crypto Exchange) โ€” Data Engineer (Full-time, Remote)

Sep 2023 โ€“ Dec 2024

  • Built ETL/data pipelines with Apache Airflow & MageAI
  • Deployed PostgreSQL-based warehouse cutting analytics cost by 30%
  • Developed data quality monitoring via Prometheus + Grafana

Edgecom Energy (Canada) โ€” Data Engineer (Contract, Remote)

Sep 2023 โ€“ Jan 2024

  • Refactored pipelines to reduce latency and improve I/O throughput
  • Designed gRPC + SQS integrations for data flow between services

Digid (Canada) โ€” Data Engineer (Contract, Remote)

May 2023 โ€“ Sep 2023

  • Built ingestion pipelines reducing data lag by 2 hours
  • Used Azure OpenAI to generate SWOT analyses from embedded documents

ArzDigital (Iran) โ€” Data Engineer (Full-time, On-site)

Aug 2021 โ€“ May 2023

  • Developed trading assistant (Atlas) using Kafka & PostgreSQL
  • Reduced validation time by 40%, enabling low-latency trading insights

๐Ÿ› ๏ธ Tech Stack

Languages: Python, SQL, Shell, Rust (Intermediate)
Frameworks: FastAPI, gRPC, AsyncIO, Celery
Data Tools: ClickHouse, Kafka, DuckDB, Polars, Prefect, Airflow, MageAI
Storage & Lakehouse: MinIO, Iceberg, DuckLake, Dremio, Nessie
Databases: PostgreSQL, MongoDB, Neo4j, Qdrant, Elasticsearch
Monitoring: Prometheus, Grafana, OpenTelemetry
DevOps: Docker, GitHub Actions, GitLab CI, Kubernetes
Visualization: Metabase, Superset, Grafana


๐Ÿงพ Certifications

  • ๐ŸŽ“ Dremio Verified Lakehouse Associate
  • ๐Ÿงฑ Data Engineering Essentials
  • ๐ŸŒ€ ETL and Data Pipelines with Shell, Airflow, and Kafka
  • ๐Ÿ—ƒ๏ธ Relational Database Administration Essentials
  • ๐Ÿ Python Project for Data Engineering
  • ๐Ÿงฎ LangChain for LLM Application Development

๐ŸŽ“ Education

Bachelorโ€™s Degree in Software Engineering
Islamic Azad University (2014 โ€“ 2018) | GPA: 3.05/4.0


๐ŸŒ Connect with Me

LinkedIn
GitHub
Medium


๐Ÿ“Š โ€œData without context is noise โ€” my mission is to turn that noise into signal.โ€

About

Config files for my GitHub profile.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published