# Duyet Le | Résumé me@duyet.net · https://duyet.net · https://github.com/duyet · https://linkedin.com/in/duyet Data Engineer with 6+ years of experience in modern data warehousing, distributed systems, and cloud computing. Proficient in ClickHouse, Spark, Airflow, Python, Rust. ## Experience ### Sr. Data Engineer Cartrack (Oct 2023 - Present) • Deprecated old stack (Spark, Iceberg, Trino) replaced by ClickHouse. • Migrated 350TB+ Iceberg Data Lake to ClickHouse on Kubernetes. • Enhanced ClickHouse for 300% better data compression and 2x-100x faster queries, compared with Trino + Iceberg • Automated operations with Airflow: data replication, data processing, healthchecks, etc. • Multi-agent LLM + RAG (LlamaIndex, Qdrant, ClickHouse text2sql, Nextjs, etc) ### Sr. Data Engineer Fossil Group Inc (Oct 2018 - Jul 2023) • Optimize monthly costs from $45,000 to $20,000 (GCP and AWS Cloud). • Managed a team of 4 data engineers and 2 data analysts to provide end-to-end analytics solutions to stakeholders. Raised data-driven awareness throughout the organization and encouraged everyone to take a more data-driven approach to problem-solving. • Designed next-gen Data Platform in Rust ↗︎ • Developed tools for Data Monitoring, Data Catalog, and Self-service Analytics for internal teams with everything deployed on Kubernetes. ### Sr. Data Engineer FPT Software (Jun 2017 - Oct 2018) • Built data pipelines processing 2TB/day with AWS for a Recommendation System • Ingested and transformed 1TB+/day into Data Lake using Azure Cloud and Databricks ### Data Engineer John von Neumann Institute (Sep 2015 - Jun 2017) • Developed data pipelines, data cleaning and visualizations for adhoc problems. • Train and deployed ML models: customer lifetime value, churn prediction, sales optimization, recruitment optimization, etc. ## Education ### Bachelor's degree, Information System University of Information Technology Thesis: Network of career skills and support an optimal job search Thesis URL: https://arxiv.org/pdf/1707.09751 ## Skills **Data Engineering:** ClickHouse, Spark, Kafka, Airflow, AWS, BigQuery, Data Studio, Python, Rust, Typescript. **DevOps:** CI/CD, Kubernetes, Helm. --- This file follows the llms.txt standard for providing CV information to Large Language Models and AI assistants. Generated from: https://duyet.net