Machine Learning Engineer: 1st July 2025
🔧 Company Engineering Blogs
Normalizing Flows Are Capable Generative Models (machinelearning.apple.com). TarFlow, a Transformer-based Normalizing Flows model, achieves state-of-the-art results in likelihood estimation and image generation with advanced techniques for improved quality
Using generative AI to do multimodal information retrieval (amazon.science). GENIUS model enhances multimodal information retrieval using generative AI, outperforming traditional methods in speed and accuracy with innovative techniques like semantic quantization
🎓 ML Research Insights & Advances
Researchers Uncover Hidden Ingredients Behind AI Creativity (quantamagazine.org). Study reveals diffusion models' creativity stems from architectural imperfections in denoising, linking AI creativity with local patch processing techniques
Lessons Learned After 6.5 Years Of Machine Learning (towardsdatascience.com). Lessons on deep work, trend management, and data analysis from 6.5 years in machine learning, covering personal insights and practical research experiences
Normalizing Flows Are Capable Generative Models (machinelearning.apple.com). TarFlow, a Transformer-based Normalizing Flows model, achieves state-of-the-art results in likelihood estimation and image generation with advanced techniques for improved quality
🚀 Data Engineering & Performance Tools
Polars Boosted My Algorithm's Speed by 25x (john.soban.ski). Polars library enhances algorithm speed by 25x, outperforming Pandas with parallel processing and Apache Arrow in Reduced Columb Energy classification
The Small Data Showdown ‘25: Is it Time to Ditch Spark Yet?? (mwc360.github.io). An exploration of Spark, DuckDB, Polars, and Daft performance for small data workloads in Microsoft Fabric; assessing if Spark remains viable
Thank you, Databend; Hello, LanceDB (xuanwo.io). Xuanwo transitions from Databend to LanceDB, focusing on open-source multimodal databases, Rust optimizations, and personal milestones in tech and life
Pipelining AI/ML Training Workloads with CUDA Streams (towardsdatascience.com). Explore CUDA streams for pipelining AI/ML training workloads, enhancing performance in PyTorch models through concurrent GPU operations and data augmentation techniques
🔍 Embeddings & Information Retrieval
Predicting Average IMDb Movie Ratings Using Text Embeddings of Movie Metadata (minimaxir.com). Predicting IMDb movie ratings using text embeddings from metadata; explores neural networks, LLMs, and data wrangling with Polars
What Are Vector Embeddings? And Why They Matter in AI (sefiks.com). Explore vector embeddings, their role in AI, applications in facial recognition, image search, LLMs, and concepts like PCA for dimensionality reduction
MUVERA: Making multi-vector retrieval as fast as single-vector search (research.google). MUVERA enhances efficiency in multi-vector retrieval using fixed dimensional encodings, transforming complex queries into single-vector maximum inner product searches
🏭 Production ML & Deployment Systems
Building a dynamic inventory optimisation system: A deep dive (engineering.zalando.com). Dynamic inventory optimization system at Zalando using AI, MLForecast, LightGBM, and AWS tools to enhance e-commerce supply chain efficiency
#511: From Notebooks to Production Data Science Systems (talkpython.fm). Catherine Nelson discusses transitions from exploratory data science to production workflows, emphasizing code quality, software engineering practices, and MLOps platforms
Pareto ML Deployments (gojiberries.io). Instance-level routing enhances ML model deployment, preventing regressions while balancing operational complexity and accuracy improvements on datasets like Adult income
Training 10,000 Anomaly Detection Models on One Billion Records with Explainable Predictions (databricks.com). DAXS methodology utilizes scalable ECOD algorithm on Databricks for anomaly detection in manufacturing, improving predictive maintenance and operational efficiency with explainable insights
🔬 Applied ML & Domain Research
Another Another Place. My NOVA 2025 Demo. (benjamin.computer). NOVA 2025 demo tribute to Anthony Gormley, using photogrammetry, machine learning, shaders, and techniques for creating 3D visuals on Linux
Patients want to know what information an AI model considers (ehudreiter.com). Adarsa Sivaprasad's research reveals patients prioritize understanding AI model feature considerations over operational mechanics, highlighting a need for tailored explanations in healthcare AI
New Dataset Using Deep Learning to Predict Permafrost Thaw Damage in the Arctic with Elias Manos (arcticdata.io). Elias Manos develops a deep learning model for predicting permafrost thaw damage in Arctic communities, addressing infrastructure risk and environmental changes
Unlocking the Secrets of Bulk Metallic Glass with Graph Computing (themultidisciplinarian.com). Explore bulk metallic glass properties, design challenges, and the role of graph computing and machine learning in optimizing advanced materials
📈 Financial ML & Quantitative Analysis
PCA analysis of Futures returns for fun and profit, part #1 (qoppac.blogspot.com). Explore PCA analysis of futures returns, uncover risk factors, trading strategies, and the relationship between risk and return in systematic trading
Can We Profit from Disagreements Between Machine Learning and Trend-Following Models? (quantpedia.com). Exploration of using ML and trend-following models for equity returns, identifying profitable discrepancies for tactical allocation
Logistic regression - a simple briefing (blog.engora.com). Exploration of logistic regression for binary classification, odds ratios, sigmoid function, maximum likelihood estimation using Python's scikit-learn, and extensions to multiple classifications
📐 Mathematical Foundations & Advanced Theory
Vectorize ROC Curve for Bayesian Models (juanitorduz.github.io). Vectorizing ROC curve computation for Bayesian models using Gaussian processes and the moons dataset with PyMC, ArviZ, and scikit-learn
Calculus Phobic’s Introduction to Differentiable Programming (andersource.dev). Introduction to differentiable programming using JAX, focusing on optimization techniques, gradient descent, and automatic differentiation principles for solving computational problems
False Causes, Meet Attractor Dimension (methodsblog.com). Yair Daon explores causation in public health using attractor dimensions, proposing BCAD for rejecting false causal links in ecological data analysis
Get ready for the PDE-Transformer (ge.in.tum.de). Introducing PDE-Transformer: a novel architecture for physics simulations, combining UDiT and SWin features, outperforming existing methods in diverse PDE datasets
Random Vector Functional Link (RVFL) artificial neural network with 2 regularization parameters successfully used for forecasting/synthetic simulation in professional settings: Extensions (including Bayesian) (thierrymoudiki.github.io). RVFL artificial neural network utilizes Bayesian and Ridge2 techniques for effective forecasting and simulation in professional domains using Python and R implementations
(Don't) Forget About It: Toward Pareto-Improving GD (gojiberries.io). Exploration of catastrophic forgetting in machine learning and methods to reduce regression during model training with a focus on loss function modifications
📚 Academic Research
DiLoCoX: A Low-Communication Large-Scale Training Framework for Decentralized Cluster (arxiv:cs). DiLoCoX framework enables decentralized training of 100B+ parameter models over slow networks, utilizing Pipeline Parallelism and Adaptive Gradient Compression for efficiency
Interpretable Representation Learning for Additive Rule Ensembles (arxiv:cs). Enhanced interpretable additive ensembles leverage learnable sparse linear transformations for rule conditions, improving model complexity and performance in predictive tasks across various datasets
dreaMLearning: Data Compression Assisted Machine Learning (arxiv:cs). dreaMLearning framework utilizes Entropy-based Generalized Deduplication for efficient machine learning, accelerating training, reducing memory usage, and enhancing scalability for diverse applications
MPipeMoE: Memory Efficient MoE for Pre-trained Models with Adaptive Pipeline Parallelism (arxiv:cs). MPipeMoE enhances Mixture-of-Experts training with adaptive pipeline parallelism for efficiency, achieving 2.8x speedup and reducing memory footprint by 47%
Demystifying Distributed Training of Graph Neural Networks for Link Prediction (arxiv:cs). Distributed GNN training for link prediction faces performance degradation due to graph partitioning and negative sampling; SpLPG mitigates this with reduced communication costs
Feature-Wise Mixing for Mitigating Contextual Bias in Predictive Supervised Learning (arxiv:stat). Feature-wise mixing mitigates contextual bias in predictive ML, achieving 43.35% bias reduction and MSE improvements across classifiers without needing explicit bias identification
Towards Interpretable and Efficient Feature Selection in Trajectory Datasets: A Taxonomic Approach (arxiv:cs). Taxonomy-based feature selection improves interpretability and efficiency in trajectory datasets by organizing features into geometric and kinematic categories, enhancing predictive performance
SuperSONIC: Cloud-Native Infrastructure for ML Inferencing (arxiv:cs). SuperSONIC enables scalable ML inference using Kubernetes and GPUs, optimizing resource utilization for high energy physics and astrophysics research at major observatories
Efficient and Reuseable Cloud Configuration Search Using Discovery Spaces (arxiv:cs). Discovery Space abstraction improves cloud resource configuration search efficiency, enabling knowledge transfer and reducing optimization time by over 90% for diverse workloads
Hierarchical Modeling and Architecture Optimization: Review and Unified Framework (arxiv:stat). Review of hierarchical, conditional, and mixed-variable input modeling; introduces meta and partially-decreed variables; features design space graphs; implemented in Surrogate Modeling Toolbox for Bayesian optimization
CLoVE: Personalized Federated Learning through Clustering of Loss Vector Embeddings (arxiv:cs). CLoVE is a robust algorithm for Clustered Federated Learning, optimizing model accuracy through clustering client loss vector embeddings for personalized learning across diverse datasets
Data Uniformity Improves Training Efficiency and More, with a Convergence Framework Beyond the NTK Regime (arxiv:cs). Uniform data selection enhances training efficiency and performance in LLMs, proving smaller pairwise distances slow gradient descent and decrease approximation errors
Unimodal Strategies in Density-Based Clustering (arxiv:cs). Density-based clustering strategies improve parameter tuning for high-dimensional data using unimodal properties and Ternary Search algorithms across NLP, Audio, and Computer Vision tasks
The kernel of graph indices for vector search (arxiv:cs). Introduction of Support Vector Graph (SVG) for vector search in metric/non-metric spaces, leveraging kernel methods, enhancing graph indices like HNSW and DiskANN
👋 Before you go
I've got a big favor to ask - keeping Blaze running isn't expensive, but it does all add up, so I'm asking readers like you to help, if you can.
That's why I'm launching a Patreon page!. Nothing flashy, just a way for folks who find value in these newsletters to chip in a little each month. In return, you'll get:
- Real say in how Blaze evolves — vote on new topics, features, topic curation ideas
- First dibs on merch (details still cooking)
- That warm fuzzy feeling knowing you're supporting something that saves you time and keeps you plugged into great tech writing
If you are getting value from blaze, checking this out would mean the world. And if you can't contribute, no worries—the newsletters keep coming either way, and you can follow along on patreon for free.
Thanks for reading and being part of this nerdy corner of the internet. All the best - Alastair.
You may also like
About Machine Learning Engineer
Our Machine Learning Engineer newsletter covers the latest developments, research papers, tools, and techniques in ML engineering and deployment. Each week, we curate the most important content so you don't have to spend hours searching.
Whether you're a beginner or expert in machine learning engineering, our newsletter provides valuable information to keep you informed and ahead of the curve in this technically challenging field.
Subscribe now to join thousands of professionals who receive our weekly updates!