New progress of generative modeling – flow matching

In a previous post, we discussed an earlier generative modeling called Normalizing Flows [1]. However, Normalizing Flows has its own limitations: (1) it requires the flow mapping function to be invertible. This limits choices of potential neural network architectures to instantiate , because being invertible means that hidden layers must have the exact same dimensionality …

Learn GPU Optimization

It has been a while since I learned GPU knowledge. I am going to keep updating more recent materials for ramping up my GPU knowledge. FlashAttention [1] We start from recapping the standard Self-Attention mechanism, which is computed in 3-passes: Notes: The shape and represent the sequence length and internal dimension, respectively. , , , …