📓 nodes/20230628213013-diffusion_train.org by @tekakutli-org ☆

RLCM

IDEAS

REMEMBER

BEFORE-AFTER

ONLY ONCE

CONTEXT

PRIORS

STRUCTURE

VAE TRAINING

3D INCORPORATED

DISTRIBUTED TRAINING

DIFFUSION QUANTIZATION

ACADEMIC

CLIP RELATED

CHEAPER TRAINING

FASTER TOO

DIFFERENT ARCHITECTURE

DATASET MANIPULATION

BATCH STRUCTURE

ATLAS

MASKS

MATHEMATICAL (COPY PASTED COMMENT YET TO ANALYZE)

I have recently written a paper on understanding transformer learning via the lens of coinduction & Hopf algebra. https://arxiv.org/abs/2302.01834

The learning mechanism of transformer models was poorly understood however it turns out that a transformer is like a circuit with a feedback.

I argue that autodiff can be replaced with what I call in the paper Hopf coherence which happens within the single layer as opposed to across the whole graph.

Furthermore, if we view transformers as Hopf algebras, one can bring convolutional models, diffusion models and transformers under a single umbrella.

I'm working on a next gen Hopf algebra based machine learning framework.

Join my discord if you want to discuss this further https://discord.gg/mr9TAhpyBW