Skip to content
/ delta-x Public

Δx is a collection of ml paper summaries of the form: 'it's just x but with y'.

Notifications You must be signed in to change notification settings

p-doom/delta-x

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

21 Commits
 
 

Repository files navigation

Δx

Δx is a collection of ml paper summaries of the form: 'it's just x but with y'.

Paper title Model name Summary
Semi-supervised Learning by Entropy Minimization Entropy Minimisation Supervision but with the additional goal of maximising prediction confidencen on the unlabeled data
Fast R-CNN Fast R-CNN R-CNN but with RoI pooling
Faster R-CNN Faster R-CNN Fast R-CNN but with a region proposal network
Focal Loss for Dense Object Detection RetinaNet FPN but with a one-stage detector
Leveraging Pre-trained Checkpoints for Sequence Generation Tasks - Encoder-Decoder Transformer, but with pre-trained BERT/GPT-style encoder/decoder checkpoints
Swin Transformer: Hierarchical Vision Transformer using Shifted Windows Swin Transformer ViT but with a patch size hierarchy
Per-Pixel Classification is Not All You Need for Semantic Segmentation MaskFormer DETR but with per-instance binary masks
Masked-attention Mask Transformer for Universal Image Segmentation Mask2Former MaskFormer but with masked attention around objects/regions and multi-scale pixel decoder features
Momentum Contrast for Unsupervised Visual Representation Learning MoCo Contrastive learning but with one momentum-updated encoder
Training Transitive and Commutative Multimodal Transformers with LoReTTa LoReTTa ImageBind but with Transformers
ProtST: Multi-Modality Learning of Protein Sequences and Biomedical Texts ProtST CLIP but with proteins and functional annotations
When Scaling Meets LLM Finetuning: The Effect of Data, Model and Finetuning Method - Chinchilla scaling laws but for finetuning
Scaling Data-Constrained Language Models Datablation Chinchilla scaling laws but for the data-constrained regime
Online normalizer calculation for softmax - Softmax calculation but online/ streaming/ blockwise
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness FlashAttention Online softmax but with attention checkpointing
Blockwise Parallel Transformer for Large Context Models BPT FlashAttention but fusing blockwise attention with FFN
Ring Attention with Blockwise Transformers for Near-Infinite Context Ring Attention BPT but distribute blocks across devices & overlap computation and communication
Emergent Complexity and Zero-shot Transfer via Unsupervised Environment Design PAIRED Minimax (adversarial) environment design but with an antagonist agent that needs to maximize reward
Behaviour Distillation HaDES Dataset distillation but for RL
Structured State Space Models for In-Context Reinforcement Learning - RL^2 but with S5
Scheduled Sampling for Sequence Prediction with Recurrent Neural Networks - Language modeling but only using teacher-forcing with probability eps (and annealing eps)
Sequence Level Training with Recurrent Neural Networks End-to-End Backprop Language modeling but top-k last word predictions instead of gt as input with prob eps (+ annealing eps)
Sequence Level Training with Recurrent Neural Networks MIXER Language modeling but sequence-level reward-based post-training using REINFORCE
Genie: Generative Interactive Environments Genie Video Pretraining but with a learned latent action model
Learning to act without actions LAPO Video Pretraining but with unsupervised latent action IDM learned via predictive consistency between IDM and FDM
Behavioral Cloning Transformer BCT BC but with a causal transformer, where the context contains the state-action pair sequence
Decision Transformer DT BCT but with state-action-'return-to-go' triplet sequence instead of state-action pair sequence
Self-Consistency Preference Optimization ScPO Expert Iteration but with Self-Consistency
Self-Rewarding Language Models DPO but with LLM-as-a-Judge

About

Δx is a collection of ml paper summaries of the form: 'it's just x but with y'.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published