RetentionLabs
Popular repositories Loading
-
RWKV-LM
RWKV-LM PublicForked from BlinkDL/RWKV-LM
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RN…
Python
-
titans-pytorch
titans-pytorch PublicForked from lucidrains/titans-pytorch
Unofficial implementation of Titans, SOTA memory for transformers, in Pytorch
Python
-
-
MLP-Mixer-CIFAR
MLP-Mixer-CIFAR PublicForked from omihub777/MLP-Mixer-CIFAR
PyTorch implementation of Mixer-nano (#parameters is 0.67M, originally Mixer-S/16 has 18M) with 90.83 % acc. on CIFAR-10. Training from scratch.
Python
-
ragflow
ragflow PublicForked from infiniflow/ragflow
RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.
Python
-
ART
ART PublicForked from OpenPipe/ART
Agent Reinforcement Trainer: train multi-step agents for real-world tasks using GRPO. Give your agents on-the-job training. Reinforcement learning for Qwen2.5, Qwen3, Llama, Kimi, and more!
Python
Repositories
- transformers-ttt Public Forked from huggingface/transformers
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
retentionlabs/transformers-ttt’s past year of commit activity - microxcaling Public Forked from microsoft/microxcaling
PyTorch emulation library for Microscaling (MX)-compatible data formats
retentionlabs/microxcaling’s past year of commit activity - ttt-lm-pytorch Public Forked from test-time-training/ttt-lm-pytorch
Official PyTorch implementation of Learning to (Learn at Test Time): RNNs with Expressive Hidden States
retentionlabs/ttt-lm-pytorch’s past year of commit activity - ART Public Forked from OpenPipe/ART
Agent Reinforcement Trainer: train multi-step agents for real-world tasks using GRPO. Give your agents on-the-job training. Reinforcement learning for Qwen2.5, Qwen3, Llama, Kimi, and more!
retentionlabs/ART’s past year of commit activity - titans-pytorch Public Forked from lucidrains/titans-pytorch
Unofficial implementation of Titans, SOTA memory for transformers, in Pytorch
retentionlabs/titans-pytorch’s past year of commit activity - ragflow Public Forked from infiniflow/ragflow
RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.
retentionlabs/ragflow’s past year of commit activity - mxfp4-llm Public Forked from amazon-science/mxfp4-llm
Official implementation for Training LLMs with MXFP4
retentionlabs/mxfp4-llm’s past year of commit activity - RWKV-LM Public Forked from BlinkDL/RWKV-LM
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RNN and transformer - great performance, linear time, constant space (no kv-cache), fast training, infinite ctx_len, and free sentence embedding.
retentionlabs/RWKV-LM’s past year of commit activity - MLP-Mixer-CIFAR Public Forked from omihub777/MLP-Mixer-CIFAR
PyTorch implementation of Mixer-nano (#parameters is 0.67M, originally Mixer-S/16 has 18M) with 90.83 % acc. on CIFAR-10. Training from scratch.
retentionlabs/MLP-Mixer-CIFAR’s past year of commit activity
Top languages
Loading…
Most used topics
Loading…