moe
Here are 9 public repositories matching this topic...
[ICLR 2025] Drop-Upcycling: Training Sparse Mixture of Experts with Partial Re-initialization
-
Updated
Oct 5, 2025 - Shell
HydraNet is a state-of-the-art transformer architecture that combines Multi-Query Attention (MQA), Mixture of Experts (MoE), and continuous learning capabilities.
-
Updated
Mar 28, 2026 - Shell
GPU-tuned Docker images for LLM inference on consumer hardware. Auto-detects your GPU, downloads the model, serves an OpenAI-compatible API.
-
Updated
Mar 2, 2026 - Shell
Full-stack AI Chatbot deployment for universities. Includes RAG, secure Code Interpreter, vLLM inference, and educational prompt libraries.
-
Updated
Oct 17, 2025 - Shell
Improve this page
Add a description, image, and links to the moe topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the moe topic, visit your repo's landing page and select "manage topics."