| 1 |
vllm |
76811 |
15658 |
Python |
1825 |
A high-throughput and memory-efficient inference and serving engine for LLMs |
2026-04-16T05:26:21Z |
| 2 |
LlamaFactory |
70154 |
8586 |
Python |
934 |
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024) |
2026-04-12T00:32:15Z |
| 3 |
sglang |
25876 |
5387 |
Python |
607 |
SGLang is a high-performance serving framework for large language models and multimodal models. |
2026-04-16T05:30:15Z |
| 4 |
ms-swift |
13742 |
1352 |
Python |
970 |
Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 600+ LLMs (Qwen3.5, DeepSeek-R1, GLM-5, InternLM3, Llama4, …) and 300+ MLLMs (Qwen3-VL, Qwen3-Omni, InternVL3.5, Ovis2.5, GLM4.5v, Llava, Phi4, …) (AAAI 2025). |
2026-04-15T10:54:41Z |
| 5 |
TensorRT-LLM |
13377 |
2294 |
Python |
580 |
TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way. |
2026-04-16T05:16:23Z |
| 6 |
MoeKoeMusic |
5477 |
358 |
Vue |
1 |
一款开源简洁高颜值的酷狗第三方客户端 An open-source, concise, and aesthetically pleasing third-party client for KuGou that supports Windows / macOS / Linux / Web :electron: |
2026-04-15T06:06:27Z |
| 7 |
flashinfer |
5402 |
899 |
Python |
327 |
FlashInfer: Kernel Library for LLM Serving |
2026-04-16T03:18:24Z |
| 8 |
Bangumi |
5376 |
159 |
TypeScript |
30 |
:electron: An unofficial https://bgm.tv ui first app client for Android and iOS, built with React Native. 一个无广告、以爱好为驱动、不以盈利为目的、专门做 ACG 的类似豆瓣的追番记录,bgm.tv 第三方客户端。为移动端重新设计,内置大量加强的网页端难以实现的功能,且提供了相当的自定义选项。 目前已适配 iOS / Android。 |
2026-04-14T14:43:26Z |
| 9 |
xtuner |
5119 |
415 |
Python |
237 |
A Next-Generation Training Engine Built for Ultra-Large MoE Models |
2026-04-16T04:32:20Z |
| 10 |
trace.moe |
4971 |
260 |
None |
0 |
Trace back an anime scene with a screenshot |
2026-04-06T15:47:36Z |
| 11 |
GLM-4.5 |
4313 |
450 |
Python |
24 |
GLM-4.5: Agentic, Reasoning, and Coding (ARC) Foundation Models |
2026-02-01T08:28:10Z |
| 12 |
fastllm |
4189 |
418 |
C++ |
324 |
fastllm是后端无依赖的高性能大模型推理库。同时支持张量并行推理稠密模型和混合模式推理MOE模型,任意10G以上显卡即可推理满血DeepSeek。双路9004/9005服务器+单显卡部署DeepSeek满血满精度原版模型,单并发20tps;INT4量化模型单并发30tps,多并发可达60+。 |
2026-04-10T02:29:23Z |
| 13 |
Moeditor |
4115 |
273 |
JavaScript |
106 |
(discontinued) Your all-purpose markdown editor. |
2020-07-07T01:08:32Z |
| 14 |
flash-moe |
3643 |
442 |
Objective-C |
9 |
Running a big model on a small laptop |
2026-03-19T17:21:57Z |
| 15 |
Moe-Counter |
2857 |
289 |
JavaScript |
3 |
Moe counter badge with multiple themes! - 多种风格可选的萌萌计数器 |
2026-04-16T03:39:37Z |
| 16 |
MoeGoe |
2418 |
245 |
Python |
28 |
Executable file for VITS inference |
2023-08-22T07:17:37Z |
| 17 |
moemail |
2411 |
2092 |
TypeScript |
43 |
A cute temporary email service built with NextJS + Cloudflare technology stack 🎉 | 一个基于 NextJS + Cloudflare 技术栈构建的可爱临时邮箱服务🎉 |
2026-03-30T09:35:05Z |
| 18 |
MoE-LLaVA |
2315 |
142 |
Python |
65 |
【TMM 2025🔥】 Mixture-of-Experts for Large Vision-Language Models |
2025-07-15T07:59:33Z |
| 19 |
MoBA |
2093 |
141 |
Python |
10 |
MoBA: Mixture of Block Attention for Long-Context LLMs |
2025-04-03T07:28:06Z |
| 20 |
ICEdit |
2092 |
115 |
Python |
23 |
[NeurIPS 2025] Image editing is worth a single LoRA! 0.1% training data for fantastic image editing! Surpasses GPT-4o in ID persistence~ MoE ckpt released! Only 4GB VRAM is enough to run! |
2025-12-19T19:08:02Z |
| 21 |
DeepSeek-MoE |
1909 |
303 |
Python |
17 |
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models |
2024-01-16T12:18:10Z |
| 22 |
fastmoe |
1848 |
205 |
Python |
27 |
A fast MoE impl for PyTorch |
2025-02-10T06:04:33Z |
| 23 |
OpenMoE |
1674 |
85 |
Python |
6 |
A family of open-sourced Mixture-of-Experts (MoE) Large Language Models |
2024-03-08T15:08:26Z |
| 24 |
paimon-moe |
1510 |
281 |
JavaScript |
310 |
Your best Genshin Impact companion! Help you plan what to farm with ascension calculator and database. Also track your progress with todo and wish counter. |
2026-04-08T05:58:38Z |
| 25 |
moepush |
1344 |
426 |
TypeScript |
14 |
一个基于 NextJS + Cloudflare 技术栈构建的可爱消息推送服务, 支持多种消息推送渠道✨ |
2025-05-10T11:42:44Z |
| 26 |
MOE |
1319 |
140 |
C++ |
170 |
A global, black box optimization engine for real world metric optimization. |
2023-03-24T11:00:32Z |
| 27 |
uccl |
1302 |
136 |
C++ |
44 |
UCCL is an efficient communication library for GPUs, covering collectives, P2P (e.g., KV cache transfer, RL weight transfer), and EP (e.g., GPU-driven) |
2026-04-16T05:19:35Z |
| 28 |
SpikingBrain-7B |
1301 |
183 |
Python |
8 |
Spiking Brain-inspired Large Models, integrating hybrid efficient attention, MoE modules and spike encoding into its architecture |
2025-12-01T11:13:32Z |
| 29 |
mixture-of-experts |
1243 |
111 |
Python |
6 |
PyTorch Re-Implementation of “The Sparsely-Gated Mixture-of-Experts Layer” by Noam Shazeer et al. https://arxiv.org/abs/1701.06538 |
2024-04-19T08:22:39Z |
| 30 |
Uni-MoE |
1095 |
68 |
Python |
26 |
Uni-MoE: Lychee’s Large Multimodal Model Family. |
2025-12-22T02:32:34Z |
| 31 |
Aria |
1084 |
89 |
Jupyter Notebook |
31 |
Codebase for Aria - an Open Multimodal Native MoE |
2025-01-22T03:25:37Z |
| 32 |
MoeMemosAndroid |
1060 |
119 |
Kotlin |
88 |
An app to help you capture thoughts and ideas |
2026-04-06T18:20:14Z |
| 33 |
SmartImage |
1008 |
53 |
C# |
4 |
Reverse image search tool (SauceNao, IQDB, Ascii2D, trace.moe, and more) |
2026-03-21T15:38:53Z |
| 34 |
llama-moe |
1000 |
60 |
Python |
6 |
⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training (EMNLP 2024) |
2024-12-06T04:47:07Z |
| 35 |
MoeTTS |
996 |
75 |
None |
0 |
Speech synthesis model /inference GUI repo for galgame characters based on Tacotron2, Hifigan, VITS and Diff-svc |
2023-03-03T07:30:05Z |
| 36 |
Tutel |
985 |
107 |
C |
54 |
Tutel MoE: Optimized Mixture-of-Experts Library, Support GptOss/DeepSeek/Kimi-K2/Qwen3 using FP8/NVFP4/MXFP4 |
2026-04-11T18:05:17Z |
| 37 |
Time-MoE |
948 |
110 |
Python |
15 |
[ICLR 2025 Spotlight] Official implementation of “Time-MoE: Billion-Scale Time Series Foundation Models with Mixture of Experts” |
2026-03-21T16:00:55Z |
| 38 |
MiniMind-in-Depth |
927 |
79 |
None |
6 |
轻量级大语言模型MiniMind的源码解读,包含tokenizer、RoPE、MoE、KV Cache、pretraining、SFT、LoRA、DPO等完整流程 |
2025-06-16T14:13:15Z |
| 39 |
moebius |
925 |
51 |
JavaScript |
40 |
Modern ANSI & ASCII Art Editor |
2024-05-02T15:54:35Z |
| 40 |
Hunyuan-A13B |
813 |
118 |
Python |
24 |
Tencent Hunyuan A13B (short as Hunyuan-A13B), an innovative and open-source LLM built on a fine-grained MoE architecture. |
2025-07-08T08:45:27Z |
| 41 |
Adan |
813 |
71 |
Python |
6 |
Adan: Adaptive Nesterov Momentum Algorithm for Faster Optimizing Deep Models |
2025-06-08T14:35:41Z |
| 42 |
DeepSeek-671B-SFT-Guide |
802 |
95 |
Python |
1 |
An open-source solution for full parameter fine-tuning of DeepSeek-V3/R1 671B, including complete code and scripts from training to inference, as well as some practical experiences and conclusions. (DeepSeek-V3/R1 满血版 671B 全参数微调的开源解决方案,包含从训练到推理的完整代码和脚本,以及实践中积累一些经验和结论。) |
2025-03-13T03:51:33Z |
| 43 |
MixtralKit |
772 |
75 |
Python |
12 |
A toolkit for inference and evaluation of ‘mixtral-8x7b-32kseqlen’ from Mistral AI |
2023-12-15T19:10:55Z |
| 44 |
moe-theme.el |
769 |
66 |
Emacs Lisp |
15 |
A customizable colorful eye-candy theme for Emacser. Moe, moe, kyun! |
2026-03-04T15:21:30Z |
| 45 |
MoeMemos |
756 |
65 |
Swift |
61 |
An app to help you capture thoughts and ideas |
2026-03-13T18:10:17Z |
| 46 |
moe |
706 |
33 |
Nim |
40 |
A command line based editor inspired by Vim. Written in Nim. |
2026-04-15T16:00:22Z |
| 47 |
Awesome-Mixture-of-Experts-Papers |
663 |
46 |
None |
3 |
A curated reading list of research in Mixture-of-Experts(MoE). |
2024-10-30T07:48:14Z |
| 48 |
moedict-webkit |
645 |
99 |
Objective-C |
102 |
萌典網站 |
2026-04-15T11:09:20Z |
| 49 |
MoeList |
636 |
22 |
Kotlin |
31 |
Another unofficial Android MAL client |
2026-04-13T14:18:10Z |
| 50 |
sonic-moe |
634 |
71 |
Python |
13 |
Accelerating MoE with IO and Tile-aware Optimizations |
2026-04-15T04:43:25Z |
| 51 |
vtbs.moe |
630 |
36 |
Vue |
33 |
Virtual YouTubers in bilibili |
2025-07-31T13:39:09Z |
| 52 |
satania.moe |
614 |
54 |
HTML |
3 |
Satania IS the BEST waifu, no really, she is, if you don’t believe me, this website will convince you |
2022-10-09T23:19:01Z |
| 53 |
moebius |
610 |
42 |
Elixir |
3 |
A functional query tool for Elixir |
2024-10-23T18:55:45Z |
| 54 |
Chinese-Mixtral |
610 |
43 |
Python |
0 |
中文Mixtral混合专家大模型(Chinese Mixtral MoE LLMs) |
2024-04-30T04:29:06Z |
| 55 |
MoePeek |
609 |
37 |
Swift |
6 |
A lightweight macOS selection translator built with pure Swift 6, featuring on-device Apple Translate for privacy, only 5MB install size and stable ~50MB memory usage. 一款轻量级 macOS 划词翻译工具,纯 Swift 6 开发,设备端 Apple 翻译保护隐私,安装体积仅 5MB,后台运行内存稳定约 50MB |
2026-04-03T05:56:57Z |
| 56 |
moebooru |
596 |
80 |
Ruby |
29 |
Moebooru, a fork of danbooru1 that has been heavily modified |
2026-04-08T06:13:32Z |
| 57 |
MoeGoe_GUI |
572 |
69 |
C# |
8 |
GUI for MoeGoe |
2023-08-22T07:32:08Z |
| 58 |
trace.moe-telegram-bot |
551 |
79 |
TypeScript |
0 |
This Telegram Bot can tell the anime when you send an screenshot to it |
2026-04-13T14:03:54Z |
| 59 |
diy-llm |
540 |
67 |
Jupyter Notebook |
3 |
🎓 系统性大语言模型构建课程|🛠️ 覆盖预训练数据工程、Tokenizer、Transformer、MoE、GPU 编程 (CUDA/Triton)、分布式训练、Scaling Laws、推理优化及对齐 (SFT/RLHF/GRPO)|🚀 6 个渐进式作业 + 代码驱动,建立 LLM 全栈认知体系 |
2026-04-15T08:26:02Z |
| 60 |
moerail |
526 |
42 |
JavaScript |
21 |
铁路车站代码查询 × 动车组交路查询 |
2025-08-13T12:55:25Z |
| 61 |
LPLB |
498 |
33 |
Python |
1 |
An early research stage expert-parallel load balancer for MoE models based on linear programming. |
2025-11-19T07:20:35Z |
| 62 |
step_into_llm |
480 |
127 |
Jupyter Notebook |
27 |
MindSpore online courses: Step into LLM |
2025-12-22T11:46:46Z |
| 63 |
YOLO-Master |
473 |
53 |
Python |
15 |
[CVPR2026]🚀🚀🚀Official code for the paper “YOLO-Master: MOE-Accelerated with Specialized Transformers for Enhanced Real-time Detection.” (YOLO = You Only Look Once) 🔥🔥🔥 |
2026-04-15T09:27:10Z |
| 64 |
MOE |
423 |
78 |
Java |
18 |
Make Opensource Easy - tools for synchronizing repositories |
2022-06-20T22:41:08Z |
| 65 |
DiT-MoE |
420 |
19 |
Python |
7 |
Scaling Diffusion Transformers with Mixture of Experts |
2024-09-09T02:12:12Z |
| 66 |
hydra-moe |
415 |
16 |
Python |
10 |
None |
2023-11-02T22:53:15Z |
| 67 |
MoeLoaderP |
407 |
28 |
C# |
12 |
🖼二次元图片下载器 Pics downloader for booru sites,Pixiv.net,Bilibili.com,Konachan.com,Yande.re , behoimi.org, safebooru, danbooru,Gelbooru,SankakuComplex,Kawainyan,MiniTokyo,e-shuushuu,Zerochan,WorldCosplay ,Yuriimg etc. |
2025-05-19T13:20:58Z |
| 68 |
Awesome-Efficient-Arch |
403 |
33 |
None |
0 |
Speed Always Wins: A Survey on Efficient Architectures for Large Language Models |
2025-11-11T09:47:37Z |
| 69 |
BitSoulStockSkill |
403 |
32 |
Python |
0 |
由BitSoul出品的A股市场全能Skill,自带免费历史数据,内置100+行业主流因子,完整的回测框架,基于MOE架构的股票筛选与买卖判断,更提供因子挖矿等趣味接口,欢迎安装试用,也欢共同开发交流! |
2026-03-21T08:19:00Z |
| 70 |
MoeSR |
400 |
9 |
JavaScript |
8 |
An application specialized in image super-resolution for ACGN illustrations and Visual Novel CG. 专注于插画/Galgame CG等ACGN领域的图像超分辨率的应用 |
2026-03-09T14:07:16Z |
| 71 |
moe-sticker-bot |
397 |
49 |
Go |
36 |
A Telegram bot that imports LINE/kakao stickers or creates/manages new sticker set. |
2024-06-06T15:28:28Z |
| 72 |
nmoe |
382 |
32 |
Python |
2 |
MoE training for Me and You and maybe other people |
2026-03-15T22:23:47Z |
| 73 |
st-moe-pytorch |
381 |
33 |
Python |
4 |
Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch |
2024-06-17T00:48:47Z |
| 74 |
awesome-moe-inference |
372 |
15 |
None |
0 |
Curated collection of papers in MoE model inference |
2026-03-12T01:59:19Z |
| 75 |
WThermostatBeca |
371 |
72 |
C++ |
4 |
Open Source firmware replacement for Tuya Wifi Thermostate from Beca and Moes with Home Assistant Autodiscovery |
2023-08-26T22:10:38Z |
| 76 |
pixiv.moe |
370 |
42 |
TypeScript |
0 |
😘 A pinterest-style layout site, shows illusts on pixiv.net order by popularity. |
2023-03-08T06:54:34Z |
| 77 |
MOEAFramework |
355 |
129 |
Java |
0 |
A Free and Open Source Java Framework for Multiobjective Optimization |
2026-01-21T16:26:02Z |
| 78 |
notify.moe |
351 |
46 |
Go |
86 |
:dancer: Anime tracker, database and community. Moved to https://git.akyoto.dev/web/notify.moe |
2022-09-26T07:15:05Z |
| 79 |
soft-moe-pytorch |
345 |
10 |
Python |
4 |
Implementation of Soft MoE, proposed by Brain’s Vision team, in Pytorch |
2025-04-02T12:47:40Z |
| 80 |
SwiftLM |
345 |
19 |
Swift |
1 |
⚡ Native MLX Swift LLM inference server for Apple Silicon. OpenAI-compatible API, SSD streaming for 100B+ MoE models, TurboQuant KV cache compression, MACOS + iOS iPhone app. |
2026-04-16T05:11:27Z |
| 81 |
dialogue.moe |
344 |
11 |
Python |
1 |
None |
2022-12-14T14:50:38Z |
| 82 |
FreeMoe |
333 |
8 |
None |
13 |
Unlock App Vip |
2026-01-30T05:50:54Z |
| 83 |
Lvllm |
324 |
30 |
Python |
1 |
LvLLM is a special NUMA extension of vllm that makes full use of CPU and memory resources, reduces GPU memory requirements, and features an efficient GPU parallel and NUMA parallel architecture, supporting hybrid inference for MOE large models. |
2026-04-15T13:51:41Z |
| 84 |
MoH |
311 |
15 |
Python |
5 |
MoH: Multi-Head Attention as Mixture-of-Head Attention |
2024-10-29T15:22:54Z |
| 85 |
moell-blog |
302 |
79 |
PHP |
2 |
基于 Laravel 开发,支持 Markdown 语法的博客 |
2022-07-31T11:51:54Z |
| 86 |
moeSS |
298 |
107 |
PHP |
11 |
moe SS Front End for https://github.com/mengskysama/shadowsocks/tree/manyuser |
2015-02-27T08:44:30Z |
| 87 |
android-app |
297 |
28 |
Kotlin |
5 |
Official LISTEN.moe Android app |
2026-04-12T08:59:25Z |
| 88 |
MoE-Infinity |
295 |
25 |
Python |
8 |
PyTorch library for cost-effective, fast and easy serving of MoE models. |
2026-04-14T14:14:06Z |
| 89 |
kimi-agent-internals |
283 |
89 |
Fluent |
0 |
Extracted artifacts from Kimi OK-Computer (and other agents) system for AI studies in agentic architecture. |
2026-03-29T01:36:16Z |
| 90 |
WAM-Diff |
279 |
51 |
Python |
1 |
WAM-Diff: A Masked Diffusion VLA Framework with MoE and Online Reinforcement Learning for Autonomous Driving |
2026-02-01T03:59:10Z |
| 91 |
moe |
277 |
47 |
Scala |
18 |
An -OFun prototype of an Ultra Modern Perl 5 |
2013-09-27T18:39:18Z |
| 92 |
parameter-efficient-moe |
276 |
17 |
Python |
1 |
None |
2023-10-31T19:21:15Z |
| 93 |
Cornell-MOE |
275 |
65 |
C++ |
25 |
A Python library for the state-of-the-art Bayesian optimization algorithms, with the core implemented in C++. |
2020-02-04T18:39:37Z |
| 94 |
MoE-Adapters4CL |
270 |
26 |
Python |
7 |
Code for paper “Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts Adapters” CVPR2024 |
2025-09-18T08:38:29Z |
| 95 |
GRIN-MoE |
264 |
14 |
None |
0 |
GRadient-INformed MoE |
2024-09-25T18:46:48Z |
| 96 |
MoE-plus-plus |
265 |
13 |
Python |
1 |
[ICLR 2025] MoE++: Accelerating Mixture-of-Experts Methods with Zero-Computation Experts |
2024-10-16T06:21:31Z |
| 97 |
fiddler |
264 |
35 |
Python |
3 |
[ICLR’25] Fast Inference of MoE Models with CPU-GPU Orchestration |
2024-11-18T00:25:45Z |
| 98 |
Ling-V2 |
264 |
18 |
Python |
4 |
Ling-V2 is a MoE LLM provided and open-sourced by InclusionAI. |
2025-10-04T06:15:38Z |
| 99 |
transformers-qwen3-moe-fused |
251 |
14 |
Python |
3 |
Fused Qwen3 MoE layer for faster training, compatible with Transformers, LoRA, bnb 4-bit quant, Unsloth. Also possible to train LoRA over GGUF |
2026-02-19T06:15:21Z |
| 100 |
MoeQuest |
250 |
76 |
Java |
1 |
The meizi of a material design style welfare App. |
2017-02-14T14:13:53Z |