1 |
LLaMA-Factory |
56146 |
6887 |
Python |
561 |
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024) |
2025-08-13T15:05:21Z |
2 |
sglang |
16884 |
2611 |
Python |
542 |
SGLang is a fast serving framework for large language models and vision language models. |
2025-08-14T03:37:45Z |
3 |
TensorRT-LLM |
11322 |
1675 |
C++ |
707 |
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in performant way. |
2025-08-14T02:39:26Z |
4 |
ms-swift |
9285 |
813 |
Python |
737 |
Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 500+ LLMs (Qwen3, Qwen3-MoE, Llama4, GLM4.5, InternLM3, DeepSeek-R1, …) and 200+ MLLMs (Qwen2.5-VL, Qwen2.5-Omni, Qwen2-Audio, Ovis2, InternVL3, Llava, GLM4v, Phi4, …) (AAAI 2025). |
2025-08-13T15:10:19Z |
5 |
trace.moe |
4750 |
249 |
None |
0 |
Anime Scene Search by Image |
2024-10-13T03:00:58Z |
6 |
Bangumi |
4653 |
151 |
TypeScript |
24 |
:electron: An unofficial https://bgm.tv ui first app client for Android and iOS, built with React Native. 一个无广告、以爱好为驱动、不以盈利为目的、专门做 ACG 的类似豆瓣的追番记录,bgm.tv 第三方客户端。为移动端重新设计,内置大量加强的网页端难以实现的功能,且提供了相当的自定义选项。 目前已适配 iOS / Android / WSA、mobile / 简单 pad、light / dark theme、移动端网页。 |
2025-08-13T15:31:24Z |
7 |
Moeditor |
4133 |
274 |
JavaScript |
106 |
(discontinued) Your all-purpose markdown editor. |
2020-07-07T01:08:32Z |
8 |
fastllm |
3862 |
393 |
C++ |
264 |
fastllm是后端无依赖的高性能大模型推理库。同时支持张量并行推理稠密模型和混合模式推理MOE模型,任意10G以上显卡即可推理满血DeepSeek。双路9004/9005服务器+单显卡部署DeepSeek满血满精度原版模型,单并发20tps;INT4量化模型单并发30tps,多并发可达60+。 |
2025-08-08T09:41:01Z |
9 |
flashinfer |
3543 |
435 |
Cuda |
157 |
FlashInfer: Kernel Library for LLM Serving |
2025-08-14T03:52:29Z |
10 |
MoeKoeMusic |
3131 |
205 |
Vue |
7 |
一款开源简洁高颜值的酷狗第三方客户端 An open-source, concise, and aesthetically pleasing third-party client for KuGou that supports Windows / macOS / Linux / Web :electron: |
2025-08-12T07:48:19Z |
11 |
Moe-Counter |
2470 |
249 |
JavaScript |
4 |
Moe counter badge with multiple themes! - 多种风格可选的萌萌计数器 |
2025-08-12T08:16:18Z |
12 |
MoeGoe |
2393 |
248 |
Python |
27 |
Executable file for VITS inference |
2023-08-22T07:17:37Z |
13 |
MoE-LLaVA |
2214 |
138 |
Python |
64 |
【TMM 2025🔥】 Mixture-of-Experts for Large Vision-Language Models |
2025-07-15T07:59:33Z |
14 |
GLM-4.5 |
2003 |
188 |
Python |
20 |
GLM-4.5: An open-source large language model designed for intelligent agents by Z.ai |
2025-08-12T03:48:33Z |
15 |
MoBA |
1858 |
113 |
Python |
8 |
MoBA: Mixture of Block Attention for Long-Context LLMs |
2025-04-03T07:28:06Z |
16 |
fastmoe |
1775 |
196 |
Python |
27 |
A fast MoE impl for PyTorch |
2025-02-10T06:04:33Z |
17 |
DeepSeek-MoE |
1767 |
289 |
Python |
17 |
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models |
2024-01-16T12:18:10Z |
18 |
OpenMoE |
1578 |
80 |
Python |
6 |
A family of open-sourced Mixture-of-Experts (MoE) Large Language Models |
2024-03-08T15:08:26Z |
19 |
moemail |
1507 |
844 |
TypeScript |
25 |
一个基于 NextJS + Cloudflare 技术栈构建的可爱临时邮箱服务🎉 |
2025-06-22T16:38:30Z |
20 |
paimon-moe |
1465 |
276 |
JavaScript |
287 |
Your best Genshin Impact companion! Help you plan what to farm with ascension calculator and database. Also track your progress with todo and wish counter. |
2025-07-29T22:20:35Z |
21 |
MOE |
1319 |
138 |
C++ |
170 |
A global, black box optimization engine for real world metric optimization. |
2023-03-24T11:00:32Z |
22 |
mixture-of-experts |
1154 |
109 |
Python |
5 |
PyTorch Re-Implementation of “The Sparsely-Gated Mixture-of-Experts Layer” by Noam Shazeer et al. https://arxiv.org/abs/1701.06538 |
2024-04-19T08:22:39Z |
23 |
moepush |
1122 |
279 |
TypeScript |
10 |
一个基于 NextJS + Cloudflare 技术栈构建的可爱消息推送服务, 支持多种消息推送渠道✨ |
2025-05-10T11:42:44Z |
24 |
Aria |
1067 |
85 |
Jupyter Notebook |
31 |
Codebase for Aria - an Open Multimodal Native MoE |
2025-01-22T03:25:37Z |
25 |
MoeTTS |
988 |
77 |
None |
0 |
Speech synthesis model /inference GUI repo for galgame characters based on Tacotron2, Hifigan, VITS and Diff-svc |
2023-03-03T07:30:05Z |
26 |
llama-moe |
980 |
62 |
Python |
6 |
⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training (EMNLP 2024) |
2024-12-06T04:47:07Z |
27 |
Tutel |
887 |
99 |
C |
50 |
Tutel MoE: Optimized Mixture-of-Experts Library, Support GptOss/DeepSeek/Kimi-K2/Qwen3 FP8/NVFP4/MXFP4 |
2025-08-12T20:39:26Z |
28 |
moebius |
828 |
48 |
JavaScript |
39 |
Modern ANSI & ASCII Art Editor |
2024-05-02T15:54:35Z |
29 |
MoeMemosAndroid |
814 |
84 |
Kotlin |
81 |
An app to help you capture thoughts and ideas |
2025-06-07T11:05:26Z |
30 |
Adan |
798 |
69 |
Python |
3 |
Adan: Adaptive Nesterov Momentum Algorithm for Faster Optimizing Deep Models |
2025-06-08T14:35:41Z |
31 |
MixtralKit |
768 |
79 |
Python |
12 |
A toolkit for inference and evaluation of ‘mixtral-8x7b-32kseqlen’ from Mistral AI |
2023-12-15T19:10:55Z |
32 |
moe-theme.el |
759 |
64 |
Emacs Lisp |
14 |
A customizable colorful eye-candy theme for Emacser. Moe, moe, kyun! |
2025-05-27T06:12:05Z |
33 |
UMOE-Scaling-Unified-Multimodal-LLMs |
751 |
44 |
Python |
13 |
The codes about “Uni-MoE: Scaling Unified Multimodal Models with Mixture of Experts” |
2025-08-06T10:54:43Z |
34 |
DeepSeek-671B-SFT-Guide |
744 |
93 |
Python |
2 |
An open-source solution for full parameter fine-tuning of DeepSeek-V3/R1 671B, including complete code and scripts from training to inference, as well as some practical experiences and conclusions. (DeepSeek-V3/R1 满血版 671B 全参数微调的开源解决方案,包含从训练到推理的完整代码和脚本,以及实践中积累一些经验和结论。) |
2025-03-13T03:51:33Z |
35 |
Hunyuan-A13B |
732 |
87 |
Python |
12 |
Tencent Hunyuan A13B (short as Hunyuan-A13B), an innovative and open-source LLM built on a fine-grained MoE architecture. |
2025-07-08T08:45:27Z |
36 |
Time-MoE |
719 |
73 |
Python |
10 |
[ICLR 2025 Spotlight] Official implementation of “Time-MoE: Billion-Scale Time Series Foundation Models with Mixture of Experts” |
2025-06-17T02:47:16Z |
37 |
moe |
684 |
35 |
Nim |
83 |
A command line based editor inspired by Vim. Written in Nim. |
2025-08-12T14:08:56Z |
38 |
Awesome-Mixture-of-Experts-Papers |
641 |
45 |
None |
1 |
A curated reading list of research in Mixture-of-Experts(MoE). |
2024-10-30T07:48:14Z |
39 |
SmartImage |
630 |
28 |
C# |
7 |
Reverse image search tool (SauceNao, IQDB, Ascii2D, trace.moe, and more) |
2025-08-07T08:48:50Z |
40 |
moedict-webkit |
627 |
99 |
Objective-C |
102 |
萌典網站 |
2025-08-10T01:48:58Z |
41 |
vtbs.moe |
627 |
36 |
Vue |
32 |
Virtual YouTubers in bilibili |
2025-07-31T13:39:09Z |
42 |
satania.moe |
609 |
56 |
HTML |
3 |
Satania IS the BEST waifu, no really, she is, if you don’t believe me, this website will convince you |
2022-10-09T23:19:01Z |
43 |
moebius |
609 |
43 |
Elixir |
3 |
A functional query tool for Elixir |
2024-10-23T18:55:45Z |
44 |
Chinese-Mixtral |
608 |
44 |
Python |
0 |
中文Mixtral混合专家大模型(Chinese Mixtral MoE LLMs) |
2024-04-30T04:29:06Z |
45 |
MoeMemos |
591 |
52 |
Swift |
70 |
An app to help you capture thoughts and ideas |
2025-05-05T19:00:42Z |
46 |
MoeList |
587 |
20 |
Kotlin |
29 |
Another unofficial Android MAL client |
2025-08-10T13:06:08Z |
47 |
MoeGoe_GUI |
570 |
68 |
C# |
8 |
GUI for MoeGoe |
2023-08-22T07:32:08Z |
48 |
moebooru |
563 |
82 |
Ruby |
25 |
Moebooru, a fork of danbooru1 that has been heavily modified |
2025-08-14T00:12:11Z |
49 |
trace.moe-telegram-bot |
539 |
76 |
JavaScript |
0 |
This Telegram Bot can tell the anime when you send an screenshot to it |
2025-08-13T17:01:05Z |
50 |
step_into_llm |
476 |
122 |
Jupyter Notebook |
28 |
MindSpore online courses: Step into LLM |
2025-07-15T14:47:52Z |
51 |
moerail |
473 |
38 |
JavaScript |
16 |
铁路车站代码查询 × 动车组交路查询 |
2025-08-13T12:55:25Z |
52 |
MOE |
423 |
76 |
Java |
18 |
Make Opensource Easy - tools for synchronizing repositories |
2022-06-20T22:41:08Z |
53 |
hydra-moe |
416 |
16 |
Python |
10 |
None |
2023-11-02T22:53:15Z |
54 |
MoeLoaderP |
376 |
26 |
C# |
11 |
🖼二次元图片下载器 Pics downloader for booru sites,Pixiv.net,Bilibili.com,Konachan.com,Yande.re , behoimi.org, safebooru, danbooru,Gelbooru,SankakuComplex,Kawainyan,MiniTokyo,e-shuushuu,Zerochan,WorldCosplay ,Yuriimg etc. |
2025-05-19T13:20:58Z |
55 |
pixiv.moe |
365 |
42 |
TypeScript |
0 |
😘 A pinterest-style layout site, shows illusts on pixiv.net order by popularity. |
2023-03-08T06:54:34Z |
56 |
DiT-MoE |
359 |
18 |
Python |
6 |
Scaling Diffusion Transformers with Mixture of Experts |
2024-09-09T02:12:12Z |
57 |
WThermostatBeca |
356 |
70 |
C++ |
2 |
Open Source firmware replacement for Tuya Wifi Thermostate from Beca and Moes with Home Assistant Autodiscovery |
2023-08-26T22:10:38Z |
58 |
st-moe-pytorch |
355 |
32 |
Python |
4 |
Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch |
2024-06-17T00:48:47Z |
59 |
notify.moe |
353 |
45 |
Go |
86 |
:dancer: Anime tracker, database and community. Moved to https://git.akyoto.dev/web/notify.moe |
2022-09-26T07:15:05Z |
60 |
moe-sticker-bot |
350 |
35 |
Go |
30 |
A Telegram bot that imports LINE/kakao stickers or creates/manages new sticker set. |
2024-06-06T15:28:28Z |
61 |
MOEAFramework |
341 |
129 |
Java |
1 |
A Free and Open Source Java Framework for Multiobjective Optimization |
2025-07-30T12:40:46Z |
62 |
dialogue.moe |
334 |
9 |
Python |
1 |
None |
2022-12-14T14:50:38Z |
63 |
soft-moe-pytorch |
312 |
9 |
Python |
4 |
Implementation of Soft MoE, proposed by Brain’s Vision team, in Pytorch |
2025-04-02T12:47:40Z |
64 |
MoeSR |
311 |
9 |
JavaScript |
7 |
An application specialized in image super-resolution for ACGN illustrations and Visual Novel CG. 专注于插画/Galgame CG等ACGN领域的图像超分辨率的应用 |
2025-08-06T14:15:38Z |
65 |
moell-blog |
301 |
80 |
PHP |
2 |
基于 Laravel 开发,支持 Markdown 语法的博客 |
2022-07-31T11:51:54Z |
66 |
moeSS |
298 |
107 |
PHP |
11 |
moe SS Front End for https://github.com/mengskysama/shadowsocks/tree/manyuser |
2015-02-27T08:44:30Z |
67 |
moe |
279 |
46 |
Scala |
18 |
An -OFun prototype of an Ultra Modern Perl 5 |
2013-09-27T18:39:18Z |
68 |
MoH |
273 |
13 |
Python |
4 |
MoH: Multi-Head Attention as Mixture-of-Head Attention |
2024-10-29T15:22:54Z |
69 |
android-app |
271 |
25 |
Kotlin |
5 |
Official LISTEN.moe Android app |
2025-08-10T19:54:51Z |
70 |
Cornell-MOE |
270 |
63 |
C++ |
25 |
A Python library for the state-of-the-art Bayesian optimization algorithms, with the core implemented in C++. |
2020-02-04T18:39:37Z |
71 |
parameter-efficient-moe |
269 |
16 |
Python |
1 |
None |
2023-10-31T19:21:15Z |
72 |
GRIN-MoE |
265 |
14 |
None |
0 |
GRadient-INformed MoE |
2024-09-25T18:46:48Z |
73 |
MoeQuest |
251 |
76 |
Java |
1 |
The meizi of a material design style welfare App. |
2017-02-14T14:13:53Z |
74 |
MoeLoader-Delta |
248 |
37 |
C# |
52 |
Improved branching version of MoeLoader |
2021-07-22T20:47:41Z |
75 |
inferflow |
246 |
25 |
C++ |
8 |
Inferflow is an efficient and highly configurable inference engine for large language models (LLMs). |
2024-03-15T06:52:33Z |
76 |
MoE-plus-plus |
244 |
11 |
Python |
1 |
[ICLR 2025] MoE++: Accelerating Mixture-of-Experts Methods with Zero-Computation Experts |
2024-10-16T06:21:31Z |
77 |
moeins |
241 |
69 |
PHP |
2 |
萌音影视 - 在线影视应用 |
2018-10-31T01:47:27Z |
78 |
moebius |
233 |
4 |
PHP |
4 |
True coroutines for PHP>=8.1 without worrying about event loops and callbacks. |
2022-06-08T23:18:45Z |
79 |
gdx-pay |
232 |
89 |
Java |
8 |
A libGDX cross-platform API for InApp purchasing. |
2025-07-26T21:09:56Z |
80 |
MoE-Adapters4CL |
231 |
18 |
Python |
6 |
Code for paper “Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts Adapters” CVPR2024 |
2024-11-17T05:47:00Z |
81 |
awesome-moe-inference |
230 |
10 |
None |
0 |
Curated collection of papers in MoE model inference |
2025-08-01T03:34:19Z |
82 |
fiddler |
226 |
24 |
Python |
2 |
[ICLR’25] Fast Inference of MoE Models with CPU-GPU Orchestration |
2024-11-18T00:25:45Z |
83 |
ModuleFormer |
223 |
12 |
Python |
2 |
ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward experts. We released a collection of ModuleFormer-based Language Models (MoLM) ranging in scale from 4 billion to 8 billion parameters. |
2024-04-10T18:16:32Z |
84 |
MoE-Infinity |
221 |
18 |
Python |
10 |
PyTorch library for cost-effective, fast and easy serving of MoE models. |
2025-07-07T12:54:02Z |
85 |
CoE |
220 |
27 |
Python |
3 |
Chain of Experts (CoE) enables communication between experts within Mixture-of-Experts (MoE) models |
2025-06-25T23:14:46Z |
86 |
moe |
201 |
22 |
None |
1 |
Misspelling Oblivious Word Embeddings |
2019-08-06T12:42:31Z |
87 |
MoePhoto |
192 |
23 |
Python |
5 |
MoePhoto Image Toolbox萌图工具箱 |
2024-09-23T06:35:27Z |
88 |
LLaVA-MoD |
191 |
12 |
Python |
5 |
[ICLR 2025] LLaVA-MoD: Making LLaVA Tiny via MoE-Knowledge Distillation |
2025-03-31T09:41:38Z |
89 |
Yuan2.0-M32 |
189 |
40 |
Python |
6 |
Mixture-of-Experts (MoE) Language Model |
2024-09-09T09:14:15Z |
90 |
Ling |
187 |
15 |
Python |
2 |
Ling is a MoE LLM provided and open-sourced by InclusionAI. |
2025-05-14T06:34:57Z |
91 |
MixLoRA |
177 |
18 |
Python |
3 |
State-of-the-art Parameter-Efficient MoE Fine-tuning Method |
2024-08-22T08:02:04Z |
92 |
MOELoRA-peft |
175 |
19 |
Python |
7 |
[SIGIR’24] The official implementation code of MOELoRA. |
2024-07-22T07:32:43Z |
93 |
MOEAD |
172 |
42 |
Python |
2 |
MOEAD.多目标差分进化算法的学习,Python实现&动态展示过程 |
2022-06-22T02:07:23Z |
94 |
MoeRanker |
170 |
14 |
JavaScript |
9 |
Moe attribute sorter |
2025-06-30T03:16:02Z |
95 |
SMoE-Stereo |
165 |
12 |
Python |
2 |
[ICCV 2025 Highlight] 🌟🌟🌟 Learning Robust Stereo Matching in the Wild with Selective Mixture-of-Experts |
2025-07-24T04:31:01Z |
96 |
ghost-theme-Moegi |
164 |
26 |
Handlebars |
3 |
An elegant & fresh ghost theme. |
2023-10-16T16:09:28Z |
97 |
Frequency_Aug_VAE_MoESR |
162 |
4 |
Python |
12 |
Latent-based SR using MoE and frequency augmented VAE decoder |
2023-11-26T10:33:36Z |
98 |
moedict-desktop |
156 |
15 |
C++ |
6 |
MoeDict Desktop For MacOSX / Linux / Windows |
2016-10-14T06:49:17Z |
99 |
guide.encode.moe |
155 |
20 |
Markdown |
9 |
A guide for fansubbing |
2025-08-11T16:20:33Z |
100 |
awesome-adaptive-computation |
151 |
10 |
None |
1 |
A curated reading list of research in Adaptive Computation, Inference-Time Computation & Mixture of Experts (MoE). |
2025-01-01T13:49:09Z |