Compare commits

...

76 Commits

Author SHA1 Message Date
fujie
f882997337 feat(github-copilot-sdk): v0.3.0 - unified tool bridge & dynamic MCP discovery
Major enhancements:
- Zero-config OpenWebUI Tool Bridge: automatically converts WebUI Functions to Copilot-compatible tools
- Dynamic MCP Discovery: seamlessly reads MCP servers from Admin Settings -> Connections
- High-performance async engine with optimized event-driven streaming
- Robust interoperability via dynamic Pydantic model generation
- Simplified token acquisition (web-based PAT only, removed CLI method)
- Updated configuration valves (renamed, removed legacy parameters)
- Comprehensive bilingual documentation sync
2026-02-07 12:36:46 +08:00
github-actions[bot]
8e2c1b467e chore: update community stats - followers increased (202 -> 203) 2026-02-07 01:38:50 +00:00
github-actions[bot]
db2433afa1 chore: update community stats - points increased (245 -> 246) 2026-02-06 21:12:47 +00:00
github-actions[bot]
fd125c9dea chore: update community stats - followers increased (201 -> 202) 2026-02-06 17:20:41 +00:00
github-actions[bot]
0a726aacb6 chore: update community stats - followers increased (200 -> 201) 2026-02-06 14:21:48 +00:00
github-actions[bot]
bf82e093d4 chore: update community stats - points increased (244 -> 245), followers increased (199 -> 200) 2026-02-06 10:19:17 +00:00
github-actions[bot]
53ae7ef003 chore: update community stats - followers increased (197 -> 199) 2026-02-06 04:39:50 +00:00
github-actions[bot]
c2c7e3b2d3 chore: update community stats - followers increased (196 -> 197) 2026-02-06 01:39:25 +00:00
github-actions[bot]
26df5f5144 chore: update community stats - followers increased (195 -> 196) 2026-02-05 20:13:20 +00:00
github-actions[bot]
aebeb3677c chore: update community stats - followers increased (194 -> 195) 2026-02-05 17:22:50 +00:00
github-actions[bot]
076c1d62c6 chore: update community stats - points increased (243 -> 244) 2026-02-05 12:20:12 +00:00
github-actions[bot]
21cf6ecc1d chore: update community stats - followers increased (193 -> 194) 2026-02-05 11:18:53 +00:00
github-actions[bot]
e6bdab54c9 chore: update community stats - followers increased (192 -> 193) 2026-02-05 10:20:57 +00:00
github-actions[bot]
2232e5adb3 chore: update community stats - followers increased (191 -> 192) 2026-02-05 09:22:17 +00:00
github-actions[bot]
d17089972a chore: update community stats - followers increased (190 -> 191) 2026-02-05 07:27:09 +00:00
github-actions[bot]
a7991147be chore: update community stats - followers increased (189 -> 190) 2026-02-05 04:39:47 +00:00
github-actions[bot]
be6eb35567 chore: update community stats - followers increased (188 -> 189) 2026-02-05 00:43:05 +00:00
github-actions[bot]
90bc295871 chore: update community stats - followers increased (187 -> 188) 2026-02-04 18:22:31 +00:00
github-actions[bot]
310ad5d1d3 chore: update community stats - points increased (223 -> 243), followers increased (186 -> 187) 2026-02-04 07:23:53 +00:00
github-actions[bot]
00ce724430 chore: update community stats - points increased (204 -> 223), followers increased (184 -> 186) 2026-02-04 05:25:30 +00:00
github-actions[bot]
43b68b3ff0 chore: update community stats - points increased (202 -> 204), followers increased (183 -> 184) 2026-02-04 00:39:22 +00:00
github-actions[bot]
b6c1335335 chore: update community stats - points increased (200 -> 202) 2026-02-03 16:23:48 +00:00
github-actions[bot]
8b0b33015f chore: update community stats - followers increased (182 -> 183) 2026-02-03 10:19:31 +00:00
github-actions[bot]
11ee9086b2 chore: update community stats - followers increased (181 -> 182) 2026-02-03 08:17:37 +00:00
github-actions[bot]
3578ffc543 chore: update community stats - points increased (198 -> 200) 2026-02-02 23:11:34 +00:00
github-actions[bot]
b8531f1979 chore: update community stats - followers increased (180 -> 181) 2026-02-02 14:21:14 +00:00
github-actions[bot]
8607afd4c3 chore: update community stats - points increased (197 -> 198), followers increased (179 -> 180) 2026-02-02 12:19:30 +00:00
github-actions[bot]
41d9963d35 chore: update community stats - followers increased (178 -> 179) 2026-02-02 09:24:12 +00:00
github-actions[bot]
bd5f3d3f7c chore: update community stats - followers increased (177 -> 178) 2026-02-02 04:46:56 +00:00
github-actions[bot]
dda8262bc0 chore: update community stats - followers increased (176 -> 177) 2026-02-01 10:09:57 +00:00
github-actions[bot]
0271013ec2 chore: update community stats - points increased (196 -> 197) 2026-01-31 21:07:54 +00:00
github-actions[bot]
e1f210d600 chore: update community stats - followers increased (175 -> 176) 2026-01-31 10:08:48 +00:00
github-actions[bot]
eac8f6f355 chore: update community stats - points increased (195 -> 196) 2026-01-31 09:11:17 +00:00
github-actions[bot]
3998b93034 chore: update community stats - points increased (194 -> 195) 2026-01-31 04:36:15 +00:00
github-actions[bot]
ae04e95e13 chore: update community stats - points increased (193 -> 194) 2026-01-30 22:09:21 +00:00
github-actions[bot]
12d638e134 chore: update community stats - followers increased (173 -> 175) 2026-01-30 21:11:14 +00:00
github-actions[bot]
4a9eb8ed3d chore: update community stats - points increased (190 -> 193), followers increased (172 -> 173) 2026-01-30 19:16:44 +00:00
github-actions[bot]
af127bbfd5 chore: update community stats - points increased (189 -> 190) 2026-01-30 18:17:05 +00:00
github-actions[bot]
db7bb6250a chore: update community stats - points increased (188 -> 189) 2026-01-30 17:17:25 +00:00
github-actions[bot]
24b029e617 chore: update community stats - points increased (187 -> 188) 2026-01-30 15:15:33 +00:00
github-actions[bot]
ff63ab3118 chore: update community stats - points increased (186 -> 187) 2026-01-30 03:03:29 +00:00
github-actions[bot]
07e7e74fe1 chore: update community stats - followers increased (171 -> 172) 2026-01-30 00:41:11 +00:00
github-actions[bot]
55456775c1 chore: update community stats - points increased (185 -> 186), followers increased (170 -> 171) 2026-01-29 22:11:07 +00:00
github-actions[bot]
05bb2e4644 chore: update community stats - points increased (184 -> 185) 2026-01-29 21:11:30 +00:00
github-actions[bot]
83fa20ed08 chore: update community stats - points increased (183 -> 184) 2026-01-29 20:11:59 +00:00
fujie
ec69524357 feat: add Open WebUI Prompt Plus to extensions list and documentation 2026-01-30 02:16:38 +08:00
github-actions[bot]
829361da63 chore: update community stats - points increased (182 -> 183), followers increased (169 -> 170) 2026-01-29 17:17:44 +00:00
github-actions[bot]
af05ecec6a chore: update community stats - points increased (180 -> 182) 2026-01-29 16:18:22 +00:00
github-actions[bot]
1e08ae7d10 chore: update community stats - points increased (179 -> 180) 2026-01-29 14:20:30 +00:00
github-actions[bot]
b24233ee07 chore: update community stats - followers increased (168 -> 169) 2026-01-29 11:15:47 +00:00
github-actions[bot]
e5d1550986 chore: update community stats - points increased (170 -> 179), followers increased (167 -> 168) 2026-01-29 09:19:59 +00:00
fujie
7f5deb603e feat: 添加支持暂存插件源码更新的功能 2026-01-29 13:21:18 +08:00
fujie
55b2a28f79 fix: 移除标题中的图标,简化信息图插件的标题 2026-01-29 11:12:32 +08:00
fujie
db9bcb2c31 feat: 添加 README 文件同步工具到 OpenWebUI 社区 2026-01-29 10:58:40 +08:00
fujie
ad2773e8f1 Update plugin documentation for various filters and actions
- Updated README.md and README_CN.md for the infographic plugin to reflect new features and bug fixes in version 1.5.0, including context-aware generation and language synchronization.
- Revised README.md and README_CN.md for the smart mind map plugin to include support for user feedback and a changelog.
- Enhanced README.md and README_CN.md for the async context compression filter with critical fixes and improved compatibility details.
- Introduced initial release notes for the folder memory filter, detailing its core features and installation instructions.
- Updated markdown normalizer documentation to synchronize version numbers and improve clarity on configuration options.
- Revised GitHub Copilot SDK documentation to enhance installation instructions and troubleshooting sections, including a new changelog.
2026-01-29 03:22:21 +08:00
github-actions[bot]
7d4da3be8a chore: update community stats - followers increased (166 -> 167) 2026-01-28 18:14:59 +00:00
github-actions[bot]
28166728a4 chore: update community stats - points increased (169 -> 170) 2026-01-28 15:14:05 +00:00
github-actions[bot]
4dfca903c4 chore: update community stats - points increased (168 -> 169) 2026-01-28 10:11:40 +00:00
github-actions[bot]
b619d3f402 chore: update community stats - points increased (167 -> 168), followers increased (165 -> 166) 2026-01-28 06:14:04 +00:00
github-actions[bot]
1e68f985fb chore: update community stats - new plugin added (19 -> 20), plugin version updated 2026-01-28 05:13:56 +00:00
github-actions[bot]
596b571887 chore: update community stats - plugin version updated 2026-01-28 03:38:17 +00:00
fujie
c4d36c32a0 docs(pipes): fix copilot sdk tools link
Point tools usage link to GitHub source
2026-01-28 11:15:23 +08:00
fujie
6adbcd8d42 fix(scripts): resolve plugin scan NameError
Define metadata per file before use
2026-01-28 11:11:38 +08:00
fujie
89c039fe33 fix(actions): bump smart mind map to 0.9.2
Align mind map language rule with input text

Update plugin docs and README versions
2026-01-28 11:09:18 +08:00
github-actions[bot]
3a73ccfaa7 chore: update community stats - points increased (166 -> 167) 2026-01-28 01:37:45 +00:00
github-actions[bot]
7eff265e1c chore: update community stats - points increased (161 -> 166) 2026-01-27 21:07:14 +00:00
github-actions[bot]
989b45fc16 chore: update community stats - points increased (157 -> 161) 2026-01-27 20:09:28 +00:00
fujie
163d8ce8bd fix(scripts): exclude debug directory from release scanning 2026-01-28 02:30:38 +08:00
fujie
4e32e1a1da fix(scripts): normalize plugin paths in version extraction to prevent false positives in release diffs 2026-01-28 02:28:20 +08:00
github-actions[bot]
070e9f2456 chore: update community stats - plugin version updated 2026-01-27 18:15:04 +00:00
fujie
219ba83df3 feat(infographic): release v1.5.0 with smart language detection & organize debug tools 2026-01-28 02:14:30 +08:00
github-actions[bot]
e412aeb93d chore: update community stats - followers increased (164 -> 165) 2026-01-27 17:12:55 +00:00
github-actions[bot]
38102ca0c4 chore: update community stats - followers increased (163 -> 164) 2026-01-27 13:23:17 +00:00
github-actions[bot]
6ab69fba1c chore: update community stats - plugin version updated, followers increased (162 -> 163) 2026-01-26 21:09:49 +00:00
fujie
e0c0f69dc8 fix: update Git operations rules to allow direct pushes to main branch 2026-01-27 04:39:37 +08:00
fujie
7921b14dae fix: remove obsolete openwebui_id from SDK metadata 2026-01-27 04:32:22 +08:00
70 changed files with 8837 additions and 998 deletions

View File

@@ -42,11 +42,15 @@ plugins/actions/export_to_docx/
- 格式: `**Author:** [Fu-Jie](https://github.com/Fu-Jie) | **Version:** x.x.x | **Project:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)` - 格式: `**Author:** [Fu-Jie](https://github.com/Fu-Jie) | **Version:** x.x.x | **Project:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)`
- **注意**: Author 和 Project 为固定值,仅需更新 Version 版本号 - **注意**: Author 和 Project 为固定值,仅需更新 Version 版本号
3. **描述 (Description)**: 一句话功能介绍 3. **描述 (Description)**: 一句话功能介绍
4. **最新更新 (What's New)**: **必须**放在描述之后,显著展示最新版本的变更点 (仅展示最近 3 次更新) 4. **最新更新 (What's New)**: **必须**放在描述之后,仅展示**最近 1**更新
5. **核心特性 (Key Features)**: 使用 Emoji + 粗体标题 + 描述格式 5. **核心特性 (Key Features)**: 使用 Emoji + 粗体标题 + 描述格式
6. **使用方法 (How to Use)**: 按步骤说明 6. **使用方法 (How to Use)**: 按步骤说明
7. **配置参数 (Configuration/Valves)**: 使用表格格式,包含参数名、默认值、描述 7. **配置参数 (Configuration/Valves)**: 使用表格格式,包含参数名、默认值、描述
8. **其他 (Others)**: 支持的模板类型、语法示例、故障排除 8. **支持 (Support)**: **必须**包含,放在配置参数之后、故障排除之前
- English: `If this plugin has been useful, a star on [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) is a big motivation for me. Thank you for the support.`
- 中文: `如果这个插件对你有帮助,欢迎到 [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) 点个 Star这将是我持续改进的动力感谢支持。`
9. **其他 (Others)**: 支持的模板类型、语法示例、故障排除等
- **Changelog**: 统一指向 GitHub 项目历史,不在 README 中列出具体变更
### 2. 插件目录结构 (Plugin Directory Structure) ### 2. 插件目录结构 (Plugin Directory Structure)
@@ -62,18 +66,41 @@ plugins/
│ ├── ACTION_PLUGIN_TEMPLATE_CN.py # Chinese template │ ├── ACTION_PLUGIN_TEMPLATE_CN.py # Chinese template
│ └── README.md │ └── README.md
├── filters/ # Filter 插件 (输入处理) ├── filters/ # Filter 插件 (输入处理)
── my_filter/ ── ...
│ │ ├── my_filter.py
│ │ ├── 我的过滤器.py
│ │ ├── README.md
│ │ └── README_CN.md
│ └── README.md
├── pipes/ # Pipe 插件 (输出处理) ├── pipes/ # Pipe 插件 (输出处理)
│ └── ... │ └── ...
── pipelines/ # Pipeline 插件 ── pipelines/ # Pipeline 插件
└── ... └── ...
├── debug/ # 调试与开发工具 (Debug & Development Tools)
│ ├── my_debug_tool/
│ │ ├── debug_script.py
│ │ └── notes.md
│ └── ...
``` ```
#### 调试目录规范 (Debug Directory Standards)
`plugins/debug/` 目录用于存放调试用的脚本、临时验证代码或开发笔记。
**目录结构 (Directory Structure)**:
应根据调试工具所属的插件或功能模块进行子目录分类,而非将文件散落在根目录。
```
plugins/debug/
├── my_plugin_name/ # 特定插件的调试文件 (Debug files for specific plugin)
│ ├── debug_script.py
│ └── guides/
├── common_tools/ # 通用调试工具 (General debug tools)
│ └── ...
└── ...
```
**规范说明 (Guidelines)**:
- **不强制要求 README**: 该目录下的子项目不需要包含 `README.md`
- **发布豁免**: 该目录下的内容**绝不会**被发布脚本处理。
- **内容灵活性**: 可以包含 Python 脚本、Markdown 文档、JSON 数据等。
- **分类存放**: 任何调试产物(如 `test_*.py`, `inspect_*.py`)都不应直接存放在项目根目录,必须移动到此目录下相应的子文件夹中。
### 3. 文档字符串规范 (Docstring Standard) ### 3. 文档字符串规范 (Docstring Standard)
每个插件文件必须以标准化的文档字符串开头: 每个插件文件必须以标准化的文档字符串开头:
@@ -409,6 +436,51 @@ async def long_running_task_with_notification(self, event_emitter, ...):
return task_future.result() return task_future.result()
``` ```
### 7. 前端数据获取与交互规范 (Frontend Data Access & Interaction)
#### 获取前端信息 (Retrieving Frontend Info)
当需要获取用户浏览器的上下文信息如语言、时区、LocalStorage**必须**使用 `__event_call__``execute` 类型,而不是通过文件上传或复杂的 API 请求。
```python
async def _get_frontend_value(self, js_code: str) -> str:
"""Helper to execute JS and get return value."""
try:
response = await __event_call__(
{
"type": "execute",
"data": {
"code": js_code,
},
}
)
return str(response)
except Exception as e:
logger.error(f"Failed to execute JS: {e}")
return ""
# 示例:获取界面语言 (Get UI Language)
async def get_user_language(self):
js_code = """
return (
localStorage.getItem('locale') ||
localStorage.getItem('language') ||
navigator.language ||
'en-US'
);
"""
return await self._get_frontend_value(js_code)
```
#### 适用场景与引导 (Usage Guidelines)
- **语言适配**: 动态获取界面语言 (`ru-RU`, `zh-CN`) 自动切换输出语言。
- **时区处理**: 获取 `Intl.DateTimeFormat().resolvedOptions().timeZone` 处理时间。
- **客户端存储**: 读取 `localStorage` 中的用户偏好设置。
- **硬件能力**: 获取 `navigator.clipboard``navigator.geolocation` (需授权)。
**注意**: 即使插件有 `Valves` 配置,也应优先尝试自动探测,提升用户体验。
--- ---
## ⚡ Action 插件规范 (Action Plugin Standards) ## ⚡ Action 插件规范 (Action Plugin Standards)
@@ -855,8 +927,8 @@ Filter 实例是**单例 (Singleton)**。
### 4. 🤖 Git Operations (Agent Rules) ### 4. 🤖 Git Operations (Agent Rules)
- **允许**: 直接推送到 `main` 分支并发布。
- **允许**: 创建功能分支 (`feature/xxx`),推送到功能分支。 - **允许**: 创建功能分支 (`feature/xxx`),推送到功能分支。
- **禁止**: 直接推送到 `main`,自动合并到 `main`
### 5. 🤝 贡献者认可规范 (Contributor Recognition) ### 5. 🤝 贡献者认可规范 (Contributor Recognition)

View File

@@ -6,6 +6,7 @@ on:
- main - main
paths: paths:
- 'plugins/**/*.py' - 'plugins/**/*.py'
- '!plugins/debug/**'
release: release:
types: [published] types: [published]
workflow_dispatch: workflow_dispatch:

View File

@@ -10,28 +10,28 @@ A collection of enhancements, plugins, and prompts for [OpenWebUI](https://githu
<!-- STATS_START --> <!-- STATS_START -->
## 📊 Community Stats ## 📊 Community Stats
> 🕐 Auto-updated: 2026-01-27 02:13 > 🕐 Auto-updated: 2026-02-07 09:38
| 👤 Author | 👥 Followers | ⭐ Points | 🏆 Contributions | | 👤 Author | 👥 Followers | ⭐ Points | 🏆 Contributions |
|:---:|:---:|:---:|:---:| |:---:|:---:|:---:|:---:|
| [Fu-Jie](https://openwebui.com/u/Fu-Jie) | **162** | **157** | **33** | | [Fu-Jie](https://openwebui.com/u/Fu-Jie) | **203** | **246** | **40** |
| 📝 Posts | ⬇️ Downloads | 👁️ Views | 👍 Upvotes | 💾 Saves | | 📝 Posts | ⬇️ Downloads | 👁️ Views | 👍 Upvotes | 💾 Saves |
|:---:|:---:|:---:|:---:|:---:| |:---:|:---:|:---:|:---:|:---:|
| **19** | **2438** | **28054** | **141** | **188** | | **20** | **3675** | **43326** | **213** | **253** |
### 🔥 Top 6 Popular Plugins ### 🔥 Top 6 Popular Plugins
> 🕐 Auto-updated: 2026-01-27 02:13 > 🕐 Auto-updated: 2026-02-07 09:38
| Rank | Plugin | Version | Downloads | Views | Updated | | Rank | Plugin | Version | Downloads | Views | Updated |
|:---:|------|:---:|:---:|:---:|:---:| |:---:|------|:---:|:---:|:---:|:---:|
| 🥇 | [Smart Mind Map](https://openwebui.com/posts/turn_any_text_into_beautiful_mind_maps_3094c59a) | 0.9.1 | 639 | 5682 | 2026-01-17 | | 🥇 | [Smart Mind Map](https://openwebui.com/posts/turn_any_text_into_beautiful_mind_maps_3094c59a) | 0.9.2 | 877 | 7765 | 2026-01-28 |
| 🥈 | [Smart Infographic](https://openwebui.com/posts/smart_infographic_ad6f0c7f) | 1.4.9 | 415 | 3740 | 2026-01-25 | | 🥈 | [Smart Infographic](https://openwebui.com/posts/smart_infographic_ad6f0c7f) | 1.5.0 | 631 | 5825 | 2026-01-30 |
| 🥉 | [Export to Excel](https://openwebui.com/posts/export_mulit_table_to_excel_244b8f9d) | 0.3.7 | 262 | 1067 | 2026-01-07 | | 🥉 | [Export to Word (Enhanced)](https://openwebui.com/posts/export_to_word_enhanced_formatting_fca6a315) | 0.4.3 | 343 | 2659 | 2026-01-28 |
| 4⃣ | [Export to Word (Enhanced)](https://openwebui.com/posts/export_to_word_enhanced_formatting_fca6a315) | 0.4.3 | 235 | 1872 | 2026-01-17 | | 4⃣ | [Export to Excel](https://openwebui.com/posts/export_mulit_table_to_excel_244b8f9d) | 0.3.7 | 327 | 1519 | 2026-01-29 |
| 5⃣ | [Async Context Compression](https://openwebui.com/posts/async_context_compression_b1655bc8) | 1.2.2 | 232 | 2499 | 2026-01-21 | | 5⃣ | [Async Context Compression](https://openwebui.com/posts/async_context_compression_b1655bc8) | 1.2.2 | 327 | 3412 | 2026-01-28 |
| 6⃣ | [Flash Card](https://openwebui.com/posts/flash_card_65a2ea8f) | 0.2.4 | 169 | 2708 | 2026-01-17 | | 6⃣ | [Markdown Normalizer](https://openwebui.com/posts/markdown_normalizer_baaa8732) | 1.2.4 | 293 | 4243 | 2026-01-29 |
*See full stats in [Community Stats Report](./docs/community-stats.md)* *See full stats in [Community Stats Report](./docs/community-stats.md)*
<!-- STATS_END --> <!-- STATS_END -->
@@ -73,6 +73,12 @@ Located in the `prompts/` directory, containing fine-tuned System Prompts:
- **Coding**: Programming assistance prompts. - **Coding**: Programming assistance prompts.
- **Marketing**: Marketing and copywriting prompts. - **Marketing**: Marketing and copywriting prompts.
## 🛠️ Extensions
Standalone frontend extensions to supercharge your Open WebUI:
- **[Open WebUI Prompt Plus](https://github.com/Fu-Jie/open-webui-prompt-plus)**: An all-in-one prompt management suite featuring AI-powered prompt generation, spotlight-style quick search, and advanced category organization.
## 📖 Documentation ## 📖 Documentation
Located in the `docs/en/` directory: Located in the `docs/en/` directory:

View File

@@ -7,28 +7,28 @@ OpenWebUI 增强功能集合。包含个人开发与收集的插件、提示词
<!-- STATS_START --> <!-- STATS_START -->
## 📊 社区统计 ## 📊 社区统计
> 🕐 自动更新于 2026-01-27 02:13 > 🕐 自动更新于 2026-02-07 09:38
| 👤 作者 | 👥 粉丝 | ⭐ 积分 | 🏆 贡献 | | 👤 作者 | 👥 粉丝 | ⭐ 积分 | 🏆 贡献 |
|:---:|:---:|:---:|:---:| |:---:|:---:|:---:|:---:|
| [Fu-Jie](https://openwebui.com/u/Fu-Jie) | **162** | **157** | **33** | | [Fu-Jie](https://openwebui.com/u/Fu-Jie) | **203** | **246** | **40** |
| 📝 发布 | ⬇️ 下载 | 👁️ 浏览 | 👍 点赞 | 💾 收藏 | | 📝 发布 | ⬇️ 下载 | 👁️ 浏览 | 👍 点赞 | 💾 收藏 |
|:---:|:---:|:---:|:---:|:---:| |:---:|:---:|:---:|:---:|:---:|
| **19** | **2438** | **28054** | **141** | **188** | | **20** | **3675** | **43326** | **213** | **253** |
### 🔥 热门插件 Top 6 ### 🔥 热门插件 Top 6
> 🕐 自动更新于 2026-01-27 02:13 > 🕐 自动更新于 2026-02-07 09:38
| 排名 | 插件 | 版本 | 下载 | 浏览 | 更新日期 | | 排名 | 插件 | 版本 | 下载 | 浏览 | 更新日期 |
|:---:|------|:---:|:---:|:---:|:---:| |:---:|------|:---:|:---:|:---:|:---:|
| 🥇 | [Smart Mind Map](https://openwebui.com/posts/turn_any_text_into_beautiful_mind_maps_3094c59a) | 0.9.1 | 639 | 5682 | 2026-01-17 | | 🥇 | [Smart Mind Map](https://openwebui.com/posts/turn_any_text_into_beautiful_mind_maps_3094c59a) | 0.9.2 | 877 | 7765 | 2026-01-28 |
| 🥈 | [Smart Infographic](https://openwebui.com/posts/smart_infographic_ad6f0c7f) | 1.4.9 | 415 | 3740 | 2026-01-25 | | 🥈 | [Smart Infographic](https://openwebui.com/posts/smart_infographic_ad6f0c7f) | 1.5.0 | 631 | 5825 | 2026-01-30 |
| 🥉 | [Export to Excel](https://openwebui.com/posts/export_mulit_table_to_excel_244b8f9d) | 0.3.7 | 262 | 1067 | 2026-01-07 | | 🥉 | [Export to Word (Enhanced)](https://openwebui.com/posts/export_to_word_enhanced_formatting_fca6a315) | 0.4.3 | 343 | 2659 | 2026-01-28 |
| 4⃣ | [Export to Word (Enhanced)](https://openwebui.com/posts/export_to_word_enhanced_formatting_fca6a315) | 0.4.3 | 235 | 1872 | 2026-01-17 | | 4⃣ | [Export to Excel](https://openwebui.com/posts/export_mulit_table_to_excel_244b8f9d) | 0.3.7 | 327 | 1519 | 2026-01-29 |
| 5⃣ | [Async Context Compression](https://openwebui.com/posts/async_context_compression_b1655bc8) | 1.2.2 | 232 | 2499 | 2026-01-21 | | 5⃣ | [Async Context Compression](https://openwebui.com/posts/async_context_compression_b1655bc8) | 1.2.2 | 327 | 3412 | 2026-01-28 |
| 6⃣ | [Flash Card](https://openwebui.com/posts/flash_card_65a2ea8f) | 0.2.4 | 169 | 2708 | 2026-01-17 | | 6⃣ | [Markdown Normalizer](https://openwebui.com/posts/markdown_normalizer_baaa8732) | 1.2.4 | 293 | 4243 | 2026-01-29 |
*完整统计请查看 [社区统计报告](./docs/community-stats.zh.md)* *完整统计请查看 [社区统计报告](./docs/community-stats.zh.md)*
<!-- STATS_END --> <!-- STATS_END -->
@@ -76,6 +76,12 @@ OpenWebUI 增强功能集合。包含个人开发与收集的插件、提示词
每个提示词都独立保存为 Markdown 文件,可直接在 OpenWebUI 中使用。 每个提示词都独立保存为 Markdown 文件,可直接在 OpenWebUI 中使用。
## 🛠️ 扩展 (Extensions)
Open WebUI 的前端增强扩展:
- **[Open WebUI Prompt Plus](https://github.com/Fu-Jie/open-webui-prompt-plus)**: 一站式提示词管理套件,支持 AI 提示词生成、Spotlight 风格快速搜索及高级分类管理。
## 📖 开发文档 ## 📖 开发文档
位于 `docs/zh/` 目录: 位于 `docs/zh/` 目录:

View File

@@ -10,9 +10,10 @@ The Chinese version (README_CN.md) MUST be translated based on this English vers
## What's New ## What's New
<!-- Keep the changelog for the last 3 versions here. Remove this section for the initial release. --> <!-- Keep only the latest update here. Remove this section for the initial release. -->
### v1.0.0 ### v1.0.0
- **Initial Release**: Released the first version of the plugin. - **Initial Release**: Released the first version of the plugin.
- **[Feature Name]**: [Brief description of the feature]. - **[Feature Name]**: [Brief description of the feature].
@@ -36,9 +37,17 @@ The Chinese version (README_CN.md) MUST be translated based on this English vers
| `VALVE_NAME` | `Default Value` | Description of what this setting does. | | `VALVE_NAME` | `Default Value` | Description of what this setting does. |
| `ANOTHER_VALVE` | `True` | Another setting description. | | `ANOTHER_VALVE` | `True` | Another setting description. |
## ⭐ Support
If this plugin has been useful, a star on [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) is a big motivation for me. Thank you for the support.
## Troubleshooting ❓ ## Troubleshooting ❓
- **Plugin not working?**: Check if the filter/action is enabled in the model settings. - **Plugin not working?**: Check if the filter/action is enabled in the model settings.
- **Debug Logs**: Enable `SHOW_DEBUG_LOG` in Valves and check the browser console (F12) for detailed logs. - **Debug Logs**: Enable `SHOW_DEBUG_LOG` in Valves and check the browser console (F12) for detailed logs.
- **Error Messages**: If you see an error, please copy the full error message and report it. - **Error Messages**: If you see an error, please copy the full error message and report it.
- **Submit an Issue**: If you encounter any problems, please submit an issue on GitHub: [Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues) - **Submit an Issue**: If you encounter any problems, please submit an issue on GitHub: [Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
## Changelog
See the full history on GitHub: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)

View File

@@ -1,7 +1,7 @@
{ {
"schemaVersion": 1, "schemaVersion": 1,
"label": "downloads", "label": "downloads",
"message": "2.4k", "message": "3.7k",
"color": "blue", "color": "blue",
"namedLogo": "openwebui" "namedLogo": "openwebui"
} }

View File

@@ -1,6 +1,6 @@
{ {
"schemaVersion": 1, "schemaVersion": 1,
"label": "followers", "label": "followers",
"message": "162", "message": "203",
"color": "blue" "color": "blue"
} }

View File

@@ -1,6 +1,6 @@
{ {
"schemaVersion": 1, "schemaVersion": 1,
"label": "plugins", "label": "plugins",
"message": "19", "message": "20",
"color": "green" "color": "green"
} }

View File

@@ -1,6 +1,6 @@
{ {
"schemaVersion": 1, "schemaVersion": 1,
"label": "points", "label": "points",
"message": "157", "message": "246",
"color": "orange" "color": "orange"
} }

View File

@@ -1,6 +1,6 @@
{ {
"schemaVersion": 1, "schemaVersion": 1,
"label": "upvotes", "label": "upvotes",
"message": "141", "message": "213",
"color": "brightgreen" "color": "brightgreen"
} }

View File

@@ -1,15 +1,14 @@
{ {
"total_posts": 19, "total_posts": 20,
"total_downloads": 2438, "total_downloads": 3675,
"total_views": 28054, "total_views": 43326,
"total_upvotes": 141, "total_upvotes": 213,
"total_downvotes": 2, "total_downvotes": 2,
"total_saves": 188, "total_saves": 253,
"total_comments": 37, "total_comments": 46,
"by_type": { "by_type": {
"pipe": 1, "action": 15,
"action": 14, "unknown": 4,
"unknown": 3,
"filter": 1 "filter": 1
}, },
"posts": [ "posts": [
@@ -17,50 +16,34 @@
"title": "Smart Mind Map", "title": "Smart Mind Map",
"slug": "turn_any_text_into_beautiful_mind_maps_3094c59a", "slug": "turn_any_text_into_beautiful_mind_maps_3094c59a",
"type": "action", "type": "action",
"version": "0.9.1", "version": "0.9.2",
"author": "Fu-Jie", "author": "Fu-Jie",
"description": "Intelligently analyzes text content and generates interactive mind maps to help users structure and visualize knowledge.", "description": "Intelligently analyzes text content and generates interactive mind maps to help users structure and visualize knowledge.",
"downloads": 639, "downloads": 877,
"views": 5682, "views": 7765,
"upvotes": 17, "upvotes": 21,
"saves": 38, "saves": 47,
"comments": 11, "comments": 13,
"created_at": "2025-12-30", "created_at": "2025-12-30",
"updated_at": "2026-01-17", "updated_at": "2026-01-28",
"url": "https://openwebui.com/posts/turn_any_text_into_beautiful_mind_maps_3094c59a" "url": "https://openwebui.com/posts/turn_any_text_into_beautiful_mind_maps_3094c59a"
}, },
{ {
"title": "Smart Infographic", "title": "Smart Infographic",
"slug": "smart_infographic_ad6f0c7f", "slug": "smart_infographic_ad6f0c7f",
"type": "action", "type": "action",
"version": "1.4.9", "version": "1.5.0",
"author": "Fu-Jie", "author": "Fu-Jie",
"description": "AI-powered infographic generator based on AntV Infographic. Supports professional templates, auto-icon matching, and SVG/PNG downloads.", "description": "AI-powered infographic generator based on AntV Infographic. Supports professional templates, auto-icon matching, and SVG/PNG downloads.",
"downloads": 415, "downloads": 631,
"views": 3740, "views": 5825,
"upvotes": 18, "upvotes": 23,
"saves": 27, "saves": 33,
"comments": 10, "comments": 10,
"created_at": "2025-12-28", "created_at": "2025-12-28",
"updated_at": "2026-01-25", "updated_at": "2026-01-30",
"url": "https://openwebui.com/posts/smart_infographic_ad6f0c7f" "url": "https://openwebui.com/posts/smart_infographic_ad6f0c7f"
}, },
{
"title": "Export to Excel",
"slug": "export_mulit_table_to_excel_244b8f9d",
"type": "action",
"version": "0.3.7",
"author": "Fu-Jie",
"description": "Extracts tables from chat messages and exports them to Excel (.xlsx) files with smart formatting.",
"downloads": 262,
"views": 1067,
"upvotes": 4,
"saves": 6,
"comments": 0,
"created_at": "2025-05-30",
"updated_at": "2026-01-07",
"url": "https://openwebui.com/posts/export_mulit_table_to_excel_244b8f9d"
},
{ {
"title": "Export to Word (Enhanced)", "title": "Export to Word (Enhanced)",
"slug": "export_to_word_enhanced_formatting_fca6a315", "slug": "export_to_word_enhanced_formatting_fca6a315",
@@ -68,15 +51,31 @@
"version": "0.4.3", "version": "0.4.3",
"author": "Fu-Jie", "author": "Fu-Jie",
"description": "Export current conversation from Markdown to Word (.docx) with Mermaid diagrams rendered client-side (Mermaid.js, SVG+PNG), LaTeX math, real hyperlinks, improved tables, syntax highlighting, and blockquote support.", "description": "Export current conversation from Markdown to Word (.docx) with Mermaid diagrams rendered client-side (Mermaid.js, SVG+PNG), LaTeX math, real hyperlinks, improved tables, syntax highlighting, and blockquote support.",
"downloads": 235, "downloads": 343,
"views": 1872, "views": 2659,
"upvotes": 8, "upvotes": 12,
"saves": 21, "saves": 26,
"comments": 0, "comments": 2,
"created_at": "2026-01-03", "created_at": "2026-01-03",
"updated_at": "2026-01-17", "updated_at": "2026-01-28",
"url": "https://openwebui.com/posts/export_to_word_enhanced_formatting_fca6a315" "url": "https://openwebui.com/posts/export_to_word_enhanced_formatting_fca6a315"
}, },
{
"title": "Export to Excel",
"slug": "export_mulit_table_to_excel_244b8f9d",
"type": "action",
"version": "0.3.7",
"author": "Fu-Jie",
"description": "Extracts tables from chat messages and exports them to Excel (.xlsx) files with smart formatting.",
"downloads": 327,
"views": 1519,
"upvotes": 7,
"saves": 6,
"comments": 0,
"created_at": "2025-05-30",
"updated_at": "2026-01-29",
"url": "https://openwebui.com/posts/export_mulit_table_to_excel_244b8f9d"
},
{ {
"title": "Async Context Compression", "title": "Async Context Compression",
"slug": "async_context_compression_b1655bc8", "slug": "async_context_compression_b1655bc8",
@@ -84,31 +83,15 @@
"version": "1.2.2", "version": "1.2.2",
"author": "Fu-Jie", "author": "Fu-Jie",
"description": "Reduces token consumption in long conversations while maintaining coherence through intelligent summarization and message compression.", "description": "Reduces token consumption in long conversations while maintaining coherence through intelligent summarization and message compression.",
"downloads": 232, "downloads": 327,
"views": 2499, "views": 3412,
"upvotes": 9, "upvotes": 13,
"saves": 27, "saves": 33,
"comments": 0, "comments": 0,
"created_at": "2025-11-08", "created_at": "2025-11-08",
"updated_at": "2026-01-21", "updated_at": "2026-01-28",
"url": "https://openwebui.com/posts/async_context_compression_b1655bc8" "url": "https://openwebui.com/posts/async_context_compression_b1655bc8"
}, },
{
"title": "Flash Card",
"slug": "flash_card_65a2ea8f",
"type": "action",
"version": "0.2.4",
"author": "Fu-Jie",
"description": "Quickly generates beautiful flashcards from text, extracting key points and categories.",
"downloads": 169,
"views": 2708,
"upvotes": 11,
"saves": 13,
"comments": 2,
"created_at": "2025-12-30",
"updated_at": "2026-01-17",
"url": "https://openwebui.com/posts/flash_card_65a2ea8f"
},
{ {
"title": "Markdown Normalizer", "title": "Markdown Normalizer",
"slug": "markdown_normalizer_baaa8732", "slug": "markdown_normalizer_baaa8732",
@@ -116,15 +99,47 @@
"version": "1.2.4", "version": "1.2.4",
"author": "Fu-Jie", "author": "Fu-Jie",
"description": "A content normalizer filter that fixes common Markdown formatting issues in LLM outputs, such as broken code blocks, LaTeX formulas, and list formatting.", "description": "A content normalizer filter that fixes common Markdown formatting issues in LLM outputs, such as broken code blocks, LaTeX formulas, and list formatting.",
"downloads": 153, "downloads": 293,
"views": 2810, "views": 4243,
"upvotes": 10, "upvotes": 17,
"saves": 20, "saves": 27,
"comments": 5, "comments": 5,
"created_at": "2026-01-12", "created_at": "2026-01-12",
"updated_at": "2026-01-19", "updated_at": "2026-01-29",
"url": "https://openwebui.com/posts/markdown_normalizer_baaa8732" "url": "https://openwebui.com/posts/markdown_normalizer_baaa8732"
}, },
{
"title": "Flash Card",
"slug": "flash_card_65a2ea8f",
"type": "action",
"version": "0.2.4",
"author": "Fu-Jie",
"description": "Quickly generates beautiful flashcards from text, extracting key points and categories.",
"downloads": 213,
"views": 3239,
"upvotes": 13,
"saves": 14,
"comments": 2,
"created_at": "2025-12-30",
"updated_at": "2026-01-28",
"url": "https://openwebui.com/posts/flash_card_65a2ea8f"
},
{
"title": "AI Task Instruction Generator",
"slug": "ai_task_instruction_generator_9bab8b37",
"type": "unknown",
"version": "",
"author": "",
"description": "",
"downloads": 141,
"views": 2180,
"upvotes": 8,
"saves": 3,
"comments": 0,
"created_at": "2026-01-28",
"updated_at": "2026-01-28",
"url": "https://openwebui.com/posts/ai_task_instruction_generator_9bab8b37"
},
{ {
"title": "Deep Dive", "title": "Deep Dive",
"slug": "deep_dive_c0b846e4", "slug": "deep_dive_c0b846e4",
@@ -132,10 +147,10 @@
"version": "1.0.0", "version": "1.0.0",
"author": "Fu-Jie", "author": "Fu-Jie",
"description": "A comprehensive thinking lens that dives deep into any content - from context to logic, insights, and action paths.", "description": "A comprehensive thinking lens that dives deep into any content - from context to logic, insights, and action paths.",
"downloads": 93, "downloads": 132,
"views": 850, "views": 1155,
"upvotes": 4, "upvotes": 6,
"saves": 8, "saves": 11,
"comments": 0, "comments": 0,
"created_at": "2026-01-08", "created_at": "2026-01-08",
"updated_at": "2026-01-08", "updated_at": "2026-01-08",
@@ -148,29 +163,29 @@
"version": "0.4.3", "version": "0.4.3",
"author": "Fu-Jie", "author": "Fu-Jie",
"description": "将对话导出为 Word (.docx),支持 Mermaid 图表 (客户端渲染 SVG+PNG)、LaTeX 数学公式、真实超链接、增强表格格式、代码高亮和引用块。", "description": "将对话导出为 Word (.docx),支持 Mermaid 图表 (客户端渲染 SVG+PNG)、LaTeX 数学公式、真实超链接、增强表格格式、代码高亮和引用块。",
"downloads": 87, "downloads": 119,
"views": 1635, "views": 2104,
"upvotes": 11, "upvotes": 13,
"saves": 4, "saves": 6,
"comments": 4, "comments": 4,
"created_at": "2026-01-04", "created_at": "2026-01-04",
"updated_at": "2026-01-17", "updated_at": "2026-01-28",
"url": "https://openwebui.com/posts/导出为_word_支持公式流程图表格和代码块_8a6306c0" "url": "https://openwebui.com/posts/导出为_word_支持公式流程图表格和代码块_8a6306c0"
}, },
{ {
"title": "📊 智能信息图 (AntV Infographic)", "title": "智能信息图",
"slug": "智能信息图_e04a48ff", "slug": "智能信息图_e04a48ff",
"type": "action", "type": "action",
"version": "1.4.9", "version": "1.5.0",
"author": "Fu-Jie", "author": "Fu-Jie",
"description": "基于 AntV Infographic 的智能信息图生成插件。支持多种专业模板,自动图标匹配,并提供 SVG/PNG 下载功能。", "description": "基于 AntV Infographic 的智能信息图生成插件。支持多种专业模板,自动图标匹配,并提供 SVG/PNG 下载功能。",
"downloads": 47, "downloads": 56,
"views": 792, "views": 1003,
"upvotes": 6, "upvotes": 10,
"saves": 0, "saves": 1,
"comments": 0, "comments": 0,
"created_at": "2025-12-28", "created_at": "2025-12-28",
"updated_at": "2026-01-17", "updated_at": "2026-01-29",
"url": "https://openwebui.com/posts/智能信息图_e04a48ff" "url": "https://openwebui.com/posts/智能信息图_e04a48ff"
}, },
{ {
@@ -180,29 +195,45 @@
"version": "0.1.0", "version": "0.1.0",
"author": "Fu-Jie", "author": "Fu-Jie",
"description": "Automatically extracts project rules from conversations and injects them into the folder's system prompt.", "description": "Automatically extracts project rules from conversations and injects them into the folder's system prompt.",
"downloads": 27, "downloads": 55,
"views": 760, "views": 1195,
"upvotes": 4, "upvotes": 6,
"saves": 5, "saves": 8,
"comments": 0, "comments": 0,
"created_at": "2026-01-20", "created_at": "2026-01-20",
"updated_at": "2026-01-20", "updated_at": "2026-01-20",
"url": "https://openwebui.com/posts/folder_memory_auto_evolving_project_context_4a9875b2" "url": "https://openwebui.com/posts/folder_memory_auto_evolving_project_context_4a9875b2"
}, },
{
"title": "GitHub Copilot Official SDK Pipe",
"slug": "github_copilot_official_sdk_pipe_ce96f7b4",
"type": "action",
"version": "0.2.3",
"author": "Fu-Jie",
"description": "Integrate GitHub Copilot SDK. Supports dynamic models, multi-turn conversation, streaming, multimodal input, infinite sessions, and frontend debug logging.",
"downloads": 49,
"views": 1835,
"upvotes": 11,
"saves": 5,
"comments": 1,
"created_at": "2026-01-26",
"updated_at": "2026-01-29",
"url": "https://openwebui.com/posts/github_copilot_official_sdk_pipe_ce96f7b4"
},
{ {
"title": "思维导图", "title": "思维导图",
"slug": "智能生成交互式思维导图帮助用户可视化知识_8d4b097b", "slug": "智能生成交互式思维导图帮助用户可视化知识_8d4b097b",
"type": "action", "type": "action",
"version": "0.9.1", "version": "0.9.2",
"author": "Fu-Jie", "author": "Fu-Jie",
"description": "智能分析文本内容,生成交互式思维导图,帮助用户结构化和可视化知识。", "description": "智能分析文本内容,生成交互式思维导图,帮助用户结构化和可视化知识。",
"downloads": 27, "downloads": 37,
"views": 457, "views": 563,
"upvotes": 4, "upvotes": 6,
"saves": 1, "saves": 2,
"comments": 0, "comments": 0,
"created_at": "2025-12-31", "created_at": "2025-12-31",
"updated_at": "2026-01-17", "updated_at": "2026-01-28",
"url": "https://openwebui.com/posts/智能生成交互式思维导图帮助用户可视化知识_8d4b097b" "url": "https://openwebui.com/posts/智能生成交互式思维导图帮助用户可视化知识_8d4b097b"
}, },
{ {
@@ -212,13 +243,13 @@
"version": "1.2.2", "version": "1.2.2",
"author": "Fu-Jie", "author": "Fu-Jie",
"description": "通过智能摘要和消息压缩,降低长对话的 token 消耗,同时保持对话连贯性。", "description": "通过智能摘要和消息压缩,降低长对话的 token 消耗,同时保持对话连贯性。",
"downloads": 22, "downloads": 31,
"views": 502, "views": 624,
"upvotes": 5, "upvotes": 7,
"saves": 2, "saves": 4,
"comments": 0, "comments": 0,
"created_at": "2025-11-08", "created_at": "2025-11-08",
"updated_at": "2026-01-21", "updated_at": "2026-01-28",
"url": "https://openwebui.com/posts/异步上下文压缩_5c0617cb" "url": "https://openwebui.com/posts/异步上下文压缩_5c0617cb"
}, },
{ {
@@ -228,13 +259,13 @@
"version": "0.2.4", "version": "0.2.4",
"author": "Fu-Jie", "author": "Fu-Jie",
"description": "快速将文本提炼为精美的学习记忆卡片,支持核心要点提取与分类。", "description": "快速将文本提炼为精美的学习记忆卡片,支持核心要点提取与分类。",
"downloads": 19, "downloads": 27,
"views": 514, "views": 659,
"upvotes": 6, "upvotes": 8,
"saves": 1, "saves": 1,
"comments": 0, "comments": 0,
"created_at": "2025-12-30", "created_at": "2025-12-30",
"updated_at": "2026-01-17", "updated_at": "2026-01-28",
"url": "https://openwebui.com/posts/闪记卡生成插件_4a31eac3" "url": "https://openwebui.com/posts/闪记卡生成插件_4a31eac3"
}, },
{ {
@@ -244,31 +275,15 @@
"version": "1.0.0", "version": "1.0.0",
"author": "Fu-Jie", "author": "Fu-Jie",
"description": "全方位的思维透镜 —— 从背景全景到逻辑脉络,从深度洞察到行动路径。", "description": "全方位的思维透镜 —— 从背景全景到逻辑脉络,从深度洞察到行动路径。",
"downloads": 9, "downloads": 17,
"views": 311, "views": 400,
"upvotes": 3, "upvotes": 5,
"saves": 1, "saves": 1,
"comments": 0, "comments": 0,
"created_at": "2026-01-08", "created_at": "2026-01-08",
"updated_at": "2026-01-08", "updated_at": "2026-01-08",
"url": "https://openwebui.com/posts/精读_99830b0f" "url": "https://openwebui.com/posts/精读_99830b0f"
}, },
{
"title": "GitHub Copilot Official SDK Pipe",
"slug": "github_copilot_official_sdk_pipe_ce96f7b4",
"type": "pipe",
"version": "0.1.1",
"author": "Fu-Jie",
"description": "Integrate GitHub Copilot SDK. Supports dynamic models, multi-turn conversation, streaming, multimodal input, and infinite sessions (context compaction).",
"downloads": 2,
"views": 81,
"upvotes": 2,
"saves": 0,
"comments": 0,
"created_at": "2026-01-26",
"updated_at": "2026-01-26",
"url": "https://openwebui.com/posts/github_copilot_official_sdk_pipe_ce96f7b4"
},
{ {
"title": "🚀 Open WebUI Prompt Plus: AI-Powered Prompt Manager", "title": "🚀 Open WebUI Prompt Plus: AI-Powered Prompt Manager",
"slug": "open_webui_prompt_plus_ai_powered_prompt_manager_s_15fa060e", "slug": "open_webui_prompt_plus_ai_powered_prompt_manager_s_15fa060e",
@@ -277,12 +292,12 @@
"author": "", "author": "",
"description": "", "description": "",
"downloads": 0, "downloads": 0,
"views": 401, "views": 1421,
"upvotes": 6, "upvotes": 11,
"saves": 6, "saves": 16,
"comments": 3, "comments": 7,
"created_at": "2026-01-25", "created_at": "2026-01-25",
"updated_at": "2026-01-25", "updated_at": "2026-01-28",
"url": "https://openwebui.com/posts/open_webui_prompt_plus_ai_powered_prompt_manager_s_15fa060e" "url": "https://openwebui.com/posts/open_webui_prompt_plus_ai_powered_prompt_manager_s_15fa060e"
}, },
{ {
@@ -293,8 +308,8 @@
"author": "", "author": "",
"description": "", "description": "",
"downloads": 0, "downloads": 0,
"views": 95, "views": 134,
"upvotes": 1, "upvotes": 2,
"saves": 0, "saves": 0,
"comments": 0, "comments": 0,
"created_at": "2026-01-14", "created_at": "2026-01-14",
@@ -309,9 +324,9 @@
"author": "", "author": "",
"description": "", "description": "",
"downloads": 0, "downloads": 0,
"views": 1278, "views": 1391,
"upvotes": 12, "upvotes": 14,
"saves": 8, "saves": 9,
"comments": 2, "comments": 2,
"created_at": "2026-01-10", "created_at": "2026-01-10",
"updated_at": "2026-01-10", "updated_at": "2026-01-10",
@@ -323,11 +338,11 @@
"name": "Fu-Jie", "name": "Fu-Jie",
"profile_url": "https://openwebui.com/u/Fu-Jie", "profile_url": "https://openwebui.com/u/Fu-Jie",
"profile_image": "https://community.s3.openwebui.com/uploads/users/b15d1348-4347-42b4-b815-e053342d6cb0/profile_d9510745-4bd4-4f8f-a997-4a21847d9300.webp", "profile_image": "https://community.s3.openwebui.com/uploads/users/b15d1348-4347-42b4-b815-e053342d6cb0/profile_d9510745-4bd4-4f8f-a997-4a21847d9300.webp",
"followers": 162, "followers": 203,
"following": 3, "following": 4,
"total_points": 157, "total_points": 246,
"post_points": 139, "post_points": 211,
"comment_points": 18, "comment_points": 35,
"contributions": 33 "contributions": 40
} }
} }

View File

@@ -1,45 +1,45 @@
# 📊 OpenWebUI Community Stats Report # 📊 OpenWebUI Community Stats Report
> 📅 Updated: 2026-01-27 02:13 > 📅 Updated: 2026-02-07 09:38
## 📈 Overview ## 📈 Overview
| Metric | Value | | Metric | Value |
|------|------| |------|------|
| 📝 Total Posts | 19 | | 📝 Total Posts | 20 |
| ⬇️ Total Downloads | 2438 | | ⬇️ Total Downloads | 3675 |
| 👁️ Total Views | 28054 | | 👁️ Total Views | 43326 |
| 👍 Total Upvotes | 141 | | 👍 Total Upvotes | 213 |
| 💾 Total Saves | 188 | | 💾 Total Saves | 253 |
| 💬 Total Comments | 37 | | 💬 Total Comments | 46 |
## 📂 By Type ## 📂 By Type
- **pipe**: 1 - **action**: 15
- **action**: 14 - **unknown**: 4
- **unknown**: 3
- **filter**: 1 - **filter**: 1
## 📋 Posts List ## 📋 Posts List
| Rank | Title | Type | Version | Downloads | Views | Upvotes | Saves | Updated | | Rank | Title | Type | Version | Downloads | Views | Upvotes | Saves | Updated |
|:---:|------|:---:|:---:|:---:|:---:|:---:|:---:|:---:| |:---:|------|:---:|:---:|:---:|:---:|:---:|:---:|:---:|
| 1 | [Smart Mind Map](https://openwebui.com/posts/turn_any_text_into_beautiful_mind_maps_3094c59a) | action | 0.9.1 | 639 | 5682 | 17 | 38 | 2026-01-17 | | 1 | [Smart Mind Map](https://openwebui.com/posts/turn_any_text_into_beautiful_mind_maps_3094c59a) | action | 0.9.2 | 877 | 7765 | 21 | 47 | 2026-01-28 |
| 2 | [Smart Infographic](https://openwebui.com/posts/smart_infographic_ad6f0c7f) | action | 1.4.9 | 415 | 3740 | 18 | 27 | 2026-01-25 | | 2 | [Smart Infographic](https://openwebui.com/posts/smart_infographic_ad6f0c7f) | action | 1.5.0 | 631 | 5825 | 23 | 33 | 2026-01-30 |
| 3 | [Export to Excel](https://openwebui.com/posts/export_mulit_table_to_excel_244b8f9d) | action | 0.3.7 | 262 | 1067 | 4 | 6 | 2026-01-07 | | 3 | [Export to Word (Enhanced)](https://openwebui.com/posts/export_to_word_enhanced_formatting_fca6a315) | action | 0.4.3 | 343 | 2659 | 12 | 26 | 2026-01-28 |
| 4 | [Export to Word (Enhanced)](https://openwebui.com/posts/export_to_word_enhanced_formatting_fca6a315) | action | 0.4.3 | 235 | 1872 | 8 | 21 | 2026-01-17 | | 4 | [Export to Excel](https://openwebui.com/posts/export_mulit_table_to_excel_244b8f9d) | action | 0.3.7 | 327 | 1519 | 7 | 6 | 2026-01-29 |
| 5 | [Async Context Compression](https://openwebui.com/posts/async_context_compression_b1655bc8) | action | 1.2.2 | 232 | 2499 | 9 | 27 | 2026-01-21 | | 5 | [Async Context Compression](https://openwebui.com/posts/async_context_compression_b1655bc8) | action | 1.2.2 | 327 | 3412 | 13 | 33 | 2026-01-28 |
| 6 | [Flash Card](https://openwebui.com/posts/flash_card_65a2ea8f) | action | 0.2.4 | 169 | 2708 | 11 | 13 | 2026-01-17 | | 6 | [Markdown Normalizer](https://openwebui.com/posts/markdown_normalizer_baaa8732) | action | 1.2.4 | 293 | 4243 | 17 | 27 | 2026-01-29 |
| 7 | [Markdown Normalizer](https://openwebui.com/posts/markdown_normalizer_baaa8732) | action | 1.2.4 | 153 | 2810 | 10 | 20 | 2026-01-19 | | 7 | [Flash Card](https://openwebui.com/posts/flash_card_65a2ea8f) | action | 0.2.4 | 213 | 3239 | 13 | 14 | 2026-01-28 |
| 8 | [Deep Dive](https://openwebui.com/posts/deep_dive_c0b846e4) | action | 1.0.0 | 93 | 850 | 4 | 8 | 2026-01-08 | | 8 | [AI Task Instruction Generator](https://openwebui.com/posts/ai_task_instruction_generator_9bab8b37) | unknown | | 141 | 2180 | 8 | 3 | 2026-01-28 |
| 9 | [导出为 Word (增强版)](https://openwebui.com/posts/导出为_word_支持公式流程图表格和代码块_8a6306c0) | action | 0.4.3 | 87 | 1635 | 11 | 4 | 2026-01-17 | | 9 | [Deep Dive](https://openwebui.com/posts/deep_dive_c0b846e4) | action | 1.0.0 | 132 | 1155 | 6 | 11 | 2026-01-08 |
| 10 | [📊 智能信息图 (AntV Infographic)](https://openwebui.com/posts/智能信息图_e04a48ff) | action | 1.4.9 | 47 | 792 | 6 | 0 | 2026-01-17 | | 10 | [导出为 Word (增强版)](https://openwebui.com/posts/导出为_word_支持公式流程图表格和代码块_8a6306c0) | action | 0.4.3 | 119 | 2104 | 13 | 6 | 2026-01-28 |
| 11 | [📂 Folder Memory Auto-Evolving Project Context](https://openwebui.com/posts/folder_memory_auto_evolving_project_context_4a9875b2) | filter | 0.1.0 | 27 | 760 | 4 | 5 | 2026-01-20 | | 11 | [智能信息图](https://openwebui.com/posts/智能信息图_e04a48ff) | action | 1.5.0 | 56 | 1003 | 10 | 1 | 2026-01-29 |
| 12 | [思维导图](https://openwebui.com/posts/智能生成交互式思维导图帮助用户可视化知识_8d4b097b) | action | 0.9.1 | 27 | 457 | 4 | 1 | 2026-01-17 | | 12 | [📂 Folder Memory Auto-Evolving Project Context](https://openwebui.com/posts/folder_memory_auto_evolving_project_context_4a9875b2) | filter | 0.1.0 | 55 | 1195 | 6 | 8 | 2026-01-20 |
| 13 | [异步上下文压缩](https://openwebui.com/posts/异步上下文压缩_5c0617cb) | action | 1.2.2 | 22 | 502 | 5 | 2 | 2026-01-21 | | 13 | [GitHub Copilot Official SDK Pipe](https://openwebui.com/posts/github_copilot_official_sdk_pipe_ce96f7b4) | action | 0.2.3 | 49 | 1835 | 11 | 5 | 2026-01-29 |
| 14 | [闪记卡 (Flash Card)](https://openwebui.com/posts/闪记卡生成插件_4a31eac3) | action | 0.2.4 | 19 | 514 | 6 | 1 | 2026-01-17 | | 14 | [思维导图](https://openwebui.com/posts/智能生成交互式思维导图帮助用户可视化知识_8d4b097b) | action | 0.9.2 | 37 | 563 | 6 | 2 | 2026-01-28 |
| 15 | [精读](https://openwebui.com/posts/精读_99830b0f) | action | 1.0.0 | 9 | 311 | 3 | 1 | 2026-01-08 | | 15 | [异步上下文压缩](https://openwebui.com/posts/异步上下文压缩_5c0617cb) | action | 1.2.2 | 31 | 624 | 7 | 4 | 2026-01-28 |
| 16 | [GitHub Copilot Official SDK Pipe](https://openwebui.com/posts/github_copilot_official_sdk_pipe_ce96f7b4) | pipe | 0.1.1 | 2 | 81 | 2 | 0 | 2026-01-26 | | 16 | [闪记卡 (Flash Card)](https://openwebui.com/posts/闪记卡生成插件_4a31eac3) | action | 0.2.4 | 27 | 659 | 8 | 1 | 2026-01-28 |
| 17 | [🚀 Open WebUI Prompt Plus: AI-Powered Prompt Manager](https://openwebui.com/posts/open_webui_prompt_plus_ai_powered_prompt_manager_s_15fa060e) | unknown | | 0 | 401 | 6 | 6 | 2026-01-25 | | 17 | [精读](https://openwebui.com/posts/精读_99830b0f) | action | 1.0.0 | 17 | 400 | 5 | 1 | 2026-01-08 |
| 18 | [Review of Claude Haiku 4.5](https://openwebui.com/posts/review_of_claude_haiku_45_41b0db39) | unknown | | 0 | 95 | 1 | 0 | 2026-01-14 | | 18 | [🚀 Open WebUI Prompt Plus: AI-Powered Prompt Manager](https://openwebui.com/posts/open_webui_prompt_plus_ai_powered_prompt_manager_s_15fa060e) | unknown | | 0 | 1421 | 11 | 16 | 2026-01-28 |
| 19 | [ 🛠️ Debug Open WebUI Plugins in Your Browser](https://openwebui.com/posts/debug_open_webui_plugins_in_your_browser_81bf7960) | unknown | | 0 | 1278 | 12 | 8 | 2026-01-10 | | 19 | [Review of Claude Haiku 4.5](https://openwebui.com/posts/review_of_claude_haiku_45_41b0db39) | unknown | | 0 | 134 | 2 | 0 | 2026-01-14 |
| 20 | [ 🛠️ Debug Open WebUI Plugins in Your Browser](https://openwebui.com/posts/debug_open_webui_plugins_in_your_browser_81bf7960) | unknown | | 0 | 1391 | 14 | 9 | 2026-01-10 |

View File

@@ -1,45 +1,45 @@
# 📊 OpenWebUI 社区统计报告 # 📊 OpenWebUI 社区统计报告
> 📅 更新时间: 2026-01-27 02:13 > 📅 更新时间: 2026-02-07 09:38
## 📈 总览 ## 📈 总览
| 指标 | 数值 | | 指标 | 数值 |
|------|------| |------|------|
| 📝 发布数量 | 19 | | 📝 发布数量 | 20 |
| ⬇️ 总下载量 | 2438 | | ⬇️ 总下载量 | 3675 |
| 👁️ 总浏览量 | 28054 | | 👁️ 总浏览量 | 43326 |
| 👍 总点赞数 | 141 | | 👍 总点赞数 | 213 |
| 💾 总收藏数 | 188 | | 💾 总收藏数 | 253 |
| 💬 总评论数 | 37 | | 💬 总评论数 | 46 |
## 📂 按类型分类 ## 📂 按类型分类
- **pipe**: 1 - **action**: 15
- **action**: 14 - **unknown**: 4
- **unknown**: 3
- **filter**: 1 - **filter**: 1
## 📋 发布列表 ## 📋 发布列表
| 排名 | 标题 | 类型 | 版本 | 下载 | 浏览 | 点赞 | 收藏 | 更新日期 | | 排名 | 标题 | 类型 | 版本 | 下载 | 浏览 | 点赞 | 收藏 | 更新日期 |
|:---:|------|:---:|:---:|:---:|:---:|:---:|:---:|:---:| |:---:|------|:---:|:---:|:---:|:---:|:---:|:---:|:---:|
| 1 | [Smart Mind Map](https://openwebui.com/posts/turn_any_text_into_beautiful_mind_maps_3094c59a) | action | 0.9.1 | 639 | 5682 | 17 | 38 | 2026-01-17 | | 1 | [Smart Mind Map](https://openwebui.com/posts/turn_any_text_into_beautiful_mind_maps_3094c59a) | action | 0.9.2 | 877 | 7765 | 21 | 47 | 2026-01-28 |
| 2 | [Smart Infographic](https://openwebui.com/posts/smart_infographic_ad6f0c7f) | action | 1.4.9 | 415 | 3740 | 18 | 27 | 2026-01-25 | | 2 | [Smart Infographic](https://openwebui.com/posts/smart_infographic_ad6f0c7f) | action | 1.5.0 | 631 | 5825 | 23 | 33 | 2026-01-30 |
| 3 | [Export to Excel](https://openwebui.com/posts/export_mulit_table_to_excel_244b8f9d) | action | 0.3.7 | 262 | 1067 | 4 | 6 | 2026-01-07 | | 3 | [Export to Word (Enhanced)](https://openwebui.com/posts/export_to_word_enhanced_formatting_fca6a315) | action | 0.4.3 | 343 | 2659 | 12 | 26 | 2026-01-28 |
| 4 | [Export to Word (Enhanced)](https://openwebui.com/posts/export_to_word_enhanced_formatting_fca6a315) | action | 0.4.3 | 235 | 1872 | 8 | 21 | 2026-01-17 | | 4 | [Export to Excel](https://openwebui.com/posts/export_mulit_table_to_excel_244b8f9d) | action | 0.3.7 | 327 | 1519 | 7 | 6 | 2026-01-29 |
| 5 | [Async Context Compression](https://openwebui.com/posts/async_context_compression_b1655bc8) | action | 1.2.2 | 232 | 2499 | 9 | 27 | 2026-01-21 | | 5 | [Async Context Compression](https://openwebui.com/posts/async_context_compression_b1655bc8) | action | 1.2.2 | 327 | 3412 | 13 | 33 | 2026-01-28 |
| 6 | [Flash Card](https://openwebui.com/posts/flash_card_65a2ea8f) | action | 0.2.4 | 169 | 2708 | 11 | 13 | 2026-01-17 | | 6 | [Markdown Normalizer](https://openwebui.com/posts/markdown_normalizer_baaa8732) | action | 1.2.4 | 293 | 4243 | 17 | 27 | 2026-01-29 |
| 7 | [Markdown Normalizer](https://openwebui.com/posts/markdown_normalizer_baaa8732) | action | 1.2.4 | 153 | 2810 | 10 | 20 | 2026-01-19 | | 7 | [Flash Card](https://openwebui.com/posts/flash_card_65a2ea8f) | action | 0.2.4 | 213 | 3239 | 13 | 14 | 2026-01-28 |
| 8 | [Deep Dive](https://openwebui.com/posts/deep_dive_c0b846e4) | action | 1.0.0 | 93 | 850 | 4 | 8 | 2026-01-08 | | 8 | [AI Task Instruction Generator](https://openwebui.com/posts/ai_task_instruction_generator_9bab8b37) | unknown | | 141 | 2180 | 8 | 3 | 2026-01-28 |
| 9 | [导出为 Word (增强版)](https://openwebui.com/posts/导出为_word_支持公式流程图表格和代码块_8a6306c0) | action | 0.4.3 | 87 | 1635 | 11 | 4 | 2026-01-17 | | 9 | [Deep Dive](https://openwebui.com/posts/deep_dive_c0b846e4) | action | 1.0.0 | 132 | 1155 | 6 | 11 | 2026-01-08 |
| 10 | [📊 智能信息图 (AntV Infographic)](https://openwebui.com/posts/智能信息图_e04a48ff) | action | 1.4.9 | 47 | 792 | 6 | 0 | 2026-01-17 | | 10 | [导出为 Word (增强版)](https://openwebui.com/posts/导出为_word_支持公式流程图表格和代码块_8a6306c0) | action | 0.4.3 | 119 | 2104 | 13 | 6 | 2026-01-28 |
| 11 | [📂 Folder Memory Auto-Evolving Project Context](https://openwebui.com/posts/folder_memory_auto_evolving_project_context_4a9875b2) | filter | 0.1.0 | 27 | 760 | 4 | 5 | 2026-01-20 | | 11 | [智能信息图](https://openwebui.com/posts/智能信息图_e04a48ff) | action | 1.5.0 | 56 | 1003 | 10 | 1 | 2026-01-29 |
| 12 | [思维导图](https://openwebui.com/posts/智能生成交互式思维导图帮助用户可视化知识_8d4b097b) | action | 0.9.1 | 27 | 457 | 4 | 1 | 2026-01-17 | | 12 | [📂 Folder Memory Auto-Evolving Project Context](https://openwebui.com/posts/folder_memory_auto_evolving_project_context_4a9875b2) | filter | 0.1.0 | 55 | 1195 | 6 | 8 | 2026-01-20 |
| 13 | [异步上下文压缩](https://openwebui.com/posts/异步上下文压缩_5c0617cb) | action | 1.2.2 | 22 | 502 | 5 | 2 | 2026-01-21 | | 13 | [GitHub Copilot Official SDK Pipe](https://openwebui.com/posts/github_copilot_official_sdk_pipe_ce96f7b4) | action | 0.2.3 | 49 | 1835 | 11 | 5 | 2026-01-29 |
| 14 | [闪记卡 (Flash Card)](https://openwebui.com/posts/闪记卡生成插件_4a31eac3) | action | 0.2.4 | 19 | 514 | 6 | 1 | 2026-01-17 | | 14 | [思维导图](https://openwebui.com/posts/智能生成交互式思维导图帮助用户可视化知识_8d4b097b) | action | 0.9.2 | 37 | 563 | 6 | 2 | 2026-01-28 |
| 15 | [精读](https://openwebui.com/posts/精读_99830b0f) | action | 1.0.0 | 9 | 311 | 3 | 1 | 2026-01-08 | | 15 | [异步上下文压缩](https://openwebui.com/posts/异步上下文压缩_5c0617cb) | action | 1.2.2 | 31 | 624 | 7 | 4 | 2026-01-28 |
| 16 | [GitHub Copilot Official SDK Pipe](https://openwebui.com/posts/github_copilot_official_sdk_pipe_ce96f7b4) | pipe | 0.1.1 | 2 | 81 | 2 | 0 | 2026-01-26 | | 16 | [闪记卡 (Flash Card)](https://openwebui.com/posts/闪记卡生成插件_4a31eac3) | action | 0.2.4 | 27 | 659 | 8 | 1 | 2026-01-28 |
| 17 | [🚀 Open WebUI Prompt Plus: AI-Powered Prompt Manager](https://openwebui.com/posts/open_webui_prompt_plus_ai_powered_prompt_manager_s_15fa060e) | unknown | | 0 | 401 | 6 | 6 | 2026-01-25 | | 17 | [精读](https://openwebui.com/posts/精读_99830b0f) | action | 1.0.0 | 17 | 400 | 5 | 1 | 2026-01-08 |
| 18 | [Review of Claude Haiku 4.5](https://openwebui.com/posts/review_of_claude_haiku_45_41b0db39) | unknown | | 0 | 95 | 1 | 0 | 2026-01-14 | | 18 | [🚀 Open WebUI Prompt Plus: AI-Powered Prompt Manager](https://openwebui.com/posts/open_webui_prompt_plus_ai_powered_prompt_manager_s_15fa060e) | unknown | | 0 | 1421 | 11 | 16 | 2026-01-28 |
| 19 | [ 🛠️ Debug Open WebUI Plugins in Your Browser](https://openwebui.com/posts/debug_open_webui_plugins_in_your_browser_81bf7960) | unknown | | 0 | 1278 | 12 | 8 | 2026-01-10 | | 19 | [Review of Claude Haiku 4.5](https://openwebui.com/posts/review_of_claude_haiku_45_41b0db39) | unknown | | 0 | 134 | 2 | 0 | 2026-01-14 |
| 20 | [ 🛠️ Debug Open WebUI Plugins in Your Browser](https://openwebui.com/posts/debug_open_webui_plugins_in_your_browser_81bf7960) | unknown | | 0 | 1391 | 14 | 9 | 2026-01-10 |

22
docs/extensions/index.md Normal file
View File

@@ -0,0 +1,22 @@
---
title: Extensions
---
# Extensions
Standalone frontend extensions to supercharge your Open WebUI.
---
## 🚀 Open WebUI Prompt Plus
**An all-in-one prompt management suite for power users.**
[Open WebUI Prompt Plus](https://github.com/Fu-Jie/open-webui-prompt-plus) elevates your experience with:
- **🤖 AI-Powered Prompt Generator**: Turn natural language into structured prompts.
- **⚡ Quick Insert Panel**: Spotlight-style search (`Cmd/Ctrl + Shift + P`).
- **📂 Advanced Management**: Dynamic categories, favorites, and usage stats.
- **📝 Native Variable Support**: Visual form rendering for template variables.
[:fontawesome-brands-github: View on GitHub](https://github.com/Fu-Jie/open-webui-prompt-plus){ .md-button .md-button--primary }

View File

@@ -0,0 +1,22 @@
---
title: 扩展 (Extensions)
---
# 扩展 (Extensions)
Open WebUI 的独立前端增强扩展。
---
## 🚀 Open WebUI Prompt Plus
**专为高级用户打造的一站式提示词管理套件。**
[Open WebUI Prompt Plus](https://github.com/Fu-Jie/open-webui-prompt-plus) 通过以下功能提升您的体验:
- **🤖 AI 提示词生成器**: 将自然语言转化为结构化提示词。
- **⚡ 快速插入面板**: Spotlight 风格的快速搜索 (`Cmd/Ctrl + Shift + P`)。
- **📂 高级管理**: 动态分类、收藏夹及使用统计。
- **📝 原生变量支持**: 模板变量的可视化表单渲染。
[:fontawesome-brands-github: 查看 GitHub](https://github.com/Fu-Jie/open-webui-prompt-plus){ .md-button .md-button--primary }

View File

@@ -349,6 +349,53 @@ await __event_emitter__(
) )
``` ```
#### Advanced Use Case: Retrieving Frontend Data
One of the most powerful capabilities of the `execute` event type is the ability to fetch data from the browser environment (JavaScript) and return it to your Python backend. This allows plugins to access information like:
- `localStorage` items (user preferences, tokens)
- `navigator` properties (language, geolocation, platform)
- `document` properties (cookies, URL parameters)
**How it works:**
The JavaScript code you provide in the `"code"` field is executed in the browser. If your JS code includes a `return` statement, that value is sent back to Python as the result of `await __event_call__`.
**Example: Getting the User's UI Language**
```python
try:
# Execute JS on the frontend to get language settings
response = await __event_call__(
{
"type": "execute",
"data": {
# This JS code runs in the browser.
# The 'return' value is sent back to Python.
"code": """
return (
localStorage.getItem('locale') ||
localStorage.getItem('language') ||
navigator.language ||
'en-US'
);
""",
},
}
)
# 'response' will contain the string returned by JS (e.g., "en-US", "zh-CN")
# Note: Wrap in try-except to handle potential timeouts or JS errors
logger.info(f"Frontend Language: {response}")
except Exception as e:
logger.error(f"Failed to get frontend data: {e}")
```
**Key capabilities unlocked:**
- **Context Awareness:** Adapt responses based on user time zone or language.
- **Client-Side Storage:** Use `localStorage` to persist simple plugin settings without a database.
- **Hardware Access:** Request geolocation or clipboard access (requires user permission).
--- ---
## 🏗️ When & Where to Use Events ## 🏗️ When & Where to Use Events
@@ -421,4 +468,4 @@ Refer to this document for common event types and structures, and explore Open W
--- ---
**Happy event-driven coding in Open WebUI! 🚀** **Happy event-driven coding in Open WebUI! 🚀**

View File

@@ -25,7 +25,7 @@ hide:
<div class="grid cards" markdown> <div class="grid cards" markdown>
- :material-puzzle:{ .lg .middle } **Plugin Center** - :material-puzzle:{ .lg .middle } **Plugin Center**
--- ---
@@ -33,7 +33,7 @@ hide:
[:octicons-arrow-right-24: Explore Plugins](plugins/index.md) [:octicons-arrow-right-24: Explore Plugins](plugins/index.md)
- :material-message-text:{ .lg .middle } **Prompt Library** - :material-message-text:{ .lg .middle } **Prompt Library**
--- ---
@@ -41,7 +41,7 @@ hide:
[:octicons-arrow-right-24: Browse Prompts](prompts/index.md) [:octicons-arrow-right-24: Browse Prompts](prompts/index.md)
- :material-tools:{ .lg .middle } **Enhancements** - :material-tools:{ .lg .middle } **Enhancements**
--- ---
@@ -49,7 +49,15 @@ hide:
[:octicons-arrow-right-24: View Guides](enhancements/index.md) [:octicons-arrow-right-24: View Guides](enhancements/index.md)
- :material-book-open-page-variant:{ .lg .middle } **Development** - :material-rocket-launch:{ .lg .middle } **Extensions**
---
Standalone frontend extensions to supercharge your Open WebUI interface.
[:octicons-arrow-right-24: Browse Extensions](extensions/index.md)
- :material-book-open-page-variant:{ .lg .middle } **Development**
--- ---
@@ -65,7 +73,7 @@ hide:
<div class="grid cards" markdown> <div class="grid cards" markdown>
- :material-brain:{ .lg .middle } **Smart Mind Map** - :material-brain:{ .lg .middle } **Smart Mind Map**
--- ---
@@ -73,7 +81,7 @@ hide:
[:octicons-arrow-right-24: Learn More](plugins/actions/smart-mind-map.md) [:octicons-arrow-right-24: Learn More](plugins/actions/smart-mind-map.md)
- :material-card-text:{ .lg .middle } **Flash Card** - :material-card-text:{ .lg .middle } **Flash Card**
--- ---
@@ -81,7 +89,7 @@ hide:
[:octicons-arrow-right-24: Learn More](plugins/actions/flash-card.md) [:octicons-arrow-right-24: Learn More](plugins/actions/flash-card.md)
- :material-arrow-collapse-vertical:{ .lg .middle } **Async Context Compression** - :material-arrow-collapse-vertical:{ .lg .middle } **Async Context Compression**
--- ---

View File

@@ -25,7 +25,7 @@ hide:
<div class="grid cards" markdown> <div class="grid cards" markdown>
- :material-puzzle:{ .lg .middle } **插件中心** - :material-puzzle:{ .lg .middle } **插件中心**
--- ---
@@ -33,7 +33,7 @@ hide:
[:octicons-arrow-right-24: 探索插件](plugins/index.md) [:octicons-arrow-right-24: 探索插件](plugins/index.md)
- :material-message-text:{ .lg .middle } **提示词库** - :material-message-text:{ .lg .middle } **提示词库**
--- ---
@@ -41,7 +41,7 @@ hide:
[:octicons-arrow-right-24: 浏览提示词](prompts/index.md) [:octicons-arrow-right-24: 浏览提示词](prompts/index.md)
- :material-tools:{ .lg .middle } **增强功能** - :material-tools:{ .lg .middle } **增强功能**
--- ---
@@ -49,7 +49,15 @@ hide:
[:octicons-arrow-right-24: 查看指南](enhancements/index.md) [:octicons-arrow-right-24: 查看指南](enhancements/index.md)
- :material-book-open-page-variant:{ .lg .middle } **开发指南** - :material-rocket-launch:{ .lg .middle } **扩展 (Extensions)**
---
独立的 OpenWebUI 前端增强扩展,全面提升交互体验。
[:octicons-arrow-right-24: 浏览扩展](extensions/index.md)
- :material-book-open-page-variant:{ .lg .middle } **开发指南**
--- ---
@@ -65,7 +73,7 @@ hide:
<div class="grid cards" markdown> <div class="grid cards" markdown>
- :material-brain:{ .lg .middle } **智能思维导图** - :material-brain:{ .lg .middle } **智能思维导图**
--- ---
@@ -73,7 +81,7 @@ hide:
[:octicons-arrow-right-24: 了解更多](plugins/actions/smart-mind-map.md) [:octicons-arrow-right-24: 了解更多](plugins/actions/smart-mind-map.md)
- :material-card-text:{ .lg .middle } **Flash Card闪记卡** - :material-card-text:{ .lg .middle } **Flash Card闪记卡**
--- ---
@@ -81,7 +89,7 @@ hide:
[:octicons-arrow-right-24: 了解更多](plugins/actions/flash-card.md) [:octicons-arrow-right-24: 了解更多](plugins/actions/flash-card.md)
- :material-arrow-collapse-vertical:{ .lg .middle } **异步上下文压缩** - :material-arrow-collapse-vertical:{ .lg .middle } **异步上下文压缩**
--- ---

View File

@@ -1,7 +1,7 @@
# Smart Mind Map # Smart Mind Map
<span class="category-badge action">Action</span> <span class="category-badge action">Action</span>
<span class="version-badge">v0.9.1</span> <span class="version-badge">v0.9.2</span>
Intelligently analyzes text content and generates interactive mind maps for better visualization and understanding. Intelligently analyzes text content and generates interactive mind maps for better visualization and understanding.
@@ -17,7 +17,7 @@ The Smart Mind Map plugin transforms text content into beautiful, interactive mi
- :material-gesture-swipe: **Rich Controls**: Zoom, reset view, expand level selector (All/2/3) and fullscreen - :material-gesture-swipe: **Rich Controls**: Zoom, reset view, expand level selector (All/2/3) and fullscreen
- :material-palette: **Theme Aware**: Auto-detects OpenWebUI light/dark theme with manual toggle - :material-palette: **Theme Aware**: Auto-detects OpenWebUI light/dark theme with manual toggle
- :material-download: **One-Click Export**: Download high-res PNG, copy SVG, or copy Markdown source - :material-download: **One-Click Export**: Download high-res PNG, copy SVG, or copy Markdown source
- :material-translate: **Multi-language**: Adapts output language to the user context - :material-translate: **Multi-language**: Matches output language to the input text
--- ---

View File

@@ -1,7 +1,7 @@
# Smart Mind Map智能思维导图 # Smart Mind Map智能思维导图
<span class="category-badge action">Action</span> <span class="category-badge action">Action</span>
<span class="version-badge">v0.9.1</span> <span class="version-badge">v0.9.2</span>
智能分析文本内容,生成交互式思维导图,帮助你更直观地理解信息结构。 智能分析文本内容,生成交互式思维导图,帮助你更直观地理解信息结构。
@@ -17,7 +17,7 @@ Smart Mind Map 会将文本转换成漂亮的交互式思维导图。插件会
- :material-gesture-swipe: **丰富控制**:缩放/重置、展开层级(全部/2/3 级)与全屏 - :material-gesture-swipe: **丰富控制**:缩放/重置、展开层级(全部/2/3 级)与全屏
- :material-palette: **主题感知**:自动检测 OpenWebUI 亮/暗色主题并支持手动切换 - :material-palette: **主题感知**:自动检测 OpenWebUI 亮/暗色主题并支持手动切换
- :material-download: **一键导出**:下载高分辨率 PNG、复制 SVG 或 Markdown - :material-download: **一键导出**:下载高分辨率 PNG、复制 SVG 或 Markdown
- :material-translate: **多语言**根据用户语言自动输出 - :material-translate: **多语言**:输出语言与输入文本一致
--- ---

View File

@@ -48,7 +48,7 @@ OpenWebUI supports four types of plugins, each serving a different purpose:
| Plugin | Type | Description | Version | | Plugin | Type | Description | Version |
|--------|------|-------------|---------| |--------|------|-------------|---------|
| [Smart Mind Map](actions/smart-mind-map.md) | Action | Generate interactive mind maps from text | 0.9.1 | | [Smart Mind Map](actions/smart-mind-map.md) | Action | Generate interactive mind maps from text | 0.9.2 |
| [Smart Infographic](actions/smart-infographic.md) | Action | Transform text into professional infographics | 1.4.9 | | [Smart Infographic](actions/smart-infographic.md) | Action | Transform text into professional infographics | 1.4.9 |
| [Flash Card](actions/flash-card.md) | Action | Create beautiful learning flashcards | 0.2.4 | | [Flash Card](actions/flash-card.md) | Action | Create beautiful learning flashcards | 0.2.4 |
| [Export to Excel](actions/export-to-excel.md) | Action | Export chat history to Excel files | 0.3.7 | | [Export to Excel](actions/export-to-excel.md) | Action | Export chat history to Excel files | 0.3.7 |

View File

@@ -48,7 +48,7 @@ OpenWebUI 支持四种类型的插件,每种都有不同的用途:
| 插件 | 类型 | 描述 | 版本 | | 插件 | 类型 | 描述 | 版本 |
|--------|------|-------------|---------| |--------|------|-------------|---------|
| [Smart Mind Map智能思维导图](actions/smart-mind-map.md) | Action | 从文本生成交互式思维导图 | 0.9.1 | | [Smart Mind Map智能思维导图](actions/smart-mind-map.md) | Action | 从文本生成交互式思维导图 | 0.9.2 |
| [Smart Infographic智能信息图](actions/smart-infographic.md) | Action | 将文本转成专业信息图 | 1.4.9 | | [Smart Infographic智能信息图](actions/smart-infographic.md) | Action | 将文本转成专业信息图 | 1.4.9 |
| [Flash Card闪记卡](actions/flash-card.md) | Action | 生成精美学习卡片 | 0.2.4 | | [Flash Card闪记卡](actions/flash-card.md) | Action | 生成精美学习卡片 | 0.2.4 |
| [Export to Excel导出到 Excel](actions/export-to-excel.md) | Action | 导出聊天记录为 Excel | 0.3.7 | | [Export to Excel导出到 Excel](actions/export-to-excel.md) | Action | 导出聊天记录为 Excel | 0.3.7 |

View File

@@ -1,26 +1,24 @@
# GitHub Copilot SDK Pipe for OpenWebUI # GitHub Copilot SDK Pipe for OpenWebUI
**Author:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **Version:** 0.1.0 | **Project:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **License:** MIT **Author:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **Version:** 0.3.0 | **Project:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **License:** MIT
This is an advanced Pipe function for [OpenWebUI](https://github.com/open-webui/open-webui) that allows you to use GitHub Copilot models (such as `gpt-5`, `gpt-5-mini`, `claude-sonnet-4.5`) directly within OpenWebUI. It is built upon the official [GitHub Copilot SDK for Python](https://github.com/github/copilot-sdk), providing a native integration experience. This is an advanced Pipe function for [OpenWebUI](https://github.com/open-webui/open-webui) that allows you to use GitHub Copilot models (such as `gpt-5`, `gpt-5-mini`, `claude-sonnet-4.5`) directly within OpenWebUI. It is built upon the official [GitHub Copilot SDK for Python](https://github.com/github/copilot-sdk), providing a native integration experience.
## 🚀 What's New (v0.1.0) ## 🚀 What's New (v0.3.0) - The Power of "Unified Ecosystem"
* **♾️ Infinite Sessions**: Automatic context compaction for long-running conversations. No more context limit errors! * **🔌 Zero-Config Tool Bridge**: Automatically transforms your existing OpenWebUI Functions (Tools) into Copilot-compatible tools. **Copilot now has total access to your entire WebUI toolset!**
* **🧠 Thinking Process**: Real-time display of model reasoning/thinking process (for supported models). * **🔗 Dynamic MCP Discovery**: Seamlessly connects to MCP servers defined in **Admin Settings -> Connections**. No configuration files required—it just works.
* **📂 Workspace Control**: Restricted workspace directory for secure file operations. * **⚡ High-Performance Async Engine**: Background CLI updates and optimized event-driven streaming ensure lightning-fast responses without UI lag.
* **🔍 Model Filtering**: Exclude specific models using keywords (e.g., `codex`, `haiku`). * **🛡️ Robust Interoperability**: Advanced sanitization and dynamic Pydantic model generation ensure smooth integration even with complex third-party tools.
* **💾 Session Persistence**: Improved session resume logic using OpenWebUI chat ID mapping.
## ✨ Core Features ## ✨ Key Capabilities
* **🚀 Official SDK Integration**: Built on the official SDK for stability and reliability. * **🌉 The Ultimate Bridge**: The first and only plugin that creates a seamless bridge between **OpenWebUI Tools** and **GitHub Copilot SDK**.
* **💬 Multi-turn Conversation**: Automatically concatenates history context so Copilot understands your previous messages. * **🚀 Official & Native**: Built directly on the official Python SDK, providing the most stable and authentic Copilot experience.
* **🌊 Streaming Output**: Supports typewriter effect for fast responses. * **🌊 Advanced Streaming (Thought Process)**: Supports full model reasoning/thinking display with typewriter effects.
* **🖼️ Multimodal Support**: Supports image uploads, automatically converting them to attachments for Copilot (requires model support). * **🖼️ Intelligent Multimodal**: Full support for images and attachments, enabling Copilot to "see" your workspace.
* **🛠️ Zero-config Installation**: Automatically detects and downloads the GitHub Copilot CLI, ready to use out of the box. * **🛠️ Effortless Setup**: Automatic CLI detection, version enforcement, and dependency management.
* **🔑 Secure Authentication**: Supports Fine-grained Personal Access Tokens for minimized permissions. * **🔑 Dual-Layer Security**: Supports secure OAuth flow for Chat and standard PAT for extended MCP capabilities.
* **🐛 Debug Mode**: Built-in detailed log output for easy connection troubleshooting.
## 📦 Installation & Usage ## 📦 Installation & Usage
@@ -38,27 +36,49 @@ Find "GitHub Copilot" in the function list and click the **⚙️ (Valves)** ico
| Parameter | Description | Default | | Parameter | Description | Default |
| :--- | :--- | :--- | | :--- | :--- | :--- |
| **GH_TOKEN** | **(Required)** Your GitHub Token. | - | | **GH_TOKEN** | **(Required)** GitHub Access Token (PAT or OAuth Token). Access to Chat. | - |
| **MODEL_ID** | The model name to use. Recommended `gpt-5-mini` or `gpt-5`. | `gpt-5-mini` | | **DEBUG** | Whether to enable debug logs (output to browser console). | `False` |
| **CLI_PATH** | Path to the Copilot CLI. Will download automatically if not found. | `/usr/local/bin/copilot` | | **LOG_LEVEL** | Copilot CLI log level: none, error, warning, info, debug, all. | `error` |
| **DEBUG** | Whether to enable debug logs (output to chat). | `True` | | **SHOW_THINKING** | Show model reasoning/thinking process (requires streaming + model support). | `True` |
| **SHOW_THINKING** | Show model reasoning/thinking process. | `True` | | **COPILOT_CLI_VERSION** | Specific Copilot CLI version to install/enforce. | `0.0.405` |
| **EXCLUDE_KEYWORDS** | Exclude models containing these keywords (comma separated). | - | | **EXCLUDE_KEYWORDS** | Exclude models containing these keywords (comma separated). | - |
| **WORKSPACE_DIR** | Restricted workspace directory for file operations. | - | | **WORKSPACE_DIR** | Restricted workspace directory for file operations. | - |
| **INFINITE_SESSION** | Enable Infinite Sessions (automatic context compaction). | `True` | | **INFINITE_SESSION** | Enable Infinite Sessions (automatic context compaction). | `True` |
| **COMPACTION_THRESHOLD** | Background compaction threshold (0.0-1.0). | `0.8` | | **COMPACTION_THRESHOLD** | Background compaction threshold (0.0-1.0). | `0.8` |
| **BUFFER_THRESHOLD** | Buffer exhaustion threshold (0.0-1.0). | `0.95` | | **BUFFER_THRESHOLD** | Buffer exhaustion threshold (0.0-1.0). | `0.95` |
| **TIMEOUT** | Timeout for each stream chunk (seconds). | `300` |
| **CUSTOM_ENV_VARS** | Custom environment variables (JSON format). | - |
| **REASONING_EFFORT** | Reasoning effort level: low, medium, high. `xhigh` is supported for some models. | `medium` |
| **ENFORCE_FORMATTING** | Add formatting instructions to system prompt for better readability. | `True` |
| **ENABLE_MCP_SERVER** | Enable Direct MCP Client connection (Recommended). | `True` |
| **ENABLE_OPENWEBUI_TOOLS** | Enable OpenWebUI Tools (includes defined and server tools). | `True` |
### 3. Get GH_TOKEN #### User Valves (per-user overrides)
For security, it is recommended to use a **Fine-grained Personal Access Token**: These optional settings can be set per user (overrides global Valves):
| Parameter | Description | Default |
| :--- | :--- | :--- |
| **GH_TOKEN** | Personal GitHub Token (overrides global setting). | - |
| **REASONING_EFFORT** | Reasoning effort level (low/medium/high/xhigh). | - |
| **DEBUG** | Enable technical debug logs. | `False` |
| **SHOW_THINKING** | Show model reasoning/thinking process. | `True` |
| **ENABLE_OPENWEBUI_TOOLS** | Enable OpenWebUI Tools (overrides global). | `True` |
| **ENABLE_MCP_SERVER** | Enable MCP server loading (overrides global). | `True` |
| **ENFORCE_FORMATTING** | Enforce formatting guidelines (overrides global). | `True` |
### 3. Get Token
To use GitHub Copilot, you need a GitHub Personal Access Token (PAT) with appropriate permissions.
**Steps to generate your token:**
1. Visit [GitHub Token Settings](https://github.com/settings/tokens?type=beta). 1. Visit [GitHub Token Settings](https://github.com/settings/tokens?type=beta).
2. Click **Generate new token**. 2. Click **Generate new token (fine-grained)**.
3. **Repository access**: Select `All repositories` or `Public Repositories`. 3. **Repository access**: Select **Public Repositories** (simplest) or **All repositories**.
4. **Permissions**: 4. **Permissions**:
* Click **Account permissions**. * If you chose **All repositories**, you must click **Account permissions**.
* Find **Copilot Requests**, select **Read and write** (or Access). * Find **Copilot Requests**, and select **Access**.
5. Generate and copy the Token. 5. Generate and copy the Token.
## 📋 Dependencies ## 📋 Dependencies
@@ -68,16 +88,12 @@ This Pipe will automatically attempt to install the following dependencies:
* `github-copilot-sdk` (Python package) * `github-copilot-sdk` (Python package)
* `github-copilot-cli` (Binary file, installed via official script) * `github-copilot-cli` (Binary file, installed via official script)
## ⚠️ FAQ ## Troubleshooting ❓
* **Stuck on "Waiting..."**: * **Images and Multimodal Usage**:
* Check if `GH_TOKEN` is correct and has `Copilot Requests` permission.
* Try changing `MODEL_ID` to `gpt-4o` or `copilot-chat`.
* **Images not recognized**:
* Ensure `MODEL_ID` is a model that supports multimodal input. * Ensure `MODEL_ID` is a model that supports multimodal input.
* **CLI Installation Failed**: * **Thinking not shown**:
* Ensure the OpenWebUI container has internet access. * Ensure **streaming is enabled** and the selected model supports reasoning output.
* You can manually download the CLI and specify `CLI_PATH` in Valves.
## 📄 License ## 📄 License

View File

@@ -1,26 +1,24 @@
# GitHub Copilot SDK 官方管道 # GitHub Copilot SDK 官方管道
**作者:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **版本:** 0.1.0 | **项目:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **许可证:** MIT **作者:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **版本:** 0.3.0 | **项目:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **许可证:** MIT
这是一个用于 [OpenWebUI](https://github.com/open-webui/open-webui) 的高级 Pipe 函数,允许你直接在 OpenWebUI 中使用 GitHub Copilot 模型(如 `gpt-5`, `gpt-5-mini`, `claude-sonnet-4.5`)。它基于官方 [GitHub Copilot SDK for Python](https://github.com/github/copilot-sdk) 构建,提供了原生级的集成体验。 这是一个用于 [OpenWebUI](https://github.com/open-webui/open-webui) 的高级 Pipe 函数,允许你直接在 OpenWebUI 中使用 GitHub Copilot 模型(如 `gpt-5`, `gpt-5-mini`, `claude-sonnet-4.5`)。它基于官方 [GitHub Copilot SDK for Python](https://github.com/github/copilot-sdk) 构建,提供了原生级的集成体验。
## 🚀 最新特性 (v0.1.0) ## 🚀 最新特性 (v0.3.0) - “统一生态”的力量
* **♾️ 无限会话 (Infinite Sessions)**:支持长对话的自动上下文压缩,告别上下文超限错误! * **🔌 零配置工具桥接 (Unified Tool Bridge)**: 自动将您现有的 OpenWebUI Functions (工具) 转换为 Copilot 兼容工具。**Copilot 现在可以无缝调用您手头所有的 WebUI 工具!**
* **🧠 思考过程展示**:实时显示模型的推理/思考过程(需模型支持) * **🔗 动态 MCP 自动发现**: 直接联动 OpenWebUI **管理面板 -> 连接**。无需编写任何配置文件,即插即用,瞬间扩展 Copilot 能力边界
* **📂 工作目录控制**:支持设置受限工作目录,确保文件操作安全 * **⚡ 高性能异步引擎**: 异步 CLI 更新检查与高度优化的事件驱动流式处理,确保对话毫秒级响应
* **🔍 模型过滤**:支持通过关键词排除特定模型(如 `codex`, `haiku` * **🛡️ 卓越的兼容性**: 独创的动态 Pydantic 模型生成技术,确保复杂工具参数在 Copilot 端也能得到精准验证
* **💾 会话持久化**: 改进的会话恢复逻辑,直接关联 OpenWebUI 聊天 ID连接更稳定。
## ✨ 核心特性 ## ✨ 核心能力
* **🚀 官方 SDK 集成**:基于官方 SDK稳定可靠 * **🌉 强大的生态桥接**: 首个且唯一完美打通 **OpenWebUI Tools****GitHub Copilot SDK** 的插件
* **💬 多轮对话支持**:自动拼接历史上下文,Copilot 能理解你的前文 * **🚀 官方原生产体验**: 基于官方 Python SDK 构建,提供最稳定、最纯正的 Copilot 交互体验
* **🌊 流式输出 (Streaming)**:支持打字机效果,响应迅速 * **🌊 深度推理展示**: 完整支持模型思考过程 (Thinking Process) 的流式渲染
* **🖼️ 多模态支持**:支持上传图片,自动转换为附件发送给 Copilot(需模型支持) * **🖼️ 智能多模态**: 支持图像识别与附件上传,让 Copilot 拥有视觉能力
* **🛠️ 零配置安装**:自动检测并下载 GitHub Copilot CLI开箱即用。 * **🛠️ 极简部署流程**: 自动检测环境、自动下载 CLI、自动管理依赖全自动化开箱即用。
* **🔑 安全认证**:支持 Fine-grained Personal Access Tokens权限最小化 * **🔑 安全认证体系**: 完美支持 OAuth 授权与 PAT 模式,兼顾便捷与安全性
* **🐛 调试模式**:内置详细的日志输出,方便排查连接问题。
## 📦 安装与使用 ## 📦 安装与使用
@@ -38,28 +36,50 @@
| 参数 | 说明 | 默认值 | | 参数 | 说明 | 默认值 |
| :--- | :--- | :--- | | :--- | :--- | :--- |
| **GH_TOKEN** | **(必填)** 你的 GitHub Token。 | - | | **GH_TOKEN** | **(必填)** GitHub 访问令牌 (PAT 或 OAuth Token)。用于聊天。 | - |
| **MODEL_ID** | 使用的模型名称。推荐 `gpt-5-mini``gpt-5`。 | `gpt-5-mini` | | **DEBUG** | 是否开启调试日志(输出到浏览器控制台)。 | `False` |
| **CLI_PATH** | Copilot CLI 的路径。如果未找到会自动下载。 | `/usr/local/bin/copilot` | | **LOG_LEVEL** | Copilot CLI 日志级别: none, error, warning, info, debug, all。 | `error` |
| **DEBUG** | 是否开启调试日志(输出到对话框)。 | `True` | | **SHOW_THINKING** | 是否显示模型推理/思考过程(需开启流式 + 模型支持)。 | `True` |
| **SHOW_THINKING** | 是否显示模型推理/思考过程。 | `True` | | **COPILOT_CLI_VERSION** | 指定安装/强制使用的 Copilot CLI 版本。 | `0.0.405` |
| **EXCLUDE_KEYWORDS** | 排除包含这些关键词的模型 (逗号分隔)。 | - | | **EXCLUDE_KEYWORDS** | 排除包含这些关键词的模型逗号分隔。 | - |
| **WORKSPACE_DIR** | 文件操作的受限工作目录。 | - | | **WORKSPACE_DIR** | 文件操作的受限工作目录。 | - |
| **INFINITE_SESSION** | 启用无限会话 (自动上下文压缩)。 | `True` | | **INFINITE_SESSION** | 启用无限会话自动上下文压缩。 | `True` |
| **COMPACTION_THRESHOLD** | 后台压缩阈值 (0.0-1.0)。 | `0.8` | | **COMPACTION_THRESHOLD** | 后台压缩阈值 (0.0-1.0)。 | `0.8` |
| **BUFFER_THRESHOLD** | 缓冲耗尽阈值 (0.0-1.0)。 | `0.95` | | **BUFFER_THRESHOLD** | 缓冲耗尽阈值 (0.0-1.0)。 | `0.95` |
| **TIMEOUT** | 每个流式分块超时(秒)。 | `300` |
| **CUSTOM_ENV_VARS** | 自定义环境变量 (JSON 格式)。 | - |
| **REASONING_EFFORT** | 推理强度级别: low, medium, high. `xhigh` 仅部分模型支持。 | `medium` |
| **ENFORCE_FORMATTING** | 在系统提示词中添加格式化指导。 | `True` |
| **ENABLE_MCP_SERVER** | 启用直接 MCP 客户端连接 (建议)。 | `True` |
| **ENABLE_OPENWEBUI_TOOLS** | 启用 OpenWebUI 工具 (包括自定义和服务器工具)。 | `True` |
### 3. 获取 GH_TOKEN #### 用户 Valves按用户覆盖
为了安全起见,推荐使用 **Fine-grained Personal Access Token** 以下设置可按用户单独配置(覆盖全局 Valves
1. 访问 [GitHub Token Settings](https://github.com/settings/tokens?type=beta)。 | 参数 | 说明 | 默认值 |
2. 点击 **Generate new token** | :--- | :--- | :--- |
3. **Repository access**: 选择 `All repositories``Public Repositories` | **GH_TOKEN** | 个人 GitHub Token覆盖全局设置。 | - |
| **REASONING_EFFORT** | 推理强度级别low/medium/high/xhigh。 | - |
| **DEBUG** | 是否启用技术调试日志。 | `False` |
| **SHOW_THINKING** | 是否显示思考过程。 | `True` |
| **ENABLE_OPENWEBUI_TOOLS** | 启用 OpenWebUI 工具(覆盖全局设置)。 | `True` |
| **ENABLE_MCP_SERVER** | 启用动态 MCP 服务器加载(覆盖全局设置)。 | `True` |
| **ENFORCE_FORMATTING** | 强制启用格式化指导(覆盖全局设置)。 | `True` |
### 3. 获取 Token
要使用 GitHub Copilot您需要一个具有适当权限的 GitHub 个人访问令牌 (PAT)。
**获取步骤:**
1. 访问 [GitHub 令牌设置](https://github.com/settings/tokens?type=beta)。
2. 点击 **Generate new token (fine-grained)**
3. **Repository access**: 选择 **Public Repositories** (最简单) 或 **All repositories**
4. **Permissions**: 4. **Permissions**:
* 点击 **Account permissions** * 如果您选择了 **All repositories**,则必须点击 **Account permissions**
* 找到 **Copilot Requests**,选择 **Read and write** (或 Access) * 找到 **Copilot Requests**,选择 **Access**
5. 生成并复制 Token 5. 生成并复制令牌
## 📋 依赖说明 ## 📋 依赖说明
@@ -68,16 +88,12 @@
* `github-copilot-sdk` (Python 包) * `github-copilot-sdk` (Python 包)
* `github-copilot-cli` (二进制文件,通过官方脚本安装) * `github-copilot-cli` (二进制文件,通过官方脚本安装)
## ⚠️ 常见问题 ## 故障排除 (Troubleshooting) ❓
* **一直显示 "Waiting..."** * **图片及多模态使用说明**
* 检查 `GH_TOKEN` 是否正确且拥有 `Copilot Requests` 权限。
* 尝试将 `MODEL_ID` 改为 `gpt-4o``copilot-chat`
* **图片无法识别**
* 确保 `MODEL_ID` 是支持多模态的模型。 * 确保 `MODEL_ID` 是支持多模态的模型。
* **CLI 安装失败** * **看不到思考过程**
*保 OpenWebUI 容器有外网访问权限 *认已开启**流式输出**,且所选模型支持推理输出
* 你可以手动下载 CLI 并挂载到容器中,然后在 Valves 中指定 `CLI_PATH`
## 📄 许可证 ## 📄 许可证

View File

@@ -15,7 +15,7 @@ Pipes allow you to:
## Available Pipe Plugins ## Available Pipe Plugins
- [GitHub Copilot SDK](github-copilot-sdk.md) (v0.1.1) - Official GitHub Copilot SDK integration. Supports dynamic models, multi-turn conversation, streaming, multimodal input, and infinite sessions. - [GitHub Copilot SDK](github-copilot-sdk.md) (v0.3.0) - Official GitHub Copilot SDK integration. Features **zero-config OpenWebUI Tool Bridge** and **dynamic MCP discovery**. Supports streaming, multimodal, and infinite sessions.
--- ---

View File

@@ -15,7 +15,7 @@ Pipes 可以用于:
## 可用的 Pipe 插件 ## 可用的 Pipe 插件
- [GitHub Copilot SDK](github-copilot-sdk.zh.md) (v0.1.1) - GitHub Copilot SDK 官方集成。支持动态模型、多轮对话、流式输出、图片输入及无限会话。 - [GitHub Copilot SDK](github-copilot-sdk.zh.md) (v0.3.0) - GitHub Copilot SDK 官方集成。**零配置工具桥接**与**动态 MCP 发现**。支持流式输出、多模态及无限会话。
--- ---

View File

@@ -39,6 +39,10 @@ A comprehensive thinking lens that dives deep into any content - from context to
| **Clear Previous HTML (CLEAR_PREVIOUS_HTML)** | `True` | Whether to clear previous plugin results. | | **Clear Previous HTML (CLEAR_PREVIOUS_HTML)** | `True` | Whether to clear previous plugin results. |
| **Message Count (MESSAGE_COUNT)** | `1` | Number of recent messages to analyze. | | **Message Count (MESSAGE_COUNT)** | `1` | Number of recent messages to analyze. |
## ⭐ Support
If this plugin has been useful, a star on [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) is a big motivation for me. Thank you for the support.
## 🌗 Theme Support ## 🌗 Theme Support
The plugin automatically detects and adapts to OpenWebUI's theme settings: The plugin automatically detects and adapts to OpenWebUI's theme settings:
@@ -48,7 +52,7 @@ The plugin automatically detects and adapts to OpenWebUI's theme settings:
2. Parent document `html/body` class or `data-theme` attribute 2. Parent document `html/body` class or `data-theme` attribute
3. System preference via `prefers-color-scheme: dark` 3. System preference via `prefers-color-scheme: dark`
- **Requirements**: For best results, enable **iframe Sandbox Allow Same Origin** in OpenWebUI: - **Requirements**: For best results, enable **iframe Sandbox Allow Same Origin** in OpenWebUI:
- Go to **Settings****Interface****Artifacts** → Check **iframe Sandbox Allow Same Origin** - Go to **Settings****Interface****Artifacts** → Check **iframe Sandbox Allow Same Origin**
## 🎨 Visual Preview ## 🎨 Visual Preview
@@ -88,3 +92,7 @@ The plugin generates a structured thinking timeline:
- **Debug Logs**: Enable `SHOW_STATUS` in Valves to see progress updates. - **Debug Logs**: Enable `SHOW_STATUS` in Valves to see progress updates.
- **Error Messages**: If you see an error, please copy the full error message and report it. - **Error Messages**: If you see an error, please copy the full error message and report it.
- **Submit an Issue**: If you encounter any problems, please submit an issue on GitHub: [Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues) - **Submit an Issue**: If you encounter any problems, please submit an issue on GitHub: [Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
## Changelog
See the full history on GitHub: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)

View File

@@ -39,6 +39,10 @@
| **清除旧 HTML (CLEAR_PREVIOUS_HTML)** | `True` | 是否清除之前的插件结果。 | | **清除旧 HTML (CLEAR_PREVIOUS_HTML)** | `True` | 是否清除之前的插件结果。 |
| **消息数量 (MESSAGE_COUNT)** | `1` | 要分析的最近消息数量。 | | **消息数量 (MESSAGE_COUNT)** | `1` | 要分析的最近消息数量。 |
## ⭐ 支持
如果这个插件对你有帮助,欢迎到 [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) 点个 Star这将是我持续改进的动力感谢支持。
## 🌗 主题支持 ## 🌗 主题支持
插件会自动检测并适配 OpenWebUI 的主题设置: 插件会自动检测并适配 OpenWebUI 的主题设置:
@@ -88,3 +92,7 @@
- **调试日志**: 在 Valves 中启用 `SHOW_STATUS` 以查看进度更新。 - **调试日志**: 在 Valves 中启用 `SHOW_STATUS` 以查看进度更新。
- **错误信息**: 如果看到错误,请复制完整的错误信息并报告。 - **错误信息**: 如果看到错误,请复制完整的错误信息并报告。
- **提交 Issue**: 如果遇到任何问题,请在 GitHub 上提交 Issue[Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues) - **提交 Issue**: 如果遇到任何问题,请在 GitHub 上提交 Issue[Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
## 更新日志
完整历史请查看 GitHub 项目: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)

View File

@@ -44,6 +44,10 @@ Export conversation to Word (.docx) with **syntax highlighting**, **native math
| **Mermaid PNG Scale** | `3.0` | Resolution multiplier for Mermaid images | | **Mermaid PNG Scale** | `3.0` | Resolution multiplier for Mermaid images |
| **Math Enable** | `True` | Enable LaTeX math conversion | | **Math Enable** | `True` | Enable LaTeX math conversion |
## ⭐ Support
If this plugin has been useful, a star on [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) is a big motivation for me. Thank you for the support.
## 🛠️ Supported Markdown Syntax ## 🛠️ Supported Markdown Syntax
| Syntax | Word Result | | Syntax | Word Result |
@@ -71,25 +75,13 @@ Export conversation to Word (.docx) with **syntax highlighting**, **native math
- `latex2mathml` - LaTeX to MathML conversion - `latex2mathml` - LaTeX to MathML conversion
- `mathml2omml` - MathML to Office Math (OMML) conversion - `mathml2omml` - MathML to Office Math (OMML) conversion
## 📝 Changelog
### v0.4.3
- **S3 Object Storage**: Direct S3/MinIO access via boto3 for faster image retrieval.
- **6-Level Fallback**: Robust file retrieval: DB → S3 → Local → URL → API → Attributes.
- **Better Logging**: Improved error messages for debugging file access issues.
### v0.4.1
- **Chinese Parameter Names**: Localized configuration names for Chinese version.
### v0.4.0
- **Multi-language Support**: UI language switching (English/Chinese).
- **Font & Style Configuration**: Customizable fonts and table colors.
- **Mermaid Enhancements**: Hybrid SVG+PNG rendering, background color config.
- **Performance**: Real-time progress updates for large exports.
## Troubleshooting ❓ ## Troubleshooting ❓
- **Plugin not working?**: Check if the filter/action is enabled in the model settings. - **Plugin not working?**: Check if the filter/action is enabled in the model settings.
- **Debug Logs**: Check the browser console (F12) for detailed logs if available. - **Debug Logs**: Check the browser console (F12) for detailed logs if available.
- **Error Messages**: If you see an error, please copy the full error message and report it. - **Error Messages**: If you see an error, please copy the full error message and report it.
- **Submit an Issue**: If you encounter any problems, please submit an issue on GitHub: [Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues) - **Submit an Issue**: If you encounter any problems, please submit an issue on GitHub: [Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
## 📝 Changelog
See the full history on GitHub: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)

View File

@@ -44,6 +44,10 @@
| **Mermaid_PNG缩放比例** | `3.0` | Mermaid 图片分辨率倍数 | | **Mermaid_PNG缩放比例** | `3.0` | Mermaid 图片分辨率倍数 |
| **启用数学公式** | `True` | 启用 LaTeX 公式转换 | | **启用数学公式** | `True` | 启用 LaTeX 公式转换 |
## ⭐ 支持
如果这个插件对你有帮助,欢迎到 [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) 点个 Star这将是我持续改进的动力感谢支持。
## 🛠️ 支持的 Markdown 语法 ## 🛠️ 支持的 Markdown 语法
| 语法 | Word 效果 | | 语法 | Word 效果 |
@@ -71,25 +75,13 @@
- `latex2mathml` - LaTeX 转 MathML - `latex2mathml` - LaTeX 转 MathML
- `mathml2omml` - MathML 转 Office Math (OMML) - `mathml2omml` - MathML 转 Office Math (OMML)
## 📝 更新日志
### v0.4.3
- **S3 对象存储**: 通过 boto3 直连 S3/MinIO图片获取速度更快。
- **6 级回退机制**: 稳健的文件获取:数据库 → S3 → 本地 → URL → API → 属性。
- **日志优化**: 改进错误提示,便于调试文件访问问题。
### v0.4.1
- **中文参数名**: 配置项名称和描述全部汉化。
### v0.4.0
- **多语言支持**: 界面语言切换(中文/英文)。
- **字体与样式配置**: 支持自定义中英文字体、代码字体以及表格颜色。
- **Mermaid 增强**: 混合 SVG+PNG 渲染,支持背景色配置。
- **性能优化**: 导出大型文档时提供实时进度反馈。
## 故障排除 (Troubleshooting) ❓ ## 故障排除 (Troubleshooting) ❓
- **插件不工作?**: 请检查是否在模型设置中启用了该过滤器/动作。 - **插件不工作?**: 请检查是否在模型设置中启用了该过滤器/动作。
- **调试日志**: 请查看浏览器控制台 (F12) 获取详细日志(如果可用)。 - **调试日志**: 请查看浏览器控制台 (F12) 获取详细日志(如果可用)。
- **错误信息**: 如果看到错误,请复制完整的错误信息并报告。 - **错误信息**: 如果看到错误,请复制完整的错误信息并报告。
- **提交 Issue**: 如果遇到任何问题,请在 GitHub 上提交 Issue[Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues) - **提交 Issue**: 如果遇到任何问题,请在 GitHub 上提交 Issue[Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
## 📝 更新日志
完整历史请查看 GitHub 项目: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)

View File

@@ -1,4 +1,59 @@
# Export to Excel # 📊 Export to Excel
**Author:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **Version:** 0.3.6 | **Project:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **License:** MIT
Export chat history to an Excel (.xlsx) file directly from the chat interface.
## 🔥 What's New in v0.3.6
- **OpenWebUI-Style Theme**: Modern dark header (#1f2937) with light gray zebra striping for better readability.
- **Zebra Striping**: Alternating row colors (#ffffff / #f3f4f6) for improved visual scanning.
- **Smart Data Type Conversion**: Automatically converts columns to numeric or datetime types with fallback to string.
- **Full Cell Bold/Italic**: Supports full cell Markdown bold (`**text**`) and italic (`*text*`) formatting in Excel.
- **Partial Markdown Cleanup**: Removes partial Markdown formatting symbols for cleaner Excel output.
- **Export Scope**: Added `EXPORT_SCOPE` to choose between the last message or all messages.
- **Smart Sheet Naming**: Names sheets based on Markdown headers, AI titles, or message index.
- **Multiple Tables Support**: Improved handling of multiple tables across messages.
- **Smart Filename Generation**: Supports filenames based on chat title, AI summary, or Markdown headers.
- **Configuration Options**: Added `TITLE_SOURCE` to control filename strategy.
- **AI Title Generation**: Added `MODEL_ID` to use AI for filename generation with progress notifications.
## ✨ Core Features
- 🚀 **One-Click Export**: Adds an “Export to Excel” action button to the chat.
- 🧠 **Automatic Header Extraction**: Intelligently identifies table headers from chat content.
- 📊 **Multi-Table Support**: Handles multiple tables within a single chat session.
## 🚀 How to Use
1. **Install**: Search for “Export to Excel” in the Open WebUI Community and install.
2. **Trigger**: In any chat, click the “Export to Excel” action button.
3. **Download**: The .xlsx file will be automatically downloaded.
## ⚙️ Configuration (Valves)
| Parameter | Default | Description |
| :--- | :--- | :--- |
| `TITLE_SOURCE` | `chat_title` | Filename source: `chat_title`, `ai_generated`, or `markdown_title`. |
| `EXPORT_SCOPE` | `last_message` | Export scope: `last_message` or `all_messages`. |
| `MODEL_ID` | `""` | Model ID for AI title generation. Empty uses current chat model. |
| `SHOW_STATUS` | `True` | Show operation status updates. |
| `SHOW_DEBUG_LOG` | `False` | Print debug logs in the browser console (F12). |
## ⭐ Support
If this plugin has been useful, a star on [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) is a big motivation for me. Thank you for the support.
## Troubleshooting ❓
- **Plugin not working?**: Check if the filter/action is enabled in the model settings.
- **Debug Logs**: Enable `SHOW_STATUS` and check the browser console (F12) if needed.
- **Error Messages**: If you see an error, please copy the full error message and report it.
- **Submit an Issue**: If you encounter any problems, please submit an issue on GitHub: [Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
## Changelog
See the full history on GitHub: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)# Export to Excel
This plugin allows you to export your chat history to an Excel (.xlsx) file directly from the chat interface. This plugin allows you to export your chat history to an Excel (.xlsx) file directly from the chat interface.

View File

@@ -1,4 +1,59 @@
# 导出为 Excel # 📊 导出为 Excel
**作者:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **版本:** 0.3.6 | **项目:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **许可证:** MIT
将对话历史直接导出为 Excel (.xlsx) 文件。
## 🔥 最新更新 v0.3.6
- **OpenWebUI 风格主题**:现代深灰表头(#1f2937)与浅灰斑马纹,提升可读性。
- **斑马纹效果**:隔行变色(#ffffff / #f3f4f6),方便视觉扫描。
- **智能数据类型转换**:自动将列转换为数字或日期类型,无法转换时保持字符串。
- **全单元格粗体/斜体**:支持全单元格 Markdown 粗体与斜体格式。
- **部分 Markdown 清理**:移除部分 Markdown 格式符号,输出更整洁。
- **导出范围**:新增 `EXPORT_SCOPE`,可选择导出最后一条或所有消息。
- **智能 Sheet 命名**:按 Markdown 标题、AI 标题或消息索引命名。
- **多表格支持**:优化了多表格处理能力。
- **智能文件名生成**:支持对话标题 / AI 总结 / Markdown 标题命名。
- **配置选项**:新增 `TITLE_SOURCE` 控制文件名策略。
- **AI 标题生成**:新增 `MODEL_ID`,支持 AI 标题生成与进度提示。
## ✨ 核心特性
- 🚀 **一键导出**:在聊天界面添加“导出为 Excel”按钮。
- 🧠 **自动表头提取**:智能识别聊天内容中的表格标题。
- 📊 **多表支持**:支持单次对话中的多个表格。
## 🚀 使用方法
1. **安装**:在 Open WebUI 社区搜索“导出为 Excel”并安装。
2. **触发**:在任意对话中,点击“导出为 Excel”动作按钮。
3. **下载**.xlsx 文件将自动下载到你的设备。
## ⚙️ 配置参数 (Valves)
| 参数 | 默认值 | 描述 |
| :--- | :--- | :--- |
| `TITLE_SOURCE` | `chat_title` | 文件名来源:`chat_title``ai_generated``markdown_title`。 |
| `EXPORT_SCOPE` | `last_message` | 导出范围:`last_message``all_messages`。 |
| `MODEL_ID` | `""` | AI 标题生成的模型 ID。为空则使用当前对话模型。 |
| `SHOW_STATUS` | `True` | 是否显示操作状态更新。 |
| `SHOW_DEBUG_LOG` | `False` | 是否在浏览器控制台输出调试日志 (F12)。 |
## ⭐ 支持
如果这个插件对你有帮助,欢迎到 [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) 点个 Star这将是我持续改进的动力感谢支持。
## 故障排除 (Troubleshooting) ❓
- **插件不工作?**: 请检查是否在模型设置中启用了该过滤器/动作。
- **调试日志**: 如需排查,启用 `SHOW_STATUS` 并查看浏览器控制台 (F12)。
- **错误信息**: 如果看到错误,请复制完整的错误信息并报告。
- **提交 Issue**: 如果遇到任何问题,请在 GitHub 上提交 Issue[Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
## 更新日志
完整历史请查看 GitHub 项目: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)# 导出为 Excel
此插件允许你直接从聊天界面将对话历史导出为 Excel (.xlsx) 文件。 此插件允许你直接从聊天界面将对话历史导出为 Excel (.xlsx) 文件。

View File

@@ -4,13 +4,10 @@ Generate polished learning flashcards from any text—title, summary, key points
**Author:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **Version:** 0.2.4 | **Project:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **License:** MIT **Author:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **Version:** 0.2.4 | **Project:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **License:** MIT
## Preview 📸
![Flash Card Example](flash_card.png)
## What's New ## What's New
### v0.2.4 ### v0.2.4
- **Clean Output**: Removed debug messages from output. - **Clean Output**: Removed debug messages from output.
## Key Features 🔑 ## Key Features 🔑
@@ -39,9 +36,21 @@ Generate polished learning flashcards from any text—title, summary, key points
| CLEAR_PREVIOUS_HTML | Whether to clear previous card HTML (otherwise append/merge) | false | | CLEAR_PREVIOUS_HTML | Whether to clear previous card HTML (otherwise append/merge) | false |
| MESSAGE_COUNT | Use the latest N messages to build the card | 1 | | MESSAGE_COUNT | Use the latest N messages to build the card | 1 |
## ⭐ Support
If this plugin has been useful, a star on [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) is a big motivation for me. Thank you for the support.
## Troubleshooting ❓ ## Troubleshooting ❓
- **Plugin not working?**: Check if the filter/action is enabled in the model settings. - **Plugin not working?**: Check if the filter/action is enabled in the model settings.
- **Debug Logs**: Enable `SHOW_STATUS` in Valves to see progress updates. - **Debug Logs**: Enable `SHOW_STATUS` in Valves to see progress updates.
- **Error Messages**: If you see an error, please copy the full error message and report it. - **Error Messages**: If you see an error, please copy the full error message and report it.
- **Submit an Issue**: If you encounter any problems, please submit an issue on GitHub: [Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues) - **Submit an Issue**: If you encounter any problems, please submit an issue on GitHub: [Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
## Preview 📸
![Flash Card Example](flash_card.png)
## Changelog
See the full history on GitHub: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)

View File

@@ -4,22 +4,17 @@
**作者:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **版本:** 0.2.4 | **项目:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **许可证:** MIT **作者:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **版本:** 0.2.4 | **项目:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **许可证:** MIT
## 预览 📸 ## 🔥 最新更新 v0.2.4
![闪记卡示例](flash_card_cn.png) * **输出优化**: 移除输出中的调试信息。
## 更新日志
### v0.2.4
- **输出优化**: 移除输出中的调试信息。
## 核心特性 🔑 ## 核心特性 🔑
- **一键生成**:输入任意文本,直接产出结构化卡片。 * **一键生成**:输入任意文本,直接产出结构化卡片。
- **要点聚合**:自动提取 3-5 个记忆要点与 2-4 个标签。 * **要点聚合**:自动提取 3-5 个记忆要点与 2-4 个标签。
- **多语言支持**:可设定目标语言(默认中文)。 * **多语言支持**:可设定目标语言(默认中文)。
- **渐进合并**:多次调用会将新卡片合并到同一 HTML 容器中;如需重置可启用清空选项。 * **渐进合并**:多次调用会将新卡片合并到同一 HTML 容器中;如需重置可启用清空选项。
- **状态提示**:实时推送“生成中/完成/错误”等状态与通知。 * **状态提示**:实时推送“生成中/完成/错误”等状态与通知。
## 使用方法 🛠️ ## 使用方法 🛠️
@@ -39,9 +34,21 @@
| CLEAR_PREVIOUS_HTML | 是否清空旧的卡片 HTML否则合并追加 | false | | CLEAR_PREVIOUS_HTML | 是否清空旧的卡片 HTML否则合并追加 | false |
| MESSAGE_COUNT | 取最近 N 条消息生成卡片 | 1 | | MESSAGE_COUNT | 取最近 N 条消息生成卡片 | 1 |
## ⭐ 支持
如果这个插件对你有帮助,欢迎到 [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) 点个 Star这将是我持续改进的动力感谢支持。
## 故障排除 (Troubleshooting) ❓ ## 故障排除 (Troubleshooting) ❓
- **插件不工作?**: 请检查是否在模型设置中启用了该过滤器/动作。 * **插件不工作?**: 请检查是否在模型设置中启用了该过滤器/动作。
- **调试日志**: 在 Valves 中启用 `SHOW_STATUS` 以查看进度更新。 * **调试日志**: 在 Valves 中启用 `SHOW_STATUS` 以查看进度更新。
- **错误信息**: 如果看到错误,请复制完整的错误信息并报告。 * **错误信息**: 如果看到错误,请复制完整的错误信息并报告。
- **提交 Issue**: 如果遇到任何问题,请在 GitHub 上提交 Issue[Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues) * **提交 Issue**: 如果遇到任何问题,请在 GitHub 上提交 Issue[Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
## 预览 📸
![闪记卡示例](flash_card_cn.png)
## 更新日志
完整历史请查看 GitHub 项目: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)

View File

@@ -1,21 +1,14 @@
# 📊 Smart Infographic (AntV) # Smart Infographic
**Author:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **Version:** 1.4.9 | **Project:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **License:** MIT **Author:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **Version:** 1.5.0 | **Project:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **License:** MIT
An Open WebUI plugin powered by the AntV Infographic engine. It transforms long text into professional, beautiful infographics with a single click. An Open WebUI plugin powered by the AntV Infographic engine. It transforms long text into professional, beautiful infographics with a single click.
## 🔥 What's New in v1.4.9 ## 🔥 What's New in v1.5.0
- 🎨 **70+ Official Templates**: Integrated comprehensive AntV infographic template library. - 🌐 **Smart Language Detection**: Automatically detects the accurate UI language from your browser.
- 🖼 **Iconify & unDraw Support**: Richer visuals with official icons and illustrations. - 🗣 **Context-Aware Generation**: Generated infographics now strictly follow the language of your input content (e.g., input Japanese -> output Japanese infographic).
- 📏 **Visual Optimization**: Improved text wrapping, adaptive sizing, and layout refinement. - 🐛 **Bug Fixes**: Fixed issues with language synchronization between the UI and generated content.
-**PNG Upload**: Infographics now upload as PNG format for better Word export compatibility.
- 🔧 **Canvas Conversion**: Uses browser canvas for high-quality SVG to PNG conversion (2x scale).
### Previous: v1.4.0
-**Default Mode Change**: Default output mode is now `image` (static image) for better compatibility.
- 📱 **Responsive Sizing**: Images now auto-adapt to the chat container width.
## ✨ Key Features ## ✨ Key Features
@@ -46,6 +39,10 @@ You can adjust the following parameters in the plugin settings to optimize the g
| **Message Count (MESSAGE_COUNT)** | `1` | Number of recent messages to use for analysis. Increase this for more context. | | **Message Count (MESSAGE_COUNT)** | `1` | Number of recent messages to use for analysis. Increase this for more context. |
| **Output Mode (OUTPUT_MODE)** | `image` | `image` for static image embedding (default, better compatibility), `html` for interactive chart. | | **Output Mode (OUTPUT_MODE)** | `image` | `image` for static image embedding (default, better compatibility), `html` for interactive chart. |
## ⭐ Support
If this plugin has been useful, a star on [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) is a big motivation for me. Thank you for the support.
## 🛠️ Supported Template Types ## 🛠️ Supported Template Types
| Category | Template Name | Use Case | | Category | Template Name | Use Case |
@@ -63,6 +60,9 @@ You can adjust the following parameters in the plugin settings to optimize the g
- **Error Messages**: If you see an error, please copy the full error message and report it. - **Error Messages**: If you see an error, please copy the full error message and report it.
- **Submit an Issue**: If you encounter any problems, please submit an issue on GitHub: [Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues) - **Submit an Issue**: If you encounter any problems, please submit an issue on GitHub: [Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
## Changelog
See the full history on GitHub: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)
## 📝 Syntax Example (For Advanced Users) ## 📝 Syntax Example (For Advanced Users)

View File

@@ -1,21 +1,14 @@
# 📊 智能信息图 (AntV Infographic) # 智能信息图
**作者:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **版本:** 1.4.9 | **项目:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **许可证:** MIT **作者:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **版本:** 1.5.0 | **项目:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **许可证:** MIT
基于 AntV Infographic 引擎的 Open WebUI 插件,能够将长文本内容一键转换为专业、美观的信息图表。 基于 AntV Infographic 引擎的 Open WebUI 插件,能够将长文本内容一键转换为专业、美观的信息图表。
## 🔥 v1.4.9 更新日志 ## 🔥 最新更新 v1.5.0
- 🎨 **70+ 官方模板**:全面集成 AntV 官方信息图模板库 - 🌐 **智能语言检测**:自动从浏览器准确识别当前界面语言设置
- 🖼 **图标与插图支持**:支持 Iconify 图标库与 unDraw 插图库,视觉效果更丰富 - 🗣 **上下文感知生成**:生成的信息图内容现在严格跟随用户输入内容的语言(例如:输入日语 -> 生成日语信息图)
- 📏 **视觉优化**改进文本换行逻辑,优化自适应尺寸,提升卡片布局精细度 - 🐛 **问题修复**修复了界面语言与生成内容语言不同步的问题
-**PNG 上传**:信息图现在以 PNG 格式上传,与 Word 导出完美兼容。
- 🔧 **Canvas 转换**:使用浏览器 Canvas 高质量转换 SVG 为 PNG2倍缩放
### 此前: v1.4.0
-**默认模式变更**:默认输出模式调整为 `image`(静态图片)。
- 📱 **响应式尺寸**:图片模式下自动适应聊天容器宽度。
## ✨ 核心特性 ## ✨ 核心特性
@@ -46,6 +39,10 @@
| **上下文消息数 (MESSAGE_COUNT)** | `1` | 用于分析的最近消息条数。增加此值可让 AI 参考更多对话背景。 | | **上下文消息数 (MESSAGE_COUNT)** | `1` | 用于分析的最近消息条数。增加此值可让 AI 参考更多对话背景。 |
| **输出模式 (OUTPUT_MODE)** | `image` | `image` 为静态图片嵌入(默认,兼容性好),`html` 为交互式图表。 | | **输出模式 (OUTPUT_MODE)** | `image` | `image` 为静态图片嵌入(默认,兼容性好),`html` 为交互式图表。 |
## ⭐ 支持
如果这个插件对你有帮助,欢迎到 [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) 点个 Star这将是我持续改进的动力感谢支持。
## 🛠️ 支持的模板类型 ## 🛠️ 支持的模板类型
| 分类 | 模板名称 | 适用场景 | | 分类 | 模板名称 | 适用场景 |
@@ -63,6 +60,9 @@
- **错误信息**: 如果看到错误,请复制完整的错误信息并报告。 - **错误信息**: 如果看到错误,请复制完整的错误信息并报告。
- **提交 Issue**: 如果遇到任何问题,请在 GitHub 上提交 Issue[Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues) - **提交 Issue**: 如果遇到任何问题,请在 GitHub 上提交 Issue[Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
## 更新日志
完整历史请查看 GitHub 项目: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)
## 📝 语法示例 (高级用户) ## 📝 语法示例 (高级用户)

View File

@@ -1,10 +1,10 @@
""" """
title: 📊 Smart Infographic (AntV) title: Smart Infographic
author: Fu-Jie author: Fu-Jie
author_url: https://github.com/Fu-Jie/awesome-openwebui author_url: https://github.com/Fu-Jie/awesome-openwebui
funding_url: https://github.com/open-webui funding_url: https://github.com/open-webui
icon_url: data:image/svg+xml;base64,PHN2ZyB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciIHdpZHRoPSIyNCIgaGVpZ2h0PSIyNCIgdmlld0JveD0iMCAwIDI0IDI0IiBmaWxsPSJub25lIiBzdHJva2U9ImN1cnJlbnRDb2xvciIgc3Ryb2tlLXdpZHRoPSIyIiBzdHJva2UtbGluZWNhcD0icm91bmQiIHN0cm9rZS1saW5lam9pbj0icm91bmQiPgogIDxsaW5lIHgxPSIxMiIgeTE9IjIwIiB4Mj0iMTIiIHkyPSIxMCIgLz4KICA8bGluZSB4MT0iMTgiIHkxPSIyMCIgeDI9IjE4IiB5Mj0iNCIgLz4KICA8bGluZSB4MT0iNiIgeTE9IjIwIiB4Mj0iNiIgeTI9IjE2IiAvPgo8L3N2Zz4= icon_url: data:image/svg+xml;base64,PHN2ZyB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciIHdpZHRoPSIyNCIgaGVpZ2h0PSIyNCIgdmlld0JveD0iMCAwIDI0IDI0IiBmaWxsPSJub25lIiBzdHJva2U9ImN1cnJlbnRDb2xvciIgc3Ryb2tlLXdpZHRoPSIyIiBzdHJva2UtbGluZWNhcD0icm91bmQiIHN0cm9rZS1saW5lam9pbj0icm91bmQiPgogIDxsaW5lIHgxPSIxMiIgeTE9IjIwIiB4Mj0iMTIiIHkyPSIxMCIgLz4KICA8bGluZSB4MT0iMTgiIHkxPSIyMCIgeDI9IjE4IiB5Mj0iNCIgLz4KICA8bGluZSB4MT0iNiIgeTE9IjIwIiB4Mj0iNiIgeTI9IjE2IiAvPgo8L3N2Zz4=
version: 1.4.9 version: 1.5.0
openwebui_id: ad6f0c7f-c571-4dea-821d-8e71697274cf openwebui_id: ad6f0c7f-c571-4dea-821d-8e71697274cf
description: AI-powered infographic generator based on AntV Infographic. Supports professional templates, auto-icon matching, and SVG/PNG downloads. description: AI-powered infographic generator based on AntV Infographic. Supports professional templates, auto-icon matching, and SVG/PNG downloads.
""" """
@@ -32,6 +32,10 @@ logger = logging.getLogger(__name__)
SYSTEM_PROMPT_INFOGRAPHIC_ASSISTANT = """ SYSTEM_PROMPT_INFOGRAPHIC_ASSISTANT = """
You are a professional infographic design expert who can analyze user-provided text content and convert it into AntV Infographic syntax format. You are a professional infographic design expert who can analyze user-provided text content and convert it into AntV Infographic syntax format.
## Important Language Rule
- **GENERATE CONTENT IN INPUT LANGUAGE**: You must generate the text content of the infographic in the **exact same language** as the user's input content (the text you are analyzing).
- **Format Consistency**: Even if this system prompt is in English, if the user input is in Chinese, the infographic content must be in Chinese. If input is Japanese, output Japanese.
## Infographic Syntax Specification ## Infographic Syntax Specification
Infographic syntax is a Mermaid-like declarative syntax for describing infographic templates, data, and themes. Infographic syntax is a Mermaid-like declarative syntax for describing infographic templates, data, and themes.
@@ -958,7 +962,11 @@ class Action:
def __init__(self): def __init__(self):
self.valves = self.Valves() self.valves = self.Valves()
def _get_user_context(self, __user__: Optional[Dict[str, Any]]) -> Dict[str, str]: async def _get_user_context(
self,
__user__: Optional[Dict[str, Any]],
__event_call__: Optional[Callable[[Any], Awaitable[None]]] = None,
) -> Dict[str, str]:
"""Safely extracts user context information.""" """Safely extracts user context information."""
if isinstance(__user__, (list, tuple)): if isinstance(__user__, (list, tuple)):
user_data = __user__[0] if __user__ else {} user_data = __user__[0] if __user__ else {}
@@ -967,10 +975,32 @@ class Action:
else: else:
user_data = {} user_data = {}
user_id = user_data.get("id", "unknown_user")
user_name = user_data.get("name", "User")
user_language = user_data.get("language", "en-US")
if __event_call__:
try:
js_code = """
return (
localStorage.getItem('locale') ||
localStorage.getItem('language') ||
navigator.language ||
'en-US'
);
"""
frontend_lang = await __event_call__(
{"type": "execute", "data": {"code": js_code}}
)
if frontend_lang and isinstance(frontend_lang, str):
user_language = frontend_lang
except Exception as e:
logger.warning(f"Failed to retrieve frontend language: {e}")
return { return {
"user_id": user_data.get("id", "unknown_user"), "user_id": user_id,
"user_name": user_data.get("name", "User"), "user_name": user_name,
"user_language": user_data.get("language", "en-US"), "user_language": user_language,
} }
def _get_chat_context( def _get_chat_context(
@@ -1469,18 +1499,10 @@ class Action:
logger.info("Action: Infographic started (v1.4.0)") logger.info("Action: Infographic started (v1.4.0)")
# Get user information # Get user information
if isinstance(__user__, (list, tuple)): user_ctx = await self._get_user_context(__user__, __event_call__)
user_language = __user__[0].get("language", "en") if __user__ else "en" user_name = user_ctx["user_name"]
user_name = __user__[0].get("name", "User") if __user__[0] else "User" user_id = user_ctx["user_id"]
user_id = ( user_language = user_ctx["user_language"]
__user__[0]["id"]
if __user__ and "id" in __user__[0]
else "unknown_user"
)
elif isinstance(__user__, dict):
user_language = __user__.get("language", "en")
user_name = __user__.get("name", "User")
user_id = __user__.get("id", "unknown_user")
# Get current time # Get current time
now = datetime.now() now = datetime.now()

View File

@@ -1,10 +1,10 @@
""" """
title: 📊 智能信息图 (AntV Infographic) title: 智能信息图
author: Fu-Jie author: Fu-Jie
author_url: https://github.com/Fu-Jie/awesome-openwebui author_url: https://github.com/Fu-Jie/awesome-openwebui
funding_url: https://github.com/open-webui funding_url: https://github.com/open-webui
icon_url: data:image/svg+xml;base64,PHN2ZyB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciIHdpZHRoPSIyNCIgaGVpZ2h0PSIyNCIgdmlld0JveD0iMCAwIDI0IDI0IiBmaWxsPSJub25lIiBzdHJva2U9ImN1cnJlbnRDb2xvciIgc3Ryb2tlLXdpZHRoPSIyIiBzdHJva2UtbGluZWNhcD0icm91bmQiIHN0cm9rZS1saW5lam9pbj0icm91bmQiPgogIDxsaW5lIHgxPSIxMiIgeTE9IjIwIiB4Mj0iMTIiIHkyPSIxMCIgLz4KICA8bGluZSB4MT0iMTgiIHkxPSIyMCIgeDI9IjE4IiB5Mj0iNCIgLz4KICA8bGluZSB4MT0iNiIgeTE9IjIwIiB4Mj0iNiIgeTI9IjE2IiAvPgo8L3N2Zz4= icon_url: data:image/svg+xml;base64,PHN2ZyB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciIHdpZHRoPSIyNCIgaGVpZ2h0PSIyNCIgdmlld0JveD0iMCAwIDI0IDI0IiBmaWxsPSJub25lIiBzdHJva2U9ImN1cnJlbnRDb2xvciIgc3Ryb2tlLXdpZHRoPSIyIiBzdHJva2UtbGluZWNhcD0icm91bmQiIHN0cm9rZS1saW5lam9pbj0icm91bmQiPgogIDxsaW5lIHgxPSIxMiIgeTE9IjIwIiB4Mj0iMTIiIHkyPSIxMCIgLz4KICA8bGluZSB4MT0iMTgiIHkxPSIyMCIgeDI9IjE4IiB5Mj0iNCIgLz4KICA8bGluZSB4MT0iNiIgeTE9IjIwIiB4Mj0iNiIgeTI9IjE2IiAvPgo8L3N2Zz4=
version: 1.4.9 version: 1.5.0
openwebui_id: e04a48ff-23ee-4a41-8ea7-66c19524e7c8 openwebui_id: e04a48ff-23ee-4a41-8ea7-66c19524e7c8
description: 基于 AntV Infographic 的智能信息图生成插件。支持多种专业模板,自动图标匹配,并提供 SVG/PNG 下载功能。 description: 基于 AntV Infographic 的智能信息图生成插件。支持多种专业模板,自动图标匹配,并提供 SVG/PNG 下载功能。
""" """
@@ -32,6 +32,10 @@ logger = logging.getLogger(__name__)
SYSTEM_PROMPT_INFOGRAPHIC_ASSISTANT = """ SYSTEM_PROMPT_INFOGRAPHIC_ASSISTANT = """
You are a professional infographic design expert who can analyze user-provided text content and convert it into AntV Infographic syntax format. You are a professional infographic design expert who can analyze user-provided text content and convert it into AntV Infographic syntax format.
## Important Language Rule (语言规则)
- **Priority Input Language (优先使用输入语言)**: You must generate the text content of the infographic in the **exact same language** as the user's input content.
- **Example**: If the user provides a summary in Chinese, the labels and descriptions in the infographic must be in Chinese.
## Infographic Syntax Specification ## Infographic Syntax Specification
Infographic syntax is a Mermaid-like declarative syntax for describing infographic templates, data, and themes. Infographic syntax is a Mermaid-like declarative syntax for describing infographic templates, data, and themes.
@@ -974,7 +978,11 @@ class Action:
"Sunday": "星期日", "Sunday": "星期日",
} }
def _get_user_context(self, __user__: Optional[Dict[str, Any]]) -> Dict[str, str]: async def _get_user_context(
self,
__user__: Optional[Dict[str, Any]],
__event_call__: Optional[Callable[[Any], Awaitable[None]]] = None,
) -> Dict[str, str]:
"""安全提取用户上下文信息。""" """安全提取用户上下文信息。"""
if isinstance(__user__, (list, tuple)): if isinstance(__user__, (list, tuple)):
user_data = __user__[0] if __user__ else {} user_data = __user__[0] if __user__ else {}
@@ -983,10 +991,32 @@ class Action:
else: else:
user_data = {} user_data = {}
user_id = user_data.get("id", "unknown_user")
user_name = user_data.get("name", "用户")
user_language = user_data.get("language", "zh-CN")
if __event_call__:
try:
js_code = """
return (
localStorage.getItem('locale') ||
localStorage.getItem('language') ||
navigator.language ||
'zh-CN'
);
"""
frontend_lang = await __event_call__(
{"type": "execute", "data": {"code": js_code}}
)
if frontend_lang and isinstance(frontend_lang, str):
user_language = frontend_lang
except Exception as e:
pass
return { return {
"user_id": user_data.get("id", "unknown_user"), "user_id": user_id,
"user_name": user_data.get("name", "用户"), "user_name": user_name,
"user_language": user_data.get("language", "zh-CN"), "user_language": user_language,
} }
def _get_chat_context( def _get_chat_context(
@@ -1509,20 +1539,10 @@ class Action:
logger.info("Action: 信息图启动 (v1.4.0)") logger.info("Action: 信息图启动 (v1.4.0)")
# 获取用户信息 # 获取用户信息
if isinstance(__user__, (list, tuple)): user_ctx = await self._get_user_context(__user__, __event_call__)
user_language = ( user_name = user_ctx["user_name"]
__user__[0].get("language", "zh-CN") if __user__ else "zh-CN" user_id = user_ctx["user_id"]
) user_language = user_ctx["user_language"]
user_name = __user__[0].get("name", "用户") if __user__[0] else "用户"
user_id = (
__user__[0]["id"]
if __user__ and "id" in __user__[0]
else "unknown_user"
)
elif isinstance(__user__, dict):
user_language = __user__.get("language", "zh-CN")
user_name = __user__.get("name", "用户")
user_id = __user__.get("id", "unknown_user")
# 获取当前时间 # 获取当前时间
now = datetime.now() now = datetime.now()

View File

@@ -2,17 +2,14 @@
Smart Mind Map is a powerful OpenWebUI action plugin that intelligently analyzes long-form text content and automatically generates interactive mind maps, helping users structure and visualize knowledge. Smart Mind Map is a powerful OpenWebUI action plugin that intelligently analyzes long-form text content and automatically generates interactive mind maps, helping users structure and visualize knowledge.
**Author:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **Version:** 0.9.1 | **Project:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **License:** MIT **Author:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **Version:** 0.9.2 | **Project:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **License:** MIT
## What's New in v0.9.1 ## What's New in v0.9.2
**New Feature: Image Output Mode** **Language Rule Alignment**
- **Static Image Support**: Added `OUTPUT_MODE` configuration parameter. - **Input Language First**: Mind map output now strictly matches the input text language.
- `html` (default): Interactive HTML mind map. - **Consistent Behavior**: Matches the infographic language rule for predictable multilingual output.
- `image`: Static SVG image embedded directly in Markdown (**No HTML code output**, cleaner chat history).
- **Efficient Storage**: Image mode uploads SVG to `/api/v1/files`, avoiding huge base64 strings in chat history.
- **Smart Features**: Auto-responsive width and automatic theme detection (light/dark) for generated images.
## Key Features 🔑 ## Key Features 🔑
@@ -41,6 +38,10 @@ Smart Mind Map is a powerful OpenWebUI action plugin that intelligently analyzes
| `MESSAGE_COUNT` | `1` | Number of recent messages to use for generation (1-5). | | `MESSAGE_COUNT` | `1` | Number of recent messages to use for generation (1-5). |
| `OUTPUT_MODE` | `html` | Output mode: `html` (interactive) or `image` (static). | | `OUTPUT_MODE` | `html` | Output mode: `html` (interactive) or `image` (static). |
## ⭐ Support
If this plugin has been useful, a star on [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) is a big motivation for me. Thank you for the support.
## Troubleshooting ❓ ## Troubleshooting ❓
- **Plugin not working?**: Check if the action is enabled in the chat settings. - **Plugin not working?**: Check if the action is enabled in the chat settings.
@@ -62,3 +63,7 @@ Smart Mind Map is a powerful OpenWebUI action plugin that intelligently analyzes
1. **Text Preparation**: Provide text with clear structure and distinct hierarchies. 1. **Text Preparation**: Provide text with clear structure and distinct hierarchies.
2. **Model Selection**: Use fast models like `gemini-2.5-flash` for daily use. 2. **Model Selection**: Use fast models like `gemini-2.5-flash` for daily use.
3. **Export Quality**: Use PNG for presentations and SVG for further editing. 3. **Export Quality**: Use PNG for presentations and SVG for further editing.
## Changelog
See the full history on GitHub: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)

View File

@@ -2,17 +2,14 @@
思维导图是一个强大的 OpenWebUI 动作插件,能够智能分析长篇文本内容,自动生成交互式思维导图,帮助用户结构化和可视化知识。 思维导图是一个强大的 OpenWebUI 动作插件,能够智能分析长篇文本内容,自动生成交互式思维导图,帮助用户结构化和可视化知识。
**作者:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **版本:** 0.9.1 | **项目:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **许可证:** MIT **作者:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **版本:** 0.9.2 | **项目:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **许可证:** MIT
## v0.9.1 更新亮点 ## v0.9.2 更新亮点
**新功能:图片输出模式** **语言规则对齐**
- **静态图片支持**:新增 `OUTPUT_MODE` 配置参数 - **输入语言优先**:导图输出严格与输入文本语言一致
- `html`(默认):交互式 HTML 思维导图 - **一致性提升**:与信息图语言规则保持一致,多语言输出更可预期
- `image`:静态 SVG 图片直接嵌入 Markdown**不输出 HTML 代码**,聊天记录更简洁)。
- **高效存储**:图片模式将 SVG 上传至 `/api/v1/files`,避免聊天记录中出现超长 Base64 字符串。
- **智能特性**:生成的图片支持自动响应式宽度和自动主题检测(亮色/暗色)。
## 核心特性 🔑 ## 核心特性 🔑
@@ -41,6 +38,10 @@
| `MESSAGE_COUNT` | `1` | 用于生成思维导图的最近消息数量1-5。 | | `MESSAGE_COUNT` | `1` | 用于生成思维导图的最近消息数量1-5。 |
| `OUTPUT_MODE` | `html` | 输出模式:`html`(交互式)或 `image`(静态图片)。 | | `OUTPUT_MODE` | `html` | 输出模式:`html`(交互式)或 `image`(静态图片)。 |
## ⭐ 支持
如果这个插件对你有帮助,欢迎到 [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) 点个 Star这将是我持续改进的动力感谢支持。
## 故障排除 (Troubleshooting) ❓ ## 故障排除 (Troubleshooting) ❓
- **插件无法启动**:检查 OpenWebUI 日志,确认插件已正确上传并启用。 - **插件无法启动**:检查 OpenWebUI 日志,确认插件已正确上传并启用。
@@ -62,3 +63,7 @@
1. **文本准备**:提供结构清晰、层次分明的文本内容。 1. **文本准备**:提供结构清晰、层次分明的文本内容。
2. **模型选择**:日常使用推荐 `gemini-2.5-flash` 等快速模型。 2. **模型选择**:日常使用推荐 `gemini-2.5-flash` 等快速模型。
3. **导出质量**PNG 适合演示分享SVG 适合进一步矢量编辑。 3. **导出质量**PNG 适合演示分享SVG 适合进一步矢量编辑。
## 更新日志
完整历史请查看 GitHub 项目: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)

View File

@@ -4,7 +4,7 @@ author: Fu-Jie
author_url: https://github.com/Fu-Jie/awesome-openwebui author_url: https://github.com/Fu-Jie/awesome-openwebui
funding_url: https://github.com/open-webui funding_url: https://github.com/open-webui
funding_url: https://github.com/Fu-Jie/awesome-openwebui funding_url: https://github.com/Fu-Jie/awesome-openwebui
version: 0.9.1 version: 0.9.2
openwebui_id: 3094c59a-b4dd-4e0c-9449-15e2dd547dc4 openwebui_id: 3094c59a-b4dd-4e0c-9449-15e2dd547dc4
icon_url: data:image/svg+xml;base64,PHN2ZyB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciIHdpZHRoPSIyNCIgaGVpZ2h0PSIyNCIgdmlld0JveD0iMCAwIDI0IDI0IiBmaWxsPSJub25lIiBzdHJva2U9ImN1cnJlbnRDb2xvciIgc3Ryb2tlLXdpZHRoPSIyIiBzdHJva2UtbGluZWNhcD0icm91bmQiIHN0cm9rZS1saW5lam9pbj0icm91bmQiPjxyZWN0IHg9IjE2IiB5PSIxNiIgd2lkdGg9IjYiIGhlaWdodD0iNiIgcng9IjEiLz48cmVjdCB4PSIyIiB5PSIxNiIgd2lkdGg9IjYiIGhlaWdodD0iNiIgcng9IjEiLz48cmVjdCB4PSI5IiB5PSIyIiB3aWR0aD0iNiIgaGVpZ2h0PSI2IiByeD0iMSIvPjxwYXRoIGQ9Ik01IDE2di0zYTEgMSAwIDAgMSAxLTFoMTJhMSAxIDAgMCAxIDEgMXYzIi8+PHBhdGggZD0iTTEyIDEyVjgiLz48L3N2Zz4= icon_url: data:image/svg+xml;base64,PHN2ZyB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciIHdpZHRoPSIyNCIgaGVpZ2h0PSIyNCIgdmlld0JveD0iMCAwIDI0IDI0IiBmaWxsPSJub25lIiBzdHJva2U9ImN1cnJlbnRDb2xvciIgc3Ryb2tlLXdpZHRoPSIyIiBzdHJva2UtbGluZWNhcD0icm91bmQiIHN0cm9rZS1saW5lam9pbj0icm91bmQiPjxyZWN0IHg9IjE2IiB5PSIxNiIgd2lkdGg9IjYiIGhlaWdodD0iNiIgcng9IjEiLz48cmVjdCB4PSIyIiB5PSIxNiIgd2lkdGg9IjYiIGhlaWdodD0iNiIgcng9IjEiLz48cmVjdCB4PSI5IiB5PSIyIiB3aWR0aD0iNiIgaGVpZ2h0PSI2IiByeD0iMSIvPjxwYXRoIGQ9Ik01IDE2di0zYTEgMSAwIDAgMSAxLTFoMTJhMSAxIDAgMCAxIDEgMXYzIi8+PHBhdGggZD0iTTEyIDEyVjgiLz48L3N2Zz4=
description: Intelligently analyzes text content and generates interactive mind maps to help users structure and visualize knowledge. description: Intelligently analyzes text content and generates interactive mind maps to help users structure and visualize knowledge.
@@ -33,7 +33,8 @@ SYSTEM_PROMPT_MINDMAP_ASSISTANT = """
You are a professional mind map generation assistant, capable of efficiently analyzing long-form text provided by users and structuring its core themes, key concepts, branches, and sub-branches into standard Markdown list syntax for rendering by Markmap.js. You are a professional mind map generation assistant, capable of efficiently analyzing long-form text provided by users and structuring its core themes, key concepts, branches, and sub-branches into standard Markdown list syntax for rendering by Markmap.js.
Please strictly follow these guidelines: Please strictly follow these guidelines:
- **Language**: All output must be in the language specified by the user. - **Language**: All output must be in the exact same language as the input text (the text you are analyzing).
- **Format Consistency**: Even if this system prompt is in English, if the user input is in Chinese, the mind map content must be in Chinese. If input is Japanese, output Japanese.
- **Format**: Your output must strictly be in Markdown list format, wrapped with ```markdown and ```. - **Format**: Your output must strictly be in Markdown list format, wrapped with ```markdown and ```.
- Use `#` to define the central theme (root node). - Use `#` to define the central theme (root node).
- Use `-` with two-space indentation to represent branches and sub-branches. - Use `-` with two-space indentation to represent branches and sub-branches.
@@ -811,7 +812,11 @@ class Action:
"Sunday": "Sunday", "Sunday": "Sunday",
} }
def _get_user_context(self, __user__: Optional[Dict[str, Any]]) -> Dict[str, str]: async def _get_user_context(
self,
__user__: Optional[Dict[str, Any]],
__event_call__: Optional[Callable[[Any], Awaitable[None]]] = None,
) -> Dict[str, str]:
"""Extract basic user context with safe fallbacks.""" """Extract basic user context with safe fallbacks."""
if isinstance(__user__, (list, tuple)): if isinstance(__user__, (list, tuple)):
user_data = __user__[0] if __user__ else {} user_data = __user__[0] if __user__ else {}
@@ -820,10 +825,32 @@ class Action:
else: else:
user_data = {} user_data = {}
user_id = user_data.get("id", "unknown_user")
user_name = user_data.get("name", "User")
user_language = user_data.get("language", "en-US")
if __event_call__:
try:
js_code = """
return (
localStorage.getItem('locale') ||
localStorage.getItem('language') ||
navigator.language ||
'en-US'
);
"""
frontend_lang = await __event_call__(
{"type": "execute", "data": {"code": js_code}}
)
if frontend_lang and isinstance(frontend_lang, str):
user_language = frontend_lang
except Exception as e:
logger.warning(f"Failed to retrieve frontend language: {e}")
return { return {
"user_id": user_data.get("id", "unknown_user"), "user_id": user_id,
"user_name": user_data.get("name", "User"), "user_name": user_name,
"user_language": user_data.get("language", "en-US"), "user_language": user_language,
} }
def _get_chat_context( def _get_chat_context(
@@ -1369,8 +1396,8 @@ class Action:
__metadata__: Optional[dict] = None, __metadata__: Optional[dict] = None,
__request__: Optional[Request] = None, __request__: Optional[Request] = None,
) -> Optional[dict]: ) -> Optional[dict]:
logger.info("Action: Smart Mind Map (v0.9.1) started") logger.info("Action: Smart Mind Map (v0.9.2) started")
user_ctx = self._get_user_context(__user__) user_ctx = await self._get_user_context(__user__, __event_call__)
user_language = user_ctx["user_language"] user_language = user_ctx["user_language"]
user_name = user_ctx["user_name"] user_name = user_ctx["user_name"]
user_id = user_ctx["user_id"] user_id = user_ctx["user_id"]

View File

@@ -3,7 +3,7 @@ title: 思维导图
author: Fu-Jie author: Fu-Jie
author_url: https://github.com/Fu-Jie/awesome-openwebui author_url: https://github.com/Fu-Jie/awesome-openwebui
funding_url: https://github.com/open-webui funding_url: https://github.com/open-webui
version: 0.9.1 version: 0.9.2
openwebui_id: 8d4b097b-219b-4dd2-b509-05fbe6388335 openwebui_id: 8d4b097b-219b-4dd2-b509-05fbe6388335
icon_url: data:image/svg+xml;base64,PHN2ZyB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciIHdpZHRoPSIyNCIgaGVpZ2h0PSIyNCIgdmlld0JveD0iMCAwIDI0IDI0IiBmaWxsPSJub25lIiBzdHJva2U9ImN1cnJlbnRDb2xvciIgc3Ryb2tlLXdpZHRoPSIyIiBzdHJva2UtbGluZWNhcD0icm91bmQiIHN0cm9rZS1saW5lam9pbj0icm91bmQiPjxyZWN0IHg9IjE2IiB5PSIxNiIgd2lkdGg9IjYiIGhlaWdodD0iNiIgcng9IjEiLz48cmVjdCB4PSIyIiB5PSIxNiIgd2lkdGg9IjYiIGhlaWdodD0iNiIgcng9IjEiLz48cmVjdCB4PSI5IiB5PSIyIiB3aWR0aD0iNiIgaGVpZ2h0PSI2IiByeD0iMSIvPjxwYXRoIGQ9Ik01IDE2di0zYTEgMSAwIDAgMSAxLTFoMTJhMSAxIDAgMCAxIDEgMXYzIi8+PHBhdGggZD0iTTEyIDEyVjgiLz48L3N2Zz4= icon_url: data:image/svg+xml;base64,PHN2ZyB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciIHdpZHRoPSIyNCIgaGVpZ2h0PSIyNCIgdmlld0JveD0iMCAwIDI0IDI0IiBmaWxsPSJub25lIiBzdHJva2U9ImN1cnJlbnRDb2xvciIgc3Ryb2tlLXdpZHRoPSIyIiBzdHJva2UtbGluZWNhcD0icm91bmQiIHN0cm9rZS1saW5lam9pbj0icm91bmQiPjxyZWN0IHg9IjE2IiB5PSIxNiIgd2lkdGg9IjYiIGhlaWdodD0iNiIgcng9IjEiLz48cmVjdCB4PSIyIiB5PSIxNiIgd2lkdGg9IjYiIGhlaWdodD0iNiIgcng9IjEiLz48cmVjdCB4PSI5IiB5PSIyIiB3aWR0aD0iNiIgaGVpZ2h0PSI2IiByeD0iMSIvPjxwYXRoIGQ9Ik01IDE2di0zYTEgMSAwIDAgMSAxLTFoMTJhMSAxIDAgMCAxIDEgMXYzIi8+PHBhdGggZD0iTTEyIDEyVjgiLz48L3N2Zz4=
description: 智能分析文本内容,生成交互式思维导图,帮助用户结构化和可视化知识。 description: 智能分析文本内容,生成交互式思维导图,帮助用户结构化和可视化知识。
@@ -32,7 +32,8 @@ SYSTEM_PROMPT_MINDMAP_ASSISTANT = """
你是一个专业的思维导图生成助手,能够高效地分析用户提供的长篇文本,并将其核心主题、关键概念、分支和子分支结构化为标准的Markdown列表语法,以便Markmap.js进行渲染。 你是一个专业的思维导图生成助手,能够高效地分析用户提供的长篇文本,并将其核心主题、关键概念、分支和子分支结构化为标准的Markdown列表语法,以便Markmap.js进行渲染。
请严格遵循以下指导原则: 请严格遵循以下指导原则:
- **语言**: 所有输出必须使用用户指定的语言。 - **语言**: 所有输出必须与输入文本(正在分析的文本)保持完全一致的语言。
- **格式一致性**: 即使系统提示词是中文,只要用户输入是英文,导图内容必须是英文;若输入为日文,则输出日文。
- **格式**: 你的输出必须严格为Markdown列表格式,并用```markdown 和 ``` 包裹。 - **格式**: 你的输出必须严格为Markdown列表格式,并用```markdown 和 ``` 包裹。
- 使用 `#` 定义中心主题(根节点)。 - 使用 `#` 定义中心主题(根节点)。
- 使用 `-` 和两个空格的缩进表示分支和子分支。 - 使用 `-` 和两个空格的缩进表示分支和子分支。
@@ -809,7 +810,11 @@ class Action:
"Sunday": "星期日", "Sunday": "星期日",
} }
def _get_user_context(self, __user__: Optional[Dict[str, Any]]) -> Dict[str, str]: async def _get_user_context(
self,
__user__: Optional[Dict[str, Any]],
__event_call__: Optional[Callable[[Any], Awaitable[None]]] = None,
) -> Dict[str, str]:
"""Extract basic user context with safe fallbacks.""" """Extract basic user context with safe fallbacks."""
if isinstance(__user__, (list, tuple)): if isinstance(__user__, (list, tuple)):
user_data = __user__[0] if __user__ else {} user_data = __user__[0] if __user__ else {}
@@ -818,10 +823,32 @@ class Action:
else: else:
user_data = {} user_data = {}
user_id = user_data.get("id", "unknown_user")
user_name = user_data.get("name", "User")
user_language = user_data.get("language", "en-US")
if __event_call__:
try:
js_code = """
return (
localStorage.getItem('locale') ||
localStorage.getItem('language') ||
navigator.language ||
'en-US'
);
"""
frontend_lang = await __event_call__(
{"type": "execute", "data": {"code": js_code}}
)
if frontend_lang and isinstance(frontend_lang, str):
user_language = frontend_lang
except Exception as e:
logger.warning(f"Failed to retrieve frontend language: {e}")
return { return {
"user_id": user_data.get("id", "unknown_user"), "user_id": user_id,
"user_name": user_data.get("name", "用户"), "user_name": user_name,
"user_language": user_data.get("language", "zh-CN"), "user_language": user_language,
} }
def _get_chat_context( def _get_chat_context(
@@ -1348,8 +1375,8 @@ class Action:
__metadata__: Optional[dict] = None, __metadata__: Optional[dict] = None,
__request__: Optional[Request] = None, __request__: Optional[Request] = None,
) -> Optional[dict]: ) -> Optional[dict]:
logger.info("Action: 思维导图 (v0.9.1) started") logger.info("Action: 思维导图 (v0.9.2) started")
user_ctx = self._get_user_context(__user__) user_ctx = await self._get_user_context(__user__, __event_call__)
user_language = user_ctx["user_language"] user_language = user_ctx["user_language"]
user_name = user_ctx["user_name"] user_name = user_ctx["user_name"]
user_id = user_ctx["user_id"] user_id = user_ctx["user_id"]

View File

@@ -0,0 +1,359 @@
#!/usr/bin/env python3
"""
======================================================================
Staged README Synchronizer to OpenWebUI Community
暂存 README 文件同步到 OpenWebUI 社区工具
======================================================================
PURPOSE / 用途:
--------------
This script synchronizes staged README.md/README_CN.md files to their
corresponding OpenWebUI Community posts automatically. It's designed for
batch updating documentation content without modifying plugin versions
or media attachments.
本脚本自动将暂存的 README.md/README_CN.md 文件同步到对应的 OpenWebUI
社区帖子。专为批量更新文档内容设计,不修改插件版本或媒体附件。
USAGE / 使用方法:
----------------
1. Set up environment:
配置环境:
Create a .env file in the repository root with:
在仓库根目录创建 .env 文件,包含:
OPENWEBUI_API_KEY=your_api_key_here
2. Stage README files to sync:
暂存需要同步的 README 文件:
git add plugins/actions/my_plugin/README.md
git add plugins/actions/my_plugin/README_CN.md
3. Run the script:
运行脚本:
python plugins/debug/common_tools/update_readmes_to_market.py
WORKFLOW / 工作流程:
-------------------
1. Load OPENWEBUI_API_KEY from .env file
从 .env 文件加载 OPENWEBUI_API_KEY
2. Get list of staged README.md/README_CN.md files via git
通过 git 获取暂存的 README.md/README_CN.md 文件列表
3. For each staged README:
对于每个暂存的 README
a. Locate the corresponding plugin .py file
定位对应的插件 .py 文件
b. Extract openwebui_id/post_id from plugin frontmatter
从插件前置信息中提取 openwebui_id/post_id
c. Fetch existing post data from OpenWebUI Community API
从 OpenWebUI 社区 API 获取现有帖子数据
d. Update post content with new README content
用新的 README 内容更新帖子内容
e. Push changes via API (preserves version & media)
通过 API 推送更改(保留版本和媒体)
REQUIREMENTS / 依赖要求:
-----------------------
- python-dotenv: For loading .env configuration
用于加载 .env 配置文件
- Git repository: Must be run from a git-tracked workspace
必须在 git 跟踪的工作区中运行
KEY FEATURES / 关键特性:
-----------------------
✅ Only updates content field (不仅更新内容字段)
✅ Skips files without openwebui_id (跳过没有 openwebui_id 的文件)
✅ Automatically matches CN/EN plugin files (自动匹配中英文插件文件)
✅ Supports staged plugin source code updates (支持暂存插件源码更新)
✅ Safe: Won't modify version or media fields (安全:不会修改版本或媒体字段)
NOTES / 注意事项:
---------------
- This is a DEBUG/DEVELOPMENT tool, not for production workflows
这是一个调试/开发工具,不用于生产工作流
- Always verify changes in OpenWebUI Community after sync
同步后务必在 OpenWebUI 社区中验证更改
- Requires valid API key with update permissions
需要具有更新权限的有效 API 密钥
AUTHOR / 作者:
-------------
Fu-Jie
GitHub: https://github.com/Fu-Jie/awesome-openwebui
======================================================================
"""
from __future__ import annotations
import importlib.util
import os
import re
import sys
import subprocess
from pathlib import Path
from typing import Dict, Optional, List
def _load_dotenv(repo_root: Path) -> None:
try:
from dotenv import load_dotenv # type: ignore
except Exception as exc: # pragma: no cover
print("Missing dependency: python-dotenv. Please install it and retry.")
raise SystemExit(1) from exc
env_path = repo_root / ".env"
load_dotenv(env_path)
def _get_repo_root() -> Path:
return Path(__file__).resolve().parents[3]
def _get_staged_readmes(repo_root: Path) -> List[Path]:
try:
output = subprocess.check_output(
[
"git",
"-C",
str(repo_root),
"diff",
"--cached",
"--name-only",
"--",
"*.md",
],
text=True,
)
except subprocess.CalledProcessError as exc:
print(f"Failed to read staged files: {exc}")
return []
paths = []
for line in output.splitlines():
line = line.strip()
if not line:
continue
if line.endswith("README.md") or line.endswith("README_CN.md"):
paths.append(repo_root / line)
return paths
def _get_staged_plugin_files(repo_root: Path) -> List[Path]:
try:
output = subprocess.check_output(
[
"git",
"-C",
str(repo_root),
"diff",
"--cached",
"--name-only",
"--",
"*.py",
],
text=True,
)
except subprocess.CalledProcessError as exc:
print(f"Failed to read staged files: {exc}")
return []
paths = []
for line in output.splitlines():
line = line.strip()
if not line:
continue
if "/plugins/" not in line:
continue
if line.endswith("__init__.py") or os.path.basename(line).startswith("test_"):
continue
paths.append(repo_root / line)
return paths
def _parse_frontmatter(content: str) -> Dict[str, str]:
match = re.search(r'^\s*"""\n(.*?)\n"""', content, re.DOTALL)
if not match:
match = re.search(r'"""\n(.*?)\n"""', content, re.DOTALL)
if not match:
return {}
frontmatter = match.group(1)
meta: Dict[str, str] = {}
for line in frontmatter.split("\n"):
if ":" in line:
key, value = line.split(":", 1)
meta[key.strip()] = value.strip()
return meta
def _find_plugin_file(readme_path: Path) -> Optional[Path]:
plugin_dir = readme_path.parent
is_cn = readme_path.name.lower().endswith("readme_cn.md")
py_files = [
p
for p in plugin_dir.glob("*.py")
if p.name != "__init__.py" and not p.name.startswith("test_")
]
if not py_files:
return None
cn_files = [p for p in py_files if p.stem.endswith("_cn")]
en_files = [p for p in py_files if not p.stem.endswith("_cn")]
candidates = cn_files + en_files if is_cn else en_files + cn_files
# Prefer files that contain openwebui_id/post_id in frontmatter
for candidate in candidates:
post_id = _get_post_id(candidate)
if post_id:
return candidate
return candidates[0] if candidates else None
def _get_post_id(plugin_file: Path) -> Optional[str]:
try:
content = plugin_file.read_text(encoding="utf-8")
except Exception:
return None
meta = _parse_frontmatter(content)
return meta.get("openwebui_id") or meta.get("post_id")
def _get_plugin_metadata(plugin_file: Path) -> Dict[str, str]:
try:
content = plugin_file.read_text(encoding="utf-8")
except Exception:
return {}
return _parse_frontmatter(content)
def _find_readme_for_plugin(plugin_file: Path) -> Optional[str]:
plugin_dir = plugin_file.parent
is_cn = plugin_file.stem.endswith("_cn")
readme_candidates = ["README_CN.md", "README.md"] if is_cn else ["README.md", "README_CN.md"]
for name in readme_candidates:
readme_path = plugin_dir / name
if readme_path.exists():
return readme_path.read_text(encoding="utf-8")
return None
def main() -> int:
repo_root = _get_repo_root()
_load_dotenv(repo_root)
api_key = os.environ.get("OPENWEBUI_API_KEY")
if not api_key:
print("OPENWEBUI_API_KEY is not set in environment.")
return 1
client_module_path = repo_root / "scripts" / "openwebui_community_client.py"
spec = importlib.util.spec_from_file_location(
"openwebui_community_client", client_module_path
)
if not spec or not spec.loader:
print("Failed to load openwebui_community_client module.")
return 1
module = importlib.util.module_from_spec(spec)
spec.loader.exec_module(module)
client = module.get_client(api_key)
staged_plugins = _get_staged_plugin_files(repo_root)
staged_readmes = _get_staged_readmes(repo_root)
if not staged_plugins and not staged_readmes:
print("No staged README or plugin files found.")
return 0
updated_post_ids: set[str] = set()
for plugin_file in staged_plugins:
if not plugin_file.exists():
print(f"Skipped (missing): {plugin_file}")
continue
post_id = _get_post_id(plugin_file)
if not post_id:
print(f"Skipped (no openwebui_id): {plugin_file}")
continue
try:
post_data = client.get_post(post_id)
if not post_data:
print(f"Skipped (post not found): {plugin_file}")
continue
source_code = plugin_file.read_text(encoding="utf-8")
metadata = _get_plugin_metadata(plugin_file)
readme_content = _find_readme_for_plugin(plugin_file)
ok = client.update_plugin(
post_id=post_id,
source_code=source_code,
readme_content=readme_content or metadata.get("description", ""),
metadata=metadata,
media_urls=None,
)
if ok:
updated_post_ids.add(post_id)
print(f"Updated plugin -> {plugin_file} (post_id: {post_id})")
except Exception as exc:
print(f"Failed: {plugin_file} ({exc})")
for readme_path in staged_readmes:
if not readme_path.exists():
print(f"Skipped (missing): {readme_path}")
continue
plugin_file = _find_plugin_file(readme_path)
if not plugin_file:
print(f"Skipped (no plugin file): {readme_path}")
continue
post_id = _get_post_id(plugin_file)
if not post_id:
print(f"Skipped (no openwebui_id): {readme_path}")
continue
try:
if post_id in updated_post_ids:
print(f"Skipped (already updated via plugin): {readme_path}")
continue
post_data = client.get_post(post_id)
if not post_data:
print(f"Skipped (post not found): {readme_path}")
continue
readme_content = readme_path.read_text(encoding="utf-8")
# Update README content only, keep other fields unchanged.
post_data["content"] = readme_content
ok = client.update_post(post_id, post_data)
if ok:
print(f"Updated README -> {readme_path} (post_id: {post_id})")
except Exception as exc:
print(f"Failed: {readme_path} ({exc})")
return 0
if __name__ == "__main__":
raise SystemExit(main())

View File

@@ -0,0 +1,568 @@
# GitHub Copilot SDK 自定义工具快速入门
## 🎯 目标
在 OpenWebUI Pipe 中直接使用 GitHub Copilot SDK 的自定义工具功能,无需集成 OpenWebUI Function 系统。
---
## 📖 基础概念
### Copilot SDK Tool 的三要素
```python
from copilot.types import Tool, ToolInvocation, ToolResult
# 1. Tool Definition工具定义
tool = Tool(
name="tool_name", # 工具名称
description="What it does", # 描述(给 AI 看的)
parameters={...}, # JSON Schema 参数定义
handler=handler_function # 处理函数
)
# 2. Tool Handler处理函数
async def handler_function(invocation: ToolInvocation) -> ToolResult:
# invocation 包含:
# - session_id: 会话 ID
# - tool_call_id: 调用 ID
# - tool_name: 工具名称
# - arguments: dict实际参数
result = do_something(invocation["arguments"])
return ToolResult(
textResultForLlm="结果文本",
resultType="success", # 或 "failure"
error=None,
toolTelemetry={}
)
# 3. Session Configuration会话配置
session_config = SessionConfig(
model="claude-sonnet-4.5",
tools=[tool1, tool2, tool3], # ✅ 传入工具列表
streaming=True
)
```
---
## 💻 完整实现示例
### 示例 1获取当前时间
```python
from datetime import datetime
from copilot.types import Tool, ToolInvocation, ToolResult
def create_time_tool():
"""创建获取时间的工具"""
async def get_time_handler(invocation: ToolInvocation) -> ToolResult:
"""工具处理函数"""
try:
# 获取参数
timezone = invocation["arguments"].get("timezone", "UTC")
format_str = invocation["arguments"].get("format", "%Y-%m-%d %H:%M:%S")
# 执行逻辑
current_time = datetime.now().strftime(format_str)
result_text = f"Current time: {current_time}"
# 返回结果
return ToolResult(
textResultForLlm=result_text,
resultType="success",
error=None,
toolTelemetry={"execution_time": "fast"}
)
except Exception as e:
return ToolResult(
textResultForLlm=f"Error getting time: {str(e)}",
resultType="failure",
error=str(e),
toolTelemetry={}
)
# 创建工具定义
return Tool(
name="get_current_time",
description="Get the current date and time. Useful when user asks 'what time is it' or needs to know the current date.",
parameters={
"type": "object",
"properties": {
"timezone": {
"type": "string",
"description": "Timezone name (e.g., 'UTC', 'Asia/Shanghai')",
"default": "UTC"
},
"format": {
"type": "string",
"description": "Time format string",
"default": "%Y-%m-%d %H:%M:%S"
}
}
},
handler=get_time_handler
)
```
### 示例 2数学计算器
```python
def create_calculator_tool():
"""创建计算器工具"""
async def calculate_handler(invocation: ToolInvocation) -> ToolResult:
try:
expression = invocation["arguments"].get("expression", "")
# 安全检查
allowed_chars = set("0123456789+-*/()., ")
if not all(c in allowed_chars for c in expression):
raise ValueError("Expression contains invalid characters")
# 计算(安全的 eval
result = eval(expression, {"__builtins__": {}})
return ToolResult(
textResultForLlm=f"The result of {expression} is {result}",
resultType="success",
error=None,
toolTelemetry={}
)
except Exception as e:
return ToolResult(
textResultForLlm=f"Calculation error: {str(e)}",
resultType="failure",
error=str(e),
toolTelemetry={}
)
return Tool(
name="calculate",
description="Perform mathematical calculations. Supports basic arithmetic operations (+, -, *, /).",
parameters={
"type": "object",
"properties": {
"expression": {
"type": "string",
"description": "Mathematical expression to evaluate (e.g., '2 + 2 * 3')"
}
},
"required": ["expression"]
},
handler=calculate_handler
)
```
### 示例 3随机数生成器
```python
import random
def create_random_number_tool():
"""创建随机数生成工具"""
async def random_handler(invocation: ToolInvocation) -> ToolResult:
try:
min_val = invocation["arguments"].get("min", 1)
max_val = invocation["arguments"].get("max", 100)
if min_val >= max_val:
raise ValueError("min must be less than max")
number = random.randint(min_val, max_val)
return ToolResult(
textResultForLlm=f"Generated random number: {number}",
resultType="success",
error=None,
toolTelemetry={}
)
except Exception as e:
return ToolResult(
textResultForLlm=f"Error: {str(e)}",
resultType="failure",
error=str(e),
toolTelemetry={}
)
return Tool(
name="generate_random_number",
description="Generate a random integer within a specified range.",
parameters={
"type": "object",
"properties": {
"min": {
"type": "integer",
"description": "Minimum value (inclusive)",
"default": 1
},
"max": {
"type": "integer",
"description": "Maximum value (inclusive)",
"default": 100
}
}
},
handler=random_handler
)
```
---
## 🔧 集成到 Pipe
### 完整的 Pipe 实现
```python
class Pipe:
class Valves(BaseModel):
# ... 现有 Valves ...
ENABLE_TOOLS: bool = Field(
default=False,
description="Enable custom tools (time, calculator, random)"
)
AVAILABLE_TOOLS: str = Field(
default="all",
description="Available tools: 'all' or comma-separated list (e.g., 'get_current_time,calculate')"
)
def __init__(self):
# ... 现有初始化 ...
self._custom_tools = []
def _initialize_custom_tools(self):
"""初始化自定义工具"""
if not self.valves.ENABLE_TOOLS:
return []
# 定义所有可用工具
all_tools = {
"get_current_time": create_time_tool(),
"calculate": create_calculator_tool(),
"generate_random_number": create_random_number_tool(),
}
# 根据配置过滤工具
if self.valves.AVAILABLE_TOOLS == "all":
return list(all_tools.values())
# 只启用指定的工具
enabled = [t.strip() for t in self.valves.AVAILABLE_TOOLS.split(",")]
return [all_tools[name] for name in enabled if name in all_tools]
async def pipe(
self,
body: dict,
__metadata__: Optional[dict] = None,
__event_emitter__=None,
__event_call__=None,
) -> Union[str, AsyncGenerator]:
# ... 现有代码 ...
# ✅ 初始化工具
custom_tools = self._initialize_custom_tools()
if custom_tools:
await self._emit_debug_log(
f"Enabled {len(custom_tools)} custom tools: {[t.name for t in custom_tools]}",
__event_call__
)
# ✅ 创建会话配置(传入工具)
from copilot.types import SessionConfig, InfiniteSessionConfig
session_config = SessionConfig(
session_id=chat_id if chat_id else None,
model=real_model_id,
streaming=body.get("stream", False),
tools=custom_tools, # ✅✅✅ 关键:传入工具列表
infinite_sessions=infinite_session_config if self.valves.INFINITE_SESSION else None,
)
session = await client.create_session(config=session_config)
# ... 其余代码保持不变 ...
```
---
## 📊 处理工具调用事件
### 在 stream_response 中显示工具调用
```python
async def stream_response(
self, client, session, send_payload, init_message: str = "", __event_call__=None
) -> AsyncGenerator:
# ... 现有代码 ...
def handler(event):
event_type = str(getattr(event.type, "value", event.type))
# ✅ 工具调用开始
if "tool_invocation_started" in event_type or "tool_call_started" in event_type:
tool_name = get_event_data(event, "tool_name", "")
if tool_name:
queue.put_nowait(f"\n\n🔧 **Calling tool**: `{tool_name}`\n")
# ✅ 工具调用完成
elif "tool_invocation_completed" in event_type or "tool_call_completed" in event_type:
tool_name = get_event_data(event, "tool_name", "")
result = get_event_data(event, "result", "")
if tool_name:
queue.put_nowait(f"\n✅ **Tool `{tool_name}` completed**\n")
# ✅ 工具调用失败
elif "tool_invocation_failed" in event_type or "tool_call_failed" in event_type:
tool_name = get_event_data(event, "tool_name", "")
error = get_event_data(event, "error", "")
if tool_name:
queue.put_nowait(f"\n❌ **Tool `{tool_name}` failed**: {error}\n")
# ... 其他事件处理 ...
# ... 其余代码 ...
```
---
## 🧪 测试示例
### 测试 1询问时间
```
User: "What time is it now?"
Expected Flow:
1. Copilot 识别需要调用 get_current_time 工具
2. 调用工具(无参数或默认参数)
3. 工具返回: "Current time: 2026-01-26 15:30:00"
4. Copilot 回答: "The current time is 2026-01-26 15:30:00"
Pipe Output:
---
🔧 **Calling tool**: `get_current_time`
✅ **Tool `get_current_time` completed**
The current time is 2026-01-26 15:30:00
---
```
### 测试 2数学计算
```
User: "Calculate 123 * 456"
Expected Flow:
1. Copilot 调用 calculate 工具
2. 参数: {"expression": "123 * 456"}
3. 工具返回: "The result of 123 * 456 is 56088"
4. Copilot 回答: "123 multiplied by 456 equals 56,088"
Pipe Output:
---
🔧 **Calling tool**: `calculate`
✅ **Tool `calculate` completed**
123 multiplied by 456 equals 56,088
---
```
### 测试 3生成随机数
```
User: "Give me a random number between 1 and 10"
Expected Flow:
1. Copilot 调用 generate_random_number 工具
2. 参数: {"min": 1, "max": 10}
3. 工具返回: "Generated random number: 7"
4. Copilot 回答: "I generated a random number for you: 7"
```
---
## 🔍 调试技巧
### 1. 记录所有工具事件
```python
def handler(event):
event_type = str(getattr(event.type, "value", event.type))
# 记录所有包含 "tool" 的事件
if "tool" in event_type.lower():
event_data = {}
if hasattr(event, "data"):
try:
event_data = {
"type": event_type,
"data": str(event.data)[:200] # 截断长数据
}
except:
pass
self._emit_debug_log_sync(
f"Tool Event: {json.dumps(event_data)}",
__event_call__
)
```
### 2. 验证工具注册
```python
async def pipe(...):
# ...
custom_tools = self._initialize_custom_tools()
# 调试:打印工具信息
if self.valves.DEBUG:
tool_info = [
{
"name": t.name,
"description": t.description[:50],
"has_handler": t.handler is not None
}
for t in custom_tools
]
await self._emit_debug_log(
f"Registered tools: {json.dumps(tool_info, indent=2)}",
__event_call__
)
```
### 3. 测试工具处理函数
```python
# 单独测试工具
async def test_tool():
tool = create_time_tool()
# 模拟调用
invocation = {
"session_id": "test",
"tool_call_id": "test_call",
"tool_name": "get_current_time",
"arguments": {"format": "%H:%M:%S"}
}
result = await tool.handler(invocation)
print(f"Result: {result}")
```
---
## ⚠️ 注意事项
### 1. 工具描述的重要性
工具的 `description` 字段非常重要,它告诉 AI 何时应该使用这个工具:
```python
# ❌ 差的描述
description="Get time"
# ✅ 好的描述
description="Get the current date and time. Use this when the user asks 'what time is it', 'what's the date', or needs to know the current timestamp."
```
### 2. 参数定义
使用标准的 JSON Schema 定义参数:
```python
parameters={
"type": "object",
"properties": {
"param_name": {
"type": "string", # string, integer, boolean, array, object
"description": "Clear description",
"enum": ["option1", "option2"], # 可选:枚举值
"default": "default_value" # 可选:默认值
}
},
"required": ["param_name"] # 必需参数
}
```
### 3. 错误处理
总是捕获异常并返回有意义的错误:
```python
try:
result = do_something()
return ToolResult(
textResultForLlm=f"Success: {result}",
resultType="success",
error=None,
toolTelemetry={}
)
except Exception as e:
return ToolResult(
textResultForLlm=f"Error occurred: {str(e)}",
resultType="failure",
error=str(e), # 用于调试
toolTelemetry={}
)
```
### 4. 异步 vs 同步
工具处理函数可以是同步或异步:
```python
# 同步工具
def sync_handler(invocation):
result = calculate(invocation["arguments"])
return ToolResult(...)
# 异步工具(推荐)
async def async_handler(invocation):
result = await fetch_data(invocation["arguments"])
return ToolResult(...)
```
---
## 🚀 快速开始清单
- [ ] 1. 在 Valves 中添加 `ENABLE_TOOLS` 配置
- [ ] 2. 定义 2-3 个简单的工具函数
- [ ] 3. 实现 `_initialize_custom_tools()` 方法
- [ ] 4. 修改 `SessionConfig` 传入 `tools` 参数
- [ ] 5. 在 `stream_response` 中添加工具事件处理
- [ ] 6. 测试:询问时间、计算数学、生成随机数
- [ ] 7. 添加调试日志
- [ ] 8. 同步中文版本
---
## 📚 完整的工具事件列表
根据 SDK 源码,可能的工具相关事件:
- `tool_invocation_started` / `tool_call_started`
- `tool_invocation_completed` / `tool_call_completed`
- `tool_invocation_failed` / `tool_call_failed`
- `tool_parameter_validation_failed`
实际事件名称可能因 SDK 版本而异,建议先记录所有事件类型:
```python
def handler(event):
print(f"Event type: {event.type}")
```
---
**快速实现入口:** 从示例 1获取时间开始这是最简单的工具可以快速验证整个流程
**作者:** Fu-Jie
**日期:** 2026-01-26

View File

@@ -0,0 +1,480 @@
# OpenWebUI Native Tool Call Display Implementation Guide
**Date:** 2026-01-27
**Purpose:** Analyze and implement OpenWebUI's native tool call display mechanism
---
## 📸 Current vs Native Display
### Current Implementation
```markdown
> 🔧 **Running Tool**: `search_chats`
> ✅ **Tool Completed**: {...}
```
### OpenWebUI Native Display (from screenshot)
- ✅ Collapsible panel: "查看来自 search_chats 的结果"
- ✅ Formatted JSON display
- ✅ Syntax highlighting
- ✅ Expand/collapse functionality
- ✅ Clean visual separation
---
## 🔍 Understanding OpenWebUI's Tool Call Format
### Standard OpenAI Tool Call Message Format
OpenWebUI follows the OpenAI Chat Completion API format for tool calls:
#### 1. Assistant Message with Tool Calls
```python
{
"role": "assistant",
"content": None, # or explanatory text
"tool_calls": [
{
"id": "call_abc123",
"type": "function",
"function": {
"name": "search_chats",
"arguments": '{"query": ""}'
}
}
]
}
```
#### 2. Tool Response Message
```python
{
"role": "tool",
"tool_call_id": "call_abc123",
"name": "search_chats", # Optional but recommended
"content": '{"count": 5, "results": [...]}' # JSON string
}
```
---
## 🎯 Implementation Strategy for Native Display
### Option 1: Event Emitter Approach (Recommended)
Use OpenWebUI's event emitter to send structured tool call data:
```python
async def stream_response(self, ...):
# When tool execution starts
if event_type == "tool.execution_start":
await self._emit_tool_call_start(
emitter=__event_call__,
tool_call_id=tool_call_id,
tool_name=tool_name,
arguments=arguments
)
# When tool execution completes
elif event_type == "tool.execution_complete":
await self._emit_tool_call_result(
emitter=__event_call__,
tool_call_id=tool_call_id,
tool_name=tool_name,
result=result_content
)
```
#### Helper Methods
```python
async def _emit_tool_call_start(
self,
emitter: Optional[Callable[[Any], Awaitable[None]]],
tool_call_id: str,
tool_name: str,
arguments: dict
):
"""Emit a tool call start event to OpenWebUI."""
if not emitter:
return
try:
# OpenWebUI expects tool_calls in assistant message format
await emitter({
"type": "message",
"data": {
"role": "assistant",
"content": "",
"tool_calls": [
{
"id": tool_call_id,
"type": "function",
"function": {
"name": tool_name,
"arguments": json.dumps(arguments, ensure_ascii=False)
}
}
]
}
})
except Exception as e:
logger.error(f"Failed to emit tool call start: {e}")
async def _emit_tool_call_result(
self,
emitter: Optional[Callable[[Any], Awaitable[None]]],
tool_call_id: str,
tool_name: str,
result: Any
):
"""Emit a tool call result to OpenWebUI."""
if not emitter:
return
try:
# Format result as JSON string
if isinstance(result, str):
result_content = result
else:
result_content = json.dumps(result, ensure_ascii=False, indent=2)
# OpenWebUI expects tool results in tool message format
await emitter({
"type": "message",
"data": {
"role": "tool",
"tool_call_id": tool_call_id,
"name": tool_name,
"content": result_content
}
})
except Exception as e:
logger.error(f"Failed to emit tool result: {e}")
```
### Option 2: Message History Injection
Modify the conversation history to include tool calls:
```python
# After tool execution, append to messages
messages.append({
"role": "assistant",
"content": None,
"tool_calls": [{
"id": tool_call_id,
"type": "function",
"function": {
"name": tool_name,
"arguments": json.dumps(arguments)
}
}]
})
messages.append({
"role": "tool",
"tool_call_id": tool_call_id,
"name": tool_name,
"content": json.dumps(result)
})
```
---
## ⚠️ Challenges with Current Architecture
### 1. Streaming Context
Our current implementation uses:
- **Queue-based streaming**: Events → Queue → Yield chunks
- **Text chunks only**: We yield plain text, not structured messages
OpenWebUI's native display requires:
- **Structured message events**: Not text chunks
- **Message-level control**: Need to emit complete messages
### 2. Event Emitter Compatibility
**Current usage:**
```python
# We use event_emitter for status/notifications
await event_emitter({
"type": "status",
"data": {"description": "Processing..."}
})
```
**Need for tool calls:**
```python
# Need to emit message-type events
await event_emitter({
"type": "message",
"data": {
"role": "tool",
"content": "..."
}
})
```
**Question:** Does `__event_emitter__` support `message` type events?
### 3. Session SDK Events vs OpenWebUI Messages
**Copilot SDK events:**
- `tool.execution_start` → We get tool name, arguments
- `tool.execution_complete` → We get tool result
- Designed for streaming text output
**OpenWebUI messages:**
- Expect structured message objects
- Not designed for mid-stream injection
---
## 🧪 Experimental Implementation
### Step 1: Add Valve for Native Display
```python
class Valves(BaseModel):
USE_NATIVE_TOOL_DISPLAY: bool = Field(
default=False,
description="Use OpenWebUI's native tool call display instead of markdown formatting"
)
```
### Step 2: Modify Tool Event Handling
```python
async def stream_response(self, ...):
# ...existing code...
def handler(event):
event_type = get_event_type(event)
if event_type == "tool.execution_start":
tool_name = safe_get_data_attr(event, "name")
# Get tool arguments
tool_input = safe_get_data_attr(event, "input") or {}
tool_call_id = safe_get_data_attr(event, "tool_call_id", f"call_{time.time()}")
if tool_call_id:
active_tools[tool_call_id] = {
"name": tool_name,
"arguments": tool_input
}
if self.valves.USE_NATIVE_TOOL_DISPLAY:
# Emit structured tool call
asyncio.create_task(
self._emit_tool_call_start(
__event_call__,
tool_call_id,
tool_name,
tool_input
)
)
else:
# Current markdown display
queue.put_nowait(f"\n\n> 🔧 **Running Tool**: `{tool_name}`\n\n")
elif event_type == "tool.execution_complete":
tool_call_id = safe_get_data_attr(event, "tool_call_id")
tool_info = active_tools.get(tool_call_id, {})
tool_name = tool_info.get("name", "Unknown")
# Extract result
result_obj = safe_get_data_attr(event, "result")
result_content = ""
if hasattr(result_obj, "content"):
result_content = result_obj.content
elif isinstance(result_obj, dict):
result_content = result_obj.get("content", "")
if self.valves.USE_NATIVE_TOOL_DISPLAY:
# Emit structured tool result
asyncio.create_task(
self._emit_tool_call_result(
__event_call__,
tool_call_id,
tool_name,
result_content
)
)
else:
# Current markdown display
queue.put_nowait(f"> ✅ **Tool Completed**: {result_content}\n\n")
```
---
## 🔬 Testing Plan
### Test 1: Event Emitter Message Type Support
```python
# In a test conversation, try:
await __event_emitter__({
"type": "message",
"data": {
"role": "assistant",
"content": "Test message"
}
})
```
**Expected:** Message appears in chat
**If fails:** Event emitter doesn't support message type
### Test 2: Tool Call Message Format
```python
# Send a tool call message
await __event_emitter__({
"type": "message",
"data": {
"role": "assistant",
"content": None,
"tool_calls": [{
"id": "test_123",
"type": "function",
"function": {
"name": "test_tool",
"arguments": '{"param": "value"}'
}
}]
}
})
# Send tool result
await __event_emitter__({
"type": "message",
"data": {
"role": "tool",
"tool_call_id": "test_123",
"name": "test_tool",
"content": '{"result": "success"}'
}
})
```
**Expected:** OpenWebUI displays collapsible tool panel
**If fails:** Event format doesn't match OpenWebUI expectations
### Test 3: Mid-Stream Tool Call Injection
Test if tool call messages can be injected during streaming:
```python
# Start streaming text
yield "Processing your request..."
# Mid-stream: emit tool call
await __event_emitter__({"type": "message", "data": {...}})
# Continue streaming
yield "Done!"
```
**Expected:** Tool panel appears mid-response
**Risk:** May break streaming flow
---
## 📋 Implementation Checklist
- [x] Add `REASONING_EFFORT` valve (completed)
- [ ] Add `USE_NATIVE_TOOL_DISPLAY` valve
- [ ] Implement `_emit_tool_call_start()` helper
- [ ] Implement `_emit_tool_call_result()` helper
- [ ] Modify tool event handling in `stream_response()`
- [ ] Test event emitter message type support
- [ ] Test tool call message format
- [ ] Test mid-stream injection
- [ ] Update documentation
- [ ] Add user configuration guide
---
## 🤔 Recommendation
### Hybrid Approach (Safest)
Keep both display modes:
1. **Default (Current):** Markdown-based display
- ✅ Works reliably with streaming
- ✅ No OpenWebUI API dependencies
- ✅ Consistent across versions
2. **Experimental (Native):** Structured tool messages
- ✅ Better visual integration
- ⚠️ Requires testing with OpenWebUI internals
- ⚠️ May not work in all scenarios
**Configuration:**
```python
USE_NATIVE_TOOL_DISPLAY: bool = Field(
default=False,
description="[EXPERIMENTAL] Use OpenWebUI's native tool call display"
)
```
### Why Markdown Display is Currently Better
1. **Reliability:** Always works with streaming
2. **Flexibility:** Can customize format easily
3. **Context:** Shows tools inline with reasoning
4. **Compatibility:** Works across OpenWebUI versions
### When to Use Native Display
- Non-streaming mode (easier to inject messages)
- After confirming event emitter supports message type
- For tools with large JSON results (better formatting)
---
## 📚 Next Steps
1. **Research OpenWebUI Source Code**
- Check `__event_emitter__` implementation
- Verify supported event types
- Test message injection patterns
2. **Create Proof of Concept**
- Simple test plugin
- Emit tool call messages
- Verify UI rendering
3. **Document Findings**
- Update this guide with test results
- Add code examples that work
- Create migration guide if successful
---
## 🔗 References
- [OpenAI Chat Completion API](https://platform.openai.com/docs/api-reference/chat/create)
- [OpenWebUI Plugin Development](https://docs.openwebui.com/)
- [Copilot SDK Events](https://github.com/github/copilot-sdk)
---
**Author:** Fu-Jie
**Status:** Analysis Complete - Implementation Pending Testing

View File

@@ -0,0 +1,480 @@
# OpenWebUI 原生工具调用展示实现指南
**日期:** 2026-01-27
**目的:** 分析并实现 OpenWebUI 的原生工具调用展示机制
---
## 📸 当前展示 vs 原生展示
### 当前实现
```markdown
> 🔧 **Running Tool**: `search_chats`
> ✅ **Tool Completed**: {...}
```
### OpenWebUI 原生展示(来自截图)
- ✅ 可折叠面板:"查看来自 search_chats 的结果"
- ✅ 格式化的 JSON 显示
- ✅ 语法高亮
- ✅ 展开/折叠功能
- ✅ 清晰的视觉分隔
---
## 🔍 理解 OpenWebUI 的工具调用格式
### 标准 OpenAI 工具调用消息格式
OpenWebUI 遵循 OpenAI Chat Completion API 的工具调用格式:
#### 1. 带工具调用的助手消息
```python
{
"role": "assistant",
"content": None, # 或解释性文本
"tool_calls": [
{
"id": "call_abc123",
"type": "function",
"function": {
"name": "search_chats",
"arguments": '{"query": ""}'
}
}
]
}
```
#### 2. 工具响应消息
```python
{
"role": "tool",
"tool_call_id": "call_abc123",
"name": "search_chats", # 可选但推荐
"content": '{"count": 5, "results": [...]}' # JSON 字符串
}
```
---
## 🎯 原生展示的实现策略
### 方案 1事件发射器方法推荐
使用 OpenWebUI 的事件发射器发送结构化工具调用数据:
```python
async def stream_response(self, ...):
# 工具执行开始时
if event_type == "tool.execution_start":
await self._emit_tool_call_start(
emitter=__event_call__,
tool_call_id=tool_call_id,
tool_name=tool_name,
arguments=arguments
)
# 工具执行完成时
elif event_type == "tool.execution_complete":
await self._emit_tool_call_result(
emitter=__event_call__,
tool_call_id=tool_call_id,
tool_name=tool_name,
result=result_content
)
```
#### 辅助方法
```python
async def _emit_tool_call_start(
self,
emitter: Optional[Callable[[Any], Awaitable[None]]],
tool_call_id: str,
tool_name: str,
arguments: dict
):
"""向 OpenWebUI 发射工具调用开始事件。"""
if not emitter:
return
try:
# OpenWebUI 期望 assistant 消息格式的 tool_calls
await emitter({
"type": "message",
"data": {
"role": "assistant",
"content": "",
"tool_calls": [
{
"id": tool_call_id,
"type": "function",
"function": {
"name": tool_name,
"arguments": json.dumps(arguments, ensure_ascii=False)
}
}
]
}
})
except Exception as e:
logger.error(f"发射工具调用开始事件失败: {e}")
async def _emit_tool_call_result(
self,
emitter: Optional[Callable[[Any], Awaitable[None]]],
tool_call_id: str,
tool_name: str,
result: Any
):
"""向 OpenWebUI 发射工具调用结果。"""
if not emitter:
return
try:
# 将结果格式化为 JSON 字符串
if isinstance(result, str):
result_content = result
else:
result_content = json.dumps(result, ensure_ascii=False, indent=2)
# OpenWebUI 期望 tool 消息格式的工具结果
await emitter({
"type": "message",
"data": {
"role": "tool",
"tool_call_id": tool_call_id,
"name": tool_name,
"content": result_content
}
})
except Exception as e:
logger.error(f"发射工具结果失败: {e}")
```
### 方案 2消息历史注入
修改对话历史以包含工具调用:
```python
# 工具执行后,追加到消息中
messages.append({
"role": "assistant",
"content": None,
"tool_calls": [{
"id": tool_call_id,
"type": "function",
"function": {
"name": tool_name,
"arguments": json.dumps(arguments)
}
}]
})
messages.append({
"role": "tool",
"tool_call_id": tool_call_id,
"name": tool_name,
"content": json.dumps(result)
})
```
---
## ⚠️ 当前架构的挑战
### 1. 流式上下文
我们当前的实现使用:
- **基于队列的流式传输**:事件 → 队列 → 产出块
- **仅文本块**:我们产出纯文本,而非结构化消息
OpenWebUI 的原生展示需要:
- **结构化消息事件**:不是文本块
- **消息级别控制**:需要发射完整消息
### 2. 事件发射器兼容性
**当前用法:**
```python
# 我们使用 event_emitter 发送状态/通知
await event_emitter({
"type": "status",
"data": {"description": "处理中..."}
})
```
**工具调用所需:**
```python
# 需要发射 message 类型事件
await event_emitter({
"type": "message",
"data": {
"role": "tool",
"content": "..."
}
})
```
**问题:** `__event_emitter__` 是否支持 `message` 类型事件?
### 3. Session SDK 事件 vs OpenWebUI 消息
**Copilot SDK 事件:**
- `tool.execution_start` → 获取工具名称、参数
- `tool.execution_complete` → 获取工具结果
- 为流式文本输出设计
**OpenWebUI 消息:**
- 期望结构化消息对象
- 不为中间流注入设计
---
## 🧪 实验性实现
### 步骤 1添加原生展示 Valve
```python
class Valves(BaseModel):
USE_NATIVE_TOOL_DISPLAY: bool = Field(
default=False,
description="使用 OpenWebUI 的原生工具调用展示,而非 Markdown 格式"
)
```
### 步骤 2修改工具事件处理
```python
async def stream_response(self, ...):
# ...现有代码...
def handler(event):
event_type = get_event_type(event)
if event_type == "tool.execution_start":
tool_name = safe_get_data_attr(event, "name")
# 获取工具参数
tool_input = safe_get_data_attr(event, "input") or {}
tool_call_id = safe_get_data_attr(event, "tool_call_id", f"call_{time.time()}")
if tool_call_id:
active_tools[tool_call_id] = {
"name": tool_name,
"arguments": tool_input
}
if self.valves.USE_NATIVE_TOOL_DISPLAY:
# 发射结构化工具调用
asyncio.create_task(
self._emit_tool_call_start(
__event_call__,
tool_call_id,
tool_name,
tool_input
)
)
else:
# 当前 Markdown 展示
queue.put_nowait(f"\n\n> 🔧 **运行工具**: `{tool_name}`\n\n")
elif event_type == "tool.execution_complete":
tool_call_id = safe_get_data_attr(event, "tool_call_id")
tool_info = active_tools.get(tool_call_id, {})
tool_name = tool_info.get("name", "未知")
# 提取结果
result_obj = safe_get_data_attr(event, "result")
result_content = ""
if hasattr(result_obj, "content"):
result_content = result_obj.content
elif isinstance(result_obj, dict):
result_content = result_obj.get("content", "")
if self.valves.USE_NATIVE_TOOL_DISPLAY:
# 发射结构化工具结果
asyncio.create_task(
self._emit_tool_call_result(
__event_call__,
tool_call_id,
tool_name,
result_content
)
)
else:
# 当前 Markdown 展示
queue.put_nowait(f"> ✅ **工具完成**: {result_content}\n\n")
```
---
## 🔬 测试计划
### 测试 1事件发射器消息类型支持
```python
# 在测试对话中尝试:
await __event_emitter__({
"type": "message",
"data": {
"role": "assistant",
"content": "测试消息"
}
})
```
**预期:** 消息出现在聊天中
**如果失败:** 事件发射器不支持 message 类型
### 测试 2工具调用消息格式
```python
# 发送工具调用消息
await __event_emitter__({
"type": "message",
"data": {
"role": "assistant",
"content": None,
"tool_calls": [{
"id": "test_123",
"type": "function",
"function": {
"name": "test_tool",
"arguments": '{"param": "value"}'
}
}]
}
})
# 发送工具结果
await __event_emitter__({
"type": "message",
"data": {
"role": "tool",
"tool_call_id": "test_123",
"name": "test_tool",
"content": '{"result": "success"}'
}
})
```
**预期:** OpenWebUI 显示可折叠工具面板
**如果失败:** 事件格式与 OpenWebUI 期望不符
### 测试 3中间流工具调用注入
测试是否可以在流式传输期间注入工具调用消息:
```python
# 开始流式文本
yield "正在处理您的请求..."
# 中间流:发射工具调用
await __event_emitter__({"type": "message", "data": {...}})
# 继续流式传输
yield "完成!"
```
**预期:** 工具面板出现在响应中间
**风险:** 可能破坏流式传输流程
---
## 📋 实施检查清单
- [x] 添加 `REASONING_EFFORT` valve已完成
- [ ] 添加 `USE_NATIVE_TOOL_DISPLAY` valve
- [ ] 实现 `_emit_tool_call_start()` 辅助方法
- [ ] 实现 `_emit_tool_call_result()` 辅助方法
- [ ] 修改 `stream_response()` 中的工具事件处理
- [ ] 测试事件发射器消息类型支持
- [ ] 测试工具调用消息格式
- [ ] 测试中间流注入
- [ ] 更新文档
- [ ] 添加用户配置指南
---
## 🤔 建议
### 混合方法(最安全)
保留两种展示模式:
1. **默认(当前):** 基于 Markdown 的展示
- ✅ 与流式传输可靠工作
- ✅ 无 OpenWebUI API 依赖
- ✅ 跨版本一致
2. **实验性(原生):** 结构化工具消息
- ✅ 更好的视觉集成
- ⚠️ 需要测试 OpenWebUI 内部
- ⚠️ 可能不适用于所有场景
**配置:**
```python
USE_NATIVE_TOOL_DISPLAY: bool = Field(
default=False,
description="[实验性] 使用 OpenWebUI 的原生工具调用展示"
)
```
### 为什么 Markdown 展示目前更好
1. **可靠性:** 始终与流式传输兼容
2. **灵活性:** 可以轻松自定义格式
3. **上下文:** 与推理内联显示工具
4. **兼容性:** 跨 OpenWebUI 版本工作
### 何时使用原生展示
- 非流式模式(更容易注入消息)
- 确认事件发射器支持 message 类型后
- 对于具有大型 JSON 结果的工具(更好的格式化)
---
## 📚 后续步骤
1. **研究 OpenWebUI 源代码**
- 检查 `__event_emitter__` 实现
- 验证支持的事件类型
- 测试消息注入模式
2. **创建概念验证**
- 简单测试插件
- 发射工具调用消息
- 验证 UI 渲染
3. **记录发现**
- 使用测试结果更新本指南
- 添加有效的代码示例
- 如果成功,创建迁移指南
---
## 🔗 参考资料
- [OpenAI Chat Completion API](https://platform.openai.com/docs/api-reference/chat/create)
- [OpenWebUI 插件开发](https://docs.openwebui.com/)
- [Copilot SDK 事件](https://github.com/github/copilot-sdk)
---
**作者:** Fu-Jie
**状态:** 分析完成 - 实施等待测试

View File

@@ -0,0 +1,182 @@
# Native Tool Display Usage Guide
## 🎨 What is Native Tool Display?
Native Tool Display is an experimental feature that integrates with OpenWebUI's built-in tool call visualization system. When enabled, tool calls and their results are displayed in **collapsible JSON panels** instead of plain markdown text.
### Visual Comparison
**Traditional Display (markdown):**
```
> 🔧 Running Tool: `get_current_time`
> ✅ Tool Completed: 2026-01-27 10:30:00
```
**Native Display (collapsible panels):**
- Tool call appears in a collapsible `assistant.tool_calls` panel
- Tool result appears in a separate collapsible `tool.content` panel
- JSON syntax highlighting for better readability
- Compact by default, expandable on click
## 🚀 How to Enable
1. Open the GitHub Copilot SDK Pipe configuration (Valves)
2. Set `USE_NATIVE_TOOL_DISPLAY` to `true`
3. Save the configuration
4. Start a new conversation with tool calls
## 📋 Requirements
- OpenWebUI with native tool display support
- `__event_emitter__` must support `message` type events
- Tool-enabled models (e.g., GPT-4, Claude Sonnet)
## ⚙️ How It Works
### OpenAI Standard Format
The native display uses the OpenAI standard message format:
**Tool Call (Assistant Message):**
```json
{
"role": "assistant",
"content": "",
"tool_calls": [
{
"id": "call_abc123",
"type": "function",
"function": {
"name": "get_current_time",
"arguments": "{\"timezone\":\"UTC\"}"
}
}
]
}
```
**Tool Result (Tool Message):**
```json
{
"role": "tool",
"tool_call_id": "call_abc123",
"content": "2026-01-27 10:30:00 UTC"
}
```
### Message Flow
1. **Tool Execution Start**:
- SDK emits `tool.execution_start` event
- Plugin sends `assistant` message with `tool_calls` array
- OpenWebUI displays collapsible tool call panel
2. **Tool Execution Complete**:
- SDK emits `tool.execution_complete` event
- Plugin sends `tool` message with `tool_call_id` and `content`
- OpenWebUI displays collapsible result panel
## 🔧 Troubleshooting
### Panel Not Showing
**Symptoms:** Tool calls still appear as markdown text
**Possible Causes:**
1. `__event_emitter__` doesn't support `message` type events
2. OpenWebUI version too old
3. Feature not enabled (`USE_NATIVE_TOOL_DISPLAY = false`)
**Solution:**
- Enable DEBUG mode to see error messages in browser console
- Check browser console for "Native message emission failed" warnings
- Update OpenWebUI to latest version
- Keep `USE_NATIVE_TOOL_DISPLAY = false` to use traditional markdown display
### Duplicate Tool Information
**Symptoms:** Tool calls appear in both native panels and markdown
**Cause:** Mixed display modes
**Solution:**
- Ensure `USE_NATIVE_TOOL_DISPLAY` is either `true` (native only) or `false` (markdown only)
- Restart the conversation after changing this setting
## 🧪 Experimental Status
This feature is marked as **EXPERIMENTAL** because:
1. **Event Emitter API**: The `__event_emitter__` support for `message` type events is not fully documented
2. **OpenWebUI Version Dependency**: Requires recent versions of OpenWebUI with native tool display support
3. **Streaming Architecture**: May have compatibility issues with streaming responses
### Fallback Behavior
If native message emission fails:
- Plugin automatically falls back to markdown display
- Error logged to browser console (when DEBUG is enabled)
- No interruption to conversation flow
## 📊 Performance Considerations
Native display has slightly better performance characteristics:
| Aspect | Native Display | Markdown Display |
|--------|----------------|------------------|
| **Rendering** | Native UI components | Markdown parser |
| **Interactivity** | Collapsible panels | Static text |
| **JSON Parsing** | Handled by UI | Not formatted |
| **Token Usage** | Minimal overhead | Formatting tokens |
## 🔮 Future Enhancements
Planned improvements for native tool display:
- [ ] Automatic fallback detection
- [ ] Tool call history persistence
- [ ] Rich metadata display (execution time, arguments preview)
- [ ] Copy tool call JSON button
- [ ] Tool call replay functionality
## 💡 Best Practices
1. **Enable DEBUG First**: Test with DEBUG mode before using in production
2. **Monitor Browser Console**: Check for warning messages during tool calls
3. **Test with Simple Tools**: Verify with built-in tools before custom implementations
4. **Keep Fallback Option**: Don't rely solely on native display until it exits experimental status
## 📖 Related Documentation
- [TOOLS_USAGE.md](TOOLS_USAGE.md) - How to create and use custom tools
- [NATIVE_TOOL_DISPLAY_GUIDE.md](NATIVE_TOOL_DISPLAY_GUIDE.md) - Technical implementation details
- [WORKFLOW.md](WORKFLOW.md) - Complete integration workflow
## 🐛 Reporting Issues
If you encounter issues with native tool display:
1. Enable `DEBUG` and `USE_NATIVE_TOOL_DISPLAY`
2. Open browser console (F12)
3. Trigger a tool call
4. Copy any error messages
5. Report to [GitHub Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
Include:
- OpenWebUI version
- Browser and version
- Error messages from console
- Steps to reproduce
---
**Author:** Fu-Jie | **Version:** 0.2.0 | **License:** MIT

View File

@@ -0,0 +1,182 @@
# 原生工具显示使用指南
## 🎨 什么是原生工具显示?
原生工具显示是一项实验性功能,与 OpenWebUI 的内置工具调用可视化系统集成。启用后,工具调用及其结果将以**可折叠的 JSON 面板**显示,而不是纯文本 markdown。
### 视觉对比
**传统显示 (markdown):**
```
> 🔧 正在运行工具: `get_current_time`
> ✅ 工具已完成: 2026-01-27 10:30:00
```
**原生显示 (可折叠面板):**
- 工具调用显示在可折叠的 `assistant.tool_calls` 面板中
- 工具结果显示在单独的可折叠 `tool.content` 面板中
- JSON 语法高亮,提高可读性
- 默认折叠,点击即可展开
## 🚀 如何启用
1. 打开 GitHub Copilot SDK Pipe 配置 (Valves)
2.`USE_NATIVE_TOOL_DISPLAY` 设置为 `true`
3. 保存配置
4. 开始新的对话并使用工具调用
## 📋 要求
- 支持原生工具显示的 OpenWebUI
- `__event_emitter__` 必须支持 `message` 类型事件
- 支持工具的模型(例如 GPT-4、Claude Sonnet
## ⚙️ 工作原理
### OpenAI 标准格式
原生显示使用 OpenAI 标准消息格式:
**工具调用(助手消息):**
```json
{
"role": "assistant",
"content": "",
"tool_calls": [
{
"id": "call_abc123",
"type": "function",
"function": {
"name": "get_current_time",
"arguments": "{\"timezone\":\"UTC\"}"
}
}
]
}
```
**工具结果(工具消息):**
```json
{
"role": "tool",
"tool_call_id": "call_abc123",
"content": "2026-01-27 10:30:00 UTC"
}
```
### 消息流程
1. **工具执行开始**
- SDK 发出 `tool.execution_start` 事件
- 插件发送带有 `tool_calls` 数组的 `assistant` 消息
- OpenWebUI 显示可折叠的工具调用面板
2. **工具执行完成**
- SDK 发出 `tool.execution_complete` 事件
- 插件发送带有 `tool_call_id``content``tool` 消息
- OpenWebUI 显示可折叠的结果面板
## 🔧 故障排除
### 面板未显示
**症状:** 工具调用仍以 markdown 文本形式显示
**可能原因:**
1. `__event_emitter__` 不支持 `message` 类型事件
2. OpenWebUI 版本过旧
3. 功能未启用(`USE_NATIVE_TOOL_DISPLAY = false`
**解决方案:**
- 启用 DEBUG 模式查看浏览器控制台中的错误消息
- 检查浏览器控制台的 "Native message emission failed" 警告
- 更新 OpenWebUI 到最新版本
- 保持 `USE_NATIVE_TOOL_DISPLAY = false` 使用传统 markdown 显示
### 重复的工具信息
**症状:** 工具调用同时出现在原生面板和 markdown 中
**原因:** 混合显示模式
**解决方案:**
- 确保 `USE_NATIVE_TOOL_DISPLAY``true`(仅原生)或 `false`(仅 markdown
- 更改设置后重启对话
## 🧪 实验性状态
此功能标记为**实验性**,因为:
1. **事件发射器 API**`__event_emitter__``message` 类型事件的支持未完全文档化
2. **OpenWebUI 版本依赖**:需要支持原生工具显示的较新 OpenWebUI 版本
3. **流式架构**:可能与流式响应存在兼容性问题
### 回退行为
如果原生消息发送失败:
- 插件自动回退到 markdown 显示
- 错误记录到浏览器控制台(启用 DEBUG 时)
- 不会中断对话流程
## 📊 性能考虑
原生显示具有略好的性能特征:
| 方面 | 原生显示 | Markdown 显示 |
|------|----------|---------------|
| **渲染** | 原生 UI 组件 | Markdown 解析器 |
| **交互性** | 可折叠面板 | 静态文本 |
| **JSON 解析** | 由 UI 处理 | 未格式化 |
| **Token 使用** | 最小开销 | 格式化 token |
## 🔮 未来增强
原生工具显示的计划改进:
- [ ] 自动回退检测
- [ ] 工具调用历史持久化
- [ ] 丰富的元数据显示(执行时间、参数预览)
- [ ] 复制工具调用 JSON 按钮
- [ ] 工具调用重放功能
## 💡 最佳实践
1. **先启用 DEBUG**:在生产环境使用前先在 DEBUG 模式下测试
2. **监控浏览器控制台**:在工具调用期间检查警告消息
3. **使用简单工具测试**:在自定义实现前先用内置工具验证
4. **保留回退选项**:在退出实验性状态前不要完全依赖原生显示
## 📖 相关文档
- [TOOLS_USAGE.md](TOOLS_USAGE.md) - 如何创建和使用自定义工具
- [NATIVE_TOOL_DISPLAY_GUIDE.md](NATIVE_TOOL_DISPLAY_GUIDE.md) - 技术实现细节
- [WORKFLOW.md](WORKFLOW.md) - 完整集成工作流程
## 🐛 报告问题
如果您在使用原生工具显示时遇到问题:
1. 启用 `DEBUG``USE_NATIVE_TOOL_DISPLAY`
2. 打开浏览器控制台F12
3. 触发工具调用
4. 复制任何错误消息
5. 报告到 [GitHub Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
包含:
- OpenWebUI 版本
- 浏览器和版本
- 控制台的错误消息
- 复现步骤
---
**作者:** Fu-Jie | **版本:** 0.2.0 | **许可证:** MIT

View File

@@ -0,0 +1,509 @@
# OpenWebUI Function 集成方案
## 🎯 核心挑战
在 Copilot Tool Handler 中调用 OpenWebUI Functions 的关键问题:
**问题:** Copilot SDK 的 Tool Handler 是一个独立的回调函数,如何在这个上下文中访问和执行 OpenWebUI 的 Function
---
## 🔍 OpenWebUI Function 系统分析
### 1. Function 数据结构
OpenWebUI 的 Function/Tool 传递格式:
```python
body = {
"tools": [
{
"type": "function",
"function": {
"name": "get_weather",
"description": "Get current weather",
"parameters": {
"type": "object",
"properties": {
"location": {"type": "string"}
},
"required": ["location"]
}
}
}
]
}
```
### 2. Function 执行机制
OpenWebUI Functions 的执行方式有几种可能:
#### 选项 A: 通过 Function ID 调用内部 API
```python
# 假设 OpenWebUI 提供内部 API
from open_webui.apps.webui.models.functions import Functions
function_id = "function_uuid" # 需要从配置中获取
result = await Functions.execute_function(
function_id=function_id,
arguments={"location": "Beijing"}
)
```
#### 选项 B: 通过 **event_emitter** 触发
```python
# 通过事件系统触发 function 执行
if __event_emitter__:
await __event_emitter__({
"type": "function_call",
"data": {
"name": "get_weather",
"arguments": {"location": "Beijing"}
}
})
```
#### 选项 C: 自己实现 Function 逻辑
```python
# 在 Pipe 内部实现常用功能
class Pipe:
def _builtin_get_weather(self, location: str) -> dict:
# 实现天气查询
pass
def _builtin_search_web(self, query: str) -> dict:
# 实现网页搜索
pass
```
---
## 💡 推荐方案:混合架构
### 架构设计
```
User Message
OpenWebUI UI (Functions 已配置)
Pipe.pipe(body) - body 包含 tools[]
转换为 Copilot Tools + 存储 Function Registry
Copilot 决定调用 Tool
Tool Handler 查询 Registry → 执行对应逻辑
返回结果给 Copilot
继续生成回答
```
### 核心实现
#### 1. Function Registry函数注册表
```python
class Pipe:
def __init__(self):
# ...
self._function_registry = {} # {function_name: callable}
self._function_metadata = {} # {function_name: metadata}
```
#### 2. 注册 Functions
```python
def _register_openwebui_functions(
self,
owui_functions: List[dict],
__event_emitter__=None,
__event_call__=None
):
"""
注册 OpenWebUI Functions 到内部 registry
关键:将 function 定义和执行逻辑关联起来
"""
for func_def in owui_functions:
if func_def.get("type") != "function":
continue
func_info = func_def.get("function", {})
func_name = func_info.get("name")
if not func_name:
continue
# 存储元数据
self._function_metadata[func_name] = {
"description": func_info.get("description", ""),
"parameters": func_info.get("parameters", {}),
"original_def": func_def
}
# 创建执行器(关键)
executor = self._create_function_executor(
func_name,
func_def,
__event_emitter__,
__event_call__
)
self._function_registry[func_name] = executor
```
#### 3. Function Executor 工厂
```python
def _create_function_executor(
self,
func_name: str,
func_def: dict,
__event_emitter__=None,
__event_call__=None
):
"""
为每个 function 创建执行器
策略:
1. 优先使用内置实现
2. 尝试调用 OpenWebUI API
3. 返回错误
"""
async def executor(arguments: dict) -> dict:
# 策略 1: 检查是否有内置实现
builtin_method = getattr(self, f"_builtin_{func_name}", None)
if builtin_method:
self._emit_debug_log_sync(
f"Using builtin implementation for {func_name}",
__event_call__
)
try:
result = builtin_method(arguments)
if inspect.iscoroutine(result):
result = await result
return {"success": True, "result": result}
except Exception as e:
return {"success": False, "error": str(e)}
# 策略 2: 尝试通过 Event Emitter 调用
if __event_emitter__:
try:
# 尝试触发 function_call 事件
response_queue = asyncio.Queue()
await __event_emitter__({
"type": "function_call",
"data": {
"name": func_name,
"arguments": arguments,
"response_queue": response_queue # 回调队列
}
})
# 等待结果(带超时)
result = await asyncio.wait_for(
response_queue.get(),
timeout=self.valves.TOOL_TIMEOUT
)
return {"success": True, "result": result}
except asyncio.TimeoutError:
return {"success": False, "error": "Function execution timeout"}
except Exception as e:
self._emit_debug_log_sync(
f"Event emitter call failed: {e}",
__event_call__
)
# 继续尝试其他方法
# 策略 3: 尝试调用 OpenWebUI internal API
try:
# 这需要研究 OpenWebUI 源码确定正确的调用方式
from open_webui.apps.webui.models.functions import Functions
# 需要获取 function_id这是关键问题
function_id = self._get_function_id_by_name(func_name)
if function_id:
result = await Functions.execute(
function_id=function_id,
params=arguments
)
return {"success": True, "result": result}
except ImportError:
pass
except Exception as e:
self._emit_debug_log_sync(
f"OpenWebUI API call failed: {e}",
__event_call__
)
# 策略 4: 返回"未实现"错误
return {
"success": False,
"error": f"Function '{func_name}' is not implemented. "
"Please implement it as a builtin method or ensure OpenWebUI API is available."
}
return executor
```
#### 4. Tool Handler 实现
```python
def _create_tool_handler(self, tool_name: str, __event_call__=None):
"""为 Copilot SDK 创建 Tool Handler"""
async def handler(invocation: dict) -> dict:
"""
Copilot Tool Handler
invocation: {
"session_id": str,
"tool_call_id": str,
"tool_name": str,
"arguments": dict
}
"""
try:
# 从 registry 获取 executor
executor = self._function_registry.get(invocation["tool_name"])
if not executor:
return {
"textResultForLlm": f"Function '{invocation['tool_name']}' not found.",
"resultType": "failure",
"error": "function_not_found",
"toolTelemetry": {}
}
# 执行 function
self._emit_debug_log_sync(
f"Executing function: {invocation['tool_name']}({invocation['arguments']})",
__event_call__
)
exec_result = await executor(invocation["arguments"])
# 处理结果
if exec_result.get("success"):
result_text = str(exec_result.get("result", ""))
return {
"textResultForLlm": result_text,
"resultType": "success",
"error": None,
"toolTelemetry": {}
}
else:
error_msg = exec_result.get("error", "Unknown error")
return {
"textResultForLlm": f"Function execution failed: {error_msg}",
"resultType": "failure",
"error": error_msg,
"toolTelemetry": {}
}
except Exception as e:
self._emit_debug_log_sync(
f"Tool handler error: {e}",
__event_call__
)
return {
"textResultForLlm": "An unexpected error occurred during function execution.",
"resultType": "failure",
"error": str(e),
"toolTelemetry": {}
}
return handler
```
---
## 🔌 内置 Functions 实现示例
### 示例 1: 获取当前时间
```python
def _builtin_get_current_time(self, arguments: dict) -> str:
"""内置实现:获取当前时间"""
from datetime import datetime
timezone = arguments.get("timezone", "UTC")
format_str = arguments.get("format", "%Y-%m-%d %H:%M:%S")
now = datetime.now()
return now.strftime(format_str)
```
### 示例 2: 简单计算器
```python
def _builtin_calculate(self, arguments: dict) -> str:
"""内置实现:数学计算"""
expression = arguments.get("expression", "")
try:
# 安全的数学计算(仅允许基本运算)
allowed_chars = set("0123456789+-*/()., ")
if not all(c in allowed_chars for c in expression):
raise ValueError("Invalid characters in expression")
result = eval(expression, {"__builtins__": {}})
return str(result)
except Exception as e:
raise ValueError(f"Calculation error: {e}")
```
### 示例 3: 网页搜索(需要外部 API
```python
async def _builtin_search_web(self, arguments: dict) -> str:
"""内置实现:网页搜索(使用 DuckDuckGo"""
query = arguments.get("query", "")
max_results = arguments.get("max_results", 5)
try:
# 使用 duckduckgo_search 库
from duckduckgo_search import DDGS
results = []
with DDGS() as ddgs:
for r in ddgs.text(query, max_results=max_results):
results.append({
"title": r.get("title", ""),
"url": r.get("href", ""),
"snippet": r.get("body", "")
})
# 格式化结果
formatted = "\n\n".join([
f"**{r['title']}**\n{r['url']}\n{r['snippet']}"
for r in results
])
return formatted
except Exception as e:
raise ValueError(f"Search failed: {e}")
```
---
## 🚀 完整集成流程
### pipe() 方法中的集成
```python
async def pipe(
self,
body: dict,
__metadata__: Optional[dict] = None,
__event_emitter__=None,
__event_call__=None,
) -> Union[str, AsyncGenerator]:
# ... 现有代码 ...
# ✅ Step 1: 提取 OpenWebUI Functions
owui_functions = body.get("tools", [])
# ✅ Step 2: 注册 Functions
if self.valves.ENABLE_TOOLS and owui_functions:
self._register_openwebui_functions(
owui_functions,
__event_emitter__,
__event_call__
)
# ✅ Step 3: 转换为 Copilot Tools
copilot_tools = []
for func_name in self._function_registry.keys():
metadata = self._function_metadata[func_name]
copilot_tools.append({
"name": func_name,
"description": metadata["description"],
"parameters": metadata["parameters"],
"handler": self._create_tool_handler(func_name, __event_call__)
})
# ✅ Step 4: 创建 Session 并传递 Tools
session_config = SessionConfig(
model=real_model_id,
tools=copilot_tools, # ✅ 关键
...
)
session = await client.create_session(config=session_config)
# ... 后续代码 ...
```
---
## ⚠️ 待解决问题
### 1. Function ID 映射
**问题:** OpenWebUI Functions 通常通过 UUID 标识,但 body 中只有 name
**解决思路:**
- 在 OpenWebUI 启动时建立 name → id 映射表
- 或者修改 OpenWebUI 在 body 中同时传递 id
### 2. Event Emitter 回调机制
**问题:** 不确定 **event_emitter** 是否支持 function_call 事件
**验证方法:**
```python
# 测试代码
await __event_emitter__({
"type": "function_call",
"data": {"name": "test_func", "arguments": {}}
})
```
### 3. 异步执行超时
**问题:** 某些 Functions 可能执行很慢
**解决方案:**
- 实现 timeout 机制(已在 executor 中实现)
- 对于长时间运行的任务,考虑返回"processing"状态
---
## 📝 实现清单
- [ ] 实现 _function_registry 和 _function_metadata
- [ ] 实现 _register_openwebui_functions()
- [ ] 实现 _create_function_executor()
- [ ] 实现 _create_tool_handler()
- [ ] 实现 3-5 个常用内置 Functions
- [ ] 测试 Function 注册和调用流程
- [ ] 验证 **event_emitter** 机制
- [ ] 研究 OpenWebUI Functions API
- [ ] 添加错误处理和超时机制
- [ ] 更新文档
---
**下一步行动:**
1. 实现基础的 Function Registry
2. 添加 2-3 个简单的内置 Functions如 get_time, calculate
3. 测试基本的 Tool Calling 流程
4. 根据测试结果调整架构
**作者:** Fu-Jie
**日期:** 2026-01-26

View File

@@ -0,0 +1,708 @@
# SessionConfig 完整功能集成指南
## 📋 概述
本文档详细说明如何将 GitHub Copilot SDK 的 `SessionConfig` 所有功能集成到 OpenWebUI Pipe 中。
---
## 🎯 功能清单与集成状态
| 功能 | 状态 | 优先级 | 说明 |
|------|------|--------|------|
| `session_id` | ✅ 已实现 | 高 | 使用 OpenWebUI chat_id |
| `model` | ✅ 已实现 | 高 | 从 body 动态获取 |
| `tools` | ✅ 已实现 | 高 | v0.2.0 新增示例工具 |
| `streaming` | ✅ 已实现 | 高 | 支持流式输出 |
| `infinite_sessions` | ✅ 已实现 | 高 | 自动上下文压缩 |
| `system_message` | ⚠️ 部分支持 | 中 | 可通过 Valves 添加 |
| `available_tools` | ⚠️ 部分支持 | 中 | 已有 AVAILABLE_TOOLS |
| `excluded_tools` | 🔲 未实现 | 低 | 可添加到 Valves |
| `on_permission_request` | 🔲 未实现 | 中 | 需要 UI 交互支持 |
| `provider` (BYOK) | 🔲 未实现 | 低 | 高级功能 |
| `mcp_servers` | 🔲 未实现 | 低 | MCP 协议支持 |
| `custom_agents` | 🔲 未实现 | 低 | 自定义代理 |
| `config_dir` | 🔲 未实现 | 低 | 可通过 WORKSPACE_DIR |
| `skill_directories` | 🔲 未实现 | 低 | 技能系统 |
| `disabled_skills` | 🔲 未实现 | 低 | 技能过滤 |
---
## 📖 详细集成方案
### 1. ✅ session_id已实现
**功能:** 持久化会话 ID
**当前实现:**
```python
session_config = SessionConfig(
session_id=chat_id if chat_id else None, # 使用 OpenWebUI 的 chat_id
...
)
```
**工作原理:**
- OpenWebUI 的 `chat_id` 直接映射为 Copilot 的 `session_id`
- 会话状态持久化到磁盘
- 支持跨重启恢复对话
---
### 2. ✅ model已实现
**功能:** 选择 Copilot 模型
**当前实现:**
```python
# 从用户选择的模型中提取
request_model = body.get("model", "")
if request_model.startswith(f"{self.id}-"):
real_model_id = request_model[len(f"{self.id}-"):]
```
**Valves 配置:**
```python
MODEL_ID: str = Field(
default="claude-sonnet-4.5",
description="默认模型(动态获取失败时使用)"
)
```
---
### 3. ✅ tools已实现 - v0.2.0
**功能:** 自定义工具/函数调用
**当前实现:**
```python
custom_tools = self._initialize_custom_tools()
session_config = SessionConfig(
tools=custom_tools,
...
)
```
**Valves 配置:**
```python
ENABLE_TOOLS: bool = Field(default=False)
AVAILABLE_TOOLS: str = Field(default="all")
```
**内置示例工具:**
- `get_current_time` - 获取当前时间
- `calculate` - 数学计算
- `generate_random_number` - 随机数生成
**扩展方法:** 参考 [TOOLS_USAGE.md](TOOLS_USAGE.md)
---
### 4. ⚠️ system_message部分支持
**功能:** 自定义系统提示词
**集成方案:**
#### 方案 A通过 Valves 添加(推荐)
```python
class Valves(BaseModel):
SYSTEM_MESSAGE: str = Field(
default="",
description="Custom system message (append mode)"
)
SYSTEM_MESSAGE_MODE: str = Field(
default="append",
description="System message mode: 'append' or 'replace'"
)
```
**实现:**
```python
async def pipe(self, body, ...):
system_message_config = None
if self.valves.SYSTEM_MESSAGE:
if self.valves.SYSTEM_MESSAGE_MODE == "replace":
system_message_config = {
"mode": "replace",
"content": self.valves.SYSTEM_MESSAGE
}
else:
system_message_config = {
"mode": "append",
"content": self.valves.SYSTEM_MESSAGE
}
session_config = SessionConfig(
system_message=system_message_config,
...
)
```
#### 方案 B从 OpenWebUI 系统提示词读取
```python
# 从 body 中获取系统提示词
system_prompt = body.get("system", "")
if system_prompt:
system_message_config = {
"mode": "append",
"content": system_prompt
}
```
**注意事项:**
- `append` 模式:在默认系统提示词后追加
- `replace` 模式:完全替换(移除 SDK 安全保护)
---
### 5. ⚠️ available_tools / excluded_tools
**功能:** 工具白名单/黑名单
**当前部分支持:**
```python
AVAILABLE_TOOLS: str = Field(
default="all",
description="'all' or comma-separated list"
)
```
**增强实现:**
```python
class Valves(BaseModel):
AVAILABLE_TOOLS: str = Field(
default="all",
description="Available tools (comma-separated or 'all')"
)
EXCLUDED_TOOLS: str = Field(
default="",
description="Excluded tools (comma-separated)"
)
```
**应用到 SessionConfig**
```python
session_config = SessionConfig(
tools=custom_tools,
available_tools=self._parse_tool_list(self.valves.AVAILABLE_TOOLS),
excluded_tools=self._parse_tool_list(self.valves.EXCLUDED_TOOLS),
...
)
def _parse_tool_list(self, value: str) -> list[str]:
"""解析工具列表"""
if not value or value == "all":
return []
return [t.strip() for t in value.split(",") if t.strip()]
```
---
### 6. 🔲 on_permission_request未实现
**功能:** 处理权限请求shell 命令、文件写入等)
**使用场景:**
- Copilot 需要执行 shell 命令
- 需要写入文件
- 需要访问 URL
**集成挑战:**
- 需要 OpenWebUI 前端支持实时权限弹窗
- 需要异步处理用户确认
**推荐方案:**
#### 方案 A自动批准开发/测试环境)
```python
async def auto_approve_permission_handler(
request: dict,
context: dict
) -> dict:
"""自动批准所有权限请求(危险!)"""
return {
"kind": "approved",
"rules": []
}
session_config = SessionConfig(
on_permission_request=auto_approve_permission_handler,
...
)
```
#### 方案 B基于规则的批准
```python
class Valves(BaseModel):
ALLOW_SHELL_COMMANDS: bool = Field(default=False)
ALLOW_FILE_WRITE: bool = Field(default=False)
ALLOW_URL_ACCESS: bool = Field(default=True)
async def rule_based_permission_handler(
request: dict,
context: dict
) -> dict:
kind = request.get("kind")
if kind == "shell" and not self.valves.ALLOW_SHELL_COMMANDS:
return {"kind": "denied-by-rules"}
if kind == "write" and not self.valves.ALLOW_FILE_WRITE:
return {"kind": "denied-by-rules"}
if kind == "url" and not self.valves.ALLOW_URL_ACCESS:
return {"kind": "denied-by-rules"}
return {"kind": "approved", "rules": []}
```
#### 方案 C通过 Event Emitter 请求用户确认(理想)
```python
async def interactive_permission_handler(
request: dict,
context: dict
) -> dict:
"""通过前端请求用户确认"""
if not __event_emitter__:
return {"kind": "denied-no-approval-rule-and-could-not-request-from-user"}
# 发送权限请求到前端
response_queue = asyncio.Queue()
await __event_emitter__({
"type": "permission_request",
"data": {
"kind": request.get("kind"),
"description": request.get("description"),
"response_queue": response_queue
}
})
# 等待用户响应(带超时)
try:
user_response = await asyncio.wait_for(
response_queue.get(),
timeout=30.0
)
if user_response.get("approved"):
return {"kind": "approved", "rules": []}
else:
return {"kind": "denied-interactively-by-user"}
except asyncio.TimeoutError:
return {"kind": "denied-no-approval-rule-and-could-not-request-from-user"}
```
---
### 7. 🔲 providerBYOK - Bring Your Own Key
**功能:** 使用自己的 API 密钥连接 OpenAI/Azure/Anthropic
**使用场景:**
- 不使用 GitHub Copilot 配额
- 直接连接云服务提供商
- 使用 Azure OpenAI 部署
**集成方案:**
```python
class Valves(BaseModel):
USE_CUSTOM_PROVIDER: bool = Field(default=False)
PROVIDER_TYPE: str = Field(
default="openai",
description="Provider type: openai, azure, anthropic"
)
PROVIDER_BASE_URL: str = Field(default="")
PROVIDER_API_KEY: str = Field(default="")
PROVIDER_BEARER_TOKEN: str = Field(default="")
AZURE_API_VERSION: str = Field(default="2024-10-21")
def _build_provider_config(self) -> dict | None:
"""构建 Provider 配置"""
if not self.valves.USE_CUSTOM_PROVIDER:
return None
config = {
"type": self.valves.PROVIDER_TYPE,
"base_url": self.valves.PROVIDER_BASE_URL,
}
if self.valves.PROVIDER_API_KEY:
config["api_key"] = self.valves.PROVIDER_API_KEY
if self.valves.PROVIDER_BEARER_TOKEN:
config["bearer_token"] = self.valves.PROVIDER_BEARER_TOKEN
if self.valves.PROVIDER_TYPE == "azure":
config["azure"] = {
"api_version": self.valves.AZURE_API_VERSION
}
# 自动推断 wire_api
if self.valves.PROVIDER_TYPE == "anthropic":
config["wire_api"] = "responses"
else:
config["wire_api"] = "completions"
return config
```
**应用:**
```python
session_config = SessionConfig(
provider=self._build_provider_config(),
...
)
```
---
### 8. ✅ streaming已实现
**功能:** 流式输出
**当前实现:**
```python
session_config = SessionConfig(
streaming=body.get("stream", False),
...
)
```
---
### 9. 🔲 mcp_serversMCP 协议)
**功能:** Model Context Protocol 服务器集成
**使用场景:**
- 连接外部数据源数据库、API
- 集成第三方服务
**集成方案:**
```python
class Valves(BaseModel):
MCP_SERVERS_CONFIG: str = Field(
default="{}",
description="MCP servers configuration (JSON format)"
)
def _parse_mcp_servers(self) -> dict | None:
"""解析 MCP 服务器配置"""
if not self.valves.MCP_SERVERS_CONFIG:
return None
try:
return json.loads(self.valves.MCP_SERVERS_CONFIG)
except:
return None
```
**配置示例:**
```json
{
"database": {
"type": "local",
"command": "mcp-server-sqlite",
"args": ["--db", "/path/to/db.sqlite"],
"tools": ["*"]
},
"weather": {
"type": "http",
"url": "https://weather-api.example.com/mcp",
"tools": ["get_weather", "get_forecast"]
}
}
```
---
### 10. 🔲 custom_agents
**功能:** 自定义 AI 代理
**使用场景:**
- 专门化的子代理(如代码审查、文档编写)
- 不同的提示词策略
**集成方案:**
```python
class Valves(BaseModel):
CUSTOM_AGENTS_CONFIG: str = Field(
default="[]",
description="Custom agents configuration (JSON array)"
)
def _parse_custom_agents(self) -> list | None:
"""解析自定义代理配置"""
if not self.valves.CUSTOM_AGENTS_CONFIG:
return None
try:
return json.loads(self.valves.CUSTOM_AGENTS_CONFIG)
except:
return None
```
**配置示例:**
```json
[
{
"name": "code_reviewer",
"display_name": "Code Reviewer",
"description": "Reviews code for best practices",
"prompt": "You are an expert code reviewer. Focus on security, performance, and maintainability.",
"tools": ["read_file", "write_file"],
"infer": true
}
]
```
---
### 11. 🔲 config_dir
**功能:** 自定义配置目录
**当前支持:**
- 已有 `WORKSPACE_DIR` 控制工作目录
**增强方案:**
```python
class Valves(BaseModel):
CONFIG_DIR: str = Field(
default="",
description="Custom config directory for session state"
)
session_config = SessionConfig(
config_dir=self.valves.CONFIG_DIR if self.valves.CONFIG_DIR else None,
...
)
```
---
### 12. 🔲 skill_directories / disabled_skills
**功能:** Copilot Skills 系统
**使用场景:**
- 加载自定义技能包
- 禁用特定技能
**集成方案:**
```python
class Valves(BaseModel):
SKILL_DIRECTORIES: str = Field(
default="",
description="Comma-separated skill directories"
)
DISABLED_SKILLS: str = Field(
default="",
description="Comma-separated disabled skills"
)
def _parse_skills_config(self):
"""解析技能配置"""
skill_dirs = []
if self.valves.SKILL_DIRECTORIES:
skill_dirs = [
d.strip()
for d in self.valves.SKILL_DIRECTORIES.split(",")
if d.strip()
]
disabled = []
if self.valves.DISABLED_SKILLS:
disabled = [
s.strip()
for s in self.valves.DISABLED_SKILLS.split(",")
if s.strip()
]
return skill_dirs, disabled
# 应用
skill_dirs, disabled_skills = self._parse_skills_config()
session_config = SessionConfig(
skill_directories=skill_dirs if skill_dirs else None,
disabled_skills=disabled_skills if disabled_skills else None,
...
)
```
---
### 13. ✅ infinite_sessions已实现
**功能:** 无限会话与自动上下文压缩
**当前实现:**
```python
class Valves(BaseModel):
INFINITE_SESSION: bool = Field(default=True)
COMPACTION_THRESHOLD: float = Field(default=0.8)
BUFFER_THRESHOLD: float = Field(default=0.95)
infinite_session_config = None
if self.valves.INFINITE_SESSION:
infinite_session_config = {
"enabled": True,
"background_compaction_threshold": self.valves.COMPACTION_THRESHOLD,
"buffer_exhaustion_threshold": self.valves.BUFFER_THRESHOLD,
}
session_config = SessionConfig(
infinite_sessions=infinite_session_config,
...
)
```
---
## 🎯 实施优先级建议
### 🔥 高优先级(立即实现)
1. **system_message** - 用户最常需要的功能
2. **on_permission_request (基于规则)** - 安全性需求
### 📌 中优先级(下一阶段)
3. **excluded_tools** - 完善工具管理
4. **provider (BYOK)** - 高级用户需求
5. **config_dir** - 增强会话管理
### 📋 低优先级(可选)
6. **mcp_servers** - 高级集成
7. **custom_agents** - 专业化功能
8. **skill_directories** - 生态系统功能
---
## 🚀 快速实施计划
### Phase 1: 基础增强1-2小时
```python
# 添加到 Valves
SYSTEM_MESSAGE: str = Field(default="")
SYSTEM_MESSAGE_MODE: str = Field(default="append")
EXCLUDED_TOOLS: str = Field(default="")
# 添加到 pipe() 方法
system_message_config = self._build_system_message_config()
excluded_tools = self._parse_tool_list(self.valves.EXCLUDED_TOOLS)
session_config = SessionConfig(
system_message=system_message_config,
excluded_tools=excluded_tools,
...
)
```
### Phase 2: 权限管理2-3小时
```python
# 添加权限控制 Valves
ALLOW_SHELL_COMMANDS: bool = Field(default=False)
ALLOW_FILE_WRITE: bool = Field(default=False)
ALLOW_URL_ACCESS: bool = Field(default=True)
# 实现权限处理器
session_config = SessionConfig(
on_permission_request=self._create_permission_handler(),
...
)
```
### Phase 3: BYOK 支持3-4小时
```python
# 添加 Provider Valves
USE_CUSTOM_PROVIDER: bool = Field(default=False)
PROVIDER_TYPE: str = Field(default="openai")
PROVIDER_BASE_URL: str = Field(default="")
PROVIDER_API_KEY: str = Field(default="")
# 实现 Provider 配置
session_config = SessionConfig(
provider=self._build_provider_config(),
...
)
```
---
## 📚 参考资源
- **SDK 类型定义**: `/opt/homebrew/.../copilot/types.py`
- **工具系统**: [TOOLS_USAGE.md](TOOLS_USAGE.md)
- **SDK 文档**: <https://github.com/github/copilot-sdk>
---
## ✅ 实施检查清单
使用此清单跟踪实施进度:
- [x] session_id
- [x] model
- [x] tools
- [x] streaming
- [x] infinite_sessions
- [ ] system_message
- [ ] available_tools (完善)
- [ ] excluded_tools
- [ ] on_permission_request
- [ ] provider (BYOK)
- [ ] mcp_servers
- [ ] custom_agents
- [ ] config_dir
- [ ] skill_directories
- [ ] disabled_skills
---
**作者:** Fu-Jie
**版本:** v1.0
**日期:** 2026-01-26
**更新:** 随功能实施持续更新

View File

@@ -0,0 +1,191 @@
# 🛠️ Custom Tools Usage / 自定义工具使用指南
## Overview / 概览
This pipe includes **1 example custom tool** that demonstrates how to use GitHub Copilot SDK's tool calling feature.
本 Pipe 包含 **1 个示例自定义工具**,展示如何使用 GitHub Copilot SDK 的工具调用功能。
---
## 🚀 Quick Start / 快速开始
### 1. Enable Tools / 启用工具
In Valves configuration:
在 Valves 配置中:
```
ENABLE_TOOLS: true
AVAILABLE_TOOLS: all
```
### 2. Test with Conversations / 测试对话
Try these examples:
尝试这些示例:
**English:**
- "Give me a random number between 1 and 100"
**中文:**
- "给我一个 1 到 100 之间的随机数"
---
## 📦 Included Tools / 内置工具
### 1. `generate_random_number` / 生成随机数
**Description:** Generate a random integer
**描述:** 生成随机整数
**Parameters / 参数:**
- `min` (optional): Minimum value (default: 1)
- `max` (optional): Maximum value (default: 100)
- `min` (可选): 最小值 (默认: 1)
- `max` (可选): 最大值 (默认: 100)
**Example / 示例:**
```
User: "Give me a random number between 1 and 10"
Copilot: [calls generate_random_number with min=1, max=10] "Generated random number: 7"
用户: "给我一个 1 到 10 之间的随机数"
Copilot: [调用 generate_random_number参数 min=1, max=10] "生成的随机数: 7"
```
---
## ⚙️ Advanced Configuration / 高级配置
### Select Specific Tools / 选择特定工具
Instead of enabling all tools, specify which ones to use:
不启用所有工具,而是指定要使用的工具:
```
ENABLE_TOOLS: true
AVAILABLE_TOOLS: generate_random_number
```
---
## 🔧 How Tool Calling Works / 工具调用的工作原理
```
1. User asks a question / 用户提问
2. Copilot decides if it needs a tool / Copilot 决定是否需要工具
3. If yes, Copilot calls the appropriate tool / 如果需要,调用相应工具
4. Tool executes and returns result / 工具执行并返回结果
5. Copilot uses the result to answer / Copilot 使用结果回答
```
### Visual Feedback / 可视化反馈
When tools are called, you'll see:
当工具被调用时,你会看到:
```
🔧 **Calling tool**: `generate_random_number`
✅ **Tool `generate_random_number` completed**
Generated random number: 7
```
---
## 📚 Creating Your Own Tools / 创建自定义工具
Want to add your own tools? Follow this pattern (module-level tools):
想要添加自己的工具?遵循这个模式(模块级工具):
```python
from pydantic import BaseModel, Field
from copilot import define_tool
class MyToolParams(BaseModel):
param_name: str = Field(description="Parameter description")
@define_tool(description="Clear description of what the tool does and when to use it")
async def my_tool(params: MyToolParams) -> str:
# Do something
result = do_something(params.param_name)
return f"Result: {result}"
```
Then register it in `_initialize_custom_tools()`:
然后将它添加到 `_initialize_custom_tools()`:
```python
def _initialize_custom_tools(self):
if not self.valves.ENABLE_TOOLS:
return []
all_tools = {
"generate_random_number": generate_random_number,
"my_tool": my_tool, # ✅ Add here
}
if self.valves.AVAILABLE_TOOLS == "all":
return list(all_tools.values())
enabled = [t.strip() for t in self.valves.AVAILABLE_TOOLS.split(",")]
return [all_tools[name] for name in enabled if name in all_tools]
```
---
## ⚠️ Important Notes / 重要说明
### Security / 安全性
- Tools run in the same process as the pipe
- Be careful with tools that execute code or access files
- Always validate input parameters
- 工具在与 Pipe 相同的进程中运行
- 谨慎处理执行代码或访问文件的工具
- 始终验证输入参数
### Performance / 性能
- Tool execution is synchronous during streaming
- Long-running tools may cause delays
- Consider adding timeouts for external API calls
- 工具执行在流式传输期间是同步的
- 长时间运行的工具可能导致延迟
- 考虑为外部 API 调用添加超时
### Debugging / 调试
- Enable `DEBUG: true` to see tool events in the browser console
- Check tool calls in `🔧 Calling tool` messages
- Tool errors are displayed in the response
- 启用 `DEBUG: true` 在浏览器控制台查看工具事件
-`🔧 Calling tool` 消息中检查工具调用
- 工具错误会显示在响应中
---
## 📖 References / 参考资料
- [Copilot SDK Documentation](https://github.com/github/copilot-sdk)
- [COPILOT_TOOLS_QUICKSTART.md](COPILOT_TOOLS_QUICKSTART.md) - Detailed implementation guide
- [JSON Schema](https://json-schema.org/) - For parameter definitions
---
**Version:** 0.2.3
**Last Updated:** 2026-01-27

View File

@@ -0,0 +1,431 @@
# GitHub Copilot SDK - Tool 功能实现指南
## 📋 概述
本指南介绍如何在 GitHub Copilot SDK Pipe 中实现 Function/Tool Calling 功能。
---
## 🏗️ 架构设计
### 工作流程
```
OpenWebUI Tools/Functions
↓ (转换)
Copilot SDK Tool Definition
↓ (注册)
Session Tool Handlers
↓ (调用)
Tool Execution → Result
↓ (返回)
Continue Conversation
```
### 核心接口
#### 1. Tool Definition工具定义
```python
from copilot.types import Tool
tool = Tool(
name="get_weather",
description="Get current weather for a location",
parameters={
"type": "object",
"properties": {
"location": {
"type": "string",
"description": "City name, e.g., 'San Francisco'"
},
"unit": {
"type": "string",
"enum": ["celsius", "fahrenheit"],
"description": "Temperature unit"
}
},
"required": ["location"]
},
handler=weather_handler # 处理函数
)
```
#### 2. Tool Handler处理函数
```python
from copilot.types import ToolInvocation, ToolResult
async def weather_handler(invocation: ToolInvocation) -> ToolResult:
"""
invocation 包含:
- session_id: str
- tool_call_id: str
- tool_name: str
- arguments: dict # {"location": "San Francisco", "unit": "celsius"}
"""
location = invocation["arguments"]["location"]
# 执行实际逻辑
weather_data = await fetch_weather(location)
# 返回结果
return ToolResult(
textResultForLlm=f"Weather in {location}: {weather_data['temp']}°C, {weather_data['condition']}",
resultType="success", # or "failure"
error=None,
toolTelemetry={"execution_time_ms": 150}
)
```
#### 3. Session Configuration会话配置
```python
from copilot.types import SessionConfig
session_config = SessionConfig(
model="claude-sonnet-4.5",
tools=[tool1, tool2, tool3], # ✅ 传递工具列表
available_tools=["get_weather", "search_web"], # 可选:过滤可用工具
excluded_tools=["dangerous_tool"], # 可选:排除工具
)
session = await client.create_session(config=session_config)
```
---
## 💻 实现方案
### 方案 A桥接 OpenWebUI Tools推荐
#### 1. 添加 Valves 配置
```python
class Valves(BaseModel):
ENABLE_TOOLS: bool = Field(
default=True,
description="Enable OpenWebUI tool integration"
)
TOOL_TIMEOUT: int = Field(
default=30,
description="Tool execution timeout (seconds)"
)
AVAILABLE_TOOLS: str = Field(
default="",
description="Filter specific tools (comma separated, empty = all)"
)
```
#### 2. 实现 Tool 转换器
```python
def _convert_openwebui_tools_to_copilot(
self,
owui_tools: List[dict],
__event_call__=None
) -> List[dict]:
"""
将 OpenWebUI tools 转换为 Copilot SDK 格式
OpenWebUI Tool 格式:
{
"type": "function",
"function": {
"name": "get_weather",
"description": "Get weather info",
"parameters": {...} # JSON Schema
}
}
"""
copilot_tools = []
for tool in owui_tools:
if tool.get("type") != "function":
continue
func = tool.get("function", {})
tool_name = func.get("name")
if not tool_name:
continue
# 应用过滤器
if self.valves.AVAILABLE_TOOLS:
allowed = [t.strip() for t in self.valves.AVAILABLE_TOOLS.split(",")]
if tool_name not in allowed:
continue
copilot_tools.append({
"name": tool_name,
"description": func.get("description", ""),
"parameters": func.get("parameters", {}),
"handler": self._create_tool_handler(tool_name, __event_call__)
})
self._emit_debug_log_sync(
f"Registered tool: {tool_name}",
__event_call__
)
return copilot_tools
```
#### 3. 实现动态 Tool Handler
```python
def _create_tool_handler(self, tool_name: str, __event_call__=None):
"""为每个 tool 创建 handler 函数"""
async def handler(invocation: dict) -> dict:
"""
Tool handler 实现
invocation 结构:
{
"session_id": "...",
"tool_call_id": "...",
"tool_name": "get_weather",
"arguments": {"location": "Beijing"}
}
"""
try:
self._emit_debug_log_sync(
f"Tool called: {invocation['tool_name']} with {invocation['arguments']}",
__event_call__
)
# 方法 1: 调用 OpenWebUI 内部 Function API
result = await self._execute_openwebui_function(
function_name=invocation["tool_name"],
arguments=invocation["arguments"]
)
# 方法 2: 通过 __event_emitter__ 触发(需要测试)
# 方法 3: 直接实现工具逻辑
return {
"textResultForLlm": str(result),
"resultType": "success",
"error": None,
"toolTelemetry": {}
}
except asyncio.TimeoutError:
return {
"textResultForLlm": "Tool execution timed out.",
"resultType": "failure",
"error": "timeout",
"toolTelemetry": {}
}
except Exception as e:
self._emit_debug_log_sync(
f"Tool error: {e}",
__event_call__
)
return {
"textResultForLlm": f"Tool execution failed: {str(e)}",
"resultType": "failure",
"error": str(e),
"toolTelemetry": {}
}
return handler
```
#### 4. 集成到 pipe() 方法
```python
async def pipe(
self,
body: dict,
__metadata__: Optional[dict] = None,
__event_emitter__=None,
__event_call__=None,
) -> Union[str, AsyncGenerator]:
# ... 现有代码 ...
# ✅ 提取并转换 tools
copilot_tools = []
if self.valves.ENABLE_TOOLS and body.get("tools"):
copilot_tools = self._convert_openwebui_tools_to_copilot(
body["tools"],
__event_call__
)
await self._emit_debug_log(
f"Enabled {len(copilot_tools)} tools",
__event_call__
)
# ✅ 传递给 SessionConfig
session_config = SessionConfig(
session_id=chat_id if chat_id else None,
model=real_model_id,
streaming=body.get("stream", False),
tools=copilot_tools, # ✅ 关键
infinite_sessions=infinite_session_config,
)
session = await client.create_session(config=session_config)
# ...
```
#### 5. 处理 Tool 调用事件
```python
def stream_response(...):
def handler(event):
event_type = str(event.type)
# ✅ Tool 调用开始
if "tool_invocation_started" in event_type:
tool_name = get_event_data(event, "tool_name", "")
yield f"\n🔧 **Calling tool**: `{tool_name}`\n"
# ✅ Tool 调用完成
elif "tool_invocation_completed" in event_type:
tool_name = get_event_data(event, "tool_name", "")
result = get_event_data(event, "result", "")
yield f"\n✅ **Tool result**: {result}\n"
# ✅ Tool 调用失败
elif "tool_invocation_failed" in event_type:
tool_name = get_event_data(event, "tool_name", "")
error = get_event_data(event, "error", "")
yield f"\n❌ **Tool failed**: `{tool_name}` - {error}\n"
```
---
### 方案 B自定义 Tool 实现
#### Valves 配置
```python
class Valves(BaseModel):
CUSTOM_TOOLS: str = Field(
default="[]",
description="Custom tools JSON: [{name, description, parameters, implementation}]"
)
```
#### 工具定义示例
```json
[
{
"name": "calculate",
"description": "Perform mathematical calculations",
"parameters": {
"type": "object",
"properties": {
"expression": {
"type": "string",
"description": "Math expression, e.g., '2 + 2 * 3'"
}
},
"required": ["expression"]
},
"implementation": "eval" // 或指定 Python 函数名
}
]
```
---
## 🧪 测试方案
### 1. 测试 Tool 定义
```python
# 在 OpenWebUI 中创建一个简单的 Function:
# Name: get_time
# Description: Get current time
# Parameters: {"type": "object", "properties": {}}
# 测试对话:
# User: "What time is it?"
# Expected: Copilot 调用 get_time tool返回当前时间
```
### 2. 测试 Tool 调用链
```python
# User: "Search for Python tutorials and summarize the top 3 results"
# Expected Flow:
# 1. Copilot calls search_web(query="Python tutorials")
# 2. Copilot receives search results
# 3. Copilot summarizes top 3
# 4. Returns final answer
```
### 3. 测试错误处理
```python
# User: "Call a non-existent tool"
# Expected: 返回 "Tool not supported" error
```
---
## 📊 事件监听
Tool 相关事件类型:
- `tool_invocation_started` - Tool 调用开始
- `tool_invocation_completed` - Tool 完成
- `tool_invocation_failed` - Tool 失败
- `tool_parameter_validation_failed` - 参数验证失败
---
## ⚠️ 注意事项
### 1. 安全性
- ✅ 验证 tool parameters
- ✅ 限制执行超时
- ✅ 不暴露详细错误信息给 LLM
- ❌ 禁止执行危险命令(如 `rm -rf`
### 2. 性能
- ⏱️ 设置合理的 timeout
- 🔄 考虑异步执行长时间运行的 tool
- 📊 记录 tool 执行时间toolTelemetry
### 3. 调试
- 🐛 在 DEBUG 模式下记录所有 tool 调用
- 📝 记录 arguments 和 results
- 🔍 使用前端 console 显示 tool 流程
---
## 🔗 参考资源
- [GitHub Copilot SDK 官方文档](https://github.com/github/copilot-sdk)
- [OpenWebUI Function API](https://docs.openwebui.com/features/plugin-system)
- [JSON Schema 规范](https://json-schema.org/)
---
## 📝 实现清单
- [ ] 添加 ENABLE_TOOLS Valve
- [ ] 实现 _convert_openwebui_tools_to_copilot()
- [ ] 实现 _create_tool_handler()
- [ ] 修改 SessionConfig 传递 tools
- [ ] 处理 tool 事件流
- [ ] 添加调试日志
- [ ] 测试基础 tool 调用
- [ ] 测试错误处理
- [ ] 更新文档和 README
- [ ] 同步中文版本
---
**作者:** Fu-Jie
**版本:** v1.0
**日期:** 2026-01-26

View File

@@ -0,0 +1,835 @@
# GitHub Copilot SDK Integration Workflow
**Author:** Fu-Jie
**Version:** 0.2.3
**Last Updated:** 2026-01-27
---
## Table of Contents
1. [Architecture Overview](#architecture-overview)
2. [Request Processing Flow](#request-processing-flow)
3. [Session Management](#session-management)
4. [Streaming Response Handling](#streaming-response-handling)
5. [Event Processing Mechanism](#event-processing-mechanism)
6. [Tool Execution Flow](#tool-execution-flow)
7. [System Prompt Extraction](#system-prompt-extraction)
8. [Configuration Parameters](#configuration-parameters)
9. [Key Functions Reference](#key-functions-reference)
---
## Architecture Overview
### Component Diagram
```
┌─────────────────────────────────────────────────────────────┐
│ OpenWebUI │
│ ┌───────────────────────────────────────────────────────┐ │
│ │ Pipe Interface (Entry Point) │ │
│ └─────────────────────┬─────────────────────────────────┘ │
│ │ │
│ ▼ │
│ ┌───────────────────────────────────────────────────────┐ │
│ │ _pipe_impl (Main Logic) │ │
│ │ ┌──────────────────────────────────────────────────┐ │ │
│ │ │ 1. Environment Setup (_setup_env) │ │ │
│ │ │ 2. Model Selection (request_model parsing) │ │ │
│ │ │ 3. Chat Context Extraction │ │ │
│ │ │ 4. System Prompt Extraction │ │ │
│ │ │ 5. Session Management (create/resume) │ │ │
│ │ │ 6. Streaming/Non-streaming Response │ │ │
│ │ └──────────────────────────────────────────────────┘ │ │
│ └─────────────────────┬─────────────────────────────────┘ │
│ │ │
│ ▼ │
│ ┌───────────────────────────────────────────────────────┐ │
│ │ GitHub Copilot Client │ │
│ │ ┌──────────────────────────────────────────────────┐ │ │
│ │ │ • CopilotClient (SDK instance) │ │ │
│ │ │ • Session (conversation context) │ │ │
│ │ │ • Event Stream (async events) │ │ │
│ │ └──────────────────────────────────────────────────┘ │ │
│ └─────────────────────┬─────────────────────────────────┘ │
│ │ │
└────────────────────────┼─────────────────────────────────────┘
┌──────────────────────┐
│ Copilot CLI Process │
│ (Backend Agent) │
└──────────────────────┘
```
### Key Components
1. **Pipe Interface**: OpenWebUI's standard entry point
2. **Environment Manager**: CLI setup, token validation, environment variables
3. **Session Manager**: Persistent conversation state with automatic compaction
4. **Event Processor**: Asynchronous streaming event handler
5. **Tool System**: Custom tool registration and execution
6. **Debug Logger**: Frontend console logging for troubleshooting
---
## Request Processing Flow
### Complete Request Lifecycle
```mermaid
graph TD
A[OpenWebUI Request] --> B[pipe Entry Point]
B --> C[_pipe_impl]
C --> D{Setup Environment}
D --> E[Parse Model ID]
E --> F[Extract Chat Context]
F --> G[Extract System Prompt]
G --> H{Session Exists?}
H -->|Yes| I[Resume Session]
H -->|No| J[Create New Session]
I --> K[Initialize Tools]
J --> K
K --> L[Process Images]
L --> M{Streaming Mode?}
M -->|Yes| N[stream_response]
M -->|No| O[send_and_wait]
N --> P[Async Event Stream]
O --> Q[Direct Response]
P --> R[Return to OpenWebUI]
Q --> R
```
### Step-by-Step Breakdown
#### 1. Environment Setup (`_setup_env`)
```python
def _setup_env(self, __event_call__=None):
"""
Priority:
1. Check VALVES.CLI_PATH
2. Search system PATH
3. Auto-install via curl (if not found)
4. Set GH_TOKEN environment variables
"""
```
**Actions:**
- Locate Copilot CLI binary
- Set `COPILOT_CLI_PATH` environment variable
- Configure `GH_TOKEN` for authentication
- Apply custom environment variables
#### 2. Model Selection
```python
# Input: body["model"] = "copilotsdk-claude-sonnet-4.5"
request_model = body.get("model", "")
if request_model.startswith(f"{self.id}-"):
real_model_id = request_model[len(f"{self.id}-"):] # "claude-sonnet-4.5"
```
#### 3. Chat Context Extraction (`_get_chat_context`)
```python
# Priority order for chat_id:
# 1. __metadata__ (most reliable)
# 2. body["chat_id"]
# 3. body["metadata"]["chat_id"]
chat_ctx = self._get_chat_context(body, __metadata__, __event_call__)
chat_id = chat_ctx.get("chat_id")
```
#### 4. System Prompt Extraction (`_extract_system_prompt`)
Multi-source fallback strategy:
1. `metadata.model.params.system`
2. Model database lookup (by model_id)
3. `body.params.system`
4. Messages with `role="system"`
#### 5. Session Creation/Resumption
**New Session:**
```python
session_config = SessionConfig(
session_id=chat_id,
model=real_model_id,
streaming=is_streaming,
tools=custom_tools,
system_message={"mode": "append", "content": system_prompt_content},
infinite_sessions=InfiniteSessionConfig(
enabled=True,
background_compaction_threshold=0.8,
buffer_exhaustion_threshold=0.95
)
)
session = await client.create_session(config=session_config)
```
**Resume Session:**
```python
try:
session = await client.resume_session(chat_id)
# Session state preserved: history, tools, workspace
except Exception:
# Fallback to creating new session
```
---
## Session Management
### Infinite Sessions Architecture
```
┌─────────────────────────────────────────────────────────┐
│ Session Lifecycle │
│ │
│ ┌──────────┐ create ┌──────────┐ resume ┌───────┴───┐
│ │ Chat ID │─────────▶ │ Session │ ◀────────│ OpenWebUI │
│ └──────────┘ │ State │ └───────────┘
│ └─────┬────┘ │
│ │ │
│ ▼ │
│ ┌─────────────────────────────────────────────────────┐ │
│ │ Context Window Management │ │
│ │ ┌──────────────────────────────────────────────┐ │ │
│ │ │ Messages [user, assistant, tool_results...] │ │ │
│ │ │ Token Usage: ████████████░░░░ (80%) │ │ │
│ │ └──────────────────────────────────────────────┘ │ │
│ │ │ │ │
│ │ ▼ │ │
│ │ ┌──────────────────────────────────────────────┐ │ │
│ │ │ Threshold Reached (0.8) │ │ │
│ │ │ → Background Compaction Triggered │ │ │
│ │ └──────────────────────────────────────────────┘ │ │
│ │ │ │ │
│ │ ▼ │ │
│ │ ┌──────────────────────────────────────────────┐ │ │
│ │ │ Compacted Summary + Recent Messages │ │ │
│ │ │ Token Usage: ██████░░░░░░░░░░░ (40%) │ │ │
│ │ └──────────────────────────────────────────────┘ │ │
│ └─────────────────────────────────────────────────────┘ │
└─────────────────────────────────────────────────────────┘
```
### Configuration Parameters
```python
InfiniteSessionConfig(
enabled=True, # Enable infinite sessions
background_compaction_threshold=0.8, # Start compaction at 80% token usage
buffer_exhaustion_threshold=0.95 # Emergency threshold at 95%
)
```
**Behavior:**
- **< 80%**: Normal operation, no compaction
- **80-95%**: Background compaction (summarize older messages)
- **> 95%**: Force compaction before next request
---
## Streaming Response Handling
### Event-Driven Architecture
```python
async def stream_response(
self, client, session, send_payload, init_message: str = "", __event_call__=None
) -> AsyncGenerator:
"""
Asynchronous event processing with queue-based buffering.
Flow:
1. Start async send task
2. Register event handler
3. Process events via queue
4. Yield chunks to OpenWebUI
5. Clean up resources
"""
```
### Event Processing Pipeline
```
┌────────────────────────────────────────────────────────────┐
│ Copilot SDK Event Stream │
└────────────────────┬───────────────────────────────────────┘
┌────────────────────────┐
│ Event Handler │
│ (Sync Callback) │
└────────┬───────────────┘
┌────────────────────────┐
│ Async Queue │
│ (Thread-safe) │
└────────┬───────────────┘
┌────────────────────────┐
│ Consumer Loop │
│ (async for) │
└────────┬───────────────┘
┌────────────────────────┐
│ yield to OpenWebUI │
└────────────────────────┘
```
### State Management During Streaming
```python
state = {
"thinking_started": False, # <think> tags opened
"content_sent": False # Main content has started
}
active_tools = {} # Track concurrent tool executions
```
**State Transitions:**
1. `reasoning_delta` arrives → `thinking_started = True` → Output: `<think>\n{reasoning}`
2. `message_delta` arrives → Close `</think>` if open → `content_sent = True` → Output: `{content}`
3. `tool.execution_start` → Output tool indicator (inside/outside `<think>`)
4. `session.complete` → Finalize stream
---
## Event Processing Mechanism
### Event Type Reference
Following official SDK patterns (from `copilot.SessionEventType`):
| Event Type | Description | Key Data Fields | Handler Action |
|-----------|-------------|-----------------|----------------|
| `assistant.message_delta` | Main content streaming | `delta_content` | Yield text chunk |
| `assistant.reasoning_delta` | Chain-of-thought | `delta_content` | Wrap in `<think>` tags |
| `tool.execution_start` | Tool call initiated | `name`, `tool_call_id` | Display tool indicator |
| `tool.execution_complete` | Tool finished | `result.content` | Show completion status |
| `session.compaction_start` | Context compaction begins | - | Log debug info |
| `session.compaction_complete` | Compaction done | - | Log debug info |
| `session.error` | Error occurred | `error`, `message` | Emit error notification |
### Event Handler Implementation
```python
def handler(event):
"""Process streaming events following official SDK patterns."""
event_type = get_event_type(event) # Handle enum/string types
# Extract data using safe_get_data_attr (handles dict/object)
if event_type == "assistant.message_delta":
delta = safe_get_data_attr(event, "delta_content")
if delta:
queue.put_nowait(delta) # Thread-safe enqueue
```
### Official SDK Pattern Compliance
```python
def safe_get_data_attr(event, attr: str, default=None):
"""
Official pattern: event.data.delta_content
Handles both dict and object access patterns.
"""
if not hasattr(event, "data") or event.data is None:
return default
data = event.data
# Dict access (JSON-like)
if isinstance(data, dict):
return data.get(attr, default)
# Object attribute (Python SDK)
return getattr(data, attr, default)
```
---
## Tool Execution Flow
### Tool Registration
```python
# 1. Define tool at module level
@define_tool(description="Generate a random integer within a specified range.")
async def generate_random_number(params: RandomNumberParams) -> str:
number = random.randint(params.min, params.max)
return f"Generated random number: {number}"
# 2. Register in _initialize_custom_tools
def _initialize_custom_tools(self):
if not self.valves.ENABLE_TOOLS:
return []
all_tools = {
"generate_random_number": generate_random_number,
}
# Filter based on AVAILABLE_TOOLS valve
if self.valves.AVAILABLE_TOOLS == "all":
return list(all_tools.values())
enabled = [t.strip() for t in self.valves.AVAILABLE_TOOLS.split(",")]
return [all_tools[name] for name in enabled if name in all_tools]
```
### Tool Execution Timeline
```
User Message: "Generate a random number between 1 and 100"
Model Decision: Use tool `generate_random_number`
Event: tool.execution_start
│ → Display: "🔧 Running Tool: generate_random_number"
Tool Function Execution (async)
Event: tool.execution_complete
│ → Result: "Generated random number: 42"
│ → Display: "✅ Tool Completed: 42"
Model generates response using tool result
Event: assistant.message_delta
│ → "I generated the number 42 for you."
Stream Complete
```
### Visual Indicators
**Before Content:**
```markdown
<think>
Running Tool: generate_random_number...
Tool `generate_random_number` Completed. Result: 42
</think>
I generated the number 42 for you.
```
**After Content Started:**
```markdown
The number is
> 🔧 **Running Tool**: `generate_random_number`
> ✅ **Tool Completed**: 42
actually 42.
```
---
## System Prompt Extraction
### Multi-Source Priority System
```python
async def _extract_system_prompt(self, body, messages, request_model, real_model_id):
"""
Priority order:
1. metadata.model.params.system (highest)
2. Model database lookup
3. body.params.system
4. messages[role="system"] (fallback)
"""
```
### Source 1: Metadata Model Params
```python
# OpenWebUI injects model configuration
metadata = body.get("metadata", {})
meta_model = metadata.get("model", {})
meta_params = meta_model.get("params", {})
system_prompt = meta_params.get("system") # Priority 1
```
### Source 2: Model Database
```python
from open_webui.models.models import Models
# Try multiple model ID variations
model_ids_to_try = [
request_model, # "copilotsdk-claude-sonnet-4.5"
request_model.removeprefix(...), # "claude-sonnet-4.5"
real_model_id, # From valves
]
for mid in model_ids_to_try:
model_record = Models.get_model_by_id(mid)
if model_record and hasattr(model_record, "params"):
system_prompt = model_record.params.get("system")
if system_prompt:
break
```
### Source 3: Body Params
```python
body_params = body.get("params", {})
system_prompt = body_params.get("system")
```
### Source 4: System Message
```python
for msg in messages:
if msg.get("role") == "system":
system_prompt = self._extract_text_from_content(msg.get("content"))
break
```
### Configuration in SessionConfig
```python
system_message_config = {
"mode": "append", # Append to conversation context
"content": system_prompt_content
}
session_config = SessionConfig(
system_message=system_message_config,
# ... other params
)
```
---
## Configuration Parameters
### Valve Definitions
| Parameter | Type | Default | Description |
|-----------|------|---------|-------------|
| `GH_TOKEN` | str | `""` | GitHub Fine-grained Token (requires 'Copilot Requests' permission) |
| `MODEL_ID` | str | `"claude-sonnet-4.5"` | Default model when dynamic fetching fails |
| `CLI_PATH` | str | `"/usr/local/bin/copilot"` | Path to Copilot CLI binary |
| `DEBUG` | bool | `False` | Enable frontend console debug logging |
| `LOG_LEVEL` | str | `"error"` | CLI log level: none, error, warning, info, debug, all |
| `SHOW_THINKING` | bool | `True` | Display model reasoning in `<think>` tags |
| `SHOW_WORKSPACE_INFO` | bool | `True` | Show session workspace path in debug mode |
| `EXCLUDE_KEYWORDS` | str | `""` | Comma-separated keywords to exclude models |
| `WORKSPACE_DIR` | str | `""` | Restricted workspace directory (empty = process cwd) |
| `INFINITE_SESSION` | bool | `True` | Enable automatic context compaction |
| `COMPACTION_THRESHOLD` | float | `0.8` | Background compaction at 80% token usage |
| `BUFFER_THRESHOLD` | float | `0.95` | Emergency threshold at 95% |
| `TIMEOUT` | int | `300` | Stream chunk timeout (seconds) |
| `CUSTOM_ENV_VARS` | str | `""` | JSON string of custom environment variables |
| `ENABLE_TOOLS` | bool | `False` | Enable custom tool system |
| `AVAILABLE_TOOLS` | str | `"all"` | Available tools: "all" or comma-separated list |
### Environment Variables
```bash
# Set by _setup_env
export COPILOT_CLI_PATH="/usr/local/bin/copilot"
export GH_TOKEN="ghp_xxxxxxxxxxxxxxxxxxxx"
export GITHUB_TOKEN="ghp_xxxxxxxxxxxxxxxxxxxx"
# Custom variables (from CUSTOM_ENV_VARS valve)
export CUSTOM_VAR_1="value1"
export CUSTOM_VAR_2="value2"
```
---
## Key Functions Reference
### Entry Points
#### `pipe(body, __metadata__, __event_emitter__, __event_call__)`
- **Purpose**: OpenWebUI stable entry point
- **Returns**: Delegates to `_pipe_impl`
#### `_pipe_impl(body, __metadata__, __event_emitter__, __event_call__)`
- **Purpose**: Main request processing logic
- **Flow**: Setup → Extract → Session → Response
- **Returns**: `str` (non-streaming) or `AsyncGenerator` (streaming)
#### `pipes()`
- **Purpose**: Dynamic model list fetching
- **Returns**: List of available models with multiplier info
- **Caching**: Uses `_model_cache` to avoid repeated API calls
### Session Management
#### `_build_session_config(chat_id, real_model_id, custom_tools, system_prompt_content, is_streaming)`
- **Purpose**: Construct SessionConfig object
- **Returns**: `SessionConfig` with infinite sessions and tools
#### `_get_chat_context(body, __metadata__, __event_call__)`
- **Purpose**: Extract chat_id with priority fallback
- **Returns**: `{"chat_id": str}`
### Streaming
#### `stream_response(client, session, send_payload, init_message, __event_call__)`
- **Purpose**: Async streaming event processor
- **Yields**: Text chunks to OpenWebUI
- **Resources**: Auto-cleanup client and session
#### `handler(event)`
- **Purpose**: Sync event callback (inside `stream_response`)
- **Action**: Parse event → Enqueue chunks → Update state
### Helpers
#### `_emit_debug_log(message, __event_call__)`
- **Purpose**: Send debug logs to frontend console
- **Condition**: Only when `DEBUG=True`
#### `_setup_env(__event_call__)`
- **Purpose**: Locate CLI, set environment variables
- **Side Effects**: Modifies `os.environ`
#### `_extract_system_prompt(body, messages, request_model, real_model_id, __event_call__)`
- **Purpose**: Multi-source system prompt extraction
- **Returns**: `(system_prompt_content, source_name)`
#### `_process_images(messages, __event_call__)`
- **Purpose**: Extract text and images from multimodal messages
- **Returns**: `(text_content, attachments_list)`
#### `_initialize_custom_tools()`
- **Purpose**: Register and filter custom tools
- **Returns**: List of tool functions
### Utility Functions
#### `get_event_type(event) -> str`
- **Purpose**: Extract event type string from enum/string
- **Handles**: `SessionEventType` enum → `.value` extraction
#### `safe_get_data_attr(event, attr: str, default=None)`
- **Purpose**: Safe attribute extraction from event.data
- **Handles**: Both dict access and object attribute access
---
## Troubleshooting Guide
### Enable Debug Mode
```python
# In OpenWebUI Valves UI:
DEBUG = True
SHOW_WORKSPACE_INFO = True
LOG_LEVEL = "debug"
```
### Debug Output Location
**Frontend Console:**
```javascript
// Open browser DevTools (F12)
// Look for logs with prefix: [Copilot Pipe]
console.debug("[Copilot Pipe] Extracted ChatID: abc123 (Source: __metadata__)")
```
**Backend Logs:**
```python
# Python logging output
logger.debug(f"[Copilot Pipe] Session resumed: {chat_id}")
```
### Common Issues
#### 1. Session Not Resuming
**Symptom:** New session created every request
**Causes:**
- `chat_id` not extracted correctly
- Session expired on Copilot side
- `INFINITE_SESSION=False` (sessions not persistent)
**Solution:**
```python
# Check debug logs for:
"Extracted ChatID: <id> (Source: ...)"
"Session <id> not found (...), creating new."
```
#### 2. System Prompt Not Applied
**Symptom:** Model ignores configured system prompt
**Causes:**
- Not found in any of 4 sources
- Session resumed (system prompt only set on creation)
**Solution:**
```python
# Check debug logs for:
"Extracted system prompt from <source> (length: X)"
"Configured system message (mode: append)"
```
#### 3. Tools Not Available
**Symptom:** Model can't use custom tools
**Causes:**
- `ENABLE_TOOLS=False`
- Tool not registered in `_initialize_custom_tools`
- Wrong `AVAILABLE_TOOLS` filter
**Solution:**
```python
# Check debug logs for:
"Enabled X custom tools: ['tool1', 'tool2']"
```
---
## Performance Optimization
### Model List Caching
```python
# First request: Fetch from API
models = await client.list_models()
self._model_cache = [...] # Cache result
# Subsequent requests: Use cache
if self._model_cache:
return self._model_cache
```
### Session Persistence
**Impact:** Eliminates redundant model initialization on every request
```python
# Without session:
# Each request: Initialize model → Load context → Generate → Discard
# With session (chat_id):
# First request: Initialize model → Load context → Generate → Save
# Later: Resume → Generate (instant)
```
### Streaming vs Non-streaming
**Streaming:**
- Lower perceived latency (first token faster)
- Better UX for long responses
- Resource cleanup via generator exit
**Non-streaming:**
- Simpler error handling
- Atomic response (no partial output)
- Use for short responses
---
## Security Considerations
### Token Protection
```python
# ❌ Never log tokens
logger.debug(f"Token: {self.valves.GH_TOKEN}") # DON'T DO THIS
# ✅ Mask sensitive data
logger.debug(f"Token configured: {'*' * 10}")
```
### Workspace Isolation
```python
# Set WORKSPACE_DIR to restrict file access
WORKSPACE_DIR = "/safe/sandbox/path"
# Copilot CLI respects this directory
client_config["cwd"] = WORKSPACE_DIR
```
### Input Validation
```python
# Validate chat_id format
if chat_id and not re.match(r'^[a-zA-Z0-9_-]+$', chat_id):
logger.warning(f"Invalid chat_id format: {chat_id}")
chat_id = None
```
---
## Future Enhancements
### Planned Features
1. **Multi-Session Management**: Support multiple parallel sessions per user
2. **Session Analytics**: Track token usage, compaction frequency
3. **Tool Result Caching**: Avoid redundant tool calls
4. **Custom Event Filters**: User-configurable event handling
5. **Workspace Templates**: Pre-configured workspace environments
6. **Streaming Abort**: Graceful cancellation of long-running requests
### API Evolution
Monitoring Copilot SDK updates for:
- New event types (e.g., `assistant.function_call`)
- Enhanced tool capabilities
- Improved session serialization
---
## References
- [GitHub Copilot SDK Documentation](https://github.com/github/copilot-sdk)
- [OpenWebUI Pipe Development](https://docs.openwebui.com/)
- [Awesome OpenWebUI Project](https://github.com/Fu-Jie/awesome-openwebui)
---
**License:** MIT
**Maintainer:** Fu-Jie ([@Fu-Jie](https://github.com/Fu-Jie))

View File

@@ -0,0 +1,835 @@
# GitHub Copilot SDK 集成工作流程
**作者:** Fu-Jie
**版本:** 0.2.3
**最后更新:** 2026-01-27
---
## 目录
1. [架构概览](#架构概览)
2. [请求处理流程](#请求处理流程)
3. [会话管理](#会话管理)
4. [流式响应处理](#流式响应处理)
5. [事件处理机制](#事件处理机制)
6. [工具执行流程](#工具执行流程)
7. [系统提示词提取](#系统提示词提取)
8. [配置参数](#配置参数)
9. [核心函数参考](#核心函数参考)
---
## 架构概览
### 组件图
```
┌─────────────────────────────────────────────────────────────┐
│ OpenWebUI │
│ ┌───────────────────────────────────────────────────────┐ │
│ │ Pipe 接口 (入口点) │ │
│ └─────────────────────┬─────────────────────────────────┘ │
│ │ │
│ ▼ │
│ ┌───────────────────────────────────────────────────────┐ │
│ │ _pipe_impl (主逻辑) │ │
│ │ ┌──────────────────────────────────────────────────┐ │ │
│ │ │ 1. 环境设置 (_setup_env) │ │ │
│ │ │ 2. 模型选择 (request_model 解析) │ │ │
│ │ │ 3. 聊天上下文提取 │ │ │
│ │ │ 4. 系统提示词提取 │ │ │
│ │ │ 5. 会话管理 (创建/恢复) │ │ │
│ │ │ 6. 流式/非流式响应 │ │ │
│ │ └──────────────────────────────────────────────────┘ │ │
│ └─────────────────────┬─────────────────────────────────┘ │
│ │ │
│ ▼ │
│ ┌───────────────────────────────────────────────────────┐ │
│ │ GitHub Copilot 客户端 │ │
│ │ ┌──────────────────────────────────────────────────┐ │ │
│ │ │ • CopilotClient (SDK 实例) │ │ │
│ │ │ • Session (对话上下文) │ │ │
│ │ │ • Event Stream (异步事件流) │ │ │
│ │ └──────────────────────────────────────────────────┘ │ │
│ └─────────────────────┬─────────────────────────────────┘ │
│ │ │
└────────────────────────┼─────────────────────────────────────┘
┌──────────────────────┐
│ Copilot CLI 进程 │
│ (后端代理) │
└──────────────────────┘
```
### 核心组件
1. **Pipe 接口**OpenWebUI 的标准入口点
2. **环境管理器**CLI 设置、令牌验证、环境变量
3. **会话管理器**:持久化对话状态,自动压缩
4. **事件处理器**:异步流式事件处理器
5. **工具系统**:自定义工具注册和执行
6. **调试日志器**:前端控制台日志,用于故障排除
---
## 请求处理流程
### 完整请求生命周期
```mermaid
graph TD
A[OpenWebUI 请求] --> B[pipe 入口点]
B --> C[_pipe_impl]
C --> D{设置环境}
D --> E[解析模型 ID]
E --> F[提取聊天上下文]
F --> G[提取系统提示词]
G --> H{会话存在?}
H -->|是| I[恢复会话]
H -->|否| J[创建新会话]
I --> K[初始化工具]
J --> K
K --> L[处理图片]
L --> M{流式模式?}
M -->|是| N[stream_response]
M -->|否| O[send_and_wait]
N --> P[异步事件流]
O --> Q[直接响应]
P --> R[返回到 OpenWebUI]
Q --> R
```
### 逐步分解
#### 1. 环境设置 (`_setup_env`)
```python
def _setup_env(self, __event_call__=None):
"""
优先级:
1. 检查 VALVES.CLI_PATH
2. 搜索系统 PATH
3. 自动通过 curl 安装(如果未找到)
4. 设置 GH_TOKEN 环境变量
"""
```
**操作:**
- 定位 Copilot CLI 二进制文件
- 设置 `COPILOT_CLI_PATH` 环境变量
- 配置 `GH_TOKEN` 进行身份验证
- 应用自定义环境变量
#### 2. 模型选择
```python
# 输入body["model"] = "copilotsdk-claude-sonnet-4.5"
request_model = body.get("model", "")
if request_model.startswith(f"{self.id}-"):
real_model_id = request_model[len(f"{self.id}-"):] # "claude-sonnet-4.5"
```
#### 3. 聊天上下文提取 (`_get_chat_context`)
```python
# chat_id 的优先级顺序:
# 1. __metadata__最可靠
# 2. body["chat_id"]
# 3. body["metadata"]["chat_id"]
chat_ctx = self._get_chat_context(body, __metadata__, __event_call__)
chat_id = chat_ctx.get("chat_id")
```
#### 4. 系统提示词提取 (`_extract_system_prompt`)
多源回退策略:
1. `metadata.model.params.system`
2. 模型数据库查询(按 model_id
3. `body.params.system`
4. 包含 `role="system"` 的消息
#### 5. 会话创建/恢复
**新会话:**
```python
session_config = SessionConfig(
session_id=chat_id,
model=real_model_id,
streaming=is_streaming,
tools=custom_tools,
system_message={"mode": "append", "content": system_prompt_content},
infinite_sessions=InfiniteSessionConfig(
enabled=True,
background_compaction_threshold=0.8,
buffer_exhaustion_threshold=0.95
)
)
session = await client.create_session(config=session_config)
```
**恢复会话:**
```python
try:
session = await client.resume_session(chat_id)
# 会话状态保留:历史、工具、工作区
except Exception:
# 回退到创建新会话
```
---
## 会话管理
### 无限会话架构
```
┌─────────────────────────────────────────────────────────┐
│ 会话生命周期 │
│ │
│ ┌──────────┐ 创建 ┌──────────┐ 恢复 ┌───────────┐ │
│ │ Chat ID │─────▶ │ Session │ ◀────────│ OpenWebUI │ │
│ └──────────┘ │ State │ └───────────┘ │
│ └─────┬────┘ │
│ │ │
│ ▼ │
│ ┌─────────────────────────────────────────────────┐ │
│ │ 上下文窗口管理 │ │
│ │ ┌──────────────────────────────────────────┐ │ │
│ │ │ 消息 [user, assistant, tool_results...] │ │ │
│ │ │ Token 使用率: ████████████░░░░ (80%) │ │ │
│ │ └──────────────────────────────────────────┘ │ │
│ │ │ │ │
│ │ ▼ │ │
│ │ ┌──────────────────────────────────────────┐ │ │
│ │ │ 达到阈值 (0.8) │ │ │
│ │ │ → 后台压缩触发 │ │ │
│ │ └──────────────────────────────────────────┘ │ │
│ │ │ │ │
│ │ ▼ │ │
│ │ ┌──────────────────────────────────────────┐ │ │
│ │ │ 压缩摘要 + 最近消息 │ │ │
│ │ │ Token 使用率: ██████░░░░░░░░░░░ (40%) │ │ │
│ │ └──────────────────────────────────────────┘ │ │
│ └─────────────────────────────────────────────────┘ │
└─────────────────────────────────────────────────────────┘
```
### 配置参数
```python
InfiniteSessionConfig(
enabled=True, # 启用无限会话
background_compaction_threshold=0.8, # 在 80% token 使用率时开始压缩
buffer_exhaustion_threshold=0.95 # 95% 紧急阈值
)
```
**行为:**
- **< 80%**:正常操作,无压缩
- **80-95%**:后台压缩(总结旧消息)
- **> 95%**:在下一个请求前强制压缩
---
## 流式响应处理
### 事件驱动架构
```python
async def stream_response(
self, client, session, send_payload, init_message: str = "", __event_call__=None
) -> AsyncGenerator:
"""
使用基于队列的缓冲进行异步事件处理。
流程:
1. 启动异步发送任务
2. 注册事件处理器
3. 通过队列处理事件
4. 向 OpenWebUI 产出块
5. 清理资源
"""
```
### 事件处理管道
```
┌────────────────────────────────────────────────────────────┐
│ Copilot SDK 事件流 │
└────────────────────┬───────────────────────────────────────┘
┌────────────────────────┐
│ 事件处理器 │
│ (同步回调) │
└────────┬───────────────┘
┌────────────────────────┐
│ 异步队列 │
│ (线程安全) │
└────────┬───────────────┘
┌────────────────────────┐
│ 消费者循环 │
│ (async for) │
└────────┬───────────────┘
┌────────────────────────┐
│ yield 到 OpenWebUI │
└────────────────────────┘
```
### 流式传输期间的状态管理
```python
state = {
"thinking_started": False, # <think> 标签已打开
"content_sent": False # 主内容已开始
}
active_tools = {} # 跟踪并发工具执行
```
**状态转换:**
1. `reasoning_delta` 到达 → `thinking_started = True` → 输出:`<think>\n{reasoning}`
2. `message_delta` 到达 → 如果打开则关闭 `</think>``content_sent = True` → 输出:`{content}`
3. `tool.execution_start` → 输出工具指示器(在 `<think>` 内部/外部)
4. `session.complete` → 完成流
---
## 事件处理机制
### 事件类型参考
遵循官方 SDK 模式(来自 `copilot.SessionEventType`
| 事件类型 | 描述 | 关键数据字段 | 处理器操作 |
|---------|------|-------------|-----------|
| `assistant.message_delta` | 主内容流式传输 | `delta_content` | 产出文本块 |
| `assistant.reasoning_delta` | 思维链 | `delta_content` | 用 `<think>` 标签包装 |
| `tool.execution_start` | 工具调用启动 | `name`, `tool_call_id` | 显示工具指示器 |
| `tool.execution_complete` | 工具完成 | `result.content` | 显示完成状态 |
| `session.compaction_start` | 上下文压缩开始 | - | 记录调试信息 |
| `session.compaction_complete` | 压缩完成 | - | 记录调试信息 |
| `session.error` | 发生错误 | `error`, `message` | 发出错误通知 |
### 事件处理器实现
```python
def handler(event):
"""遵循官方 SDK 模式处理流式事件。"""
event_type = get_event_type(event) # 处理枚举/字符串类型
# 使用 safe_get_data_attr 提取数据(处理 dict/object
if event_type == "assistant.message_delta":
delta = safe_get_data_attr(event, "delta_content")
if delta:
queue.put_nowait(delta) # 线程安全入队
```
### 官方 SDK 模式合规性
```python
def safe_get_data_attr(event, attr: str, default=None):
"""
官方模式event.data.delta_content
处理 dict 和对象访问模式。
"""
if not hasattr(event, "data") or event.data is None:
return default
data = event.data
# Dict 访问(类似 JSON
if isinstance(data, dict):
return data.get(attr, default)
# 对象属性Python SDK
return getattr(data, attr, default)
```
---
## 工具执行流程
### 工具注册
```python
# 1. 在模块级别定义工具
@define_tool(description="在指定范围内生成随机整数。")
async def generate_random_number(params: RandomNumberParams) -> str:
number = random.randint(params.min, params.max)
return f"生成的随机数: {number}"
# 2. 在 _initialize_custom_tools 中注册
def _initialize_custom_tools(self):
if not self.valves.ENABLE_TOOLS:
return []
all_tools = {
"generate_random_number": generate_random_number,
}
# 根据 AVAILABLE_TOOLS valve 过滤
if self.valves.AVAILABLE_TOOLS == "all":
return list(all_tools.values())
enabled = [t.strip() for t in self.valves.AVAILABLE_TOOLS.split(",")]
return [all_tools[name] for name in enabled if name in all_tools]
```
### 工具执行时间线
```
用户消息:生成一个 1 到 100 之间的随机数
模型决策:使用工具 `generate_random_number`
事件tool.execution_start
│ → 显示:"🔧 运行工具generate_random_number"
工具函数执行(异步)
事件tool.execution_complete
│ → 结果:"生成的随机数42"
│ → 显示:"✅ 工具完成42"
模型使用工具结果生成响应
事件assistant.message_delta
│ → "我为你生成了数字 42。"
流完成
```
### 视觉指示器
**内容前:**
```markdown
<think>
运行工具generate_random_number...
工具 `generate_random_number` 完成。结果42
</think>
我为你生成了数字 42。
```
**内容开始后:**
```markdown
数字是
> 🔧 **运行工具**`generate_random_number`
> ✅ **工具完成**42
实际上是 42。
```
---
## 系统提示词提取
### 多源优先级系统
```python
async def _extract_system_prompt(self, body, messages, request_model, real_model_id):
"""
优先级顺序:
1. metadata.model.params.system最高
2. 模型数据库查询
3. body.params.system
4. messages[role="system"](回退)
"""
```
### 来源 1元数据模型参数
```python
# OpenWebUI 注入模型配置
metadata = body.get("metadata", {})
meta_model = metadata.get("model", {})
meta_params = meta_model.get("params", {})
system_prompt = meta_params.get("system") # 优先级 1
```
### 来源 2模型数据库
```python
from open_webui.models.models import Models
# 尝试多个模型 ID 变体
model_ids_to_try = [
request_model, # "copilotsdk-claude-sonnet-4.5"
request_model.removeprefix(...), # "claude-sonnet-4.5"
real_model_id, # 来自 valves
]
for mid in model_ids_to_try:
model_record = Models.get_model_by_id(mid)
if model_record and hasattr(model_record, "params"):
system_prompt = model_record.params.get("system")
if system_prompt:
break
```
### 来源 3Body 参数
```python
body_params = body.get("params", {})
system_prompt = body_params.get("system")
```
### 来源 4系统消息
```python
for msg in messages:
if msg.get("role") == "system":
system_prompt = self._extract_text_from_content(msg.get("content"))
break
```
### SessionConfig 中的配置
```python
system_message_config = {
"mode": "append", # 追加到对话上下文
"content": system_prompt_content
}
session_config = SessionConfig(
system_message=system_message_config,
# ... 其他参数
)
```
---
## 配置参数
### Valve 定义
| 参数 | 类型 | 默认值 | 描述 |
|-----|------|--------|------|
| `GH_TOKEN` | str | `""` | GitHub 精细化令牌(需要 'Copilot Requests' 权限) |
| `MODEL_ID` | str | `"claude-sonnet-4.5"` | 动态获取失败时的默认模型 |
| `CLI_PATH` | str | `"/usr/local/bin/copilot"` | Copilot CLI 二进制文件路径 |
| `DEBUG` | bool | `False` | 启用前端控制台调试日志 |
| `LOG_LEVEL` | str | `"error"` | CLI 日志级别none、error、warning、info、debug、all |
| `SHOW_THINKING` | bool | `True` | 在 `<think>` 标签中显示模型推理 |
| `SHOW_WORKSPACE_INFO` | bool | `True` | 在调试模式下显示会话工作区路径 |
| `EXCLUDE_KEYWORDS` | str | `""` | 逗号分隔的关键字,用于排除模型 |
| `WORKSPACE_DIR` | str | `""` | 限制的工作区目录(空 = 进程 cwd |
| `INFINITE_SESSION` | bool | `True` | 启用自动上下文压缩 |
| `COMPACTION_THRESHOLD` | float | `0.8` | 80% token 使用率时后台压缩 |
| `BUFFER_THRESHOLD` | float | `0.95` | 95% 紧急阈值 |
| `TIMEOUT` | int | `300` | 流块超时(秒) |
| `CUSTOM_ENV_VARS` | str | `""` | 自定义环境变量的 JSON 字符串 |
| `ENABLE_TOOLS` | bool | `False` | 启用自定义工具系统 |
| `AVAILABLE_TOOLS` | str | `"all"` | 可用工具:"all" 或逗号分隔列表 |
### 环境变量
```bash
# 由 _setup_env 设置
export COPILOT_CLI_PATH="/usr/local/bin/copilot"
export GH_TOKEN="ghp_xxxxxxxxxxxxxxxxxxxx"
export GITHUB_TOKEN="ghp_xxxxxxxxxxxxxxxxxxxx"
# 自定义变量(来自 CUSTOM_ENV_VARS valve
export CUSTOM_VAR_1="value1"
export CUSTOM_VAR_2="value2"
```
---
## 核心函数参考
### 入口点
#### `pipe(body, __metadata__, __event_emitter__, __event_call__)`
- **目的**OpenWebUI 稳定入口点
- **返回**:委托给 `_pipe_impl`
#### `_pipe_impl(body, __metadata__, __event_emitter__, __event_call__)`
- **目的**:主请求处理逻辑
- **流程**:设置 → 提取 → 会话 → 响应
- **返回**`str`(非流式)或 `AsyncGenerator`(流式)
#### `pipes()`
- **目的**:动态模型列表获取
- **返回**:带有倍数信息的可用模型列表
- **缓存**:使用 `_model_cache` 避免重复 API 调用
### 会话管理
#### `_build_session_config(chat_id, real_model_id, custom_tools, system_prompt_content, is_streaming)`
- **目的**:构建 SessionConfig 对象
- **返回**:带有无限会话和工具的 `SessionConfig`
#### `_get_chat_context(body, __metadata__, __event_call__)`
- **目的**:使用优先级回退提取 chat_id
- **返回**`{"chat_id": str}`
### 流式传输
#### `stream_response(client, session, send_payload, init_message, __event_call__)`
- **目的**:异步流式事件处理器
- **产出**:文本块到 OpenWebUI
- **资源**:自动清理客户端和会话
#### `handler(event)`
- **目的**:同步事件回调(在 `stream_response` 内)
- **操作**:解析事件 → 入队块 → 更新状态
### 辅助函数
#### `_emit_debug_log(message, __event_call__)`
- **目的**:将调试日志发送到前端控制台
- **条件**:仅当 `DEBUG=True`
#### `_setup_env(__event_call__)`
- **目的**:定位 CLI设置环境变量
- **副作用**:修改 `os.environ`
#### `_extract_system_prompt(body, messages, request_model, real_model_id, __event_call__)`
- **目的**:多源系统提示词提取
- **返回**`(system_prompt_content, source_name)`
#### `_process_images(messages, __event_call__)`
- **目的**:从多模态消息中提取文本和图片
- **返回**`(text_content, attachments_list)`
#### `_initialize_custom_tools()`
- **目的**:注册和过滤自定义工具
- **返回**:工具函数列表
### 实用函数
#### `get_event_type(event) -> str`
- **目的**:从枚举/字符串提取事件类型字符串
- **处理**`SessionEventType` 枚举 → `.value` 提取
#### `safe_get_data_attr(event, attr: str, default=None)`
- **目的**:从 event.data 安全提取属性
- **处理**dict 访问和对象属性访问
---
## 故障排除指南
### 启用调试模式
```python
# 在 OpenWebUI Valves UI 中:
DEBUG = True
SHOW_WORKSPACE_INFO = True
LOG_LEVEL = "debug"
```
### 调试输出位置
**前端控制台:**
```javascript
// 打开浏览器开发工具 (F12)
// 查找前缀为 [Copilot Pipe] 的日志
console.debug("[Copilot Pipe] 提取的 ChatIDabc123来源__metadata__")
```
**后端日志:**
```python
# Python 日志输出
logger.debug(f"[Copilot Pipe] 会话已恢复:{chat_id}")
```
### 常见问题
#### 1. 会话未恢复
**症状**:每次请求都创建新会话
**原因**
- `chat_id` 提取不正确
- Copilot 端会话过期
- `INFINITE_SESSION=False`(会话不持久)
**解决方案**
```python
# 检查调试日志中的:
"提取的 ChatID<id>(来源:..."
"会话 <id> 未找到(...),正在创建新会话。"
```
#### 2. 系统提示词未应用
**症状**:模型忽略配置的系统提示词
**原因**
- 在 4 个来源中均未找到
- 会话已恢复(系统提示词仅在创建时设置)
**解决方案**
```python
# 检查调试日志中的:
"从 <source> 提取系统提示词长度X"
"配置系统消息模式append"
```
#### 3. 工具不可用
**症状**:模型无法使用自定义工具
**原因**
- `ENABLE_TOOLS=False`
- 工具未在 `_initialize_custom_tools` 中注册
- 错误的 `AVAILABLE_TOOLS` 过滤器
**解决方案**
```python
# 检查调试日志中的:
"已启用 X 个自定义工具:['tool1', 'tool2']"
```
---
## 性能优化
### 模型列表缓存
```python
# 第一次请求:从 API 获取
models = await client.list_models()
self._model_cache = [...] # 缓存结果
# 后续请求:使用缓存
if self._model_cache:
return self._model_cache
```
### 会话持久化
**影响**:消除每次请求的冗余模型初始化
```python
# 没有会话:
# 每次请求:初始化模型 → 加载上下文 → 生成 → 丢弃
# 有会话chat_id
# 第一次请求:初始化模型 → 加载上下文 → 生成 → 保存
# 后续:恢复 → 生成(即时)
```
### 流式 vs 非流式
**流式:**
- 降低感知延迟(首个 token 更快)
- 长响应的更好用户体验
- 通过生成器退出进行资源清理
**非流式:**
- 更简单的错误处理
- 原子响应(无部分输出)
- 用于短响应
---
## 安全考虑
### 令牌保护
```python
# ❌ 永远不要记录令牌
logger.debug(f"令牌:{self.valves.GH_TOKEN}") # 不要这样做
# ✅ 屏蔽敏感数据
logger.debug(f"令牌已配置:{'*' * 10}")
```
### 工作区隔离
```python
# 设置 WORKSPACE_DIR 以限制文件访问
WORKSPACE_DIR = "/safe/sandbox/path"
# Copilot CLI 遵守此目录
client_config["cwd"] = WORKSPACE_DIR
```
### 输入验证
```python
# 验证 chat_id 格式
if chat_id and not re.match(r'^[a-zA-Z0-9_-]+$', chat_id):
logger.warning(f"无效的 chat_id 格式:{chat_id}")
chat_id = None
```
---
## 未来增强
### 计划功能
1. **多会话管理**:支持每个用户的多个并行会话
2. **会话分析**:跟踪 token 使用率、压缩频率
3. **工具结果缓存**:避免冗余工具调用
4. **自定义事件过滤器**:用户可配置的事件处理
5. **工作区模板**:预配置的工作区环境
6. **流式中止**:优雅取消长时间运行的请求
### API 演进
监控 Copilot SDK 更新:
- 新事件类型(例如 `assistant.function_call`
- 增强的工具功能
- 改进的会话序列化
---
## 参考资料
- [GitHub Copilot SDK 文档](https://github.com/github/copilot-sdk)
- [OpenWebUI Pipe 开发](https://docs.openwebui.com/)
- [Awesome OpenWebUI 项目](https://github.com/Fu-Jie/awesome-openwebui)
---
**许可证**MIT
**维护者**Fu-Jie ([@Fu-Jie](https://github.com/Fu-Jie))

View File

@@ -0,0 +1,124 @@
import asyncio
import os
import json
import sys
from copilot import CopilotClient, define_tool
from copilot.types import SessionConfig
from pydantic import BaseModel, Field
# Define a simple tool for testing
class RandomNumberParams(BaseModel):
min: int = Field(description="Minimum value")
max: int = Field(description="Maximum value")
@define_tool(description="Generate a random integer within a range.")
async def generate_random_number(params: RandomNumberParams) -> str:
import random
return f"Result: {random.randint(params.min, params.max)}"
async def main():
print(f"Running tests with Python: {sys.executable}")
# 1. Setup Client
client = CopilotClient({"log_level": "error"})
await client.start()
try:
print("\n=== Test 1: Session Creation & Formatting Injection ===")
# Use gpt-4o or similar capable model
model_id = "gpt-5-mini"
system_message_config = {
"mode": "append",
"content": "You are a test assistant. Always start your response with 'TEST_PREFIX: '.",
}
session_config = SessionConfig(
model=model_id,
system_message=system_message_config,
tools=[generate_random_number],
)
session = await client.create_session(config=session_config)
session_id = session.session_id
print(f"Session Created: {session_id}")
# Test 1.1: Check system prompt effect
resp = await session.send_and_wait(
{"prompt": "Say hello.", "mode": "immediate"}
)
content = resp.data.content
print(f"Response 1: {content}")
if "TEST_PREFIX:" in content:
print("✅ System prompt injection active.")
else:
print("⚠️ System prompt injection NOT detected.")
print("\n=== Test 2: Tool Execution ===")
# Test Tool Usage
prompt_with_tool = (
"Generate a random number between 100 and 200 using the tool."
)
print(f"Sending: {prompt_with_tool}")
# We need to listen to events to verify tool execution,
# but send_and_wait handles it internally and returns the final answer.
# We check if the final answer mentions the result.
resp_tool = await session.send_and_wait(
{"prompt": prompt_with_tool, "mode": "immediate"}
)
tool_content = resp_tool.data.content
print(f"Response 2: {tool_content}")
if "Result:" in tool_content or any(char.isdigit() for char in tool_content):
print("✅ Tool likely executed (numbers found).")
else:
print("⚠️ Tool execution uncertain.")
print("\n=== Test 3: Context Retention (Memory) ===")
# Store a fact
await session.send_and_wait(
{"prompt": "My secret code is 'BLUE-42'. Remember it.", "mode": "immediate"}
)
print("Fact sent.")
# Retrieve it
resp_mem = await session.send_and_wait(
{"prompt": "What is my secret code?", "mode": "immediate"}
)
mem_content = resp_mem.data.content
print(f"Response 3: {mem_content}")
if "BLUE-42" in mem_content:
print("✅ Context retention successful.")
else:
print("⚠️ Context retention failed.")
# Cleanup
await session.destroy()
print("\n=== Test 4: Resume Session (Simulation) ===")
# Note: Actual resuming depends on backend persistence.
# The SDK's client.resume_session(id) tries to find it.
# Since we destroyed it above, we expect failure or new session logic in real app.
# But let's create a new one to persist, close client, and try to resume if process was same?
# Actually persistence usually requires the Copilot Agent/Extension host to keep state or file backed.
# The Python SDK defaults to file-based workspace in standard generic usage?
# Let's just skip complex resume testing for this simple script as it depends on environment (vscode-chat-session vs file).
print("Skipping complex resume test in script.")
except Exception as e:
print(f"Test Failed: {e}")
finally:
await client.stop()
print("\nTests Completed.")
if __name__ == "__main__":
asyncio.run(main())

View File

@@ -0,0 +1,94 @@
import asyncio
import os
import sys
import json
from copilot import CopilotClient
from copilot.types import SessionConfig
# Define the formatting instruction exactly as in the plugin
FORMATTING_INSTRUCTION = (
"\n\n[Formatting Guidelines]\n"
"When providing explanations or descriptions:\n"
"- Use clear paragraph breaks (double line breaks)\n"
"- Break long sentences into multiple shorter ones\n"
"- Use bullet points or numbered lists for multiple items\n"
"- Add headings (##, ###) for major sections\n"
"- Ensure proper spacing between different topics"
)
async def main():
print(f"Python executable: {sys.executable}")
# Check for GH_TOKEN
token = os.environ.get("GH_TOKEN")
if token:
print("GH_TOKEN is set.")
else:
print(
"Warning: GH_TOKEN not found in environment variables. Relying on CLI auth."
)
client_config = {"log_level": "debug"}
client = CopilotClient(client_config)
try:
print("Starting client...")
await client.start()
# Test 1: Check available models
try:
models = await client.list_models()
print(f"Connection successful. Found {len(models)} models.")
model_id = "gpt-5-mini" # User requested model
except Exception as e:
print(f"Failed to list models: {e}")
return
print(f"\nCreating session with model {model_id} and system injection...")
system_message_config = {
"mode": "append",
"content": "You are a helpful assistant." + FORMATTING_INSTRUCTION,
}
session_config = SessionConfig(
model=model_id, system_message=system_message_config
)
session = await client.create_session(config=session_config)
print(f"Session created: {session.session_id}")
# Test 2: Ask the model to summarize its instructions
prompt = "Please summarize the [Formatting Guidelines] you have been given in a list."
print(f"\nSending prompt: '{prompt}'")
response = await session.send_and_wait({"prompt": prompt, "mode": "immediate"})
print("\n--- Model Response ---")
content = response.data.content if response and response.data else "No content"
print(content)
print("----------------------")
required_keywords = ["paragraph", "break", "heading", "spacing", "bullet"]
found_keywords = [kw for kw in required_keywords if kw in content.lower()]
if len(found_keywords) >= 3:
print(
f"\n✅ SUCCESS: Model summarized the guidelines correctly. Found match for: {found_keywords}"
)
else:
print(
f"\n⚠️ UNCERTAIN: Summary might be generic. Found keywords: {found_keywords}"
)
except Exception as e:
print(f"\nError: {e}")
finally:
await client.stop()
print("\nClient stopped.")
if __name__ == "__main__":
asyncio.run(main())

View File

@@ -0,0 +1,359 @@
"""
title: UI Language Debugger
author: Fu-Jie
author_url: https://github.com/Fu-Jie/awesome-openwebui
funding_url: https://github.com/open-webui
version: 0.1.0
icon_url: data:image/svg+xml;base64,PHN2ZyB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciIHdpZHRoPSIyNCIgaGVpZ2h0PSIyNCIgdmlld0JveD0iMCAwIDI0IDI0IiBmaWxsPSJub25lIiBzdHJva2U9ImN1cnJlbnRDb2xvciIgc3Ryb2tlLXdpZHRoPSIyIiBzdHJva2UtbGluZWNhcD0icm91bmQiIHN0cm9rZS1saW5lam9pbj0icm91bmQiPgogIDxwYXRoIGQ9Im01IDggNiA2Ii8+CiAgPHBhdGggZD0ibTQgMTQgNi02IDItMiIvPgogIDxwYXRoIGQ9Ik0yIDVoMTIiLz4KICA8cGF0aCBkPSJNNyAyaDEiLz4KICA8cGF0aCBkPSJtMjIgMjItNS0xMC01IDEwIi8+CiAgPHBhdGggZD0iTTE0IDE4aDYiLz4KPC9zdmc+Cg==
description: Debug UI language detection in the browser console and on-page panel.
"""
import json
import logging
from typing import Optional, Dict, Any, Callable, Awaitable
from pydantic import BaseModel, Field
logging.basicConfig(
level=logging.INFO,
format="%(asctime)s - %(name)s - %(levelname)s - %(message)s",
)
logger = logging.getLogger(__name__)
HTML_WRAPPER_TEMPLATE = """
<!-- OPENWEBUI_PLUGIN_OUTPUT -->
<!DOCTYPE html>
<html lang="{user_language}">
<head>
<meta charset="UTF-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<style>
body {
font-family: -apple-system, BlinkMacSystemFont, "Segoe UI", Roboto, Helvetica, Arial, sans-serif;
margin: 0;
padding: 10px;
background-color: transparent;
}
#main-container {
display: flex;
flex-direction: column;
gap: 16px;
width: 100%;
max-width: 100%;
}
/* STYLES_INSERTION_POINT */
</style>
</head>
<body>
<div id="main-container">
<!-- CONTENT_INSERTION_POINT -->
</div>
<!-- SCRIPTS_INSERTION_POINT -->
</body>
</html>
"""
CONTENT_TEMPLATE = """
<div class="lang-debug-card" id="lang-debug-card-{unique_id}">
<div class="lang-debug-header">
🧭 UI Language Debugger
</div>
<div class="lang-debug-body">
<div class="lang-debug-row"><span>python.ui_language</span><code id="lang-py-{unique_id}">{python_language}</code></div>
<div class="lang-debug-row"><span>document.documentElement.lang</span><code id="lang-html-{unique_id}">-</code></div>
<div class="lang-debug-row"><span>document.documentElement.getAttribute('lang')</span><code id="lang-attr-{unique_id}">-</code></div>
<div class="lang-debug-row"><span>document.documentElement.dir</span><code id="lang-dir-{unique_id}">-</code></div>
<div class="lang-debug-row"><span>document.body.lang</span><code id="lang-body-{unique_id}">-</code></div>
<div class="lang-debug-row"><span>navigator.language</span><code id="lang-nav-{unique_id}">-</code></div>
<div class="lang-debug-row"><span>navigator.languages</span><code id="lang-navs-{unique_id}">-</code></div>
<div class="lang-debug-row"><span>localStorage.language</span><code id="lang-store-{unique_id}">-</code></div>
<div class="lang-debug-row"><span>localStorage.locale</span><code id="lang-locale-{unique_id}">-</code></div>
<div class="lang-debug-row"><span>localStorage.i18n</span><code id="lang-i18n-{unique_id}">-</code></div>
<div class="lang-debug-row"><span>localStorage.settings</span><code id="lang-settings-{unique_id}">-</code></div>
<div class="lang-debug-row"><span>document.documentElement.dataset</span><code id="lang-dataset-{unique_id}">-</code></div>
</div>
</div>
"""
STYLE_TEMPLATE = """
.lang-debug-card {
border: 1px solid #e2e8f0;
border-radius: 12px;
background: #ffffff;
overflow: hidden;
box-shadow: 0 2px 10px rgba(15, 23, 42, 0.06);
}
.lang-debug-header {
padding: 12px 16px;
background: linear-gradient(135deg, #6366f1, #8b5cf6);
color: #fff;
font-weight: 600;
}
.lang-debug-body {
padding: 12px 16px;
display: flex;
flex-direction: column;
gap: 8px;
}
.lang-debug-row {
display: flex;
justify-content: space-between;
gap: 12px;
font-size: 0.9em;
color: #1f2937;
}
.lang-debug-row code {
background: #f8fafc;
border: 1px solid #e2e8f0;
padding: 2px 6px;
border-radius: 6px;
color: #0f172a;
}
"""
SCRIPT_TEMPLATE = """
<script>
(function() {{
const uniqueId = "{unique_id}";
const get = (id) => document.getElementById(id + '-' + uniqueId);
const safe = (value) => {
if (value === undefined || value === null || value === "") return "-";
if (Array.isArray(value)) return value.join(", ");
if (typeof value === "object") return JSON.stringify(value);
return String(value);
};
const safeJson = (value) => {
try {
return value ? JSON.stringify(JSON.parse(value)) : "-";
} catch (e) {
return value ? String(value) : "-";
}
};
const settingsRaw = localStorage.getItem('settings');
const i18nRaw = localStorage.getItem('i18n');
const localeRaw = localStorage.getItem('locale');
const payload = {{
htmlLang: document.documentElement.lang,
htmlAttr: document.documentElement.getAttribute('lang'),
htmlDir: document.documentElement.dir,
bodyLang: document.body ? document.body.lang : "",
navigatorLanguage: navigator.language,
navigatorLanguages: navigator.languages,
localStorageLanguage: localStorage.getItem('language'),
localStorageLocale: localeRaw,
localStorageI18n: i18nRaw,
localStorageSettings: settingsRaw,
htmlDataset: document.documentElement.dataset,
}};
get('lang-html').textContent = safe(payload.htmlLang);
get('lang-attr').textContent = safe(payload.htmlAttr);
get('lang-dir').textContent = safe(payload.htmlDir);
get('lang-body').textContent = safe(payload.bodyLang);
get('lang-nav').textContent = safe(payload.navigatorLanguage);
get('lang-navs').textContent = safe(payload.navigatorLanguages);
get('lang-store').textContent = safe(payload.localStorageLanguage);
get('lang-locale').textContent = safe(payload.localStorageLocale);
get('lang-i18n').textContent = safeJson(payload.localStorageI18n);
get('lang-settings').textContent = safeJson(payload.localStorageSettings);
get('lang-dataset').textContent = safe(payload.htmlDataset);
console.group('🧭 UI Language Debugger');
console.log(payload);
console.groupEnd();
}})();
</script>
"""
class Action:
class Valves(BaseModel):
SHOW_STATUS: bool = Field(
default=True,
description="Whether to show operation status updates.",
)
SHOW_DEBUG_LOG: bool = Field(
default=True,
description="Whether to print debug logs in the browser console.",
)
def __init__(self):
self.valves = self.Valves()
def _get_user_context(self, __user__: Optional[Dict[str, Any]]) -> Dict[str, str]:
if isinstance(__user__, (list, tuple)):
user_data = __user__[0] if __user__ else {}
elif isinstance(__user__, dict):
user_data = __user__
else:
user_data = {}
return {
"user_id": user_data.get("id", "unknown_user"),
"user_name": user_data.get("name", "User"),
"user_language": user_data.get("language", ""),
}
def _get_chat_context(
self, body: dict, __metadata__: Optional[dict] = None
) -> Dict[str, str]:
chat_id = ""
message_id = ""
if isinstance(body, dict):
chat_id = body.get("chat_id", "")
message_id = body.get("id", "")
if not chat_id or not message_id:
body_metadata = body.get("metadata", {})
if isinstance(body_metadata, dict):
if not chat_id:
chat_id = body_metadata.get("chat_id", "")
if not message_id:
message_id = body_metadata.get("message_id", "")
if __metadata__ and isinstance(__metadata__, dict):
if not chat_id:
chat_id = __metadata__.get("chat_id", "")
if not message_id:
message_id = __metadata__.get("message_id", "")
return {
"chat_id": str(chat_id).strip(),
"message_id": str(message_id).strip(),
}
async def _emit_status(
self,
emitter: Optional[Callable[[Any], Awaitable[None]]],
description: str,
done: bool = False,
):
if self.valves.SHOW_STATUS and emitter:
await emitter(
{"type": "status", "data": {"description": description, "done": done}}
)
async def _emit_debug_log(
self,
emitter: Optional[Callable[[Any], Awaitable[None]]],
title: str,
data: dict,
):
if not self.valves.SHOW_DEBUG_LOG or not emitter:
return
try:
js_code = f"""
(async function() {{
console.group("🛠️ {title}");
console.log({json.dumps(data, ensure_ascii=False)});
console.groupEnd();
}})();
"""
await emitter({"type": "execute", "data": {"code": js_code}})
except Exception as e:
logger.error("Error emitting debug log: %s", e, exc_info=True)
def _merge_html(
self,
existing_html: str,
new_content: str,
new_styles: str = "",
new_scripts: str = "",
user_language: str = "en-US",
) -> str:
if not existing_html:
base_html = HTML_WRAPPER_TEMPLATE.replace("{user_language}", user_language)
else:
base_html = existing_html
if "<!-- CONTENT_INSERTION_POINT -->" in base_html:
base_html = base_html.replace(
"<!-- CONTENT_INSERTION_POINT -->",
f"{new_content}\n <!-- CONTENT_INSERTION_POINT -->",
)
if new_styles and "/* STYLES_INSERTION_POINT */" in base_html:
base_html = base_html.replace(
"/* STYLES_INSERTION_POINT */",
f"{new_styles}\n /* STYLES_INSERTION_POINT */",
)
if new_scripts and "<!-- SCRIPTS_INSERTION_POINT -->" in base_html:
base_html = base_html.replace(
"<!-- SCRIPTS_INSERTION_POINT -->",
f"{new_scripts}\n <!-- SCRIPTS_INSERTION_POINT -->",
)
return base_html
async def action(
self,
body: dict,
__user__: Optional[Dict[str, Any]] = None,
__event_emitter__: Optional[Any] = None,
__event_call__: Optional[Callable[[Any], Awaitable[None]]] = None,
__metadata__: Optional[dict] = None,
__request__: Optional[Any] = None,
) -> Optional[dict]:
await self._emit_status(__event_emitter__, "Detecting UI language...", False)
user_ctx = self._get_user_context(__user__)
await self._emit_debug_log(
__event_emitter__,
"Language Debugger: user context",
user_ctx,
)
ui_language = ""
if __event_call__:
try:
response = await __event_call__(
{
"type": "execute",
"data": {
"code": "return (localStorage.getItem('locale') || localStorage.getItem('language') || (navigator.languages && navigator.languages[0]) || navigator.language || document.documentElement.lang || '')",
},
}
)
await self._emit_debug_log(
__event_emitter__,
"Language Debugger: execute response",
{"response": response},
)
if isinstance(response, dict) and "value" in response:
ui_language = response.get("value", "") or ""
elif isinstance(response, str):
ui_language = response
except Exception as e:
logger.error(
"Failed to read UI language from frontend: %s", e, exc_info=True
)
unique_id = f"lang_{int(__import__('time').time() * 1000)}"
content_html = CONTENT_TEMPLATE.replace("{unique_id}", unique_id).replace(
"{python_language}", ui_language or "-"
)
script_html = SCRIPT_TEMPLATE.replace("{unique_id}", unique_id)
script_html = script_html.replace("{{", "{").replace("}}", "}")
final_html = self._merge_html(
"",
content_html,
STYLE_TEMPLATE,
script_html,
"en",
)
html_embed_tag = f"```html\n{final_html}\n```"
body["messages"][-1]["content"] = (
body["messages"][-1].get("content", "") + "\n\n" + html_embed_tag
)
await self._emit_status(__event_emitter__, "UI language captured.", True)
return body

View File

@@ -5,31 +5,10 @@
This filter reduces token consumption in long conversations through intelligent summarization and message compression while keeping conversations coherent. This filter reduces token consumption in long conversations through intelligent summarization and message compression while keeping conversations coherent.
## What's new in 1.2.2 ## What's new in 1.2.2
- **Critical Fix**: Resolved `TypeError: 'str' object is not callable` caused by variable name conflict in logging function. - **Critical Fix**: Resolved `TypeError: 'str' object is not callable` caused by variable name conflict in logging function.
- **Compatibility**: Enhanced `params` handling to support Pydantic objects, improving compatibility with different OpenWebUI versions. - **Compatibility**: Enhanced `params` handling to support Pydantic objects, improving compatibility with different OpenWebUI versions.
## What's new in 1.2.1
- **Smart Configuration**: Automatically detects base model settings for custom models and adds `summary_model_max_context` for independent summary limits.
- **Performance & Refactoring**: Optimized threshold parsing with caching, removed redundant code, and improved LLM response handling (JSONResponse support).
- **Bug Fixes & Modernization**: Fixed `datetime` deprecation warnings, corrected type annotations, and replaced print statements with proper logging.
## What's new in 1.2.0
- **Preflight Context Check**: Before sending to the model, validates that total tokens fit within the context window. Automatically trims or drops oldest messages if exceeded.
- **Structure-Aware Assistant Trimming**: When context exceeds the limit, long AI responses are intelligently collapsed while preserving their structure (headers H1-H6, first line, last line).
- **Native Tool Output Trimming**: Detects and trims native tool outputs (`function_calling: "native"`), extracting only the final answer. Enable via `enable_tool_output_trimming`. **Note**: Non-native tool outputs are not fully injected into context.
- **Consolidated Status Notifications**: Unified "Context Usage" and "Context Summary Updated" notifications with appended warnings (e.g., `| ⚠️ High Usage`) for clearer feedback.
- **Context Usage Warning**: Emits a warning notification when context usage exceeds 90%.
- **Enhanced Header Detection**: Optimized regex (`^#{1,6}\s+`) to avoid false positives like `#hashtag`.
- **Detailed Token Logging**: Logs now show token breakdown for System, Head, Summary, and Tail sections with total.
## What's new in 1.1.3
- **Improved Compatibility**: Changed summary injection role from `user` to `assistant` for better compatibility across different LLMs.
- **Enhanced Stability**: Fixed a race condition in state management that could cause "inlet state not found" warnings in high-concurrency scenarios.
- **Bug Fixes**: Corrected default model handling to prevent misleading logs when no model is specified.
--- ---
## Core Features ## Core Features
@@ -57,11 +36,7 @@ This filter reduces token consumption in long conversations through intelligent
### 2) Filter order ### 2) Filter order
It is recommended to keep this filter early in the chain so it runs before filters that mutate messages: - Recommended order: pre-filters (<10) → this filter (10) → post-filters (>10).
1. Pre-filters (priority < 10) — e.g., system prompt injectors.
2. This compression filter (priority = 10).
3. Post-filters (priority > 10) — e.g., output formatting.
--- ---
@@ -85,6 +60,16 @@ It is recommended to keep this filter early in the chain so it runs before filte
--- ---
## ⭐ Support
If this plugin has been useful, a star on [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) is a big motivation for me. Thank you for the support.
## Troubleshooting ❓
- **Initial system prompt is lost**: Keep `keep_first` greater than 0 to protect the initial message. - **Initial system prompt is lost**: Keep `keep_first` greater than 0 to protect the initial message.
- **Compression effect is weak**: Raise `compression_threshold_tokens` or lower `keep_first` / `keep_last` to allow more aggressive compression. - **Compression effect is weak**: Raise `compression_threshold_tokens` or lower `keep_first` / `keep_last` to allow more aggressive compression.
- **Submit an Issue**: If you encounter any problems, please submit an issue on GitHub: [Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues) - **Submit an Issue**: If you encounter any problems, please submit an issue on GitHub: [Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
## Changelog
See the full history on GitHub: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)

View File

@@ -7,31 +7,10 @@
本过滤器通过智能摘要和消息压缩技术,在保持对话连贯性的同时,显著降低长对话的 Token 消耗。 本过滤器通过智能摘要和消息压缩技术,在保持对话连贯性的同时,显著降低长对话的 Token 消耗。
## 1.2.2 版本更新 ## 1.2.2 版本更新
- **严重错误修复**: 解决了因日志函数变量名冲突导致的 `TypeError: 'str' object is not callable` 错误。 - **严重错误修复**: 解决了因日志函数变量名冲突导致的 `TypeError: 'str' object is not callable` 错误。
- **兼容性增强**: 改进了 `params` 处理逻辑以支持 Pydantic 对象,提高了对不同 OpenWebUI 版本的兼容性。 - **兼容性增强**: 改进了 `params` 处理逻辑以支持 Pydantic 对象,提高了对不同 OpenWebUI 版本的兼容性。
## 1.2.1 版本更新
- **智能配置增强**: 自动检测自定义模型的基础模型配置,并新增 `summary_model_max_context` 参数以独立控制摘要模型的上下文限制。
- **性能优化与重构**: 重构了阈值解析逻辑并增加缓存,移除了冗余的处理代码,并增强了 LLM 响应处理(支持 JSONResponse
- **稳定性改进**: 修复了 `datetime` 弃用警告,修正了类型注解,并将 print 语句替换为标准日志记录。
## 1.2.0 版本更新
- **预检上下文检查 (Preflight Context Check)**: 在发送给模型之前,验证总 Token 是否符合上下文窗口。如果超出,自动裁剪或丢弃最旧的消息。
- **结构感知助手裁剪 (Structure-Aware Assistant Trimming)**: 当上下文超出限制时,智能折叠过长的 AI 回复,同时保留其结构(标题 H1-H6、首行、尾行
- **原生工具输出裁剪 (Native Tool Output Trimming)**: 检测并裁剪原生工具输出 (`function_calling: "native"`),仅提取最终答案。通过 `enable_tool_output_trimming` 启用。**注意**:非原生工具调用输出不会完整注入上下文。
- **统一状态通知**: 统一了“上下文使用情况”和“上下文摘要更新”的通知,并附加警告(例如 `| ⚠️ 高负载`),反馈更清晰。
- **上下文使用警告**: 当上下文使用率超过 90% 时发出警告通知。
- **增强的标题检测**: 优化了正则表达式 (`^#{1,6}\s+`) 以避免误判(如 `#hashtag`)。
- **详细 Token 日志**: 日志现在显示 System、Head、Summary 和 Tail 部分的 Token 细分及总计。
## 1.1.3 版本更新
- **兼容性提升**: 将摘要注入角色从 `user` 改为 `assistant`,以提高在不同 LLM 之间的兼容性。
- **稳定性增强**: 修复了状态管理中的竞态条件,解决了高并发场景下可能出现的“无法获取 inlet 状态”警告。
- **Bug 修复**: 修正了默认模型处理逻辑,防止在未指定模型时产生误导性日志。
--- ---
## 核心特性 ## 核心特性
@@ -61,11 +40,7 @@
### 2. 过滤器顺序 ### 2. 过滤器顺序
建议将此过滤器的优先级设置得相对较高(数值较小),以确保它在其他可能修改消息内容的过滤器之前运行。一个典型的顺序可能是: - 建议顺序:前置过滤器(<10→ 本过滤器10→ 后置过滤器(>10
1. 前置过滤器 (priority < 10) —— 例如系统提示注入。
2. 本压缩过滤器 (priority = 10)。
3. 后置过滤器 (priority > 10) —— 例如最终输出格式化。
--- ---
@@ -124,6 +99,16 @@
--- ---
## ⭐ 支持
如果这个插件对你有帮助,欢迎到 [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) 点个 Star这将是我持续改进的动力感谢支持。
## 故障排除 (Troubleshooting) ❓
- **初始系统提示丢失**:将 `keep_first` 设置为大于 0。 - **初始系统提示丢失**:将 `keep_first` 设置为大于 0。
- **压缩效果不明显**:提高 `compression_threshold_tokens`,或降低 `keep_first` / `keep_last` 以增强压缩力度。 - **压缩效果不明显**:提高 `compression_threshold_tokens`,或降低 `keep_first` / `keep_last` 以增强压缩力度。
- **提交 Issue**: 如果遇到任何问题,请在 GitHub 上提交 Issue[Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues) - **提交 Issue**: 如果遇到任何问题,请在 GitHub 上提交 Issue[Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
## 更新日志
完整历史请查看 GitHub 项目: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)

View File

@@ -2,35 +2,30 @@
**Author:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **Version:** 0.1.0 | **Project:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **License:** MIT **Author:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **Version:** 0.1.0 | **Project:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **License:** MIT
--- **Folder Memory** is an intelligent context filter plugin for OpenWebUI. It automatically extracts consistent "Project Rules" from ongoing conversations within a folder and injects them back into the folder's system prompt.
## 🔥 What's New in v0.1.0
### 📌 What's new in 0.1.0
- **Initial Release**: Automated "Project Rules" management for OpenWebUI folders. - **Initial Release**: Automated "Project Rules" management for OpenWebUI folders.
- **Folder-Level Persistence**: Automatically updates folder system prompts with extracted rules. - **Folder-Level Persistence**: Automatically updates folder system prompts with extracted rules.
- **Optimized Performance**: Runs asynchronously and supports `PRIORITY` configuration for seamless integration with other filters. - **Optimized Performance**: Runs asynchronously and supports `PRIORITY` configuration for seamless integration with other filters.
--- ## ✨ Core Features
**Folder Memory** is an intelligent context filter plugin for OpenWebUI. It automatically extracts consistent "Project Rules" from ongoing conversations within a folder and injects them back into the folder's system prompt.
## ✨ Features
- **Automatic Extraction**: Analyzes chat history every N messages to extract project rules. - **Automatic Extraction**: Analyzes chat history every N messages to extract project rules.
- **Non-destructive Injection**: Updates only the specific "Project Rules" block in the system prompt, preserving other instructions. - **Non-destructive Injection**: Updates only the specific "Project Rules" block in the system prompt, preserving other instructions.
- **Async Processing**: Runs in the background without blocking the user's chat experience. - **Async Processing**: Runs in the background without blocking the user's chat experience.
- **ORM Integration**: Directly updates folder data using OpenWebUI's internal models for reliability. - **ORM Integration**: Directly updates folder data using OpenWebUI's internal models for reliability.
## ⚠️ Prerequisites ## Installation & Configuration
- **Conversations must occur inside a folder.** This plugin only triggers when a chat belongs to a folder (i.e., you need to create a folder in OpenWebUI and start a conversation within it). ### 1) Installation
## 📦 Installation
1. Copy `folder_memory.py` to your OpenWebUI `plugins/filters/` directory (or upload via Admin UI). 1. Copy `folder_memory.py` to your OpenWebUI `plugins/filters/` directory (or upload via Admin UI).
2. Enable the filter in your **Settings** -> **Filters**. 2. Enable the filter in your **Settings** -> **Filters**.
3. (Optional) Configure the triggering threshold (default: every 10 messages). 3. **Prerequisite**: Conversations must occur inside a folder (create a folder and start a chat within it).
## ⚙️ Configuration (Valves) ### 2) Configuration (Valves)
| Valve | Default | Description | | Valve | Default | Description |
| :--- | :--- | :--- | | :--- | :--- | :--- |
@@ -41,6 +36,10 @@
| `SHOW_DEBUG_LOG` | `False` | Show detailed debug logs in the browser console. | | `SHOW_DEBUG_LOG` | `False` | Show detailed debug logs in the browser console. |
| `UPDATE_ROOT_FOLDER` | `False` | If enabled, finds and updates the root folder rules instead of the current subfolder. | | `UPDATE_ROOT_FOLDER` | `False` | If enabled, finds and updates the root folder rules instead of the current subfolder. |
## ⭐ Support
If this plugin has been useful, a star on [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) is a big motivation for me. Thank you for the support.
## 🛠️ How It Works ## 🛠️ How It Works
![Folder Memory Demo](https://raw.githubusercontent.com/Fu-Jie/awesome-openwebui/main/plugins/filters/folder-memory/folder-memory-demo.png) ![Folder Memory Demo](https://raw.githubusercontent.com/Fu-Jie/awesome-openwebui/main/plugins/filters/folder-memory/folder-memory-demo.png)
@@ -58,3 +57,7 @@
## 🗺️ Roadmap ## 🗺️ Roadmap
See [ROADMAP.md](./ROADMAP.md) for future plans, including "Project Knowledge" collection. See [ROADMAP.md](./ROADMAP.md) for future plans, including "Project Knowledge" collection.
## Changelog
See the full history on GitHub: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)

View File

@@ -2,37 +2,32 @@
**作者:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **版本:** 0.1.0 | **项目:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **许可证:** MIT **作者:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **版本:** 0.1.0 | **项目:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **许可证:** MIT
---
### 📌 0.1.0 版本特性
- **首个版本发布**:专注于自动化的“项目规则”管理。
- **文件夹级持久化**:自动将提取的规则回写到文件夹系统提示词中。
- **性能优化**:采用异步处理机制,并支持 `PRIORITY` 配置,确保与其他过滤器(如上下文压缩)完美协作。
---
**文件夹记忆 (Folder Memory)** 是一个 OpenWebUI 的智能上下文过滤器插件。它能自动从文件夹内的对话中提取一致性的“项目规则”,并将其回写到文件夹的系统提示词中。 **文件夹记忆 (Folder Memory)** 是一个 OpenWebUI 的智能上下文过滤器插件。它能自动从文件夹内的对话中提取一致性的“项目规则”,并将其回写到文件夹的系统提示词中。
这确保了该文件夹内的所有未来对话都能共享相同的进化上下文和规则,无需手动更新。 这确保了该文件夹内的所有未来对话都能共享相同的进化上下文和规则,无需手动更新。
## ✨ 功能特性 ## 🔥 最新更新 v0.1.0
- **首个版本发布**:专注于自动化的“项目规则”管理。
- **文件夹级持久化**:自动将提取的规则回写到文件夹系统提示词中。
- **性能优化**:采用异步处理机制,并支持 `PRIORITY` 配置,确保与其他过滤器(如上下文压缩)完美协作。
## ✨ 核心特性
- **自动提取**:每隔 N 条消息分析一次聊天记录,提取项目规则。 - **自动提取**:每隔 N 条消息分析一次聊天记录,提取项目规则。
- **无损注入**:仅更新系统提示词中的特定“项目规则”块,保留其他指令。 - **无损注入**:仅更新系统提示词中的特定“项目规则”块,保留其他指令。
- **异步处理**:在后台运行,不阻塞用户的聊天体验。 - **异步处理**:在后台运行,不阻塞用户的聊天体验。
- **ORM 集成**:直接使用 OpenWebUI 的内部模型更新文件夹数据,确保可靠性。 - **ORM 集成**:直接使用 OpenWebUI 的内部模型更新文件夹数据,确保可靠性。
## ⚠️ 前置条件 ## 安装与配置
- **对话必须在文件夹内进行。** 此插件仅在聊天属于某个文件夹时触发(即您需要先在 OpenWebUI 中创建一个文件夹,并在其内部开始对话)。 ### 1. 安装
## 📦 安装指南 1.`folder_memory.py`(或中文版 `folder_memory_cn.py`)复制到 OpenWebUI 的 `plugins/filters/` 目录(或通过管理员 UI 上传)。
1.`folder_memory.py` (或中文版 `folder_memory_cn.py`) 复制到 OpenWebUI 的 `plugins/filters/` 目录(或通过管理员 UI 上传)。
2.**设置** -> **过滤器** 中启用该插件。 2.**设置** -> **过滤器** 中启用该插件。
3. (可选)配置触发阈值(默认:每 10 条消息)。 3. **前置条件**:对话必须在文件夹内进行(先创建文件夹并在其中开始对话)。
## ⚙️ 配置 (Valves) ### 2. 配置 (Valves)
| 参数 | 默认值 | 说明 | | 参数 | 默认值 | 说明 |
| :--- | :--- | :--- | | :--- | :--- | :--- |
@@ -43,6 +38,10 @@
| `SHOW_DEBUG_LOG` | `False` | 在浏览器控制台显示详细调试日志。 | | `SHOW_DEBUG_LOG` | `False` | 在浏览器控制台显示详细调试日志。 |
| `UPDATE_ROOT_FOLDER` | `False` | 如果启用,将向上查找并更新根文件夹的规则,而不是当前子文件夹。 | | `UPDATE_ROOT_FOLDER` | `False` | 如果启用,将向上查找并更新根文件夹的规则,而不是当前子文件夹。 |
## ⭐ 支持
如果这个插件对你有帮助,欢迎到 [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) 点个 Star这将是我持续改进的动力感谢支持。
## 🛠️ 工作原理 ## 🛠️ 工作原理
![Folder Memory Demo](https://raw.githubusercontent.com/Fu-Jie/awesome-openwebui/main/plugins/filters/folder-memory/folder-memory-demo.png) ![Folder Memory Demo](https://raw.githubusercontent.com/Fu-Jie/awesome-openwebui/main/plugins/filters/folder-memory/folder-memory-demo.png)
@@ -60,3 +59,7 @@
## 🗺️ 路线图 ## 🗺️ 路线图
查看 [ROADMAP.md](./ROADMAP.md) 了解未来计划,包括“项目知识”收集功能。 查看 [ROADMAP.md](./ROADMAP.md) 了解未来计划,包括“项目知识”收集功能。
## 更新日志
完整历史请查看 GitHub 项目: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)

View File

@@ -4,7 +4,11 @@
A content normalizer filter for Open WebUI that fixes common Markdown formatting issues in LLM outputs. It ensures that code blocks, LaTeX formulas, Mermaid diagrams, and other Markdown elements are rendered correctly. A content normalizer filter for Open WebUI that fixes common Markdown formatting issues in LLM outputs. It ensures that code blocks, LaTeX formulas, Mermaid diagrams, and other Markdown elements are rendered correctly.
## Features ## 🔥 What's New in v1.2.4
* **Documentation Sync**: Synchronized version numbers across all documentation and code files.
## ✨ Core Features
* **Details Tag Normalization**: Ensures proper spacing for `<details>` tags (used for thought chains). Adds a blank line after `</details>` and ensures a newline after self-closing `<details />` tags to prevent rendering issues. * **Details Tag Normalization**: Ensures proper spacing for `<details>` tags (used for thought chains). Adds a blank line after `</details>` and ensures a newline after self-closing `<details />` tags to prevent rendering issues.
* **Emphasis Spacing Fix**: Fixes extra spaces inside emphasis markers (e.g., `** text **` -> `**text**`) which can cause rendering failures. Includes safeguards to protect math expressions (e.g., `2 * 3 * 4`) and list variables. * **Emphasis Spacing Fix**: Fixes extra spaces inside emphasis markers (e.g., `** text **` -> `**text**`) which can cause rendering failures. Includes safeguards to protect math expressions (e.g., `2 * 3 * 4`) and list variables.
@@ -19,7 +23,7 @@ A content normalizer filter for Open WebUI that fixes common Markdown formatting
* **Table Fix**: Adds missing closing pipes in tables. * **Table Fix**: Adds missing closing pipes in tables.
* **XML Cleanup**: Removes leftover XML artifacts. * **XML Cleanup**: Removes leftover XML artifacts.
## Usage ## How to Use 🛠️
1. Install the plugin in Open WebUI. 1. Install the plugin in Open WebUI.
2. Enable the filter globally or for specific models. 2. Enable the filter globally or for specific models.
@@ -28,69 +32,38 @@ A content normalizer filter for Open WebUI that fixes common Markdown formatting
> [!WARNING] > [!WARNING]
> As this is an initial version, some "negative fixes" might occur (e.g., breaking valid Markdown). If you encounter issues, please check the console logs, copy the "Original" vs "Normalized" content, and submit an issue. > As this is an initial version, some "negative fixes" might occur (e.g., breaking valid Markdown). If you encounter issues, please check the console logs, copy the "Original" vs "Normalized" content, and submit an issue.
## Configuration (Valves) ## Configuration (Valves) ⚙️
* `priority`: Filter priority (default: 50). | Parameter | Default | Description |
* `enable_escape_fix`: Fix excessive escape characters. | :--- | :--- | :--- |
* `enable_thought_tag_fix`: Normalize thought tags. | `priority` | `50` | Filter priority. Higher runs later (recommended after other filters). |
* `enable_details_tag_fix`: Normalize details tags (default: True). | `enable_escape_fix` | `True` | Fix excessive escape characters (`\n`, `\t`, etc.). |
* `enable_code_block_fix`: Fix code block formatting. | `enable_escape_fix_in_code_blocks` | `False` | Apply escape fix inside code blocks (may affect valid code). |
* `enable_latex_fix`: Normalize LaTeX formulas. | `enable_thought_tag_fix` | `True` | Normalize thought tags (`</thought>`). |
* `enable_list_fix`: Fix list item newlines (Experimental). | `enable_details_tag_fix` | `True` | Normalize `<details>` tags and add safe spacing. |
* `enable_unclosed_block_fix`: Auto-close unclosed code blocks. | `enable_code_block_fix` | `True` | Fix code block formatting (indentation/newlines). |
* `enable_fullwidth_symbol_fix`: Fix full-width symbols in code blocks. | `enable_latex_fix` | `True` | Normalize LaTeX delimiters (`\[` -> `$$`, `\(` -> `$`). |
* `enable_mermaid_fix`: Fix Mermaid syntax errors. | `enable_list_fix` | `False` | Fix list item newlines (experimental). |
* `enable_heading_fix`: Fix missing space in headings. | `enable_unclosed_block_fix` | `True` | Auto-close unclosed code blocks. |
* `enable_table_fix`: Fix missing closing pipe in tables. | `enable_fullwidth_symbol_fix` | `False` | Fix full-width symbols in code blocks. |
* `enable_xml_tag_cleanup`: Cleanup leftover XML tags. | `enable_mermaid_fix` | `True` | Fix common Mermaid syntax errors. |
* `enable_emphasis_spacing_fix`: Fix extra spaces in emphasis (default: False). | `enable_heading_fix` | `True` | Fix missing space in headings. |
* `show_status`: Show status notification when fixes are applied. | `enable_table_fix` | `True` | Fix missing closing pipe in tables. |
* `show_debug_log`: Print debug logs to browser console. | `enable_xml_tag_cleanup` | `True` | Cleanup leftover XML tags. |
| `enable_emphasis_spacing_fix` | `False` | Fix extra spaces in emphasis. |
| `show_status` | `True` | Show status notification when fixes are applied. |
| `show_debug_log` | `True` | Print debug logs to browser console (F12). |
## Troubleshooting ❓ ## ⭐ Support
If this plugin has been useful, a star on [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) is a big motivation for me. Thank you for the support.
## 🧩 Others
### Troubleshooting ❓
* **Submit an Issue**: If you encounter any problems, please submit an issue on GitHub: [Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues) * **Submit an Issue**: If you encounter any problems, please submit an issue on GitHub: [Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
## Changelog ### Changelog
### v1.2.4 See the full history on GitHub: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)
* **Documentation Updates**: Synchronized version numbers across all documentation and code files.
### v1.2.3
* **List Marker Protection Enhancement**: Fixed a bug where list markers (`*`) followed by plain text and emphasis were having their spaces incorrectly stripped (e.g., `* U16 forward` became `*U16 forward`).
* **Placeholder Support**: Confirmed that 4 or more underscores (e.g., `____`) are correctly treated as placeholders and not modified by the emphasis fix.
### v1.2.2
* **Code Block Indentation Fix**: Fixed an issue where code blocks nested inside lists were having their indentation incorrectly stripped. Now preserves proper indentation for nested code blocks.
* **Underscore Emphasis Support**: Extended emphasis spacing fix to support `__` (double underscore for bold) and `___` (triple underscore for bold+italic) syntax.
* **List Marker Protection**: Fixed a bug where list markers (`*`) followed by emphasis markers (`**`) were incorrectly merged (e.g., `* **Yes**` became `***Yes**`). Added safeguard to prevent this.
* **Test Suite**: Added comprehensive pytest test suite with 56 test cases covering all major features.
### v1.2.1
* **Emphasis Spacing Fix**: Added a new fix for extra spaces inside emphasis markers (e.g., `** text **` -> `**text**`).
* Uses a recursive approach to handle nested emphasis (e.g., `**bold _italic _**`).
* Includes safeguards to prevent modifying math expressions (e.g., `2 * 3 * 4`) or list variables.
* Controlled by the `enable_emphasis_spacing_fix` valve (default: True).
### v1.2.0
* **Details Tag Support**: Added normalization for `<details>` tags.
* Ensures a blank line is added after `</details>` closing tags to separate thought content from the main response.
* Ensures a newline is added after self-closing `<details ... />` tags to prevent them from interfering with subsequent Markdown headings (e.g., fixing `<details/>#Heading`).
* Includes safeguard to prevent modification of `<details>` tags inside code blocks.
### v1.1.2
* **Mermaid Edge Label Protection**: Implemented comprehensive protection for edge labels (text on connecting lines) to prevent them from being incorrectly modified. Now supports all Mermaid link types including solid (`--`), dotted (`-.`), and thick (`==`) lines with or without arrows.
* **Bug Fixes**: Fixed an issue where lines without arrows (e.g., `A -- text --- B`) were not correctly protected.
### v1.1.0
* **Mermaid Fix Refinement**: Improved regex to handle nested parentheses in node labels (e.g., `ID("Label (text)")`) and avoided matching connection labels.
* **HTML Safeguard Optimization**: Refined `_contains_html` to allow common tags like `<br/>`, `<b>`, `<i>`, etc., ensuring Mermaid diagrams with these tags are still normalized.
* **Full-width Symbol Cleanup**: Fixed duplicate keys and incorrect quote mapping in `FULLWIDTH_MAP`.
* **Bug Fixes**: Fixed missing `Dict` import in Python files.

View File

@@ -4,7 +4,11 @@
这是一个用于 Open WebUI 的内容格式化过滤器,旨在修复 LLM 输出中常见的 Markdown 格式问题。它能确保代码块、LaTeX 公式、Mermaid 图表和其他 Markdown 元素被正确渲染。 这是一个用于 Open WebUI 的内容格式化过滤器,旨在修复 LLM 输出中常见的 Markdown 格式问题。它能确保代码块、LaTeX 公式、Mermaid 图表和其他 Markdown 元素被正确渲染。
## 功能特性 ## 🔥 最新更新 v1.2.4
* **文档更新**: 同步了所有文档和代码文件的版本号。
## ✨ 核心特性
* **Details 标签规范化**: 确保 `<details>` 标签(常用于思维链)有正确的间距。在 `</details>` 后添加空行,并在自闭合 `<details />` 标签后添加换行,防止渲染问题。 * **Details 标签规范化**: 确保 `<details>` 标签(常用于思维链)有正确的间距。在 `</details>` 后添加空行,并在自闭合 `<details />` 标签后添加换行,防止渲染问题。
* **强调空格修复**: 修复强调标记内部的多余空格(例如 `** 文本 **` -> `**文本**`),这会导致 Markdown 渲染失败。包含保护机制,防止误修改数学表达式(如 `2 * 3 * 4`)或列表变量。 * **强调空格修复**: 修复强调标记内部的多余空格(例如 `** 文本 **` -> `**文本**`),这会导致 Markdown 渲染失败。包含保护机制,防止误修改数学表达式(如 `2 * 3 * 4`)或列表变量。
@@ -28,69 +32,38 @@
> [!WARNING] > [!WARNING]
> 由于这是初版,可能会出现“负向修复”的情况(例如破坏了原本正确的格式)。如果您遇到问题,请务必查看控制台日志,复制“原始 (Original)”与“规范化 (Normalized)”的内容对比,并提交 Issue 反馈。 > 由于这是初版,可能会出现“负向修复”的情况(例如破坏了原本正确的格式)。如果您遇到问题,请务必查看控制台日志,复制“原始 (Original)”与“规范化 (Normalized)”的内容对比,并提交 Issue 反馈。
## 配置 (Valves) ## 配置参数 (Valves) ⚙️
* `priority`: 过滤器优先级 (默认: 50)。 | 参数 | 默认值 | 描述 |
* `enable_escape_fix`: 修复过度的转义字符。 | :--- | :--- | :--- |
* `enable_thought_tag_fix`: 规范化思维标签。 | `priority` | `50` | 过滤器优先级。数值越大越靠后(建议在其他过滤器之后运行)。 |
* `enable_details_tag_fix`: 规范化 Details 标签 (默认: True)。 | `enable_escape_fix` | `True` | 修复过度的转义字符(`\n`, `\t` 等)。 |
* `enable_code_block_fix`: 修复代码块格式。 | `enable_escape_fix_in_code_blocks` | `False` | 在代码块内应用转义修复(可能影响有效代码)。 |
* `enable_latex_fix`: 规范化 LaTeX 公式。 | `enable_thought_tag_fix` | `True` | 规范化思维标签(`</thought>`)。 |
* `enable_list_fix`: 修复列表项换行 (实验性)。 | `enable_details_tag_fix` | `True` | 规范化 `<details>` 标签并添加安全间距。 |
* `enable_unclosed_block_fix`: 自动闭合未闭合的代码块。 | `enable_code_block_fix` | `True` | 修复代码块格式(缩进/换行)。 |
* `enable_fullwidth_symbol_fix`: 修复代码块中的全角符号。 | `enable_latex_fix` | `True` | 规范化 LaTeX 定界符(`\[` -> `$$`, `\(` -> `$`)。 |
* `enable_mermaid_fix`: 修复 Mermaid 语法错误。 | `enable_list_fix` | `False` | 修复列表项换行(实验性)。 |
* `enable_heading_fix`: 修复标题中缺失的空格。 | `enable_unclosed_block_fix` | `True` | 自动闭合未闭合的代码块。 |
* `enable_table_fix`: 修复表格中缺失的闭合管道符。 | `enable_fullwidth_symbol_fix` | `False` | 修复代码块中的全角符号。 |
* `enable_xml_tag_cleanup`: 清理残留的 XML 标签。 | `enable_mermaid_fix` | `True` | 修复常见 Mermaid 语法错误。 |
* `enable_emphasis_spacing_fix`: 修复强调语法中的多余空格 (默认: True)。 | `enable_heading_fix` | `True` | 修复标题中缺失的空格。 |
* `show_status`: 应用修复时显示状态通知。 | `enable_table_fix` | `True` | 修复表格中缺失的闭合管道符。 |
* `show_debug_log`: 在浏览器控制台打印调试日志。 | `enable_xml_tag_cleanup` | `True` | 清理残留的 XML 标签。 |
| `enable_emphasis_spacing_fix` | `False` | 修复强调语法中的多余空格。 |
| `show_status` | `True` | 应用修复时显示状态通知。 |
| `show_debug_log` | `True` | 在浏览器控制台打印调试日志。 |
## 故障排除 (Troubleshooting) ❓ ## ⭐ 支持
如果这个插件对你有帮助,欢迎到 [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) 点个 Star这将是我持续改进的动力感谢支持。
## 其他
### 故障排除 (Troubleshooting) ❓
* **提交 Issue**: 如果遇到任何问题,请在 GitHub 上提交 Issue[Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues) * **提交 Issue**: 如果遇到任何问题,请在 GitHub 上提交 Issue[Awesome OpenWebUI Issues](https://github.com/Fu-Jie/awesome-openwebui/issues)
## 更新日志 ### 更新日志
### v1.2.4 完整历史请查看 GitHub 项目: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)
* **文档更新**: 同步了所有文档和代码文件的版本号。
### v1.2.3
* **列表标记保护增强**: 修复了列表标记 (`*`) 后跟普通文本和强调标记时,空格被错误剥离的问题(例如 `* U16 前锋` 变成 `*U16 前锋`)。
* **占位符支持**: 确认 4 个或更多下划线(如 `____`)会被正确视为占位符,不会被强调修复逻辑修改。
### v1.2.2
* **代码块缩进修复**: 修复了列表中嵌套代码块的缩进被错误剥离的问题。现在会正确保留嵌套代码块的缩进。
* **下划线强调语法支持**: 扩展强调空格修复以支持 `__` (双下划线加粗) 和 `___` (三下划线加粗斜体) 语法。
* **列表标记保护**: 修复了列表标记 (`*`) 后跟强调标记 (`**`) 被错误合并的 Bug例如 `* **是**` 变成 `***是**`)。添加了保护逻辑防止此问题。
* **测试套件**: 新增完整的 pytest 测试套件,包含 56 个测试用例,覆盖所有主要功能。
### v1.2.1
* **强调空格修复**: 新增了对强调标记内部多余空格的修复(例如 `** 文本 **` -> `**文本**`)。
* 采用递归方法处理嵌套强调(例如 `**加粗 _斜体 _**`)。
* 包含保护机制,防止误修改数学表达式(如 `2 * 3 * 4`)或列表变量。
* 通过 `enable_emphasis_spacing_fix` 开关控制(默认:开启)。
### v1.2.0
* **Details 标签支持**: 新增了对 `<details>` 标签的规范化支持。
* 确保在 `</details>` 闭合标签后添加空行,将思维内容与正文分隔开。
* 确保在自闭合 `<details ... />` 标签后添加换行,防止其干扰后续的 Markdown 标题(例如修复 `<details/>#标题`)。
* 包含保护机制,防止修改代码块内部的 `<details>` 标签。
### v1.1.2
* **Mermaid 连线标签保护**: 实现了全面的连线标签保护机制,防止连接线上的文字被误修改。现在支持所有 Mermaid 连线类型,包括实线 (`--`)、虚线 (`-.`) 和粗线 (`==`),无论是否带有箭头。
* **Bug 修复**: 修复了无箭头连线(如 `A -- text --- B`)未被正确保护的问题。
### v1.1.0
* **Mermaid 修复优化**: 改进了正则表达式以处理节点标签中的嵌套括号(如 `ID("标签 (文本)")`),并避免误匹配连接线上的文字。
* **HTML 保护机制优化**: 优化了 `_contains_html` 检测,允许 `<br/>`, `<b>`, `<i>` 等常见标签,确保包含这些标签的 Mermaid 图表能被正常规范化。
* **全角符号清理**: 修复了 `FULLWIDTH_MAP` 中的重复键名和错误的引号映射。
* **Bug 修复**: 修复了 Python 文件中缺失的 `Dict` 类型导入。

View File

@@ -1,30 +1,28 @@
# GitHub Copilot SDK Pipe for OpenWebUI # GitHub Copilot SDK Pipe for OpenWebUI
**Author:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **Version:** 0.2.3 | **Project:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **License:** MIT **Author:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **Version:** 0.3.0 | **Project:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **License:** MIT
This is an advanced Pipe function for [OpenWebUI](https://github.com/open-webui/open-webui) that allows you to use GitHub Copilot models (such as `gpt-5`, `gpt-5-mini`, `claude-sonnet-4.5`) directly within OpenWebUI. It is built upon the official [GitHub Copilot SDK for Python](https://github.com/github/copilot-sdk), providing a native integration experience. This is an advanced Pipe function for [OpenWebUI](https://github.com/open-webui/open-webui) that allows you to use GitHub Copilot models (such as `gpt-5`, `gpt-5-mini`, `claude-sonnet-4.5`) directly within OpenWebUI. It is built upon the official [GitHub Copilot SDK for Python](https://github.com/github/copilot-sdk), providing a native integration experience.
## 🚀 What's New (v0.2.3) ## 🚀 What's New (v0.3.0) - The Power of "Unified Ecosystem"
* **🧩 Per-user Overrides**: Added user-level overrides for `REASONING_EFFORT`, `CLI_PATH`, `DEBUG`, `SHOW_THINKING`, and `MODEL_ID`. * **🔌 Zero-Config Tool Bridge**: Automatically transforms your existing OpenWebUI Functions (Tools) into Copilot-compatible tools. **Copilot now has total access to your entire WebUI toolset!**
* **🧠 Thinking Output Reliability**: Thinking visibility now respects the user setting and is correctly passed into streaming. * **🔗 Dynamic MCP Discovery**: Seamlessly connects to MCP servers defined in **Admin Settings -> Connections**. No configuration files required—it just works.
* **📝 Formatting Enforcement**: Added automatic formatting hints to ensure outputs are well-structured (paragraphs, lists) and addressed "tight output" issues. * **⚡ High-Performance Async Engine**: Background CLI updates and optimized event-driven streaming ensure lightning-fast responses without UI lag.
* **🛡️ Robust Interoperability**: Advanced sanitization and dynamic Pydantic model generation ensure smooth integration even with complex third-party tools.
## ✨ Core Features ## ✨ Key Capabilities
* **🚀 Official SDK Integration**: Built on the official SDK for stability and reliability. * **🌉 The Ultimate Bridge**: The first and only plugin that creates a seamless bridge between **OpenWebUI Tools** and **GitHub Copilot SDK**.
* **🛠️ Custom Tools Support**: Example tools included (random number). Easy to extend with your own tools. * **🚀 Official & Native**: Built directly on the official Python SDK, providing the most stable and authentic Copilot experience.
* **💬 Multi-turn Conversation**: Automatically concatenates history context so Copilot understands your previous messages. * **🌊 Advanced Streaming (Thought Process)**: Supports full model reasoning/thinking display with typewriter effects.
* **🌊 Streaming Output**: Supports typewriter effect for fast responses. * **🖼️ Intelligent Multimodal**: Full support for images and attachments, enabling Copilot to "see" your workspace.
* **🖼 Multimodal Support**: Supports image uploads, automatically converting them to attachments for Copilot (requires model support). * **🛠 Effortless Setup**: Automatic CLI detection, version enforcement, and dependency management.
* **🛠️ Zero-config Installation**: Automatically detects and downloads the GitHub Copilot CLI, ready to use out of the box. * **🔑 Dual-Layer Security**: Supports secure OAuth flow for Chat and standard PAT for extended MCP capabilities.
* **🔑 Secure Authentication**: Supports Fine-grained Personal Access Tokens for minimized permissions.
* **🐛 Debug Mode**: Built-in detailed log output (browser console) for easy troubleshooting.
* **⚠️ Single Node Only**: Due to local session storage, this plugin currently supports single-node OpenWebUI deployment or multi-node with sticky sessions enabled.
## 📦 Installation & Usage ## Installation & Configuration
### 1. Import Function ### 1) Import Function
1. Open OpenWebUI. 1. Open OpenWebUI.
2. Go to **Workspace** -> **Functions**. 2. Go to **Workspace** -> **Functions**.
@@ -32,19 +30,17 @@ This is an advanced Pipe function for [OpenWebUI](https://github.com/open-webui/
4. Paste the content of `github_copilot_sdk.py` (or `github_copilot_sdk_cn.py` for Chinese) completely. 4. Paste the content of `github_copilot_sdk.py` (or `github_copilot_sdk_cn.py` for Chinese) completely.
5. Save. 5. Save.
### 2. Configure Valves (Settings) ### 2) Configure Valves (Settings)
Find "GitHub Copilot" in the function list and click the **⚙️ (Valves)** icon to configure: Find "GitHub Copilot" in the function list and click the **⚙️ (Valves)** icon to configure:
| Parameter | Description | Default | | Parameter | Description | Default |
| :--- | :--- | :--- | | :--- | :--- | :--- |
| **GH_TOKEN** | **(Required)** Your GitHub Token. | - | | **GH_TOKEN** | **(Required)** GitHub Access Token (PAT or OAuth Token). Access to Chat. | - |
| **MODEL_ID** | The model name to use. | `gpt-5-mini` |
| **CLI_PATH** | Path to the Copilot CLI. Will download automatically if not found. | `/usr/local/bin/copilot` |
| **DEBUG** | Whether to enable debug logs (output to browser console). | `False` | | **DEBUG** | Whether to enable debug logs (output to browser console). | `False` |
| **LOG_LEVEL** | Copilot CLI log level: none, error, warning, info, debug, all. | `error` | | **LOG_LEVEL** | Copilot CLI log level: none, error, warning, info, debug, all. | `error` |
| **SHOW_THINKING** | Show model reasoning/thinking process (requires streaming + model support). | `True` | | **SHOW_THINKING** | Show model reasoning/thinking process (requires streaming + model support). | `True` |
| **SHOW_WORKSPACE_INFO** | Show session workspace path and summary in debug mode. | `True` | | **COPILOT_CLI_VERSION** | Specific Copilot CLI version to install/enforce. | `0.0.405` |
| **EXCLUDE_KEYWORDS** | Exclude models containing these keywords (comma separated). | - | | **EXCLUDE_KEYWORDS** | Exclude models containing these keywords (comma separated). | - |
| **WORKSPACE_DIR** | Restricted workspace directory for file operations. | - | | **WORKSPACE_DIR** | Restricted workspace directory for file operations. | - |
| **INFINITE_SESSION** | Enable Infinite Sessions (automatic context compaction). | `True` | | **INFINITE_SESSION** | Enable Infinite Sessions (automatic context compaction). | `True` |
@@ -52,10 +48,10 @@ Find "GitHub Copilot" in the function list and click the **⚙️ (Valves)** ico
| **BUFFER_THRESHOLD** | Buffer exhaustion threshold (0.0-1.0). | `0.95` | | **BUFFER_THRESHOLD** | Buffer exhaustion threshold (0.0-1.0). | `0.95` |
| **TIMEOUT** | Timeout for each stream chunk (seconds). | `300` | | **TIMEOUT** | Timeout for each stream chunk (seconds). | `300` |
| **CUSTOM_ENV_VARS** | Custom environment variables (JSON format). | - | | **CUSTOM_ENV_VARS** | Custom environment variables (JSON format). | - |
| **REASONING_EFFORT** | Reasoning effort level: low, medium, high. `xhigh` is supported for gpt-5.2-codex. | `medium` | | **REASONING_EFFORT** | Reasoning effort level: low, medium, high. `xhigh` is supported for some models. | `medium` |
| **ENFORCE_FORMATTING** | Add formatting instructions to system prompt for better readability. | `True` | | **ENFORCE_FORMATTING** | Add formatting instructions to system prompt for better readability. | `True` |
| **ENABLE_TOOLS** | Enable custom tools (example: random number). | `False` | | **ENABLE_MCP_SERVER** | Enable Direct MCP Client connection (Recommended). | `True` |
| **AVAILABLE_TOOLS** | Available tools: 'all' or comma-separated list. | `all` | | **ENABLE_OPENWEBUI_TOOLS** | Enable OpenWebUI Tools (includes defined and server tools). | `True` |
#### User Valves (per-user overrides) #### User Valves (per-user overrides)
@@ -63,35 +59,30 @@ These optional settings can be set per user (overrides global Valves):
| Parameter | Description | Default | | Parameter | Description | Default |
| :--- | :--- | :--- | | :--- | :--- | :--- |
| **GH_TOKEN** | Personal GitHub Token (overrides global setting). | - |
| **REASONING_EFFORT** | Reasoning effort level (low/medium/high/xhigh). | - | | **REASONING_EFFORT** | Reasoning effort level (low/medium/high/xhigh). | - |
| **CLI_PATH** | Custom path to Copilot CLI. | - |
| **DEBUG** | Enable technical debug logs. | `False` | | **DEBUG** | Enable technical debug logs. | `False` |
| **SHOW_THINKING** | Show model reasoning/thinking process (requires streaming + model support). | `True` | | **SHOW_THINKING** | Show model reasoning/thinking process. | `True` |
| **MODEL_ID** | Custom model ID. | - | | **ENABLE_OPENWEBUI_TOOLS** | Enable OpenWebUI Tools (overrides global). | `True` |
| **ENABLE_MCP_SERVER** | Enable MCP server loading (overrides global). | `True` |
| **ENFORCE_FORMATTING** | Enforce formatting guidelines (overrides global). | `True` |
### 3. Using Custom Tools (🆕 Optional) ## ⭐ Support
This pipe includes **1 example tool** to demonstrate tool calling: If this plugin has been useful, a star on [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) is a big motivation for me. Thank you for the support.
* **🎲 generate_random_number**: Generate random integers ### Get Token
**To enable:** To use GitHub Copilot, you need a GitHub Personal Access Token (PAT) with appropriate permissions.
1. Set `ENABLE_TOOLS: true` in Valves **Steps to generate your token:**
2. Try: "Give me a random number"
**📚 For detailed usage and creating your own tools, see [TOOLS_USAGE.md](TOOLS_USAGE.md)**
### 4. Get GH_TOKEN
For security, it is recommended to use a **Fine-grained Personal Access Token**:
1. Visit [GitHub Token Settings](https://github.com/settings/tokens?type=beta). 1. Visit [GitHub Token Settings](https://github.com/settings/tokens?type=beta).
2. Click **Generate new token**. 2. Click **Generate new token (fine-grained)**.
3. **Repository access**: Select **Public repositories** (Required to access Copilot permissions). 3. **Repository access**: Select **Public Repositories** (simplest) or **All repositories**.
4. **Permissions**: 4. **Permissions**:
* Click **Account permissions**. * If you chose **All repositories**, you must click **Account permissions**.
* Find **Copilot Requests** (It defaults to **Read-only**, no selection needed). * Find **Copilot Requests**, and select **Access**.
5. Generate and copy the Token. 5. Generate and copy the Token.
## 📋 Dependencies ## 📋 Dependencies
@@ -101,14 +92,13 @@ This Pipe will automatically attempt to install the following dependencies:
* `github-copilot-sdk` (Python package) * `github-copilot-sdk` (Python package)
* `github-copilot-cli` (Binary file, installed via official script) * `github-copilot-cli` (Binary file, installed via official script)
## ⚠️ FAQ ## Troubleshooting ❓
* **Stuck on "Waiting..."**:
* Check if `GH_TOKEN` is correct and has `Copilot Requests` permission.
* **Images not recognized**: * **Images not recognized**:
* Ensure `MODEL_ID` is a model that supports multimodal input. * Ensure `MODEL_ID` is a model that supports multimodal input.
* **Thinking not shown**: * **Thinking not shown**:
* Ensure **streaming is enabled** and the selected model supports reasoning output. * Ensure **streaming is enabled** and the selected model supports reasoning output.
* **CLI Installation Failed**:
* Ensure the OpenWebUI container has internet access. ## Changelog
* You can manually download the CLI and specify `CLI_PATH` in Valves.
See the full history on GitHub: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)

View File

@@ -1,28 +1,26 @@
# GitHub Copilot SDK 官方管道 # GitHub Copilot SDK 官方管道
**作者:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **版本:** 0.2.3 | **项目:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **许可证:** MIT **作者:** [Fu-Jie](https://github.com/Fu-Jie/awesome-openwebui) | **版本:** 0.3.0 | **项目:** [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) | **许可证:** MIT
这是一个用于 [OpenWebUI](https://github.com/open-webui/open-webui) 的高级 Pipe 函数,允许你直接在 OpenWebUI 中使用 GitHub Copilot 模型(如 `gpt-5`, `gpt-5-mini`, `claude-sonnet-4.5`)。它基于官方 [GitHub Copilot SDK for Python](https://github.com/github/copilot-sdk) 构建,提供了原生级的集成体验。 这是一个用于 [OpenWebUI](https://github.com/open-webui/open-webui) 的高级 Pipe 函数,允许你直接在 OpenWebUI 中使用 GitHub Copilot 模型(如 `gpt-5`, `gpt-5-mini`, `claude-sonnet-4.5`)。它基于官方 [GitHub Copilot SDK for Python](https://github.com/github/copilot-sdk) 构建,提供了原生级的集成体验。
## 🚀 最新特性 (v0.2.3) ## 🚀 最新特性 (v0.3.0) - “统一生态”的力量
* **🧩 用户级覆盖**:新增 `REASONING_EFFORT``CLI_PATH``DEBUG``SHOW_THINKING``MODEL_ID` 的用户级覆盖。 * **🔌 零配置工具桥接 (Unified Tool Bridge)**: 自动将您现有的 OpenWebUI Functions (工具) 转换为 Copilot 兼容工具。**Copilot 现在可以无缝调用您手头所有的 WebUI 工具!**
* **🧠 思考输出可靠性**:思考显示会遵循用户设置,并正确传递到流式输出中 * **🔗 动态 MCP 自动发现**: 直接联动 OpenWebUI **管理面板 -> 连接**。无需编写任何配置文件,即插即用,瞬间扩展 Copilot 能力边界
* **📝 格式化输出增强**:自动优化输出格式(短句、段落、列表),并解决了在某些界面下显示过于紧凑的问题 * **⚡ 高性能异步引擎**: 异步 CLI 更新检查与高度优化的事件驱动流式处理,确保对话毫秒级响应
* **🛡️ 卓越的兼容性**: 独创的动态 Pydantic 模型生成技术,确保复杂工具参数在 Copilot 端也能得到精准验证。
## ✨ 核心特性 ## ✨ 核心能力
* **🚀 官方 SDK 集成**:基于官方 SDK稳定可靠 * **🌉 强大的生态桥接**: 首个且唯一完美打通 **OpenWebUI Tools****GitHub Copilot SDK** 的插件
* **🛠️ 自定义工具支持**:内置示例工具(随机数)。易于扩展自定义工具 * **🚀 官方原生产体验**: 基于官方 Python SDK 构建,提供最稳定、最纯正的 Copilot 交互体验
* **💬 多轮对话支持**自动拼接历史上下文Copilot 能理解你的前文 * **🌊 深度推理展示**: 完整支持模型思考过程 (Thinking Process) 的流式渲染
* **🌊 流式输出 (Streaming)**:支持打字机效果,响应迅速 * **🖼️ 智能多模态**: 支持图像识别与附件上传,让 Copilot 拥有视觉能力
* **🖼 多模态支持**:支持上传图片,自动转换为附件发送给 Copilot需模型支持 * **🛠 极简部署流程**: 自动检测环境、自动下载 CLI、自动管理依赖全自动化开箱即用
* **🛠️ 零配置安装**:自动检测并下载 GitHub Copilot CLI开箱即用 * **🔑 安全认证体系**: 完美支持 OAuth 授权与 PAT 模式,兼顾便捷与安全性
* **🔑 安全认证**:支持 Fine-grained Personal Access Tokens权限最小化。
* **🐛 调试模式**:内置详细的日志输出(浏览器控制台),方便排查问题。
* **⚠️ 仅支持单节点**:由于会话状态存储在本地,本插件目前仅支持 OpenWebUI 单节点部署,或开启了会话粘性 (Sticky Session) 的多节点集群。
## 📦 安装与使用 ## 安装与配置
### 1. 导入函数 ### 1. 导入函数
@@ -38,24 +36,22 @@
| 参数 | 说明 | 默认值 | | 参数 | 说明 | 默认值 |
| :--- | :--- | :--- | | :--- | :--- | :--- |
| **GH_TOKEN** | **(必填)** 你的 GitHub Token。 | - | | **GH_TOKEN** | **(必填)** GitHub 访问令牌 (PAT 或 OAuth Token)。用于聊天。 | - |
| **MODEL_ID** | 使用的模型名称。 | `gpt-5-mini` |
| **CLI_PATH** | Copilot CLI 的路径。如果未找到会自动下载。 | `/usr/local/bin/copilot` |
| **DEBUG** | 是否开启调试日志(输出到浏览器控制台)。 | `False` | | **DEBUG** | 是否开启调试日志(输出到浏览器控制台)。 | `False` |
| **LOG_LEVEL** | Copilot CLI 日志级别: none, error, warning, info, debug, all。 | `error` | | **LOG_LEVEL** | Copilot CLI 日志级别: none, error, warning, info, debug, all。 | `error` |
| **SHOW_THINKING** | 是否显示模型推理/思考过程(需开启流式 + 模型支持)。 | `True` | | **SHOW_THINKING** | 是否显示模型推理/思考过程(需开启流式 + 模型支持)。 | `True` |
| **SHOW_WORKSPACE_INFO** | 在调试模式下显示会话工作空间路径和摘要。 | `True` | | **COPILOT_CLI_VERSION** | 指定安装/强制使用的 Copilot CLI 版本。 | `0.0.405` |
| **EXCLUDE_KEYWORDS** | 排除包含这些关键词的模型 (逗号分隔)。 | - | | **EXCLUDE_KEYWORDS** | 排除包含这些关键词的模型逗号分隔。 | - |
| **WORKSPACE_DIR** | 文件操作的受限工作目录。 | - | | **WORKSPACE_DIR** | 文件操作的受限工作目录。 | - |
| **INFINITE_SESSION** | 启用无限会话 (自动上下文压缩)。 | `True` | | **INFINITE_SESSION** | 启用无限会话自动上下文压缩。 | `True` |
| **COMPACTION_THRESHOLD** | 后台压缩阈值 (0.0-1.0)。 | `0.8` | | **COMPACTION_THRESHOLD** | 后台压缩阈值 (0.0-1.0)。 | `0.8` |
| **BUFFER_THRESHOLD** | 缓冲耗尽阈值 (0.0-1.0)。 | `0.95` | | **BUFFER_THRESHOLD** | 缓冲耗尽阈值 (0.0-1.0)。 | `0.95` |
| **TIMEOUT** | 流式数据块超时时间 (秒)。 | `300` | | **TIMEOUT** | 每个流式分块超时(秒)。 | `300` |
| **CUSTOM_ENV_VARS** | 自定义环境变量 (JSON 格式)。 | - | | **CUSTOM_ENV_VARS** | 自定义环境变量 (JSON 格式)。 | - |
| **ENABLE_TOOLS** | 启用自定义工具 (示例:随机数)。 | `False` | | **REASONING_EFFORT** | 推理强度级别: low, medium, high. `xhigh` 仅部分模型支持。 | `medium` |
| **AVAILABLE_TOOLS** | 可用工具: 'all' 或逗号分隔列表。 | `all` | | **ENFORCE_FORMATTING** | 在系统提示词中添加格式化指导。 | `True` |
| **REASONING_EFFORT** | 推理强度级别low, medium, high。`gpt-5.2-codex`额外支持`xhigh`。 | `medium` | | **ENABLE_MCP_SERVER** | 启用直接 MCP 客户端连接 (建议)。 | `True` |
| **ENFORCE_FORMATTING** | 是否强制添加格式化指导,以提高输出可读性。 | `True` | | **ENABLE_OPENWEBUI_TOOLS** | 启用 OpenWebUI 工具 (包括自定义和服务器工具)。 | `True` |
#### 用户 Valves按用户覆盖 #### 用户 Valves按用户覆盖
@@ -63,36 +59,31 @@
| 参数 | 说明 | 默认值 | | 参数 | 说明 | 默认值 |
| :--- | :--- | :--- | | :--- | :--- | :--- |
| **GH_TOKEN** | 个人 GitHub Token覆盖全局设置。 | - |
| **REASONING_EFFORT** | 推理强度级别low/medium/high/xhigh。 | - | | **REASONING_EFFORT** | 推理强度级别low/medium/high/xhigh。 | - |
| **CLI_PATH** | 自定义 Copilot CLI 路径。 | - |
| **DEBUG** | 是否启用技术调试日志。 | `False` | | **DEBUG** | 是否启用技术调试日志。 | `False` |
| **SHOW_THINKING** | 是否显示思考过程(需开启流式 + 模型支持)。 | `True` | | **SHOW_THINKING** | 是否显示思考过程。 | `True` |
| **MODEL_ID** | 自定义模型 ID。 | - | | **ENABLE_OPENWEBUI_TOOLS** | 启用 OpenWebUI 工具(覆盖全局设置)。 | `True` |
| **ENABLE_MCP_SERVER** | 启用动态 MCP 服务器加载(覆盖全局设置)。 | `True` |
| **ENFORCE_FORMATTING** | 强制启用格式化指导(覆盖全局设置)。 | `True` |
### 3. 使用自定义工具 (🆕 可选) ## ⭐ 支持
本 Pipe 内置了 **1 个示例工具**来展示工具调用功能: 如果这个插件对你有帮助,欢迎到 [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui) 点个 Star这将是我持续改进的动力感谢支持。
* **🎲 generate_random_number**:生成随机整数 ### 获取 Token
**启用方法:** 要使用 GitHub Copilot您需要一个具有适当权限的 GitHub 个人访问令牌 (PAT)。
1. 在 Valves 中设置 `ENABLE_TOOLS: true` **获取步骤:**
2. 尝试问:“给我一个随机数”
**📚 详细使用说明和创建自定义工具,请参阅 [TOOLS_USAGE.md](TOOLS_USAGE.md)** 1. 访问 [GitHub 令牌设置](https://github.com/settings/tokens?type=beta)。
2. 点击 **Generate new token (fine-grained)**
### 4. 获取 GH_TOKEN 3. **Repository access**: 选择 **Public Repositories** (最简单) 或 **All repositories**
为了安全起见,推荐使用 **Fine-grained Personal Access Token**
1. 访问 [GitHub Token Settings](https://github.com/settings/tokens?type=beta)。
2. 点击 **Generate new token**
3. **Repository access**: 选择 **Public repositories** (必须选择此项才能看到 Copilot 权限)。
4. **Permissions**: 4. **Permissions**:
* 点击 **Account permissions** * 如果您选择了 **All repositories**,则必须点击 **Account permissions**
* 找到 **Copilot Requests** (默认即为 **Read-only**,无需手动修改) * 找到 **Copilot Requests**,选择 **Access**
5. 生成并复制 Token 5. 生成并复制令牌
## 📋 依赖说明 ## 📋 依赖说明
@@ -101,14 +92,13 @@
* `github-copilot-sdk` (Python 包) * `github-copilot-sdk` (Python 包)
* `github-copilot-cli` (二进制文件,通过官方脚本安装) * `github-copilot-cli` (二进制文件,通过官方脚本安装)
## ⚠️ 常见问题 ## 故障排除 (Troubleshooting) ❓
* **一直显示 "Waiting..."** * **图片及多模态使用说明**
* 检查 `GH_TOKEN` 是否正确且拥有 `Copilot Requests` 权限。
* **图片无法识别**
* 确保 `MODEL_ID` 是支持多模态的模型。 * 确保 `MODEL_ID` 是支持多模态的模型。
* **CLI 安装失败**
* 确保 OpenWebUI 容器有外网访问权限。
* 你可以手动下载 CLI 并挂载到容器中,然后在 Valves 中指定 `CLI_PATH`
* **看不到思考过程** * **看不到思考过程**
* 确认已开启**流式输出**,且所选模型支持推理输出。 * 确认已开启**流式输出**,且所选模型支持推理输出。
## 更新日志
完整历史请查看 GitHub 项目: [Awesome OpenWebUI](https://github.com/Fu-Jie/awesome-openwebui)

View File

@@ -0,0 +1,187 @@
# 🛠️ Custom Tools Usage / 自定义工具使用指南
## Overview / 概览
This pipe supports **OpenWebUI Native Tools** (Functions) and **Custom Python Tools**.
本 Pipe 支持 **OpenWebUI 原生工具** (Functions) 和 **自定义 Python 工具**
---
## 🚀 OpenWebUI Native Tools / OpenWebUI 原生工具 (v0.3.0)
**New in v0.3.0**: You can use any tool defined in OpenWebUI directly with Copilot.
**v0.3.0 新增**: 您可以直接在 Copilot 中使用 OpenWebUI 中定义的任何工具。
**How to use / 如何使用:**
1. Go to **Workspace** -> **Tools**.
2. Create a tool (e.g. `get_weather`).
3. In Copilot Chat settings (Valves), ensure `ENABLE_OPENWEBUI_TOOLS` is `True` (default).
4. Ask Copilot: "Search for the latest news" or "Check weather".
**Note / 注意:**
- Tool names are automatically sanitized to match Copilot SDK requirements (e.g. `my.tool` -> `my_tool`).
- 工具名称会自动净化以符合 Copilot SDK 要求(例如 `my.tool` 变为 `my_tool`)。
---
## 📦 Python Custom Tools / Python 自定义工具
This pipe includes **1 example custom tool** that demonstrates how to use GitHub Copilot SDK's tool calling feature directly in Python code.
本 Pipe 包含 **1 个示例自定义工具**,展示如何使用 GitHub Copilot SDK 的工具调用功能。
### 1. `generate_random_number` / 生成随机数
**Description:** Generate a random integer
**描述:** 生成随机整数
**Parameters / 参数:**
- `min` (optional): Minimum value (default: 1)
- `max` (optional): Maximum value (default: 100)
- `min` (可选): 最小值 (默认: 1)
- `max` (可选): 最大值 (默认: 100)
**Example / 示例:**
```
User: "Give me a random number between 1 and 10"
Copilot: [calls generate_random_number with min=1, max=10] "Generated random number: 7"
用户: "给我一个 1 到 10 之间的随机数"
Copilot: [调用 generate_random_number参数 min=1, max=10] "生成的随机数: 7"
```
---
## ⚙️ Configuration / 配置
### Enable Tools / 启用工具
In Valves configuration:
在 Valves 配置中:
```
ENABLE_TOOLS: true
AVAILABLE_TOOLS: all
```
### Select Specific Tools / 选择特定工具
Instead of enabling all tools, specify which ones to use:
不启用所有工具,而是指定要使用的工具:
```
ENABLE_TOOLS: true
AVAILABLE_TOOLS: generate_random_number
```
---
## 🔧 How Tool Calling Works / 工具调用的工作原理
```
1. User asks a question / 用户提问
2. Copilot decides if it needs a tool / Copilot 决定是否需要工具
3. If yes, Copilot calls the appropriate tool / 如果需要,调用相应工具
4. Tool executes and returns result / 工具执行并返回结果
5. Copilot uses the result to answer / Copilot 使用结果回答
```
### Visual Feedback / 可视化反馈
When tools are called, you'll see:
当工具被调用时,你会看到:
```
🔧 **Calling tool**: `generate_random_number`
✅ **Tool `generate_random_number` completed**
Generated random number: 7
```
---
## 📚 Creating Your Own Tools / 创建自定义工具
Want to add your own Python tools? Follow this pattern (module-level tools):
想要添加自己的 Python 工具?遵循这个模式(模块级工具):
```python
from pydantic import BaseModel, Field
from copilot import define_tool
class MyToolParams(BaseModel):
param_name: str = Field(description="Parameter description")
@define_tool(description="Clear description of what the tool does and when to use it")
async def my_tool(params: MyToolParams) -> str:
# Do something
result = do_something(params.param_name)
return f"Result: {result}"
```
Then register it in `_initialize_custom_tools()`:
然后将它添加到 `_initialize_custom_tools()`:
```python
def _initialize_custom_tools(self):
if not self.valves.ENABLE_TOOLS:
return []
all_tools = {
"generate_random_number": generate_random_number,
"my_tool": my_tool, # ✅ Add here
}
if self.valves.AVAILABLE_TOOLS == "all":
return list(all_tools.values())
enabled = [t.strip() for t in self.valves.AVAILABLE_TOOLS.split(",")]
return [all_tools[name] for name in enabled if name in all_tools]
```
---
## ⚠️ Important Notes / 重要说明
### Security / 安全性
- Tools run in the same process as the pipe
- Be careful with tools that execute code or access files
- Always validate input parameters
- 工具在与 Pipe 相同的进程中运行
- 谨慎处理执行代码或访问文件的工具
- 始终验证输入参数
### Performance / 性能
- Tool execution is synchronous during streaming
- Long-running tools may cause delays
- Consider adding timeouts for external API calls
- 工具执行在流式传输期间是同步的
- 长时间运行的工具可能导致延迟
- 考虑为外部 API 调用添加超时
### Debugging / 调试
- Enable `DEBUG: true` to see tool events in the browser console
- Check tool calls in `🔧 Calling tool` messages
- Tool errors are displayed in the response
- 启用 `DEBUG: true` 在浏览器控制台查看工具事件
-`🔧 Calling tool` 消息中检查工具调用
- 工具错误会显示在响应中
---
**Version:** 0.3.0
**Last Updated:** 2026-02-05

View File

@@ -5,11 +5,12 @@ author_url: https://github.com/Fu-Jie/awesome-openwebui
funding_url: https://github.com/open-webui funding_url: https://github.com/open-webui
openwebui_id: ce96f7b4-12fc-4ac3-9a01-875713e69359 openwebui_id: ce96f7b4-12fc-4ac3-9a01-875713e69359
description: Integrate GitHub Copilot SDK. Supports dynamic models, multi-turn conversation, streaming, multimodal input, infinite sessions, and frontend debug logging. description: Integrate GitHub Copilot SDK. Supports dynamic models, multi-turn conversation, streaming, multimodal input, infinite sessions, and frontend debug logging.
version: 0.2.3 version: 0.3.0
requirements: github-copilot-sdk requirements: github-copilot-sdk==0.1.22
""" """
import os import os
import re
import json import json
import base64 import base64
import tempfile import tempfile
@@ -17,44 +18,34 @@ import asyncio
import logging import logging
import shutil import shutil
import subprocess import subprocess
import hashlib
from pathlib import Path
from typing import Optional, Union, AsyncGenerator, List, Any, Dict from typing import Optional, Union, AsyncGenerator, List, Any, Dict
from pydantic import BaseModel, Field from types import SimpleNamespace
from pydantic import BaseModel, Field, create_model
# Import copilot SDK modules # Import copilot SDK modules
from copilot import CopilotClient, define_tool from copilot import CopilotClient, define_tool
# Import Tool Server Connections and Tool System from OpenWebUI Config
from open_webui.config import TOOL_SERVER_CONNECTIONS
from open_webui.utils.tools import get_tools as get_openwebui_tools
from open_webui.models.tools import Tools
from open_webui.models.users import Users
# Setup logger # Setup logger
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
class RandomNumberParams(BaseModel):
min: int = Field(description="Minimum value (inclusive)")
max: int = Field(description="Maximum value (inclusive)")
@define_tool(description="Generate a random integer within a specified range.")
async def generate_random_number(params: RandomNumberParams) -> str:
import random
if params.min >= params.max:
raise ValueError("min must be less than max")
number = random.randint(params.min, params.max)
return f"Generated random number: {number}"
class Pipe: class Pipe:
class Valves(BaseModel): class Valves(BaseModel):
GH_TOKEN: str = Field( GH_TOKEN: str = Field(
default="", default="",
description="GitHub Fine-grained Token (Requires 'Copilot Requests' permission)", description="GitHub Fine-grained Token (Requires 'Copilot Requests' permission)",
) )
MODEL_ID: str = Field( COPILOT_CLI_VERSION: str = Field(
default="gpt-5-mini", default="0.0.405",
description="Default Copilot model name (used when dynamic fetching fails)", description="Specific Copilot CLI version to install/enforce (e.g. '0.0.405'). Leave empty for latest.",
)
CLI_PATH: str = Field(
default="/usr/local/bin/copilot",
description="Path to Copilot CLI",
) )
DEBUG: bool = Field( DEBUG: bool = Field(
default=False, default=False,
@@ -68,10 +59,6 @@ class Pipe:
default=True, default=True,
description="Show model reasoning/thinking process", description="Show model reasoning/thinking process",
) )
SHOW_WORKSPACE_INFO: bool = Field(
default=True,
description="Show session workspace path and summary in debug mode",
)
EXCLUDE_KEYWORDS: str = Field( EXCLUDE_KEYWORDS: str = Field(
default="", default="",
description="Exclude models containing these keywords (comma separated, e.g.: codex, haiku)", description="Exclude models containing these keywords (comma separated, e.g.: codex, haiku)",
@@ -100,13 +87,14 @@ class Pipe:
default="", default="",
description='Custom environment variables (JSON format, e.g., {"VAR": "value"})', description='Custom environment variables (JSON format, e.g., {"VAR": "value"})',
) )
ENABLE_TOOLS: bool = Field(
default=False, ENABLE_OPENWEBUI_TOOLS: bool = Field(
description="Enable custom tools (example: random number)", default=True,
description="Enable OpenWebUI Tools (includes defined Tools and Tool Server Tools).",
) )
AVAILABLE_TOOLS: str = Field( ENABLE_MCP_SERVER: bool = Field(
default="all", default=True,
description="Available tools: 'all' or comma-separated list (e.g., 'generate_random_number')", description="Enable Direct MCP Client connection (Recommended).",
) )
REASONING_EFFORT: str = Field( REASONING_EFFORT: str = Field(
default="medium", default="medium",
@@ -118,14 +106,14 @@ class Pipe:
) )
class UserValves(BaseModel): class UserValves(BaseModel):
GH_TOKEN: str = Field(
default="",
description="Personal GitHub Fine-grained Token (overrides global setting)",
)
REASONING_EFFORT: str = Field( REASONING_EFFORT: str = Field(
default="", default="",
description="Reasoning effort level (low, medium, high, xhigh). Leave empty to use global setting.", description="Reasoning effort level (low, medium, high, xhigh). Leave empty to use global setting.",
) )
CLI_PATH: str = Field(
default="",
description="Custom path to Copilot CLI. Leave empty to use global setting.",
)
DEBUG: bool = Field( DEBUG: bool = Field(
default=False, default=False,
description="Enable technical debug logs (connection info, etc.)", description="Enable technical debug logs (connection info, etc.)",
@@ -134,9 +122,18 @@ class Pipe:
default=True, default=True,
description="Show model reasoning/thinking process", description="Show model reasoning/thinking process",
) )
MODEL_ID: str = Field( ENABLE_OPENWEBUI_TOOLS: bool = Field(
default="", default=True,
description="Custom model ID (e.g. gpt-4o). Leave empty to use global default.", description="Enable OpenWebUI Tools (includes defined Tools and Tool Server Tools).",
)
ENABLE_MCP_SERVER: bool = Field(
default=True,
description="Enable dynamic MCP server loading (overrides global).",
)
ENFORCE_FORMATTING: bool = Field(
default=True,
description="Enforce formatting guidelines (overrides global)",
) )
def __init__(self): def __init__(self):
@@ -147,6 +144,7 @@ class Pipe:
self.temp_dir = tempfile.mkdtemp(prefix="copilot_images_") self.temp_dir = tempfile.mkdtemp(prefix="copilot_images_")
self.thinking_started = False self.thinking_started = False
self._model_cache = [] # Model list cache self._model_cache = [] # Model list cache
self._last_update_check = 0 # Timestamp of last CLI update check
def __del__(self): def __del__(self):
try: try:
@@ -183,23 +181,269 @@ class Pipe:
# ==================== Custom Tool Examples ==================== # ==================== Custom Tool Examples ====================
# Tool registration: Add @define_tool decorated functions at module level, # Tool registration: Add @define_tool decorated functions at module level,
# then register them in _initialize_custom_tools() -> all_tools dict. # then register them in _initialize_custom_tools() -> all_tools dict.
def _initialize_custom_tools(self): async def _initialize_custom_tools(self, __user__=None, __event_call__=None):
"""Initialize custom tools based on configuration""" """Initialize custom tools based on configuration"""
if not self.valves.ENABLE_TOOLS: if not self.valves.ENABLE_OPENWEBUI_TOOLS:
return [] return []
# Define all available tools (register new tools here) # Load OpenWebUI tools dynamically
all_tools = { openwebui_tools = await self._load_openwebui_tools(
"generate_random_number": generate_random_number, __user__=__user__, __event_call__=__event_call__
)
return openwebui_tools
def _json_schema_to_python_type(self, schema: dict) -> Any:
"""Convert JSON Schema type to Python type for Pydantic models."""
if not isinstance(schema, dict):
return Any
schema_type = schema.get("type")
if isinstance(schema_type, list):
schema_type = next((t for t in schema_type if t != "null"), schema_type[0])
if schema_type == "string":
return str
if schema_type == "integer":
return int
if schema_type == "number":
return float
if schema_type == "boolean":
return bool
if schema_type == "object":
return Dict[str, Any]
if schema_type == "array":
items_schema = schema.get("items", {})
item_type = self._json_schema_to_python_type(items_schema)
return List[item_type]
return Any
def _convert_openwebui_tool(self, tool_name: str, tool_dict: dict):
"""Convert OpenWebUI tool definition to Copilot SDK tool."""
# Sanitize tool name to match pattern ^[a-zA-Z0-9_-]+$
sanitized_tool_name = re.sub(r"[^a-zA-Z0-9_-]", "_", tool_name)
# If sanitized name is empty or consists only of separators (e.g. pure Chinese name), generate a fallback name
if not sanitized_tool_name or re.match(r"^[_.-]+$", sanitized_tool_name):
hash_suffix = hashlib.md5(tool_name.encode("utf-8")).hexdigest()[:8]
sanitized_tool_name = f"tool_{hash_suffix}"
if sanitized_tool_name != tool_name:
logger.debug(
f"Sanitized tool name '{tool_name}' to '{sanitized_tool_name}'"
)
spec = tool_dict.get("spec", {}) if isinstance(tool_dict, dict) else {}
params_schema = spec.get("parameters", {}) if isinstance(spec, dict) else {}
properties = params_schema.get("properties", {})
required = params_schema.get("required", [])
if not isinstance(properties, dict):
properties = {}
if not isinstance(required, list):
required = []
required_set = set(required)
fields = {}
for param_name, param_schema in properties.items():
param_type = self._json_schema_to_python_type(param_schema)
description = ""
if isinstance(param_schema, dict):
description = param_schema.get("description", "")
if param_name in required_set:
if description:
fields[param_name] = (
param_type,
Field(..., description=description),
)
else:
fields[param_name] = (param_type, ...)
else:
optional_type = Optional[param_type]
if description:
fields[param_name] = (
optional_type,
Field(default=None, description=description),
)
else:
fields[param_name] = (optional_type, None)
if fields:
ParamsModel = create_model(f"{sanitized_tool_name}_Params", **fields)
else:
ParamsModel = create_model(f"{sanitized_tool_name}_Params")
tool_callable = tool_dict.get("callable")
tool_description = spec.get("description", "") if isinstance(spec, dict) else ""
if not tool_description and isinstance(spec, dict):
tool_description = spec.get("summary", "")
# Critical: If the tool name was sanitized (e.g. Chinese -> Hash), instructions are lost.
# We must inject the original name into the description so the model knows what it is.
if sanitized_tool_name != tool_name:
tool_description = f"Function '{tool_name}': {tool_description}"
async def _tool(params):
payload = params.model_dump() if hasattr(params, "model_dump") else {}
return await tool_callable(**payload)
_tool.__name__ = sanitized_tool_name
_tool.__doc__ = tool_description
# Debug log for tool conversion
logger.debug(
f"Converting tool '{sanitized_tool_name}': {tool_description[:50]}..."
)
# Core Fix: Explicitly pass params_type and name
return define_tool(
name=sanitized_tool_name,
description=tool_description,
params_type=ParamsModel,
)(_tool)
def _build_openwebui_request(self):
"""Build a minimal request-like object for OpenWebUI tool loading."""
app_state = SimpleNamespace(
config=SimpleNamespace(
TOOL_SERVER_CONNECTIONS=TOOL_SERVER_CONNECTIONS.value
),
TOOLS={},
)
app = SimpleNamespace(state=app_state)
request = SimpleNamespace(
app=app,
cookies={},
state=SimpleNamespace(token=SimpleNamespace(credentials="")),
)
return request
async def _load_openwebui_tools(self, __user__=None, __event_call__=None):
"""Load OpenWebUI tools and convert them to Copilot SDK tools."""
if isinstance(__user__, (list, tuple)):
user_data = __user__[0] if __user__ else {}
elif isinstance(__user__, dict):
user_data = __user__
else:
user_data = {}
if not user_data:
return []
user_id = user_data.get("id") or user_data.get("user_id")
if not user_id:
return []
user = Users.get_user_by_id(user_id)
if not user:
return []
# 1. Get User defined tools (Python scripts)
tool_items = Tools.get_tools_by_user_id(user_id, permission="read")
tool_ids = [tool.id for tool in tool_items] if tool_items else []
# 2. Get OpenAPI Tool Server tools
# We manually add enabled OpenAPI servers to the list because Tools.get_tools_by_user_id only checks the DB.
# open_webui.utils.tools.get_tools handles the actual loading and access control.
if hasattr(TOOL_SERVER_CONNECTIONS, "value"):
for server in TOOL_SERVER_CONNECTIONS.value:
# We only add 'openapi' servers here because get_tools currently only supports 'openapi' (or defaults to it).
# MCP tools are handled separately via ENABLE_MCP_SERVER.
if server.get("type") == "openapi":
# Format expected by get_tools: "server:<id>" implies types="openapi"
server_id = server.get("id")
if server_id:
tool_ids.append(f"server:{server_id}")
if not tool_ids:
return []
request = self._build_openwebui_request()
extra_params = {
"__request__": request,
"__user__": user_data,
"__event_emitter__": None,
"__event_call__": __event_call__,
"__chat_id__": None,
"__message_id__": None,
"__model_knowledge__": [],
} }
# Filter based on configuration tools_dict = await get_openwebui_tools(request, tool_ids, user, extra_params)
if self.valves.AVAILABLE_TOOLS == "all": if not tools_dict:
return list(all_tools.values()) return []
# Only enable specified tools converted_tools = []
enabled = [t.strip() for t in self.valves.AVAILABLE_TOOLS.split(",")] for tool_name, tool_def in tools_dict.items():
return [all_tools[name] for name in enabled if name in all_tools] try:
converted_tools.append(
self._convert_openwebui_tool(tool_name, tool_def)
)
except Exception as e:
await self._emit_debug_log(
f"Failed to load OpenWebUI tool '{tool_name}': {e}",
__event_call__,
)
return converted_tools
def _parse_mcp_servers(self) -> Optional[dict]:
"""
Dynamically load MCP servers from OpenWebUI TOOL_SERVER_CONNECTIONS.
Returns a dict of mcp_servers compatible with CopilotClient.
"""
if not self.valves.ENABLE_MCP_SERVER:
return None
mcp_servers = {}
# Iterate over OpenWebUI Tool Server Connections
if hasattr(TOOL_SERVER_CONNECTIONS, "value"):
connections = TOOL_SERVER_CONNECTIONS.value
else:
connections = []
for conn in connections:
if conn.get("type") == "mcp":
info = conn.get("info", {})
# Use ID from info or generate one
raw_id = info.get("id", f"mcp-server-{len(mcp_servers)}")
# Sanitize server_id (using same logic as tools)
server_id = re.sub(r"[^a-zA-Z0-9_-]", "_", raw_id)
if not server_id or re.match(r"^[_.-]+$", server_id):
hash_suffix = hashlib.md5(raw_id.encode("utf-8")).hexdigest()[:8]
server_id = f"server_{hash_suffix}"
url = conn.get("url")
if not url:
continue
# Build Headers (Handle Auth)
headers = {}
auth_type = conn.get("auth_type", "bearer")
key = conn.get("key", "")
if auth_type == "bearer" and key:
headers["Authorization"] = f"Bearer {key}"
elif auth_type == "basic" and key:
headers["Authorization"] = f"Basic {key}"
# Merge custom headers if any
custom_headers = conn.get("headers", {})
if isinstance(custom_headers, dict):
headers.update(custom_headers)
mcp_servers[server_id] = {
"type": "http",
"url": url,
"headers": headers,
"tools": ["*"], # Enable all tools by default
}
return mcp_servers if mcp_servers else None
async def _emit_debug_log(self, message: str, __event_call__=None): async def _emit_debug_log(self, message: str, __event_call__=None):
"""Emit debug log to frontend (console) when DEBUG is enabled.""" """Emit debug log to frontend (console) when DEBUG is enabled."""
@@ -390,9 +634,31 @@ class Pipe:
return system_prompt_content, system_prompt_source return system_prompt_content, system_prompt_source
def _get_workspace_dir(self) -> str:
"""Get the effective workspace directory with smart defaults."""
if self.valves.WORKSPACE_DIR:
return self.valves.WORKSPACE_DIR
# Smart default for OpenWebUI container
if os.path.exists("/app/backend/data"):
cwd = "/app/backend/data/copilot_workspace"
else:
# Local fallback: subdirectory in current working directory
cwd = os.path.join(os.getcwd(), "copilot_workspace")
# Ensure directory exists
if not os.path.exists(cwd):
try:
os.makedirs(cwd, exist_ok=True)
except Exception as e:
print(f"Error creating workspace {cwd}: {e}")
return os.getcwd() # Fallback to CWD if creation fails
return cwd
def _build_client_config(self, body: dict) -> dict: def _build_client_config(self, body: dict) -> dict:
"""Build CopilotClient config from valves and request body.""" """Build CopilotClient config from valves and request body."""
cwd = self.valves.WORKSPACE_DIR if self.valves.WORKSPACE_DIR else os.getcwd() cwd = self._get_workspace_dir()
client_config = {} client_config = {}
if os.environ.get("COPILOT_CLI_PATH"): if os.environ.get("COPILOT_CLI_PATH"):
client_config["cli_path"] = os.environ["COPILOT_CLI_PATH"] client_config["cli_path"] = os.environ["COPILOT_CLI_PATH"]
@@ -418,7 +684,6 @@ class Pipe:
custom_tools: List[Any], custom_tools: List[Any],
system_prompt_content: Optional[str], system_prompt_content: Optional[str],
is_streaming: bool, is_streaming: bool,
reasoning_effort: str = "",
): ):
"""Build SessionConfig for Copilot SDK.""" """Build SessionConfig for Copilot SDK."""
from copilot.types import SessionConfig, InfiniteSessionConfig from copilot.types import SessionConfig, InfiniteSessionConfig
@@ -470,9 +735,9 @@ class Pipe:
"infinite_sessions": infinite_session_config, "infinite_sessions": infinite_session_config,
} }
# Add reasoning_effort if not default (medium) mcp_servers = self._parse_mcp_servers()
if reasoning_effort and reasoning_effort.lower() != "medium": if mcp_servers:
session_params["reasoning_effort"] = reasoning_effort.lower() session_params["mcp_servers"] = mcp_servers
return SessionConfig(**session_params) return SessionConfig(**session_params)
@@ -628,8 +893,8 @@ class Pipe:
# Return default model on failure # Return default model on failure
return [ return [
{ {
"id": f"{self.id}-{self.valves.MODEL_ID}", "id": f"{self.id}-gpt-5-mini",
"name": f"GitHub Copilot ({self.valves.MODEL_ID})", "name": f"GitHub Copilot (gpt-5-mini)",
} }
] ]
finally: finally:
@@ -638,8 +903,8 @@ class Pipe:
await self._emit_debug_log(f"Pipes Error: {e}") await self._emit_debug_log(f"Pipes Error: {e}")
return [ return [
{ {
"id": f"{self.id}-{self.valves.MODEL_ID}", "id": f"{self.id}-gpt-5-mini",
"name": f"GitHub Copilot ({self.valves.MODEL_ID})", "name": f"GitHub Copilot (gpt-5-mini)",
} }
] ]
@@ -654,30 +919,93 @@ class Pipe:
return client return client
def _setup_env(self, __event_call__=None): def _setup_env(self, __event_call__=None):
cli_path = self.valves.CLI_PATH # Default CLI path logic
found = False cli_path = "/usr/local/bin/copilot"
if os.environ.get("COPILOT_CLI_PATH"):
cli_path = os.environ["COPILOT_CLI_PATH"]
target_version = self.valves.COPILOT_CLI_VERSION.strip()
found = False
current_version = None
# internal helper to get version
def get_cli_version(path):
try:
output = (
subprocess.check_output(
[path, "--version"], stderr=subprocess.STDOUT
)
.decode()
.strip()
)
# Copilot CLI version output format is usually just the version number or "copilot version X.Y.Z"
# We try to extract X.Y.Z
match = re.search(r"(\d+\.\d+\.\d+)", output)
return match.group(1) if match else output
except Exception:
return None
# Check default path
if os.path.exists(cli_path): if os.path.exists(cli_path):
found = True found = True
current_version = get_cli_version(cli_path)
# Check system path if not found
if not found: if not found:
sys_path = shutil.which("copilot") sys_path = shutil.which("copilot")
if sys_path: if sys_path:
cli_path = sys_path cli_path = sys_path
found = True found = True
current_version = get_cli_version(cli_path)
# Determine if we need to install or update
should_install = False
install_reason = ""
if not found: if not found:
should_install = True
install_reason = "CLI not found"
elif target_version:
# Normalize versions for comparison (remove 'v' prefix)
norm_target = target_version.lstrip("v")
norm_current = current_version.lstrip("v") if current_version else ""
if norm_target != norm_current:
should_install = True
install_reason = f"Version mismatch (Current: {current_version}, Target: {target_version})"
if should_install:
if self.valves.DEBUG:
self._emit_debug_log_sync(
f"Installing Copilot CLI: {install_reason}...", __event_call__
)
try: try:
env = os.environ.copy()
if target_version:
env["VERSION"] = target_version
subprocess.run( subprocess.run(
"curl -fsSL https://gh.io/copilot-install | bash", "curl -fsSL https://gh.io/copilot-install | bash",
shell=True, shell=True,
check=True, check=True,
env=env,
) )
if os.path.exists(self.valves.CLI_PATH):
cli_path = self.valves.CLI_PATH # Check default install location first, then system path
if os.path.exists("/usr/local/bin/copilot"):
cli_path = "/usr/local/bin/copilot"
found = True found = True
except: elif shutil.which("copilot"):
pass cli_path = shutil.which("copilot")
found = True
if found:
current_version = get_cli_version(cli_path)
except Exception as e:
if self.valves.DEBUG:
self._emit_debug_log_sync(
f"Failed to install Copilot CLI: {e}", __event_call__
)
if found: if found:
os.environ["COPILOT_CLI_PATH"] = cli_path os.environ["COPILOT_CLI_PATH"] = cli_path
@@ -687,25 +1015,9 @@ class Pipe:
if self.valves.DEBUG: if self.valves.DEBUG:
self._emit_debug_log_sync( self._emit_debug_log_sync(
f"Copilot CLI found at: {cli_path}", __event_call__ f"Copilot CLI found at: {cli_path} (Version: {current_version})",
__event_call__,
) )
try:
# Try to get version to confirm it's executable
ver = (
subprocess.check_output(
[cli_path, "--version"], stderr=subprocess.STDOUT
)
.decode()
.strip()
)
self._emit_debug_log_sync(
f"Copilot CLI Version: {ver}", __event_call__
)
except Exception as e:
self._emit_debug_log_sync(
f"Warning: Copilot CLI found but failed to run: {e}",
__event_call__,
)
else: else:
if self.valves.DEBUG: if self.valves.DEBUG:
self._emit_debug_log_sync( self._emit_debug_log_sync(
@@ -722,6 +1034,8 @@ class Pipe:
"Warning: GH_TOKEN is not set.", __event_call__ "Warning: GH_TOKEN is not set.", __event_call__
) )
self._sync_mcp_config(__event_call__)
def _process_images(self, messages, __event_call__=None): def _process_images(self, messages, __event_call__=None):
attachments = [] attachments = []
text_content = "" text_content = ""
@@ -779,8 +1093,8 @@ class Pipe:
"gpt-5.2-codex" "gpt-5.2-codex"
not in self._collect_model_ids( not in self._collect_model_ids(
body={}, body={},
request_model=self.valves.MODEL_ID, request_model=self.id,
real_model_id=self.valves.MODEL_ID, real_model_id=None,
)[0].lower() )[0].lower()
): ):
# Fallback to high if not supported # Fallback to high if not supported
@@ -823,6 +1137,53 @@ class Pipe:
except Exception as e: except Exception as e:
self._emit_debug_log_sync(f"Config sync check failed: {e}", __event_call__) self._emit_debug_log_sync(f"Config sync check failed: {e}", __event_call__)
async def _update_copilot_cli(self, cli_path: str, __event_call__=None):
"""Async task to update Copilot CLI if needed."""
import time
try:
# Check frequency (e.g., once every hour)
now = time.time()
if now - self._last_update_check < 3600:
return
self._last_update_check = now
# Simple check if "update" command is available or if we should just run it
# The user requested "async attempt to update copilot cli"
if self.valves.DEBUG:
self._emit_debug_log_sync(
"Triggering async Copilot CLI update check...", __event_call__
)
# We create a subprocess to run the update
process = await asyncio.create_subprocess_exec(
cli_path,
"update",
stdout=asyncio.subprocess.PIPE,
stderr=asyncio.subprocess.PIPE,
)
stdout, stderr = await process.communicate()
if self.valves.DEBUG:
output = stdout.decode().strip() or stderr.decode().strip()
if output:
self._emit_debug_log_sync(
f"Async CLI Update result: {output}", __event_call__
)
except Exception as e:
if self.valves.DEBUG:
self._emit_debug_log_sync(
f"Async CLI Update failed: {e}", __event_call__
)
def _sync_mcp_config(self, __event_call__=None):
"""Deprecated: MCP config is now handled dynamically via session config."""
pass
# ==================== Internal Implementation ==================== # ==================== Internal Implementation ====================
# _pipe_impl() contains the main request handling logic. # _pipe_impl() contains the main request handling logic.
# ================================================================ # ================================================================
@@ -835,12 +1196,22 @@ class Pipe:
__event_call__=None, __event_call__=None,
) -> Union[str, AsyncGenerator]: ) -> Union[str, AsyncGenerator]:
self._setup_env(__event_call__) self._setup_env(__event_call__)
cwd = self._get_workspace_dir()
if self.valves.DEBUG:
await self._emit_debug_log(f"Agent working in: {cwd}", __event_call__)
if not self.valves.GH_TOKEN: if not self.valves.GH_TOKEN:
return "Error: Please configure GH_TOKEN in Valves." return "Error: Please configure GH_TOKEN in Valves."
# Trigger async CLI update if configured
cli_path = os.environ.get("COPILOT_CLI_PATH")
if cli_path:
asyncio.create_task(self._update_copilot_cli(cli_path, __event_call__))
# Parse user selected model # Parse user selected model
request_model = body.get("model", "") request_model = body.get("model", "")
real_model_id = self.valves.MODEL_ID # Default value real_model_id = request_model
# Determine effective reasoning effort and debug setting # Determine effective reasoning effort and debug setting
if __user__: if __user__:
@@ -877,6 +1248,14 @@ class Pipe:
await self._emit_debug_log( await self._emit_debug_log(
f"Using selected model: {real_model_id}", __event_call__ f"Using selected model: {real_model_id}", __event_call__
) )
elif __metadata__ and __metadata__.get("base_model_id"):
base_model_id = __metadata__.get("base_model_id", "")
if base_model_id.startswith(f"{self.id}-"):
real_model_id = base_model_id[len(f"{self.id}-") :]
await self._emit_debug_log(
f"Using base model: {real_model_id} (derived from custom model {request_model})",
__event_call__,
)
messages = body.get("messages", []) messages = body.get("messages", [])
if not messages: if not messages:
@@ -918,26 +1297,58 @@ class Pipe:
await client.start() await client.start()
# Initialize custom tools # Initialize custom tools
custom_tools = self._initialize_custom_tools() custom_tools = await self._initialize_custom_tools(
__user__=__user__, __event_call__=__event_call__
)
if custom_tools: if custom_tools:
tool_names = [t.name for t in custom_tools] tool_names = [t.name for t in custom_tools]
await self._emit_debug_log( await self._emit_debug_log(
f"Enabled {len(custom_tools)} custom tools: {tool_names}", f"Enabled {len(custom_tools)} custom tools: {tool_names}",
__event_call__, __event_call__,
) )
if self.valves.DEBUG:
for t in custom_tools:
await self._emit_debug_log(
f"📋 Tool Detail: {t.name} - {t.description[:100]}...",
__event_call__,
)
# Check MCP Servers
mcp_servers = self._parse_mcp_servers()
mcp_server_names = list(mcp_servers.keys()) if mcp_servers else []
if mcp_server_names:
await self._emit_debug_log(
f"🔌 MCP Servers Configured: {mcp_server_names}",
__event_call__,
)
else:
await self._emit_debug_log(
" No MCP tool servers found in OpenWebUI Connections.",
__event_call__,
)
# Create or Resume Session # Create or Resume Session
session = None session = None
if chat_id: if chat_id:
try: try:
session = await client.resume_session(chat_id) # Prepare resume config
resume_params = {}
if mcp_servers:
resume_params["mcp_servers"] = mcp_servers
session = (
await client.resume_session(chat_id, resume_params)
if resume_params
else await client.resume_session(chat_id)
)
await self._emit_debug_log( await self._emit_debug_log(
f"Resumed session: {chat_id} (Note: Formatting guidelines only apply to NEW sessions. Create a new chat to use updated formatting.)", f"Resumed session: {chat_id} (Reasoning: {effective_reasoning_effort or 'default'})",
__event_call__, __event_call__,
) )
# Show workspace info if available # Show workspace info of available
if self.valves.DEBUG and self.valves.SHOW_WORKSPACE_INFO: if self.valves.DEBUG:
if session.workspace_path: if session.workspace_path:
await self._emit_debug_log( await self._emit_debug_log(
f"Session workspace: {session.workspace_path}", f"Session workspace: {session.workspace_path}",
@@ -990,7 +1401,7 @@ class Pipe:
) )
# Show workspace info for new sessions # Show workspace info for new sessions
if self.valves.DEBUG and self.valves.SHOW_WORKSPACE_INFO: if self.valves.DEBUG:
if session.workspace_path: if session.workspace_path:
await self._emit_debug_log( await self._emit_debug_log(
f"Session workspace: {session.workspace_path}", f"Session workspace: {session.workspace_path}",
@@ -1012,7 +1423,11 @@ class Pipe:
if body.get("stream", False): if body.get("stream", False):
init_msg = "" init_msg = ""
if self.valves.DEBUG: if self.valves.DEBUG:
init_msg = f"> [Debug] Agent working in: {os.getcwd()}\n" init_msg = (
f"> [Debug] Agent working in: {self._get_workspace_dir()}\n"
)
if mcp_server_names:
init_msg += f"> [Debug] 🔌 Connected MCP Servers: {', '.join(mcp_server_names)}\n"
# Transfer client ownership to stream_response # Transfer client ownership to stream_response
should_stop_client = False should_stop_client = False

View File

@@ -3,13 +3,13 @@ title: GitHub Copilot Official SDK Pipe
author: Fu-Jie author: Fu-Jie
author_url: https://github.com/Fu-Jie/awesome-openwebui author_url: https://github.com/Fu-Jie/awesome-openwebui
funding_url: https://github.com/open-webui funding_url: https://github.com/open-webui
openwebui_id: ce96f7b4-12fc-4ac3-9a01-875713e69359
description: 集成 GitHub Copilot SDK。支持动态模型、多轮对话、流式输出、多模态输入、无限会话及前端调试日志。 description: 集成 GitHub Copilot SDK。支持动态模型、多轮对话、流式输出、多模态输入、无限会话及前端调试日志。
version: 0.2.3 version: 0.3.0
requirements: github-copilot-sdk requirements: github-copilot-sdk==0.1.22
""" """
import os import os
import re
import time import time
import json import json
import base64 import base64
@@ -19,46 +19,36 @@ import logging
import shutil import shutil
import subprocess import subprocess
import sys import sys
from typing import Optional, Union, AsyncGenerator, List, Any, Dict import hashlib
from pydantic import BaseModel, Field from pathlib import Path
from typing import Optional, Union, AsyncGenerator, List, Any, Dict, Callable
from types import SimpleNamespace
from pydantic import BaseModel, Field, create_model
from datetime import datetime, timezone from datetime import datetime, timezone
import contextlib import contextlib
# 导入 Copilot SDK 模块 # 导入 Copilot SDK 模块
from copilot import CopilotClient, define_tool from copilot import CopilotClient, define_tool
# 导入 OpenWebUI 配置和工具模块
from open_webui.config import TOOL_SERVER_CONNECTIONS
from open_webui.utils.tools import get_tools as get_openwebui_tools
from open_webui.models.tools import Tools
from open_webui.models.users import Users
# Setup logger # Setup logger
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
class RandomNumberParams(BaseModel):
min: int = Field(description="最小值(包含)")
max: int = Field(description="最大值(包含)")
@define_tool(description="在指定范围内生成随机整数。")
async def generate_random_number(params: RandomNumberParams) -> str:
import random
if params.min >= params.max:
raise ValueError("min 必须小于 max")
number = random.randint(params.min, params.max)
return f"生成的随机数: {number}"
class Pipe: class Pipe:
class Valves(BaseModel): class Valves(BaseModel):
GH_TOKEN: str = Field( GH_TOKEN: str = Field(
default="", default="",
description="GitHub 细粒度 Token(需要 Copilot Requests 权限)", description="GitHub OAuth Token (来自 'gh auth token'),用于 Copilot Chat (必须)",
) )
MODEL_ID: str = Field( COPILOT_CLI_VERSION: str = Field(
default="gpt-5-mini", default="0.0.405",
description="默认 Copilot 模型名(动态获取失败时使用)", description="指定安装/强制使用的 Copilot CLI 版本 (例如 '0.0.405')。留空则使用最新版。",
)
CLI_PATH: str = Field(
default="/usr/local/bin/copilot",
description="Copilot CLI 路径",
) )
DEBUG: bool = Field( DEBUG: bool = Field(
default=False, default=False,
@@ -72,10 +62,6 @@ class Pipe:
default=True, default=True,
description="显示模型推理/思考过程", description="显示模型推理/思考过程",
) )
SHOW_WORKSPACE_INFO: bool = Field(
default=True,
description="调试模式下显示会话工作空间路径与摘要",
)
EXCLUDE_KEYWORDS: str = Field( EXCLUDE_KEYWORDS: str = Field(
default="", default="",
description="排除包含这些关键词的模型逗号分隔codex, haiku", description="排除包含这些关键词的模型逗号分隔codex, haiku",
@@ -104,13 +90,14 @@ class Pipe:
default="", default="",
description='自定义环境变量JSON 格式,例如 {"VAR": "value"}', description='自定义环境变量JSON 格式,例如 {"VAR": "value"}',
) )
ENABLE_TOOLS: bool = Field(
default=False, ENABLE_OPENWEBUI_TOOLS: bool = Field(
description="启用自定义工具(例如:随机数)", default=True,
description="启用 OpenWebUI 工具 (包括自定义工具和工具服务器工具)。",
) )
AVAILABLE_TOOLS: str = Field( ENABLE_MCP_SERVER: bool = Field(
default="all", default=True,
description="可用工具:'all' 或逗号分隔列表(例如:'generate_random_number'", description="启用直接 MCP 客户端连接 (推荐)。",
) )
REASONING_EFFORT: str = Field( REASONING_EFFORT: str = Field(
default="medium", default="medium",
@@ -122,25 +109,35 @@ class Pipe:
) )
class UserValves(BaseModel): class UserValves(BaseModel):
GH_TOKEN: str = Field(
default="",
description="个人 GitHub Fine-grained Token (覆盖全局设置)",
)
REASONING_EFFORT: str = Field( REASONING_EFFORT: str = Field(
default="", default="",
description="推理强度级别 (low, medium, high, xhigh)。留空以使用全局设置。", description="推理强度级别 (low, medium, high, xhigh)。留空以使用全局设置。",
) )
CLI_PATH: str = Field(
default="",
description="自定义 Copilot CLI 路径。留空以使用全局设置。",
)
DEBUG: bool = Field( DEBUG: bool = Field(
default=False, default=False,
description="启用技术调试日志(连接信息等)", description="启用技术调试日志(连接信息等)",
) )
SHOW_THINKING: bool = Field( SHOW_THINKING: bool = Field(
default=True, default=True,
description="显示模型推理/思考过程", description="显示模型推理/思考过程",
) )
MODEL_ID: str = Field(
default="", ENABLE_OPENWEBUI_TOOLS: bool = Field(
description="自定义模型 ID (例如 gpt-4o)。留空以使用全局默认值。", default=True,
description="启用 OpenWebUI 工具 (包括自定义工具和工具服务器工具,覆盖全局设置)。",
)
ENABLE_MCP_SERVER: bool = Field(
default=True,
description="启用动态 MCP 服务器加载 (覆盖全局设置)。",
)
ENFORCE_FORMATTING: bool = Field(
default=True,
description="强制启用格式化指导 (覆盖全局设置)",
) )
def __init__(self): def __init__(self):
@@ -151,6 +148,7 @@ class Pipe:
self.temp_dir = tempfile.mkdtemp(prefix="copilot_images_") self.temp_dir = tempfile.mkdtemp(prefix="copilot_images_")
self.thinking_started = False self.thinking_started = False
self._model_cache = [] # 模型列表缓存 self._model_cache = [] # 模型列表缓存
self._last_update_check = 0 # 上次 CLI 更新检查时间
def __del__(self): def __del__(self):
try: try:
@@ -339,9 +337,31 @@ class Pipe:
return system_prompt_content, system_prompt_source return system_prompt_content, system_prompt_source
def _get_workspace_dir(self) -> str:
"""获取具有智能默认值的有效工作空间目录。"""
if self.valves.WORKSPACE_DIR:
return self.valves.WORKSPACE_DIR
# OpenWebUI 容器的智能默认值
if os.path.exists("/app/backend/data"):
cwd = "/app/backend/data/copilot_workspace"
else:
# 本地回退:当前工作目录的子目录
cwd = os.path.join(os.getcwd(), "copilot_workspace")
# 确保目录存在
if not os.path.exists(cwd):
try:
os.makedirs(cwd, exist_ok=True)
except Exception as e:
print(f"Error creating workspace {cwd}: {e}")
return os.getcwd() # 如果创建失败回退到 CWD
return cwd
def _build_client_config(self, body: dict) -> dict: def _build_client_config(self, body: dict) -> dict:
"""根据 Valves 和请求构建 CopilotClient 配置""" """根据 Valves 和请求构建 CopilotClient 配置"""
cwd = self.valves.WORKSPACE_DIR if self.valves.WORKSPACE_DIR else os.getcwd() cwd = self._get_workspace_dir()
client_config = {} client_config = {}
if os.environ.get("COPILOT_CLI_PATH"): if os.environ.get("COPILOT_CLI_PATH"):
client_config["cli_path"] = os.environ["COPILOT_CLI_PATH"] client_config["cli_path"] = os.environ["COPILOT_CLI_PATH"]
@@ -360,6 +380,270 @@ class Pipe:
return client_config return client_config
async def _initialize_custom_tools(self, __user__=None, __event_call__=None):
"""根据配置初始化自定义工具"""
if not self.valves.ENABLE_OPENWEBUI_TOOLS:
return []
# 动态加载 OpenWebUI 工具
openwebui_tools = await self._load_openwebui_tools(
__user__=__user__, __event_call__=__event_call__
)
return openwebui_tools
def _json_schema_to_python_type(self, schema: dict) -> Any:
"""将 JSON Schema 类型转换为 Python 类型以用于 Pydantic 模型。"""
if not isinstance(schema, dict):
return Any
schema_type = schema.get("type")
if isinstance(schema_type, list):
schema_type = next((t for t in schema_type if t != "null"), schema_type[0])
if schema_type == "string":
return str
if schema_type == "integer":
return int
if schema_type == "number":
return float
if schema_type == "boolean":
return bool
if schema_type == "object":
return Dict[str, Any]
if schema_type == "array":
items_schema = schema.get("items", {})
item_type = self._json_schema_to_python_type(items_schema)
return List[item_type]
return Any
def _convert_openwebui_tool(self, tool_name: str, tool_dict: dict):
"""将 OpenWebUI 工具定义转换为 Copilot SDK 工具。"""
# 净化工具名称以匹配模式 ^[a-zA-Z0-9_-]+$
sanitized_tool_name = re.sub(r"[^a-zA-Z0-9_-]", "_", tool_name)
# 如果净化后的名称为空或仅包含分隔符(例如纯中文名称),生成回退名称
if not sanitized_tool_name or re.match(r"^[_.-]+$", sanitized_tool_name):
hash_suffix = hashlib.md5(tool_name.encode("utf-8")).hexdigest()[:8]
sanitized_tool_name = f"tool_{hash_suffix}"
if sanitized_tool_name != tool_name:
logger.debug(f"将工具名称 '{tool_name}' 净化为 '{sanitized_tool_name}'")
spec = tool_dict.get("spec", {}) if isinstance(tool_dict, dict) else {}
params_schema = spec.get("parameters", {}) if isinstance(spec, dict) else {}
properties = params_schema.get("properties", {})
required = params_schema.get("required", [])
if not isinstance(properties, dict):
properties = {}
if not isinstance(required, list):
required = []
required_set = set(required)
fields = {}
for param_name, param_schema in properties.items():
param_type = self._json_schema_to_python_type(param_schema)
description = ""
if isinstance(param_schema, dict):
description = param_schema.get("description", "")
if param_name in required_set:
if description:
fields[param_name] = (
param_type,
Field(..., description=description),
)
else:
fields[param_name] = (param_type, ...)
else:
optional_type = Optional[param_type]
if description:
fields[param_name] = (
optional_type,
Field(default=None, description=description),
)
else:
fields[param_name] = (optional_type, None)
if fields:
ParamsModel = create_model(f"{sanitized_tool_name}_Params", **fields)
else:
ParamsModel = create_model(f"{sanitized_tool_name}_Params")
tool_callable = tool_dict.get("callable")
tool_description = spec.get("description", "") if isinstance(spec, dict) else ""
if not tool_description and isinstance(spec, dict):
tool_description = spec.get("summary", "")
# 关键: 如果工具名称被净化(例如中文转哈希),语义会丢失。
# 我们必须将原始名称注入到描述中,以便模型知道它的作用。
if sanitized_tool_name != tool_name:
tool_description = f"功能 '{tool_name}': {tool_description}"
async def _tool(params):
payload = params.model_dump() if hasattr(params, "model_dump") else {}
return await tool_callable(**payload)
_tool.__name__ = sanitized_tool_name
_tool.__doc__ = tool_description
# 转换调试日志
logger.debug(
f"正在转换工具 '{sanitized_tool_name}': {tool_description[:50]}..."
)
# 核心关键点:必须显式传递 types否则 define_tool 无法推断动态函数的参数
# 显式传递 name 确保 SDK 注册的名称正确
return define_tool(
name=sanitized_tool_name,
description=tool_description,
params_type=ParamsModel,
)(_tool)
def _build_openwebui_request(self):
"""构建一个最小的 request 模拟对象用于 OpenWebUI 工具加载。"""
app_state = SimpleNamespace(
config=SimpleNamespace(
TOOL_SERVER_CONNECTIONS=TOOL_SERVER_CONNECTIONS.value
),
TOOLS={},
)
app = SimpleNamespace(state=app_state)
request = SimpleNamespace(
app=app,
cookies={},
state=SimpleNamespace(token=SimpleNamespace(credentials="")),
)
return request
async def _load_openwebui_tools(self, __user__=None, __event_call__=None):
"""动态加载 OpenWebUI 工具并转换为 Copilot SDK 工具。"""
if isinstance(__user__, (list, tuple)):
user_data = __user__[0] if __user__ else {}
elif isinstance(__user__, dict):
user_data = __user__
else:
user_data = {}
if not user_data:
return []
user_id = user_data.get("id") or user_data.get("user_id")
if not user_id:
return []
user = Users.get_user_by_id(user_id)
if not user:
return []
# 1. 获取用户自定义工具 (Python 脚本)
tool_items = Tools.get_tools_by_user_id(user_id, permission="read")
tool_ids = [tool.id for tool in tool_items] if tool_items else []
# 2. 获取 OpenAPI 工具服务器工具
# 我们手动添加已启用的 OpenAPI 服务器,因为 Tools.get_tools_by_user_id 仅检查数据库。
# open_webui.utils.tools.get_tools 会处理实际的加载和访问控制。
if hasattr(TOOL_SERVER_CONNECTIONS, "value"):
for server in TOOL_SERVER_CONNECTIONS.value:
# 我们在此处仅添加 'openapi' 服务器,因为 get_tools 目前似乎仅支持 'openapi' (默认为此)。
# MCP 工具通过 ENABLE_MCP_SERVER 单独处理。
if server.get("type") == "openapi":
# get_tools 期望的格式: "server:<id>" 隐含 type="openapi"
server_id = server.get("id")
if server_id:
tool_ids.append(f"server:{server_id}")
if not tool_ids:
return []
request = self._build_openwebui_request()
extra_params = {
"__request__": request,
"__user__": user_data,
"__event_emitter__": None,
"__event_call__": __event_call__,
"__chat_id__": None,
"__message_id__": None,
"__model_knowledge__": [],
}
tools_dict = await get_openwebui_tools(request, tool_ids, user, extra_params)
if not tools_dict:
return []
converted_tools = []
for tool_name, tool_def in tools_dict.items():
try:
converted_tools.append(
self._convert_openwebui_tool(tool_name, tool_def)
)
except Exception as e:
await self._emit_debug_log(
f"加载 OpenWebUI 工具 '{tool_name}' 失败: {e}",
__event_call__,
)
return converted_tools
def _parse_mcp_servers(self) -> Optional[dict]:
"""
从 OpenWebUI TOOL_SERVER_CONNECTIONS 动态加载 MCP 服务器配置。
返回兼容 CopilotClient 的 mcp_servers 字典。
"""
if not self.valves.ENABLE_MCP_SERVER:
return None
mcp_servers = {}
# 遍历 OpenWebUI 工具服务器连接
if hasattr(TOOL_SERVER_CONNECTIONS, "value"):
connections = TOOL_SERVER_CONNECTIONS.value
else:
connections = []
for conn in connections:
if conn.get("type") == "mcp":
info = conn.get("info", {})
# 使用 info 中的 ID 或自动生成
raw_id = info.get("id", f"mcp-server-{len(mcp_servers)}")
# 净化 server_id (使用与工具相同的逻辑)
server_id = re.sub(r"[^a-zA-Z0-9_-]", "_", raw_id)
if not server_id or re.match(r"^[_.-]+$", server_id):
hash_suffix = hashlib.md5(raw_id.encode("utf-8")).hexdigest()[:8]
server_id = f"server_{hash_suffix}"
url = conn.get("url")
if not url:
continue
# 构建 Header (处理认证)
headers = {}
auth_type = conn.get("auth_type", "bearer")
key = conn.get("key", "")
if auth_type == "bearer" and key:
headers["Authorization"] = f"Bearer {key}"
elif auth_type == "basic" and key:
headers["Authorization"] = f"Basic {key}"
# 合并自定义 headers
custom_headers = conn.get("headers", {})
if isinstance(custom_headers, dict):
headers.update(custom_headers)
mcp_servers[server_id] = {
"type": "http",
"url": url,
"headers": headers,
"tools": ["*"], # 默认启用所有工具
}
return mcp_servers if mcp_servers else None
def _build_session_config( def _build_session_config(
self, self,
chat_id: Optional[str], chat_id: Optional[str],
@@ -367,7 +651,6 @@ class Pipe:
custom_tools: List[Any], custom_tools: List[Any],
system_prompt_content: Optional[str], system_prompt_content: Optional[str],
is_streaming: bool, is_streaming: bool,
reasoning_effort: str = "",
): ):
"""构建 Copilot SDK 的 SessionConfig""" """构建 Copilot SDK 的 SessionConfig"""
from copilot.types import SessionConfig, InfiniteSessionConfig from copilot.types import SessionConfig, InfiniteSessionConfig
@@ -415,11 +698,12 @@ class Pipe:
"tools": custom_tools, "tools": custom_tools,
"system_message": system_message_config, "system_message": system_message_config,
"infinite_sessions": infinite_session_config, "infinite_sessions": infinite_session_config,
# 注册权限处理 Hook
} }
# 如果不是默认值medium添加 reasoning_effort mcp_servers = self._parse_mcp_servers()
if reasoning_effort and reasoning_effort.lower() != "medium": if mcp_servers:
session_params["reasoning_effort"] = reasoning_effort.lower() session_params["mcp_servers"] = mcp_servers
return SessionConfig(**session_params) return SessionConfig(**session_params)
@@ -546,24 +830,6 @@ class Pipe:
return system_prompt_content, system_prompt_source return system_prompt_content, system_prompt_source
def _initialize_custom_tools(self):
"""根据配置初始化自定义工具"""
if not self.valves.ENABLE_TOOLS:
return []
# 定义所有可用工具(在此注册新工具)
all_tools = {
"generate_random_number": generate_random_number,
}
# 根据配置过滤
if self.valves.AVAILABLE_TOOLS == "all":
return list(all_tools.values())
# 仅启用指定的工具
enabled = [t.strip() for t in self.valves.AVAILABLE_TOOLS.split(",")]
return [all_tools[name] for name in enabled if name in all_tools]
async def _emit_debug_log(self, message: str, __event_call__=None): async def _emit_debug_log(self, message: str, __event_call__=None):
"""在 DEBUG 开启时将日志输出到前端控制台。""" """在 DEBUG 开启时将日志输出到前端控制台。"""
if not self.valves.DEBUG: if not self.valves.DEBUG:
@@ -765,8 +1031,8 @@ class Pipe:
# 失败时返回默认模型 # 失败时返回默认模型
return [ return [
{ {
"id": f"{self.id}-{self.valves.MODEL_ID}", "id": f"{self.id}-gpt-5-mini",
"name": f"GitHub Copilot ({self.valves.MODEL_ID})", "name": f"GitHub Copilot (gpt-5-mini)",
} }
] ]
finally: finally:
@@ -775,8 +1041,8 @@ class Pipe:
await self._emit_debug_log(f"Pipes Error: {e}") await self._emit_debug_log(f"Pipes Error: {e}")
return [ return [
{ {
"id": f"{self.id}-{self.valves.MODEL_ID}", "id": f"{self.id}-gpt-5-mini",
"name": f"GitHub Copilot ({self.valves.MODEL_ID})", "name": f"GitHub Copilot (gpt-5-mini)",
} }
] ]
@@ -808,30 +1074,93 @@ class Pipe:
return client return client
def _setup_env(self, __event_call__=None): def _setup_env(self, __event_call__=None):
cli_path = self.valves.CLI_PATH cli_path = "/usr/local/bin/copilot"
found = False if os.environ.get("COPILOT_CLI_PATH"):
cli_path = os.environ["COPILOT_CLI_PATH"]
target_version = self.valves.COPILOT_CLI_VERSION.strip()
found = False
current_version = None
# 内部 helper: 获取版本
def get_cli_version(path):
try:
output = (
subprocess.check_output(
[path, "--version"], stderr=subprocess.STDOUT
)
.decode()
.strip()
)
# Copilot CLI 输出通常包含 "copilot version X.Y.Z" 或直接是版本号
match = re.search(r"(\d+\.\d+\.\d+)", output)
return match.group(1) if match else output
except Exception:
return None
# 检查默认路径
if os.path.exists(cli_path): if os.path.exists(cli_path):
found = True found = True
current_version = get_cli_version(cli_path)
# 二次检查系统路径
if not found: if not found:
sys_path = shutil.which("copilot") sys_path = shutil.which("copilot")
if sys_path: if sys_path:
cli_path = sys_path cli_path = sys_path
found = True found = True
current_version = get_cli_version(cli_path)
# 判断是否需要安装/更新
should_install = False
install_reason = ""
if not found: if not found:
should_install = True
install_reason = "CLI 未找到"
elif target_version:
# 标准化版本号 (移除 'v' 前缀)
norm_target = target_version.lstrip("v")
norm_current = current_version.lstrip("v") if current_version else ""
if norm_target != norm_current:
should_install = True
install_reason = (
f"版本不匹配 (当前: {current_version}, 目标: {target_version})"
)
if should_install:
if self.valves.DEBUG:
self._emit_debug_log_sync(
f"正在安装 Copilot CLI: {install_reason}...", __event_call__
)
try: try:
env = os.environ.copy()
if target_version:
env["VERSION"] = target_version
subprocess.run( subprocess.run(
"curl -fsSL https://gh.io/copilot-install | bash", "curl -fsSL https://gh.io/copilot-install | bash",
shell=True, shell=True,
check=True, check=True,
env=env,
) )
if os.path.exists(self.valves.CLI_PATH):
cli_path = self.valves.CLI_PATH # 优先检查默认安装路径,其次是系统路径
if os.path.exists("/usr/local/bin/copilot"):
cli_path = "/usr/local/bin/copilot"
found = True found = True
except: elif shutil.which("copilot"):
pass cli_path = shutil.which("copilot")
found = True
if found:
current_version = get_cli_version(cli_path)
except Exception as e:
if self.valves.DEBUG:
self._emit_debug_log_sync(
f"Copilot CLI 安装失败: {e}", __event_call__
)
if found: if found:
os.environ["COPILOT_CLI_PATH"] = cli_path os.environ["COPILOT_CLI_PATH"] = cli_path
@@ -841,7 +1170,14 @@ class Pipe:
if self.valves.DEBUG: if self.valves.DEBUG:
self._emit_debug_log_sync( self._emit_debug_log_sync(
f"Copilot CLI 已定位: {cli_path}", __event_call__ f"已找到 Copilot CLI: {cli_path} (版本: {current_version})",
__event_call__,
)
else:
if self.valves.DEBUG:
self._emit_debug_log_sync(
"错误: 未找到 Copilot CLI。相关 Agent 功能将被禁用。",
__event_call__,
) )
if self.valves.GH_TOKEN: if self.valves.GH_TOKEN:
@@ -851,6 +1187,8 @@ class Pipe:
if self.valves.DEBUG: if self.valves.DEBUG:
self._emit_debug_log_sync("Warning: GH_TOKEN 未设置。", __event_call__) self._emit_debug_log_sync("Warning: GH_TOKEN 未设置。", __event_call__)
self._sync_mcp_config(__event_call__)
def _process_images(self, messages, __event_call__=None): def _process_images(self, messages, __event_call__=None):
attachments = [] attachments = []
text_content = "" text_content = ""
@@ -945,9 +1283,113 @@ class Pipe:
except Exception as e: except Exception as e:
self._emit_debug_log_sync(f"配置同步检查失败: {e}", __event_call__) self._emit_debug_log_sync(f"配置同步检查失败: {e}", __event_call__)
def _sync_mcp_config(self, __event_call__=None):
"""已弃用MCP 配置现在通过 SessionConfig 动态处理。"""
pass
# ==================== 内部实现 ==================== # ==================== 内部实现 ====================
# _pipe_impl() 包含主请求处理逻辑。 # _pipe_impl() 包含主请求处理逻辑。
# ================================================ # ================================================
def _sync_copilot_config(self, reasoning_effort: str, __event_call__=None):
"""
如果设置了 REASONING_EFFORT则动态更新 ~/.copilot/config.json。
这提供了一个回退机制,以防 API 注入被服务器忽略。
"""
if not reasoning_effort:
return
effort = reasoning_effort
# 检查模型是否支持 xhigh
# 目前只有 gpt-5.2-codex 支持 xhigh
if effort == "xhigh":
# 简单检查,使用默认模型 ID
if (
"gpt-5.2-codex"
not in self._collect_model_ids(
body={},
request_model=self.id,
real_model_id=None,
)[0].lower()
):
# 如果不支持则回退到 high
effort = "high"
try:
# 目标标准路径 ~/.copilot/config.json
config_path = os.path.expanduser("~/.copilot/config.json")
config_dir = os.path.dirname(config_path)
# 仅当目录存在时才继续(避免在路径错误时创建垃圾文件)
if not os.path.exists(config_dir):
return
data = {}
# 读取现有配置
if os.path.exists(config_path):
try:
with open(config_path, "r") as f:
data = json.load(f)
except Exception:
data = {}
# 如果有变化则更新
current_val = data.get("reasoning_effort")
if current_val != effort:
data["reasoning_effort"] = effort
try:
with open(config_path, "w") as f:
json.dump(data, f, indent=4)
self._emit_debug_log_sync(
f"已动态更新 ~/.copilot/config.json: reasoning_effort='{effort}'",
__event_call__,
)
except Exception as e:
self._emit_debug_log_sync(
f"写入 config.json 失败: {e}", __event_call__
)
except Exception as e:
self._emit_debug_log_sync(f"配置同步检查失败: {e}", __event_call__)
async def _update_copilot_cli(self, cli_path: str, __event_call__=None):
"""异步任务:如果需要则更新 Copilot CLI。"""
import time
try:
# 检查频率(例如:每小时一次)
now = time.time()
if now - self._last_update_check < 3600:
return
self._last_update_check = now
if self.valves.DEBUG:
self._emit_debug_log_sync(
"触发异步 Copilot CLI 更新检查...", __event_call__
)
# 我们创建一个子进程来运行更新
process = await asyncio.create_subprocess_exec(
cli_path,
"update",
stdout=asyncio.subprocess.PIPE,
stderr=asyncio.subprocess.PIPE,
)
stdout, stderr = await process.communicate()
if self.valves.DEBUG and process.returncode == 0:
self._emit_debug_log_sync("Copilot CLI 更新检查完成", __event_call__)
elif process.returncode != 0 and self.valves.DEBUG:
self._emit_debug_log_sync(
f"Copilot CLI 更新失败: {stderr.decode()}", __event_call__
)
except Exception as e:
if self.valves.DEBUG:
self._emit_debug_log_sync(f"CLI 更新任务异常: {e}", __event_call__)
async def _pipe_impl( async def _pipe_impl(
self, self,
body: dict, body: dict,
@@ -957,12 +1399,23 @@ class Pipe:
__event_call__=None, __event_call__=None,
) -> Union[str, AsyncGenerator]: ) -> Union[str, AsyncGenerator]:
self._setup_env(__event_call__) self._setup_env(__event_call__)
cwd = self._get_workspace_dir()
if self.valves.DEBUG:
await self._emit_debug_log(f"当前工作目录: {cwd}", __event_call__)
# CLI Update Check
if os.environ.get("COPILOT_CLI_PATH"):
asyncio.create_task(
self._update_copilot_cli(os.environ["COPILOT_CLI_PATH"], __event_call__)
)
if not self.valves.GH_TOKEN: if not self.valves.GH_TOKEN:
return "Error: 请在 Valves 中配置 GH_TOKEN。" return "Error: 请在 Valves 中配置 GH_TOKEN。"
# 解析用户选择的模型 # 解析用户选择的模型
request_model = body.get("model", "") request_model = body.get("model", "")
real_model_id = self.valves.MODEL_ID # 默认值 real_model_id = request_model
# 确定有效的推理强度和调试设置 # 确定有效的推理强度和调试设置
if __user__: if __user__:
@@ -980,6 +1433,10 @@ class Pipe:
if user_valves.REASONING_EFFORT if user_valves.REASONING_EFFORT
else self.valves.REASONING_EFFORT else self.valves.REASONING_EFFORT
) )
# Sync config for reasoning effort (Legacy/Fallback)
self._sync_copilot_config(effective_reasoning_effort, __event_call__)
# 如果用户启用了 DEBUG则覆盖全局设置 # 如果用户启用了 DEBUG则覆盖全局设置
if user_valves.DEBUG: if user_valves.DEBUG:
self.valves.DEBUG = True self.valves.DEBUG = True
@@ -996,6 +1453,14 @@ class Pipe:
await self._emit_debug_log( await self._emit_debug_log(
f"使用选择的模型: {real_model_id}", __event_call__ f"使用选择的模型: {real_model_id}", __event_call__
) )
elif __metadata__ and __metadata__.get("base_model_id"):
base_model_id = __metadata__.get("base_model_id", "")
if base_model_id.startswith(f"{self.id}-"):
real_model_id = base_model_id[len(f"{self.id}-") :]
await self._emit_debug_log(
f"使用基础模型: {real_model_id} (继承自自定义模型 {request_model})",
__event_call__,
)
messages = body.get("messages", []) messages = body.get("messages", [])
if not messages: if not messages:
@@ -1020,32 +1485,66 @@ class Pipe:
is_streaming = body.get("stream", False) is_streaming = body.get("stream", False)
await self._emit_debug_log(f"请求流式传输: {is_streaming}", __event_call__) await self._emit_debug_log(f"请求流式传输: {is_streaming}", __event_call__)
# 处理多模态(图像)和提取最后的消息文本
last_text, attachments = self._process_images(messages, __event_call__)
client = CopilotClient(self._build_client_config(body)) client = CopilotClient(self._build_client_config(body))
should_stop_client = True should_stop_client = True
try: try:
await client.start() await client.start()
# 初始化自定义工具 # 初始化自定义工具
custom_tools = self._initialize_custom_tools() custom_tools = await self._initialize_custom_tools(
__user__=__user__, __event_call__=__event_call__
)
if custom_tools: if custom_tools:
tool_names = [t.name for t in custom_tools] tool_names = [t.name for t in custom_tools]
await self._emit_debug_log( await self._emit_debug_log(
f"已启用 {len(custom_tools)} 个自定义工具: {tool_names}", f"已启用 {len(custom_tools)} 个自定义工具: {tool_names}",
__event_call__, __event_call__,
) )
# 详细打印每个工具的描述 (用于调试)
if self.valves.DEBUG:
for t in custom_tools:
await self._emit_debug_log(
f"📋 工具详情: {t.name} - {t.description[:100]}...",
__event_call__,
)
# 检查 MCP 服务器
mcp_servers = self._parse_mcp_servers()
mcp_server_names = list(mcp_servers.keys()) if mcp_servers else []
if mcp_server_names:
await self._emit_debug_log(
f"🔌 MCP 服务器已配置: {mcp_server_names}",
__event_call__,
)
else:
await self._emit_debug_log(
" 未在 OpenWebUI 连接中发现 MCP 服务器。",
__event_call__,
)
session = None session = None
if chat_id: if chat_id:
try: try:
# 复用已解析的 mcp_servers
resume_config = (
{"mcp_servers": mcp_servers} if mcp_servers else None
)
# 尝试直接使用 chat_id 作为 session_id 恢复会话 # 尝试直接使用 chat_id 作为 session_id 恢复会话
session = await client.resume_session(chat_id) session = (
await client.resume_session(chat_id, resume_config)
if resume_config
else await client.resume_session(chat_id)
)
await self._emit_debug_log( await self._emit_debug_log(
f"已通过 ChatID 恢复会话: {chat_id}", __event_call__ f"已通过 ChatID 恢复会话: {chat_id}", __event_call__
) )
# 显示工作空间信息(如果可用) # 显示工作空间信息(如果可用)
if self.valves.DEBUG and self.valves.SHOW_WORKSPACE_INFO: if self.valves.DEBUG:
if session.workspace_path: if session.workspace_path:
await self._emit_debug_log( await self._emit_debug_log(
f"会话工作空间: {session.workspace_path}", f"会话工作空间: {session.workspace_path}",
@@ -1108,7 +1607,7 @@ class Pipe:
await self._emit_debug_log(f"创建了新会话: {new_sid}", __event_call__) await self._emit_debug_log(f"创建了新会话: {new_sid}", __event_call__)
# 显示新会话的工作空间信息 # 显示新会话的工作空间信息
if self.valves.DEBUG and self.valves.SHOW_WORKSPACE_INFO: if self.valves.DEBUG:
if session.workspace_path: if session.workspace_path:
await self._emit_debug_log( await self._emit_debug_log(
f"会话工作空间: {session.workspace_path}", f"会话工作空间: {session.workspace_path}",
@@ -1134,6 +1633,9 @@ class Pipe:
else: else:
init_msg = f"> [Debug] 已通过 ChatID 恢复会话: {chat_id}\n" init_msg = f"> [Debug] 已通过 ChatID 恢复会话: {chat_id}\n"
if mcp_server_names:
init_msg += f"> [Debug] 🔌 已连接 MCP 服务器: {', '.join(mcp_server_names)}\n"
return self.stream_response( return self.stream_response(
client, client,
session, session,

View File

@@ -93,7 +93,11 @@ def scan_plugins_directory(plugins_dir: str) -> list[dict[str, Any]]:
return plugins return plugins
# Walk through all subdirectories # Walk through all subdirectories
for root, _dirs, files in os.walk(plugins_path): for root, dirs, files in os.walk(plugins_path):
# Exclude debug directory from scan
if "debug" in dirs:
dirs.remove("debug")
for file in files: for file in files:
if file.endswith(".py") and not file.startswith("__"): if file.endswith(".py") and not file.startswith("__"):
# Skip specific files that should not trigger release # Skip specific files that should not trigger release
@@ -110,6 +114,11 @@ def scan_plugins_directory(plugins_dir: str) -> list[dict[str, Any]]:
if metadata: if metadata:
# Determine plugin type from directory structure # Determine plugin type from directory structure
rel_path = os.path.relpath(file_path, plugins_dir) rel_path = os.path.relpath(file_path, plugins_dir)
# Normalize file_path to always start with "plugins/" for consistent ID comparison
# regardless of where we scan from (/tmp/old_repo or ./plugins)
metadata["file_path"] = os.path.join("plugins", rel_path)
parts = rel_path.split(os.sep) parts = rel_path.split(os.sep)
if len(parts) > 0: if len(parts) > 0:
metadata["type"] = parts[0] # actions, filters, pipes, etc. metadata["type"] = parts[0] # actions, filters, pipes, etc.
@@ -139,42 +148,45 @@ def compare_versions(current: list[dict], previous_file: str) -> dict[str, list[
print(f"Error parsing {previous_file}", file=sys.stderr) print(f"Error parsing {previous_file}", file=sys.stderr)
return {"added": current, "updated": [], "removed": []} return {"added": current, "updated": [], "removed": []}
# Create lookup dictionaries by title # Create lookup dictionaries by file_path (fallback to title)
# Helper to extract title/version from either simple dict or raw post object # Helper to extract title/version/file_path from either simple dict or raw post object
def get_info(p): def get_info(p):
if "data" in p and "function" in p["data"]: if "data" in p and "function" in p["data"]:
# It's a raw post object # It's a raw post object
manifest = p["data"]["function"].get("meta", {}).get("manifest", {}) manifest = p["data"]["function"].get("meta", {}).get("manifest", {})
title = manifest.get("title") or p.get("title") title = manifest.get("title") or p.get("title")
version = manifest.get("version", "0.0.0") version = manifest.get("version", "0.0.0")
return title, version, p file_path = p.get("file_path")
return title, version, file_path, p
else: else:
# It's a simple dict # It's a simple dict
return p.get("title"), p.get("version"), p return p.get("title"), p.get("version"), p.get("file_path"), p
current_by_title = {} current_by_key = {}
for p in current: for p in current:
title, _, _ = get_info(p) title, _, file_path, _ = get_info(p)
if title: key = file_path or title
current_by_title[title] = p if key:
current_by_key[key] = p
previous_by_title = {} previous_by_key = {}
for p in previous: for p in previous:
title, _, _ = get_info(p) title, _, file_path, _ = get_info(p)
if title: key = file_path or title
previous_by_title[title] = p if key:
previous_by_key[key] = p
result = {"added": [], "updated": [], "removed": []} result = {"added": [], "updated": [], "removed": []}
# Find added and updated plugins # Find added and updated plugins
for title, plugin in current_by_title.items(): for key, plugin in current_by_key.items():
curr_title, curr_ver, _ = get_info(plugin) curr_title, curr_ver, _file_path, _ = get_info(plugin)
if title not in previous_by_title: if key not in previous_by_key:
result["added"].append(plugin) result["added"].append(plugin)
else: else:
prev_plugin = previous_by_title[title] prev_plugin = previous_by_key[key]
_, prev_ver, _ = get_info(prev_plugin) _, prev_ver, _prev_file_path, _ = get_info(prev_plugin)
if curr_ver != prev_ver: if curr_ver != prev_ver:
result["updated"].append( result["updated"].append(
@@ -185,8 +197,8 @@ def compare_versions(current: list[dict], previous_file: str) -> dict[str, list[
) )
# Find removed plugins # Find removed plugins
for title, plugin in previous_by_title.items(): for key, plugin in previous_by_key.items():
if title not in current_by_title: if key not in current_by_key:
result["removed"].append(plugin) result["removed"].append(plugin)
return result return result

View File

@@ -23,7 +23,11 @@ from openwebui_community_client import get_client
def find_existing_plugins(plugins_dir: str) -> list: def find_existing_plugins(plugins_dir: str) -> list:
"""查找所有已发布的插件文件(有 openwebui_id 的)""" """查找所有已发布的插件文件(有 openwebui_id 的)"""
plugins = [] plugins = []
for root, _, files in os.walk(plugins_dir): for root, dirs, files in os.walk(plugins_dir):
# Exclude debug directory
if "debug" in dirs:
dirs.remove("debug")
for file in files: for file in files:
if file.endswith(".py") and not file.startswith("__"): if file.endswith(".py") and not file.startswith("__"):
file_path = os.path.join(root, file) file_path = os.path.join(root, file)