mirror of
https://github.com/LiteyukiStudio/nonebot-plugin-marshoai.git
synced 2025-12-24 11:16:41 +00:00
Compare commits
26 Commits
mod/config
...
snowykami-
| Author | SHA1 | Date | |
|---|---|---|---|
| 02d465112f | |||
| d95928cab7 | |||
| 41cb287a84 | |||
| a0f2b52e59 | |||
| 75d173bed7 | |||
| f39f5cc1be | |||
| 70fd176904 | |||
| 57ea4fc10b | |||
| a1ddf40610 | |||
| dc294a257d | |||
|
|
6f085b36c6 | ||
| 8aff490aeb | |||
| b713110bcf | |||
| b495aa9490 | |||
|
|
a61d13426e | ||
| cbafaaf151 | |||
| 00605ad401 | |||
|
|
1cd60252b5 | ||
|
|
aa53643aae | ||
| 3436390f4b | |||
| e1bc81c9e1 | |||
| 5eb3c66232 | |||
| a5e72c6946 | |||
|
|
2be57309bd | ||
| 0b6ac9f73e | |||
|
|
0e72880167 |
40
.github/workflows/docs-build.yml
vendored
40
.github/workflows/docs-build.yml
vendored
@@ -1,26 +1,18 @@
|
||||
# 构建 VitePress 站点并将其部署到 GitHub Pages 的示例工作流程
|
||||
#
|
||||
name: Deploy VitePress site to Pages
|
||||
name: Deploy VitePress site to Liteyuki PaaS
|
||||
|
||||
on:
|
||||
# 在针对 `main` 分支的推送上运行。如果你
|
||||
# 使用 `master` 分支作为默认分支,请将其更改为 `master`
|
||||
push:
|
||||
branches: [main]
|
||||
on: ["push", "pull_request_target"]
|
||||
|
||||
# 允许你从 Actions 选项卡手动运行此工作流程
|
||||
workflow_dispatch:
|
||||
|
||||
# 设置 GITHUB_TOKEN 的权限,以允许部署到 GitHub Pages
|
||||
permissions:
|
||||
contents: write
|
||||
statuses: write
|
||||
|
||||
# 只允许同时进行一次部署,跳过正在运行和最新队列之间的运行队列
|
||||
# 但是,不要取消正在进行的运行,因为我们希望允许这些生产部署完成
|
||||
concurrency:
|
||||
group: pages
|
||||
cancel-in-progress: false
|
||||
|
||||
env:
|
||||
MELI_SITE: f31e3b17-c4ea-4d9d-bdce-9417d67fd30e
|
||||
|
||||
jobs:
|
||||
# 构建工作
|
||||
build:
|
||||
@@ -30,12 +22,10 @@ jobs:
|
||||
uses: actions/checkout@v4
|
||||
with:
|
||||
fetch-depth: 0 # 如果未启用 lastUpdated,则不需要
|
||||
# - uses: pnpm/action-setup@v3 # 如果使用 pnpm,请取消注释
|
||||
# - uses: oven-sh/setup-bun@v1 # 如果使用 Bun,请取消注释
|
||||
- name: Setup Python
|
||||
uses: actions/setup-python@v2
|
||||
with:
|
||||
python-version: '3.11'
|
||||
python-version: "3.11"
|
||||
|
||||
- name: Setup API markdown
|
||||
run: |-
|
||||
@@ -59,9 +49,13 @@ jobs:
|
||||
run: |-
|
||||
pnpm run docs:build
|
||||
|
||||
- name: 部署文档
|
||||
uses: JamesIves/github-pages-deploy-action@v4
|
||||
with:
|
||||
# 这是文档部署到的分支名称
|
||||
branch: docs
|
||||
folder: docs/.vitepress/dist
|
||||
- name: "发布"
|
||||
run: |
|
||||
npx -p "@getmeli/cli" meli upload docs/.vitepress/dist \
|
||||
--url "https://pages.liteyuki.icu" \
|
||||
--site "$MELI_SITE" \
|
||||
--token "$MELI_TOKEN" \
|
||||
--release "$GITHUB_SHA"
|
||||
env:
|
||||
MELI_TOKEN: ${{ secrets.MELI_TOKEN }}
|
||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||
|
||||
@@ -15,7 +15,7 @@ repos:
|
||||
args: [--config=./pyproject.toml]
|
||||
|
||||
- repo: https://github.com/PyCQA/isort
|
||||
rev: 6.0.0
|
||||
rev: 6.0.1
|
||||
hooks:
|
||||
- id: isort
|
||||
args: ["--profile", "black"]
|
||||
|
||||
13
README.md
13
README.md
@@ -10,7 +10,7 @@
|
||||
|
||||
_✨ 使用 OpenAI 标准格式 API 的聊天机器人插件 ✨_
|
||||
|
||||
[](https://qm.qq.com/q/a13iwP5kAw)
|
||||
[](https://qm.qq.com/q/a13iwP5kAw)
|
||||
[](https://registry.nonebot.dev/plugin/nonebot-plugin-marshoai:nonebot_plugin_marshoai)
|
||||
<a href="https://registry.nonebot.dev/plugin/nonebot-plugin-marshoai:nonebot_plugin_marshoai">
|
||||
<img src="https://img.shields.io/endpoint?url=https%3A%2F%2Fnbbdg.lgc2333.top%2Fplugin-adapters%2Fnonebot-plugin-marshoai&style=flat-square" alt="Supported Adapters">
|
||||
@@ -22,20 +22,21 @@ _✨ 使用 OpenAI 标准格式 API 的聊天机器人插件 ✨_
|
||||
<img src="https://img.shields.io/badge/Code%20Style-Black-121110.svg?style=flat-square" alt="codestyle">
|
||||
</div>
|
||||
|
||||
<img width="100%" src="https://starify.komoridevs.icu/api/starify?owner=LiteyukiStudio&repo=nonebot-plugin-marshoai" alt="starify" />
|
||||
|
||||
## 📖 介绍
|
||||
|
||||
通过调用 OpenAI 标准格式 API(例如 GitHub Models API) 来实现聊天的插件。
|
||||
插件内置了猫娘小棉(Marsho)的人物设定,可以进行可爱的聊天!
|
||||
通过调用 OpenAI 标准格式 API(例如 GitHub Models API),来实现聊天的插件。
|
||||
插件内置了猫娘小棉(Marsho,マルショ)的人物设定,可以进行可爱的聊天!
|
||||
_谁不喜欢回复消息快又可爱的猫娘呢?_
|
||||
**对 OneBot 以外的适配器与非 GitHub Models API 的支持未经过完全验证。**
|
||||
**对 OneBot 以外的适配器与非 GitHub Models API 的支持未完全经过验证。**
|
||||
[Melobot 实现](https://github.com/LiteyukiStudio/marshoai-melo)
|
||||
|
||||
## 🐱 设定
|
||||
|
||||
#### 基本信息
|
||||
|
||||
- 名字:小棉(Marsho)
|
||||
- 名字:小棉(Marsho,マルショ)
|
||||
- 生日:9 月 6 日
|
||||
|
||||
#### 喜好
|
||||
@@ -58,7 +59,7 @@ _谁不喜欢回复消息快又可爱的猫娘呢?_
|
||||
- [nonebot-plugin-latex](https://github.com/EillesWan/nonebot-plugin-latex)
|
||||
- [nonebot-plugin-deepseek](https://github.com/KomoriDev/nonebot-plugin-deepseek)
|
||||
|
||||
"Marsho" logo 由 [@Asankilp](https://github.com/Asankilp)绘制,基于 [CC BY-NC-SA 4.0](http://creativecommons.org/licenses/by-nc-sa/4.0/) 许可下提供。
|
||||
"Marsho" logo 由 [@Asankilp](https://github.com/Asankilp) 绘制,基于 [CC BY-NC-SA 4.0](http://creativecommons.org/licenses/by-nc-sa/4.0/) 许可下提供。
|
||||
"nonebot-plugin-marshoai" 基于 [MIT](./LICENSE-MIT) 许可下提供。
|
||||
部分指定的代码基于 [Mulan PSL v2](./LICENSE-MULAN) 许可下提供。
|
||||
|
||||
|
||||
@@ -1,81 +1,87 @@
|
||||
import { VitePressSidebarOptions } from "vitepress-sidebar/types"
|
||||
import { VitePressSidebarOptions } from "vitepress-sidebar/types";
|
||||
|
||||
export const gitea = {
|
||||
svg: '<svg t="1725391346807" class="icon" viewBox="0 0 1025 1024" version="1.1" xmlns="http://www.w3.org/2000/svg" p-id="5067" width="256" height="256"><path d="M1004.692673 466.396616l-447.094409-447.073929c-25.743103-25.763582-67.501405-25.763582-93.264987 0l-103.873521 103.873521 78.171378 78.171378c12.533635-6.00058 26.562294-9.359266 41.389666-9.359266 53.02219 0 96.00928 42.98709 96.00928 96.00928 0 14.827372-3.358686 28.856031-9.359266 41.389666l127.97824 127.97824c12.533635-6.00058 26.562294-9.359266 41.389666-9.359266 53.02219 0 96.00928 42.98709 96.00928 96.00928s-42.98709 96.00928-96.00928 96.00928-96.00928-42.98709-96.00928-96.00928c0-14.827372 3.358686-28.856031 9.359266-41.389666l-127.97824-127.97824c-3.051489 1.454065-6.184898 2.744293-9.379746 3.870681l0 266.97461c37.273227 13.188988 63.99936 48.721433 63.99936 90.520695 0 53.02219-42.98709 96.00928-96.00928 96.00928s-96.00928-42.98709-96.00928-96.00928c0-41.799262 26.726133-77.331707 63.99936-90.520695l0-266.97461c-37.273227-13.188988-63.99936-48.721433-63.99936-90.520695 0-14.827372 3.358686-28.856031 9.359266-41.389666l-78.171378-78.171378-295.892081 295.871601c-25.743103 25.784062-25.743103 67.542365 0 93.285467l447.114889 447.073929c25.743103 25.743103 67.480925 25.743103 93.264987 0l445.00547-445.00547c25.763582-25.763582 25.763582-67.542365 0-93.285467z" fill="#a2d8f4" p-id="5068"></path></svg>'
|
||||
}
|
||||
svg: '<svg t="1725391346807" class="icon" viewBox="0 0 1025 1024" version="1.1" xmlns="http://www.w3.org/2000/svg" p-id="5067" width="256" height="256"><path d="M1004.692673 466.396616l-447.094409-447.073929c-25.743103-25.763582-67.501405-25.763582-93.264987 0l-103.873521 103.873521 78.171378 78.171378c12.533635-6.00058 26.562294-9.359266 41.389666-9.359266 53.02219 0 96.00928 42.98709 96.00928 96.00928 0 14.827372-3.358686 28.856031-9.359266 41.389666l127.97824 127.97824c12.533635-6.00058 26.562294-9.359266 41.389666-9.359266 53.02219 0 96.00928 42.98709 96.00928 96.00928s-42.98709 96.00928-96.00928 96.00928-96.00928-42.98709-96.00928-96.00928c0-14.827372 3.358686-28.856031 9.359266-41.389666l-127.97824-127.97824c-3.051489 1.454065-6.184898 2.744293-9.379746 3.870681l0 266.97461c37.273227 13.188988 63.99936 48.721433 63.99936 90.520695 0 53.02219-42.98709 96.00928-96.00928 96.00928s-96.00928-42.98709-96.00928-96.00928c0-41.799262 26.726133-77.331707 63.99936-90.520695l0-266.97461c-37.273227-13.188988-63.99936-48.721433-63.99936-90.520695 0-14.827372 3.358686-28.856031 9.359266-41.389666l-78.171378-78.171378-295.892081 295.871601c-25.743103 25.784062-25.743103 67.542365 0 93.285467l447.114889 447.073929c25.743103 25.743103 67.480925 25.743103 93.264987 0l445.00547-445.00547c25.763582-25.763582 25.763582-67.542365 0-93.285467z" fill="#a2d8f4" p-id="5068"></path></svg>',
|
||||
};
|
||||
|
||||
export const defaultLang = 'zh'
|
||||
export const defaultLang = "zh";
|
||||
|
||||
const commonSidebarOptions: VitePressSidebarOptions = {
|
||||
collapsed: true,
|
||||
convertSameNameSubFileToGroupIndexPage: true,
|
||||
useTitleFromFrontmatter: true,
|
||||
useFolderTitleFromIndexFile: false,
|
||||
useFolderLinkFromIndexFile: true,
|
||||
useTitleFromFileHeading: true,
|
||||
rootGroupText: 'MARSHOAI',
|
||||
includeFolderIndexFile: true,
|
||||
sortMenusByFrontmatterOrder: true,
|
||||
}
|
||||
collapsed: true,
|
||||
convertSameNameSubFileToGroupIndexPage: true,
|
||||
useTitleFromFrontmatter: true,
|
||||
useFolderTitleFromIndexFile: false,
|
||||
useFolderLinkFromIndexFile: true,
|
||||
useTitleFromFileHeading: true,
|
||||
rootGroupText: "MARSHOAI",
|
||||
includeFolderIndexFile: true,
|
||||
sortMenusByFrontmatterOrder: true,
|
||||
};
|
||||
|
||||
export function generateSidebarConfig(): VitePressSidebarOptions[] {
|
||||
let sections = ["dev", "start"]
|
||||
let languages = ['zh', 'en']
|
||||
let ret: VitePressSidebarOptions[] = []
|
||||
for (let language of languages) {
|
||||
for (let section of sections) {
|
||||
if (language === defaultLang) {
|
||||
ret.push({
|
||||
basePath: `/${section}/`,
|
||||
scanStartPath: `docs/${language}/${section}`,
|
||||
resolvePath: `/${section}/`,
|
||||
...commonSidebarOptions
|
||||
})
|
||||
} else {
|
||||
ret.push({
|
||||
basePath: `/${language}/${section}/`,
|
||||
scanStartPath: `docs/${language}/${section}`,
|
||||
resolvePath: `/${language}/${section}/`,
|
||||
...commonSidebarOptions
|
||||
})
|
||||
}
|
||||
}
|
||||
let sections = ["dev", "start"];
|
||||
let languages = ["zh", "en"];
|
||||
let ret: VitePressSidebarOptions[] = [];
|
||||
for (let language of languages) {
|
||||
for (let section of sections) {
|
||||
if (language === defaultLang) {
|
||||
ret.push({
|
||||
basePath: `/${section}/`,
|
||||
scanStartPath: `docs/${language}/${section}`,
|
||||
resolvePath: `/${section}/`,
|
||||
...commonSidebarOptions,
|
||||
});
|
||||
} else {
|
||||
ret.push({
|
||||
basePath: `/${language}/${section}/`,
|
||||
scanStartPath: `docs/${language}/${section}`,
|
||||
resolvePath: `/${language}/${section}/`,
|
||||
...commonSidebarOptions,
|
||||
});
|
||||
}
|
||||
}
|
||||
return ret
|
||||
}
|
||||
return ret;
|
||||
}
|
||||
|
||||
export const ThemeConfig = {
|
||||
getEditLink: (editPageText: string): { pattern: (params: { filePath: string; }) => string; text: string; } => {
|
||||
return {
|
||||
pattern: ({filePath}: { filePath: string; }): string => {
|
||||
if (!filePath) {
|
||||
throw new Error("filePath is undefined");
|
||||
}
|
||||
const regex = /^(dev\/api|[^\/]+\/dev\/api)/;
|
||||
if (regex.test(filePath)) {
|
||||
filePath = filePath.replace(regex, '')
|
||||
.replace('index.md', '__init__.py')
|
||||
.replace('.md', '.py');
|
||||
const fileName = filePath.split('/').pop();
|
||||
const parentFolder = filePath.split('/').slice(-2, -1)[0];
|
||||
if (fileName && parentFolder && fileName.split('.')[0] === parentFolder) {
|
||||
filePath = filePath.split('/').slice(0, -1).join('/') + '/__init__.py';
|
||||
}
|
||||
return `https://github.com/LiteyukiStudio/nonebot-plugin-marshoai/tree/main/nonebot_plugin_marshoai/${filePath}`;
|
||||
} else {
|
||||
return `https://github.com/LiteyukiStudio/nonebot-plugin-marshoai/tree/main/docs/${filePath}`;
|
||||
}
|
||||
},
|
||||
text: editPageText
|
||||
};
|
||||
},
|
||||
getEditLink: (
|
||||
editPageText: string
|
||||
): { pattern: (params: { filePath: string }) => string; text: string } => {
|
||||
return {
|
||||
pattern: ({ filePath }: { filePath: string }): string => {
|
||||
if (!filePath) {
|
||||
throw new Error("filePath is undefined");
|
||||
}
|
||||
const regex = /^(dev\/api|[^\/]+\/dev\/api)/;
|
||||
if (regex.test(filePath)) {
|
||||
filePath = filePath
|
||||
.replace(regex, "")
|
||||
.replace("index.md", "__init__.py")
|
||||
.replace(".md", ".py");
|
||||
const fileName = filePath.split("/").pop();
|
||||
const parentFolder = filePath.split("/").slice(-2, -1)[0];
|
||||
if (
|
||||
fileName &&
|
||||
parentFolder &&
|
||||
fileName.split(".")[0] === parentFolder
|
||||
) {
|
||||
filePath =
|
||||
filePath.split("/").slice(0, -1).join("/") + "/__init__.py";
|
||||
}
|
||||
return `https://github.com/LiteyukiStudio/nonebot-plugin-marshoai/tree/main/nonebot_plugin_marshoai/${filePath}`;
|
||||
} else {
|
||||
return `https://github.com/LiteyukiStudio/nonebot-plugin-marshoai/tree/main/docs/${filePath}`;
|
||||
}
|
||||
},
|
||||
text: editPageText,
|
||||
};
|
||||
},
|
||||
|
||||
getOutLine: (label: string): { label: string; level: [number, number]; } => {
|
||||
return {
|
||||
label: label,
|
||||
level: [2, 6]
|
||||
};
|
||||
},
|
||||
|
||||
copyright: 'Copyright (C) 2020-2024 LiteyukiStudio. All Rights Reserved'
|
||||
}
|
||||
getOutLine: (label: string): { label: string; level: [number, number] } => {
|
||||
return {
|
||||
label: label,
|
||||
level: [2, 6],
|
||||
};
|
||||
},
|
||||
};
|
||||
|
||||
@@ -23,7 +23,7 @@ export const en = defineConfig({
|
||||
lightModeSwitchTitle: 'Light',
|
||||
darkModeSwitchTitle: 'Dark',
|
||||
footer: {
|
||||
message: "The document is being improved. Suggestions are welcome.",
|
||||
message: "The document is being improved. Suggestions are welcome.<br>Webpage is deployed at <a href='https://meli.liteyuki.icu' target='_blank'>Liteyuki Meli</a> and accelerated by <a href='https://cdn.liteyuki.icu' target='_blank'>Liteyukiflare</a>.",
|
||||
copyright: '© 2024 <a href="https://liteyuki.icu" target="_blank">Liteyuki Studio</a>',
|
||||
}
|
||||
},
|
||||
|
||||
@@ -8,12 +8,13 @@ import { generateSidebar } from 'vitepress-sidebar'
|
||||
// https://vitepress.dev/reference/site-config
|
||||
export default defineConfig({
|
||||
head: [
|
||||
["script", { src: "https://cdn.liteyuki.icu/js/liteyuki_footer.js" }],
|
||||
['link', { rel: 'icon', type: 'image/x-icon', href: '/favicon.ico' }],
|
||||
],
|
||||
rewrites: {
|
||||
[`${defaultLang}/:rest*`]: ":rest*",
|
||||
},
|
||||
cleanUrls: true,
|
||||
cleanUrls: false,
|
||||
themeConfig: {
|
||||
// https://vitepress.dev/reference/default-theme-config
|
||||
logo: {
|
||||
|
||||
@@ -23,7 +23,7 @@ export const ja = defineConfig({
|
||||
lightModeSwitchTitle: 'ライト',
|
||||
darkModeSwitchTitle: 'ダーク',
|
||||
footer: {
|
||||
message: "ドキュメントは改善中です。ご意見をお待ちしております。",
|
||||
message: "ドキュメントは改善中です。ご意見をお待ちしております。<br>ウェブサイトは <a href='https://meli.liteyuki.icu' target='_blank'>Liteyuki Meli</a> によってデプロイされ、<a href='https://cdn.liteyuki.icu' target='_blank'>Liteyukiflare</a> によって加速されています。",
|
||||
copyright: '© 2024 <a href="https://liteyuki.icu" target="_blank">Liteyuki Studio</a>',
|
||||
}
|
||||
},
|
||||
|
||||
@@ -23,7 +23,7 @@ export const zh = defineConfig({
|
||||
lightModeSwitchTitle: '轻色模式',
|
||||
darkModeSwitchTitle: '深色模式',
|
||||
footer: {
|
||||
message: "文档完善中,欢迎提出建议或帮助我们完善。",
|
||||
message: "文档完善中,欢迎提出建议或帮助我们完善。<br>网站部署在 <a href='https://meli.liteyuki.icu' target='_blank'>Liteyuki Meli</a> 由 <a href='https://cdn.liteyuki.icu' target='_blank'>Liteyukiflare</a> 提供加速服务。",
|
||||
copyright: '© 2024 <a href="https://liteyuki.icu" target="_blank">Liteyuki Studio</a>',
|
||||
}
|
||||
},
|
||||
|
||||
@@ -147,4 +147,5 @@ Add options in the `.env` file from the diagram below in nonebot2 project.
|
||||
| MARSHOAI_ENABLE_RICHTEXT_PARSE | `bool` | `true` | Turn on auto parse rich text feature(including image, LaTeX equation) |
|
||||
| MARSHOAI_SINGLE_LATEX_PARSE | `bool` | `false`| Render single-line equation or not |
|
||||
| MARSHOAI_FIX_TOOLCALLS | `bool` | `true` | Fix tool calls or not |
|
||||
| MARSHOAI_SEND_THINKING | `bool` | `true` | Send thinking chain or not |
|
||||
| MARSHOAI_SEND_THINKING | `bool` | `true` | Send thinking chain or not |
|
||||
| MARSHOAI_STREAM | `bool` | `false`| 是否通过流式方式请求 API **开启此项后暂无法使用函数调用,无法在 Bot 用户侧聊天界面呈现出流式效果** |
|
||||
|
||||
@@ -149,6 +149,8 @@ GitHub Models API 的限制较多,不建议使用,建议通过修改`MARSHOA
|
||||
| MARSHOAI_SINGLE_LATEX_PARSE | `bool` | `false` | 单行公式是否渲染(当消息富文本解析启用时可用)(如果单行也渲……只能说不好看) |
|
||||
| MARSHOAI_FIX_TOOLCALLS | `bool` | `true` | 是否修复工具调用(部分模型须关闭,使用 vLLM 部署的模型时须关闭) |
|
||||
| MARSHOAI_SEND_THINKING | `bool` | `true` | 是否发送思维链(部分模型不支持) |
|
||||
| MARSHOAI_STREAM | `bool` | `false`| 是否通过流式方式请求 API **开启此项后暂无法使用函数调用,无法在 Bot 用户侧聊天界面呈现出流式效果** |
|
||||
|
||||
|
||||
#### 开发及调试选项
|
||||
|
||||
|
||||
39
nonebot_plugin_marshoai/cache/decos.py
vendored
Normal file
39
nonebot_plugin_marshoai/cache/decos.py
vendored
Normal file
@@ -0,0 +1,39 @@
|
||||
from ..models import Cache
|
||||
|
||||
cache = Cache()
|
||||
|
||||
|
||||
def from_cache(key):
|
||||
"""
|
||||
当缓存中有数据时,直接返回缓存中的数据,否则执行函数并将结果存入缓存
|
||||
"""
|
||||
|
||||
def decorator(func):
|
||||
async def wrapper(*args, **kwargs):
|
||||
cached = cache.get(key)
|
||||
if cached:
|
||||
return cached
|
||||
else:
|
||||
result = await func(*args, **kwargs)
|
||||
cache.set(key, result)
|
||||
return result
|
||||
|
||||
return wrapper
|
||||
|
||||
return decorator
|
||||
|
||||
|
||||
def update_to_cache(key):
|
||||
"""
|
||||
执行函数并将结果存入缓存
|
||||
"""
|
||||
|
||||
def decorator(func):
|
||||
async def wrapper(*args, **kwargs):
|
||||
result = await func(*args, **kwargs)
|
||||
cache.set(key, result)
|
||||
return result
|
||||
|
||||
return wrapper
|
||||
|
||||
return decorator
|
||||
@@ -1,4 +1,3 @@
|
||||
import shutil
|
||||
from io import StringIO
|
||||
from pathlib import Path
|
||||
|
||||
@@ -33,6 +32,7 @@ class ConfigModel(BaseModel):
|
||||
marshoai_enable_sysasuser_prompt: bool = False
|
||||
marshoai_additional_prompt: str = ""
|
||||
marshoai_poke_suffix: str = "揉了揉你的猫耳"
|
||||
marshoai_stream: bool = False
|
||||
marshoai_enable_richtext_parse: bool = True
|
||||
"""
|
||||
是否启用自动消息富文本解析 即若包含图片链接则发送图片、若包含LaTeX公式则发送公式图。
|
||||
@@ -81,15 +81,15 @@ destination_folder = Path("config/marshoai/")
|
||||
destination_file = destination_folder / "config.yaml"
|
||||
|
||||
|
||||
def dump_config_to_yaml(config: ConfigModel):
|
||||
return yaml_.dump(config.model_dump(), allow_unicode=True, default_flow_style=False)
|
||||
def dump_config_to_yaml(cfg: ConfigModel):
|
||||
return yaml_.dump(cfg.model_dump(), allow_unicode=True, default_flow_style=False)
|
||||
|
||||
|
||||
def write_default_config(destination_file):
|
||||
def write_default_config(dest_file):
|
||||
"""
|
||||
写入默认配置
|
||||
"""
|
||||
with open(destination_file, "w", encoding="utf-8") as f:
|
||||
with open(dest_file, "w", encoding="utf-8") as f:
|
||||
with StringIO(dump_config_to_yaml(ConfigModel())) as f2:
|
||||
f.write(f2.read())
|
||||
|
||||
@@ -110,17 +110,17 @@ def check_yaml_is_changed():
|
||||
return True
|
||||
|
||||
|
||||
def merge_configs(old_config, new_config):
|
||||
def merge_configs(existing_cfg, new_cfg):
|
||||
"""
|
||||
合并配置文件
|
||||
"""
|
||||
for key, value in new_config.items():
|
||||
if key in old_config:
|
||||
for key, value in new_cfg.items():
|
||||
if key in existing_cfg:
|
||||
continue
|
||||
else:
|
||||
logger.info(f"新增配置项: {key} = {value}")
|
||||
old_config[key] = value
|
||||
return old_config
|
||||
existing_cfg[key] = value
|
||||
return existing_cfg
|
||||
|
||||
|
||||
config: ConfigModel = get_plugin_config(ConfigModel)
|
||||
|
||||
327
nonebot_plugin_marshoai/handler.py
Normal file
327
nonebot_plugin_marshoai/handler.py
Normal file
@@ -0,0 +1,327 @@
|
||||
import json
|
||||
from typing import Optional, Tuple, Union
|
||||
|
||||
from azure.ai.inference.models import (
|
||||
CompletionsFinishReason,
|
||||
ImageContentItem,
|
||||
ImageUrl,
|
||||
TextContentItem,
|
||||
ToolMessage,
|
||||
UserMessage,
|
||||
)
|
||||
from nonebot.adapters import Bot, Event
|
||||
from nonebot.log import logger
|
||||
from nonebot.matcher import (
|
||||
Matcher,
|
||||
current_bot,
|
||||
current_event,
|
||||
current_matcher,
|
||||
)
|
||||
from nonebot_plugin_alconna.uniseg import UniMessage, UniMsg
|
||||
from openai import AsyncOpenAI, AsyncStream
|
||||
from openai.types.chat import ChatCompletion, ChatCompletionChunk, ChatCompletionMessage
|
||||
from openai.types.chat.chat_completion import Choice
|
||||
|
||||
from .config import config
|
||||
from .constants import SUPPORT_IMAGE_MODELS
|
||||
from .instances import target_list
|
||||
from .models import MarshoContext
|
||||
from .plugin.func_call.caller import get_function_calls
|
||||
from .plugin.func_call.models import SessionContext
|
||||
from .util import (
|
||||
extract_content_and_think,
|
||||
get_image_b64,
|
||||
get_nickname_by_user_id,
|
||||
get_prompt,
|
||||
make_chat_openai,
|
||||
parse_richtext,
|
||||
)
|
||||
|
||||
|
||||
class MarshoHandler:
|
||||
def __init__(
|
||||
self,
|
||||
client: AsyncOpenAI,
|
||||
context: MarshoContext,
|
||||
):
|
||||
self.client = client
|
||||
self.context = context
|
||||
self.bot: Bot = current_bot.get()
|
||||
self.event: Event = current_event.get()
|
||||
# self.state: T_State = current_handler.get().state
|
||||
self.matcher: Matcher = current_matcher.get()
|
||||
self.message_id: str = UniMessage.get_message_id(self.event)
|
||||
self.target = UniMessage.get_target(self.event)
|
||||
|
||||
async def process_user_input(
|
||||
self, user_input: UniMsg, model_name: str
|
||||
) -> Union[str, list]:
|
||||
"""
|
||||
处理用户输入为可输入 API 的格式,并添加昵称提示
|
||||
"""
|
||||
is_support_image_model = (
|
||||
model_name.lower()
|
||||
in SUPPORT_IMAGE_MODELS + config.marshoai_additional_image_models
|
||||
)
|
||||
usermsg = [] if is_support_image_model else ""
|
||||
user_nickname = await get_nickname_by_user_id(self.event.get_user_id())
|
||||
if user_nickname:
|
||||
nickname_prompt = f"\n此消息的说话者为: {user_nickname}"
|
||||
else:
|
||||
nickname_prompt = ""
|
||||
for i in user_input: # type: ignore
|
||||
if i.type == "text":
|
||||
if is_support_image_model:
|
||||
usermsg += [TextContentItem(text=i.data["text"] + nickname_prompt).as_dict()] # type: ignore
|
||||
else:
|
||||
usermsg += str(i.data["text"] + nickname_prompt) # type: ignore
|
||||
elif i.type == "image":
|
||||
if is_support_image_model:
|
||||
usermsg.append( # type: ignore
|
||||
ImageContentItem(
|
||||
image_url=ImageUrl( # type: ignore
|
||||
url=str(await get_image_b64(i.data["url"])) # type: ignore
|
||||
) # type: ignore
|
||||
).as_dict() # type: ignore
|
||||
) # type: ignore
|
||||
logger.info(f"输入图片 {i.data['url']}")
|
||||
elif config.marshoai_enable_support_image_tip:
|
||||
await UniMessage(
|
||||
"*此模型不支持图片处理或管理员未启用此模型的图片支持。图片将被忽略。"
|
||||
).send()
|
||||
return usermsg # type: ignore
|
||||
|
||||
async def handle_single_chat(
|
||||
self,
|
||||
user_message: Union[str, list],
|
||||
model_name: str,
|
||||
tools_list: list | None,
|
||||
tool_message: Optional[list] = None,
|
||||
stream: bool = False,
|
||||
) -> Union[ChatCompletion, AsyncStream[ChatCompletionChunk]]:
|
||||
"""
|
||||
处理单条聊天
|
||||
"""
|
||||
|
||||
context_msg = get_prompt(model_name) + (
|
||||
self.context.build(self.target.id, self.target.private)
|
||||
)
|
||||
response = await make_chat_openai(
|
||||
client=self.client,
|
||||
msg=context_msg + [UserMessage(content=user_message).as_dict()] + (tool_message if tool_message else []), # type: ignore
|
||||
model_name=model_name,
|
||||
tools=tools_list if tools_list else None,
|
||||
stream=stream,
|
||||
)
|
||||
return response
|
||||
|
||||
async def handle_function_call(
|
||||
self,
|
||||
completion: Union[ChatCompletion, AsyncStream[ChatCompletionChunk]],
|
||||
user_message: Union[str, list],
|
||||
model_name: str,
|
||||
tools_list: list,
|
||||
):
|
||||
# function call
|
||||
# 需要获取额外信息,调用函数工具
|
||||
tool_msg = []
|
||||
if isinstance(completion, ChatCompletion):
|
||||
choice = completion.choices[0]
|
||||
else:
|
||||
raise ValueError("Unexpected completion type")
|
||||
# await UniMessage(str(response)).send()
|
||||
tool_calls = choice.message.tool_calls
|
||||
# try:
|
||||
# if tool_calls[0]["function"]["name"].startswith("$"):
|
||||
# choice.message.tool_calls[0][
|
||||
# "type"
|
||||
# ] = "builtin_function" # 兼容 moonshot AI 内置函数的临时方案
|
||||
# except:
|
||||
# pass
|
||||
tool_msg.append(choice.message)
|
||||
for tool_call in tool_calls: # type: ignore
|
||||
try:
|
||||
function_args = json.loads(tool_call.function.arguments)
|
||||
except json.JSONDecodeError:
|
||||
function_args = json.loads(
|
||||
tool_call.function.arguments.replace("'", '"')
|
||||
)
|
||||
# 删除args的placeholder参数
|
||||
if "placeholder" in function_args:
|
||||
del function_args["placeholder"]
|
||||
logger.info(
|
||||
f"调用函数 {tool_call.function.name.replace('-', '.')}\n参数:"
|
||||
+ "\n".join([f"{k}={v}" for k, v in function_args.items()])
|
||||
)
|
||||
await UniMessage(
|
||||
f"调用函数 {tool_call.function.name.replace('-', '.')}\n参数:"
|
||||
+ "\n".join([f"{k}={v}" for k, v in function_args.items()])
|
||||
).send()
|
||||
if caller := get_function_calls().get(tool_call.function.name):
|
||||
logger.debug(f"调用插件函数 {caller.full_name}")
|
||||
# 权限检查,规则检查 TODO
|
||||
# 实现依赖注入,检查函数参数及参数注解类型,对Event类型的参数进行注入
|
||||
func_return = await caller.with_ctx(
|
||||
SessionContext(
|
||||
bot=self.bot,
|
||||
event=self.event,
|
||||
matcher=self.matcher,
|
||||
state=None,
|
||||
)
|
||||
).call(**function_args)
|
||||
else:
|
||||
logger.error(f"未找到函数 {tool_call.function.name.replace('-', '.')}")
|
||||
func_return = f"未找到函数 {tool_call.function.name.replace('-', '.')}"
|
||||
tool_msg.append(
|
||||
ToolMessage(tool_call_id=tool_call.id, content=func_return).as_dict() # type: ignore
|
||||
)
|
||||
# tool_msg[0]["tool_calls"][0]["type"] = "builtin_function"
|
||||
# await UniMessage(str(tool_msg)).send()
|
||||
return await self.handle_common_chat(
|
||||
user_message=user_message,
|
||||
model_name=model_name,
|
||||
tools_list=tools_list,
|
||||
tool_message=tool_msg,
|
||||
)
|
||||
|
||||
async def handle_common_chat(
|
||||
self,
|
||||
user_message: Union[str, list],
|
||||
model_name: str,
|
||||
tools_list: list,
|
||||
stream: bool = False,
|
||||
tool_message: Optional[list] = None,
|
||||
) -> Optional[Tuple[UserMessage, ChatCompletionMessage]]:
|
||||
"""
|
||||
处理一般聊天
|
||||
"""
|
||||
global target_list
|
||||
if stream:
|
||||
response = await self.handle_stream_request(
|
||||
user_message=user_message,
|
||||
model_name=model_name,
|
||||
tools_list=tools_list,
|
||||
tools_message=tool_message,
|
||||
)
|
||||
else:
|
||||
response = await self.handle_single_chat( # type: ignore
|
||||
user_message=user_message,
|
||||
model_name=model_name,
|
||||
tools_list=tools_list,
|
||||
tool_message=tool_message,
|
||||
)
|
||||
choice = response.choices[0] # type: ignore
|
||||
# Sprint(choice)
|
||||
# 当tool_calls非空时,将finish_reason设置为TOOL_CALLS
|
||||
if choice.message.tool_calls is not None and config.marshoai_fix_toolcalls:
|
||||
choice.finish_reason = "tool_calls"
|
||||
logger.info(f"完成原因:{choice.finish_reason}")
|
||||
if choice.finish_reason == CompletionsFinishReason.STOPPED:
|
||||
|
||||
##### DeepSeek-R1 兼容部分 #####
|
||||
choice_msg_content, choice_msg_thinking, choice_msg_after = (
|
||||
extract_content_and_think(choice.message)
|
||||
)
|
||||
if choice_msg_thinking and config.marshoai_send_thinking:
|
||||
await UniMessage("思维链:\n" + choice_msg_thinking).send()
|
||||
##### 兼容部分结束 #####
|
||||
|
||||
if [self.target.id, self.target.private] not in target_list:
|
||||
target_list.append([self.target.id, self.target.private])
|
||||
|
||||
# 对话成功发送消息
|
||||
if config.marshoai_enable_richtext_parse:
|
||||
await (await parse_richtext(str(choice_msg_content))).send(
|
||||
reply_to=True
|
||||
)
|
||||
else:
|
||||
await UniMessage(str(choice_msg_content)).send(reply_to=True)
|
||||
return UserMessage(content=user_message), choice_msg_after
|
||||
elif choice.finish_reason == CompletionsFinishReason.CONTENT_FILTERED:
|
||||
|
||||
# 对话失败,消息过滤
|
||||
|
||||
await UniMessage("*已被内容过滤器过滤。请调整聊天内容后重试。").send(
|
||||
reply_to=True
|
||||
)
|
||||
return None
|
||||
elif choice.finish_reason == CompletionsFinishReason.TOOL_CALLS:
|
||||
return await self.handle_function_call(
|
||||
response, user_message, model_name, tools_list
|
||||
)
|
||||
else:
|
||||
await UniMessage(f"意外的完成原因:{choice.finish_reason}").send()
|
||||
return None
|
||||
|
||||
async def handle_stream_request(
|
||||
self,
|
||||
user_message: Union[str, list],
|
||||
model_name: str,
|
||||
tools_list: list,
|
||||
tools_message: Optional[list] = None,
|
||||
) -> Union[ChatCompletion, None]:
|
||||
"""
|
||||
处理流式请求
|
||||
"""
|
||||
response = await self.handle_single_chat(
|
||||
user_message=user_message,
|
||||
model_name=model_name,
|
||||
tools_list=None, # TODO:让流式调用支持工具调用
|
||||
tool_message=tools_message,
|
||||
stream=True,
|
||||
)
|
||||
|
||||
if isinstance(response, AsyncStream):
|
||||
reasoning_contents = ""
|
||||
answer_contents = ""
|
||||
last_chunk = None
|
||||
is_first_token_appeared = False
|
||||
is_answering = False
|
||||
async for chunk in response:
|
||||
last_chunk = chunk
|
||||
# print(chunk)
|
||||
if not is_first_token_appeared:
|
||||
logger.debug(f"{chunk.id}: 第一个 token 已出现")
|
||||
is_first_token_appeared = True
|
||||
if not chunk.choices:
|
||||
logger.info("Usage:", chunk.usage)
|
||||
else:
|
||||
delta = chunk.choices[0].delta
|
||||
if (
|
||||
hasattr(delta, "reasoning_content")
|
||||
and delta.reasoning_content is not None
|
||||
):
|
||||
reasoning_contents += delta.reasoning_content
|
||||
else:
|
||||
if not is_answering:
|
||||
logger.debug(
|
||||
f"{chunk.id}: 思维链已输出完毕或无 reasoning_content 字段输出"
|
||||
)
|
||||
is_answering = True
|
||||
if delta.content is not None:
|
||||
answer_contents += delta.content
|
||||
# print(last_chunk)
|
||||
# 创建新的 ChatCompletion 对象
|
||||
if last_chunk and last_chunk.choices:
|
||||
message = ChatCompletionMessage(
|
||||
content=answer_contents,
|
||||
role="assistant",
|
||||
tool_calls=last_chunk.choices[0].delta.tool_calls, # type: ignore
|
||||
)
|
||||
if reasoning_contents != "":
|
||||
setattr(message, "reasoning_content", reasoning_contents)
|
||||
choice = Choice(
|
||||
finish_reason=last_chunk.choices[0].finish_reason, # type: ignore
|
||||
index=last_chunk.choices[0].index,
|
||||
message=message,
|
||||
)
|
||||
return ChatCompletion(
|
||||
id=last_chunk.id,
|
||||
choices=[choice],
|
||||
created=last_chunk.created,
|
||||
model=last_chunk.model,
|
||||
system_fingerprint=last_chunk.system_fingerprint,
|
||||
object="chat.completion",
|
||||
usage=last_chunk.usage,
|
||||
)
|
||||
return None
|
||||
@@ -6,7 +6,7 @@ import nonebot_plugin_localstore as store
|
||||
from nonebot import logger
|
||||
|
||||
from .config import config
|
||||
from .instances import *
|
||||
from .instances import context, driver, target_list, tools
|
||||
from .plugin import load_plugin, load_plugins
|
||||
from .util import get_backup_context, save_context_to_json
|
||||
|
||||
|
||||
@@ -1,6 +1,4 @@
|
||||
# Marsho 的类实例以及全局变量
|
||||
from azure.ai.inference.aio import ChatCompletionsClient
|
||||
from azure.core.credentials import AzureKeyCredential
|
||||
from nonebot import get_driver
|
||||
from openai import AsyncOpenAI
|
||||
|
||||
|
||||
@@ -2,15 +2,10 @@ import contextlib
|
||||
import traceback
|
||||
from typing import Optional
|
||||
|
||||
import openai
|
||||
from arclet.alconna import Alconna, AllParam, Args
|
||||
from azure.ai.inference.models import (
|
||||
AssistantMessage,
|
||||
CompletionsFinishReason,
|
||||
ImageContentItem,
|
||||
ImageUrl,
|
||||
TextContentItem,
|
||||
ToolMessage,
|
||||
UserMessage,
|
||||
)
|
||||
from nonebot import logger, on_command, on_message
|
||||
@@ -18,15 +13,17 @@ from nonebot.adapters import Bot, Event, Message
|
||||
from nonebot.matcher import Matcher
|
||||
from nonebot.params import CommandArg
|
||||
from nonebot.permission import SUPERUSER
|
||||
from nonebot.rule import Rule, to_me
|
||||
from nonebot.rule import to_me
|
||||
from nonebot.typing import T_State
|
||||
from nonebot_plugin_alconna import MsgTarget, UniMessage, UniMsg, on_alconna
|
||||
|
||||
from .config import config
|
||||
from .constants import INTRODUCTION, SUPPORT_IMAGE_MODELS
|
||||
from .handler import MarshoHandler
|
||||
from .hooks import *
|
||||
from .instances import *
|
||||
from .instances import client, context, model_name, target_list, tools
|
||||
from .metadata import metadata
|
||||
from .plugin.func_call.caller import get_function_calls
|
||||
from .plugin.func_call.models import SessionContext
|
||||
from .util import *
|
||||
|
||||
|
||||
@@ -35,20 +32,20 @@ async def at_enable():
|
||||
|
||||
|
||||
changemodel_cmd = on_command(
|
||||
"changemodel", permission=SUPERUSER, priority=10, block=True
|
||||
"changemodel", permission=SUPERUSER, priority=96, block=True
|
||||
)
|
||||
# setprompt_cmd = on_command("prompt",permission=SUPERUSER)
|
||||
praises_cmd = on_command("praises", permission=SUPERUSER, priority=10, block=True)
|
||||
add_usermsg_cmd = on_command("usermsg", permission=SUPERUSER, priority=10, block=True)
|
||||
praises_cmd = on_command("praises", permission=SUPERUSER, priority=96, block=True)
|
||||
add_usermsg_cmd = on_command("usermsg", permission=SUPERUSER, priority=96, block=True)
|
||||
add_assistantmsg_cmd = on_command(
|
||||
"assistantmsg", permission=SUPERUSER, priority=10, block=True
|
||||
"assistantmsg", permission=SUPERUSER, priority=96, block=True
|
||||
)
|
||||
contexts_cmd = on_command("contexts", permission=SUPERUSER, priority=10, block=True)
|
||||
contexts_cmd = on_command("contexts", permission=SUPERUSER, priority=96, block=True)
|
||||
save_context_cmd = on_command(
|
||||
"savecontext", permission=SUPERUSER, priority=10, block=True
|
||||
"savecontext", permission=SUPERUSER, priority=96, block=True
|
||||
)
|
||||
load_context_cmd = on_command(
|
||||
"loadcontext", permission=SUPERUSER, priority=10, block=True
|
||||
"loadcontext", permission=SUPERUSER, priority=96, block=True
|
||||
)
|
||||
marsho_cmd = on_alconna(
|
||||
Alconna(
|
||||
@@ -56,42 +53,42 @@ marsho_cmd = on_alconna(
|
||||
Args["text?", AllParam],
|
||||
),
|
||||
aliases=tuple(config.marshoai_aliases),
|
||||
priority=10,
|
||||
priority=96,
|
||||
block=True,
|
||||
)
|
||||
resetmem_cmd = on_alconna(
|
||||
Alconna(
|
||||
config.marshoai_default_name + ".reset",
|
||||
),
|
||||
priority=10,
|
||||
priority=96,
|
||||
block=True,
|
||||
)
|
||||
marsho_help_cmd = on_alconna(
|
||||
Alconna(
|
||||
config.marshoai_default_name + ".help",
|
||||
),
|
||||
priority=10,
|
||||
priority=96,
|
||||
block=True,
|
||||
)
|
||||
marsho_status_cmd = on_alconna(
|
||||
Alconna(
|
||||
config.marshoai_default_name + ".status",
|
||||
),
|
||||
priority=10,
|
||||
priority=96,
|
||||
block=True,
|
||||
)
|
||||
|
||||
marsho_at = on_message(rule=to_me() & at_enable, priority=11)
|
||||
marsho_at = on_message(rule=to_me() & at_enable, priority=97)
|
||||
nickname_cmd = on_alconna(
|
||||
Alconna(
|
||||
"nickname",
|
||||
Args["name?", str],
|
||||
),
|
||||
priority=10,
|
||||
priority=96,
|
||||
block=True,
|
||||
)
|
||||
refresh_data_cmd = on_command(
|
||||
"refresh_data", permission=SUPERUSER, priority=10, block=True
|
||||
"refresh_data", permission=SUPERUSER, priority=96, block=True
|
||||
)
|
||||
|
||||
|
||||
@@ -230,16 +227,16 @@ async def marsho(
|
||||
# 发送说明
|
||||
# await UniMessage(metadata.usage + "\n当前使用的模型:" + model_name).send()
|
||||
await marsho_cmd.finish(INTRODUCTION)
|
||||
backup_context = await get_backup_context(target.id, target.private)
|
||||
if backup_context:
|
||||
context.set_context(
|
||||
backup_context, target.id, target.private
|
||||
) # 加载历史记录
|
||||
logger.info(f"已恢复会话 {target.id} 的上下文备份~")
|
||||
handler = MarshoHandler(client, context)
|
||||
try:
|
||||
user_id = event.get_user_id()
|
||||
nicknames = await get_nicknames()
|
||||
user_nickname = nicknames.get(user_id, "")
|
||||
if user_nickname != "":
|
||||
nickname_prompt = (
|
||||
f"\n*此消息的说话者id为:{user_id},名字为:{user_nickname}*"
|
||||
)
|
||||
else:
|
||||
nickname_prompt = ""
|
||||
user_nickname = await get_nickname_by_user_id(event.get_user_id())
|
||||
if not user_nickname:
|
||||
# 用户名无法获取,暂时注释
|
||||
# user_nickname = event.sender.nickname # 未设置昵称时获取用户名
|
||||
# nickname_prompt = f"\n*此消息的说话者:{user_nickname}"
|
||||
@@ -253,189 +250,23 @@ async def marsho(
|
||||
"※你未设置自己的昵称。推荐使用「nickname [昵称]」命令设置昵称来获得个性化(可能)回答。"
|
||||
).send()
|
||||
|
||||
is_support_image_model = (
|
||||
model_name.lower()
|
||||
in SUPPORT_IMAGE_MODELS + config.marshoai_additional_image_models
|
||||
)
|
||||
is_openai_new_model = model_name.lower() in OPENAI_NEW_MODELS
|
||||
usermsg = [] if is_support_image_model else ""
|
||||
for i in text: # type: ignore
|
||||
if i.type == "text":
|
||||
if is_support_image_model:
|
||||
usermsg += [TextContentItem(text=i.data["text"] + nickname_prompt).as_dict()] # type: ignore
|
||||
else:
|
||||
usermsg += str(i.data["text"] + nickname_prompt) # type: ignore
|
||||
elif i.type == "image":
|
||||
if is_support_image_model:
|
||||
usermsg.append( # type: ignore
|
||||
ImageContentItem(
|
||||
image_url=ImageUrl( # type: ignore
|
||||
url=str(await get_image_b64(i.data["url"])) # type: ignore
|
||||
) # type: ignore
|
||||
).as_dict() # type: ignore
|
||||
) # type: ignore
|
||||
logger.info(f"输入图片 {i.data['url']}")
|
||||
elif config.marshoai_enable_support_image_tip:
|
||||
await UniMessage(
|
||||
"*此模型不支持图片处理或管理员未启用此模型的图片支持。图片将被忽略。"
|
||||
).send()
|
||||
backup_context = await get_backup_context(target.id, target.private)
|
||||
if backup_context:
|
||||
context.set_context(
|
||||
backup_context, target.id, target.private
|
||||
) # 加载历史记录
|
||||
logger.info(f"已恢复会话 {target.id} 的上下文备份~")
|
||||
context_msg = get_prompt(model_name) + context.build(target.id, target.private)
|
||||
usermsg = await handler.process_user_input(text, model_name)
|
||||
|
||||
tools_lists = tools.tools_list + list(
|
||||
map(lambda v: v.data(), get_function_calls().values())
|
||||
)
|
||||
logger.info(f"正在获取回答,模型:{model_name}")
|
||||
# logger.info(f"上下文:{context_msg}")
|
||||
response = await make_chat_openai(
|
||||
client=client,
|
||||
model_name=model_name,
|
||||
msg=context_msg + [UserMessage(content=usermsg).as_dict()], # type: ignore
|
||||
tools=tools_lists if tools_lists else None, # TODO 临时追加函数,后期优化
|
||||
response = await handler.handle_common_chat(
|
||||
usermsg, model_name, tools_lists, config.marshoai_stream
|
||||
)
|
||||
# await UniMessage(str(response)).send()
|
||||
choice = response.choices[0]
|
||||
# Sprint(choice)
|
||||
# 当tool_calls非空时,将finish_reason设置为TOOL_CALLS
|
||||
if choice.message.tool_calls != None and config.marshoai_fix_toolcalls:
|
||||
choice.finish_reason = "tool_calls"
|
||||
logger.info(f"完成原因:{choice.finish_reason}")
|
||||
if choice.finish_reason == CompletionsFinishReason.STOPPED:
|
||||
# 当对话成功时,将dict的上下文添加到上下文类中
|
||||
context.append(
|
||||
UserMessage(content=usermsg).as_dict(), target.id, target.private # type: ignore
|
||||
)
|
||||
|
||||
##### DeepSeek-R1 兼容部分 #####
|
||||
choice_msg_content, choice_msg_thinking, choice_msg_after = (
|
||||
extract_content_and_think(choice.message)
|
||||
)
|
||||
if choice_msg_thinking and config.marshoai_send_thinking:
|
||||
await UniMessage("思维链:\n" + choice_msg_thinking).send()
|
||||
##### 兼容部分结束 #####
|
||||
|
||||
context.append(choice_msg_after.to_dict(), target.id, target.private)
|
||||
if [target.id, target.private] not in target_list:
|
||||
target_list.append([target.id, target.private])
|
||||
|
||||
# 对话成功发送消息
|
||||
if config.marshoai_enable_richtext_parse:
|
||||
await (await parse_richtext(str(choice_msg_content))).send(
|
||||
reply_to=True
|
||||
)
|
||||
else:
|
||||
await UniMessage(str(choice_msg_content)).send(reply_to=True)
|
||||
elif choice.finish_reason == CompletionsFinishReason.CONTENT_FILTERED:
|
||||
|
||||
# 对话失败,消息过滤
|
||||
|
||||
await UniMessage("*已被内容过滤器过滤。请调整聊天内容后重试。").send(
|
||||
reply_to=True
|
||||
)
|
||||
return
|
||||
elif choice.finish_reason == CompletionsFinishReason.TOOL_CALLS:
|
||||
# function call
|
||||
# 需要获取额外信息,调用函数工具
|
||||
tool_msg = []
|
||||
while choice.message.tool_calls != None:
|
||||
# await UniMessage(str(response)).send()
|
||||
tool_calls = choice.message.tool_calls
|
||||
# try:
|
||||
# if tool_calls[0]["function"]["name"].startswith("$"):
|
||||
# choice.message.tool_calls[0][
|
||||
# "type"
|
||||
# ] = "builtin_function" # 兼容 moonshot AI 内置函数的临时方案
|
||||
# except:
|
||||
# pass
|
||||
tool_msg.append(choice.message)
|
||||
for tool_call in tool_calls:
|
||||
try:
|
||||
function_args = json.loads(tool_call.function.arguments)
|
||||
except json.JSONDecodeError:
|
||||
function_args = json.loads(
|
||||
tool_call.function.arguments.replace("'", '"')
|
||||
)
|
||||
# 删除args的placeholder参数
|
||||
if "placeholder" in function_args:
|
||||
del function_args["placeholder"]
|
||||
logger.info(
|
||||
f"调用函数 {tool_call.function.name.replace('-', '.')}\n参数:"
|
||||
+ "\n".join([f"{k}={v}" for k, v in function_args.items()])
|
||||
)
|
||||
await UniMessage(
|
||||
f"调用函数 {tool_call.function.name.replace('-', '.')}\n参数:"
|
||||
+ "\n".join([f"{k}={v}" for k, v in function_args.items()])
|
||||
).send()
|
||||
# TODO 临时追加插件函数,若工具中没有则调用插件函数
|
||||
if tools.has_function(tool_call.function.name):
|
||||
logger.debug(f"调用工具函数 {tool_call.function.name}")
|
||||
func_return = await tools.call(
|
||||
tool_call.function.name, function_args
|
||||
) # 获取返回值
|
||||
else:
|
||||
if caller := get_function_calls().get(tool_call.function.name):
|
||||
logger.debug(f"调用插件函数 {caller.full_name}")
|
||||
# 权限检查,规则检查 TODO
|
||||
# 实现依赖注入,检查函数参数及参数注解类型,对Event类型的参数进行注入
|
||||
func_return = await caller.with_ctx(
|
||||
SessionContext(
|
||||
bot=bot,
|
||||
event=event,
|
||||
state=state,
|
||||
matcher=matcher,
|
||||
)
|
||||
).call(**function_args)
|
||||
else:
|
||||
logger.error(
|
||||
f"未找到函数 {tool_call.function.name.replace('-', '.')}"
|
||||
)
|
||||
func_return = f"未找到函数 {tool_call.function.name.replace('-', '.')}"
|
||||
tool_msg.append(
|
||||
ToolMessage(tool_call_id=tool_call.id, content=func_return).as_dict() # type: ignore
|
||||
)
|
||||
# tool_msg[0]["tool_calls"][0]["type"] = "builtin_function"
|
||||
# await UniMessage(str(tool_msg)).send()
|
||||
request_msg = context_msg + [UserMessage(content=usermsg).as_dict()] + tool_msg # type: ignore
|
||||
response = await make_chat_openai(
|
||||
client=client,
|
||||
model_name=model_name,
|
||||
msg=request_msg, # type: ignore
|
||||
tools=(
|
||||
tools_lists if tools_lists else None
|
||||
), # TODO 临时追加函数,后期优化
|
||||
)
|
||||
choice = response.choices[0]
|
||||
# 当tool_calls非空时,将finish_reason设置为TOOL_CALLS
|
||||
if choice.message.tool_calls != None:
|
||||
choice.finish_reason = CompletionsFinishReason.TOOL_CALLS
|
||||
if choice.finish_reason == CompletionsFinishReason.STOPPED:
|
||||
|
||||
# 对话成功 添加上下文
|
||||
context.append(
|
||||
UserMessage(content=usermsg).as_dict(), target.id, target.private # type: ignore
|
||||
)
|
||||
# context.append(tool_msg, target.id, target.private)
|
||||
choice_msg_dict = choice.message.to_dict()
|
||||
if "reasoning_content" in choice_msg_dict:
|
||||
del choice_msg_dict["reasoning_content"]
|
||||
context.append(choice_msg_dict, target.id, target.private)
|
||||
|
||||
# 发送消息
|
||||
if config.marshoai_enable_richtext_parse:
|
||||
await (await parse_richtext(str(choice.message.content))).send(
|
||||
reply_to=True
|
||||
)
|
||||
else:
|
||||
await UniMessage(str(choice.message.content)).send(reply_to=True)
|
||||
else:
|
||||
await marsho_cmd.finish(f"意外的完成原因:{choice.finish_reason}")
|
||||
if response is not None:
|
||||
context_user, context_assistant = response
|
||||
context.append(context_user.as_dict(), target.id, target.private)
|
||||
context.append(context_assistant.to_dict(), target.id, target.private)
|
||||
else:
|
||||
await marsho_cmd.finish(f"意外的完成原因:{choice.finish_reason}")
|
||||
return
|
||||
except Exception as e:
|
||||
await UniMessage(str(e) + suggest_solution(str(e))).send()
|
||||
traceback.print_exc()
|
||||
@@ -450,12 +281,10 @@ with contextlib.suppress(ImportError): # 优化先不做()
|
||||
@poke_notify.handle()
|
||||
async def poke(event: Event):
|
||||
|
||||
user_id = event.get_user_id()
|
||||
nicknames = await get_nicknames()
|
||||
user_nickname = nicknames.get(user_id, "")
|
||||
user_nickname = await get_nickname_by_user_id(event.get_user_id())
|
||||
try:
|
||||
if config.marshoai_poke_suffix != "":
|
||||
logger.info(f"收到戳一戳,用户昵称:{user_nickname},用户ID:{user_id}")
|
||||
logger.info(f"收到戳一戳,用户昵称:{user_nickname}")
|
||||
response = await make_chat_openai(
|
||||
client=client,
|
||||
model_name=model_name,
|
||||
@@ -466,7 +295,7 @@ with contextlib.suppress(ImportError): # 优化先不做()
|
||||
),
|
||||
],
|
||||
)
|
||||
choice = response.choices[0]
|
||||
choice = response.choices[0] # type: ignore
|
||||
if choice.finish_reason == CompletionsFinishReason.STOPPED:
|
||||
content = extract_content_and_think(choice.message)[0]
|
||||
await UniMessage(" " + str(content)).send(at_sender=True)
|
||||
|
||||
@@ -5,7 +5,7 @@ from .constants import USAGE
|
||||
|
||||
metadata = PluginMetadata(
|
||||
name="Marsho AI 插件",
|
||||
description="接入 Azure API 或其他 API 的 AI 聊天插件,支持图片处理,外部函数调用,兼容包括 DeepSeek-R1 在内的多个模型",
|
||||
description="接入 Azure API 或其他 API 的 AI 聊天插件,支持图片处理,外部函数调用,兼容包括 DeepSeek-R1, QwQ-32B 在内的多个模型",
|
||||
usage=USAGE,
|
||||
type="application",
|
||||
config=ConfigModel,
|
||||
|
||||
@@ -9,7 +9,25 @@ import traceback
|
||||
from nonebot import logger
|
||||
|
||||
from .config import config
|
||||
from .util import *
|
||||
|
||||
|
||||
class Cache:
|
||||
"""
|
||||
缓存类
|
||||
"""
|
||||
|
||||
def __init__(self):
|
||||
self.cache = {}
|
||||
|
||||
def get(self, key):
|
||||
if key in self.cache:
|
||||
return self.cache[key]
|
||||
else:
|
||||
self.cache[key] = None
|
||||
return None
|
||||
|
||||
def set(self, key, value):
|
||||
self.cache[key] = value
|
||||
|
||||
|
||||
class MarshoContext:
|
||||
|
||||
@@ -70,8 +70,8 @@ class Caller:
|
||||
):
|
||||
return False, "告诉用户 Permission Denied 权限不足"
|
||||
|
||||
if self.ctx.state is None:
|
||||
return False, "State is None"
|
||||
# if self.ctx.state is None:
|
||||
# return False, "State is None"
|
||||
if self._rule and not await self._rule(
|
||||
self.ctx.bot, self.ctx.event, self.ctx.state
|
||||
):
|
||||
@@ -115,6 +115,10 @@ class Caller:
|
||||
# 检查函数签名,确定依赖注入参数
|
||||
sig = inspect.signature(func)
|
||||
for name, param in sig.parameters.items():
|
||||
# if param.annotation == T_State:
|
||||
# self.di.state = name
|
||||
# continue # 防止后续判断T_State子类时报错
|
||||
|
||||
if issubclass(param.annotation, Event) or isinstance(
|
||||
param.annotation, Event
|
||||
):
|
||||
@@ -133,9 +137,6 @@ class Caller:
|
||||
):
|
||||
self.di.matcher = name
|
||||
|
||||
if param.annotation == T_State:
|
||||
self.di.state = name
|
||||
|
||||
# 检查默认值情况
|
||||
for name, param in sig.parameters.items():
|
||||
if param.default is not inspect.Parameter.empty:
|
||||
|
||||
@@ -19,7 +19,7 @@ class SessionContext(BaseModel):
|
||||
bot: Bot
|
||||
event: Event
|
||||
matcher: Matcher
|
||||
state: T_State
|
||||
state: T_State | None
|
||||
caller: Any = None
|
||||
|
||||
class Config:
|
||||
@@ -30,5 +30,5 @@ class SessionContextDepends(BaseModel):
|
||||
bot: str | None = None
|
||||
event: str | None = None
|
||||
matcher: str | None = None
|
||||
state: str | None = None
|
||||
# state: str | None = None
|
||||
caller: str | None = None
|
||||
|
||||
@@ -16,7 +16,7 @@ marsho_memory_cmd = on_alconna(
|
||||
Subcommand("view", alias={"v"}),
|
||||
Subcommand("reset", alias={"r"}),
|
||||
),
|
||||
priority=10,
|
||||
priority=96,
|
||||
block=True,
|
||||
)
|
||||
|
||||
|
||||
@@ -3,7 +3,7 @@ import json
|
||||
import mimetypes
|
||||
import re
|
||||
import uuid
|
||||
from typing import Any, Dict, List, Optional
|
||||
from typing import Any, Dict, List, Optional, Union
|
||||
|
||||
import aiofiles # type: ignore
|
||||
import httpx
|
||||
@@ -15,18 +15,19 @@ from nonebot.log import logger
|
||||
from nonebot_plugin_alconna import Image as ImageMsg
|
||||
from nonebot_plugin_alconna import Text as TextMsg
|
||||
from nonebot_plugin_alconna import UniMessage
|
||||
from openai import AsyncOpenAI, NotGiven
|
||||
from openai.types.chat import ChatCompletion, ChatCompletionMessage
|
||||
from openai import AsyncOpenAI, AsyncStream, NotGiven
|
||||
from openai.types.chat import ChatCompletion, ChatCompletionChunk, ChatCompletionMessage
|
||||
from zhDateTime import DateTime
|
||||
|
||||
from ._types import DeveloperMessage
|
||||
from .cache.decos import *
|
||||
from .config import config
|
||||
from .constants import *
|
||||
from .constants import CODE_BLOCK_PATTERN, IMG_LATEX_PATTERN, OPENAI_NEW_MODELS
|
||||
from .deal_latex import ConvertLatex
|
||||
|
||||
nickname_json = None # 记录昵称
|
||||
praises_json = None # 记录夸赞名单
|
||||
loaded_target_list = [] # 记录已恢复备份的上下文的列表
|
||||
# nickname_json = None # 记录昵称
|
||||
# praises_json = None # 记录夸赞名单
|
||||
loaded_target_list: List[str] = [] # 记录已恢复备份的上下文的列表
|
||||
|
||||
NOT_GIVEN = NotGiven()
|
||||
|
||||
@@ -108,35 +109,13 @@ async def get_image_b64(url: str, timeout: int = 10) -> Optional[str]:
|
||||
return None
|
||||
|
||||
|
||||
async def make_chat(
|
||||
client: ChatCompletionsClient,
|
||||
msg: list,
|
||||
model_name: str,
|
||||
tools: Optional[list] = None,
|
||||
):
|
||||
"""
|
||||
调用ai获取回复
|
||||
|
||||
参数:
|
||||
client: 用于与AI模型进行通信
|
||||
msg: 消息内容
|
||||
model_name: 指定AI模型名
|
||||
tools: 工具列表
|
||||
"""
|
||||
return await client.complete(
|
||||
messages=msg,
|
||||
model=model_name,
|
||||
tools=tools,
|
||||
**config.marshoai_model_args,
|
||||
)
|
||||
|
||||
|
||||
async def make_chat_openai(
|
||||
client: AsyncOpenAI,
|
||||
msg: list,
|
||||
model_name: str,
|
||||
tools: Optional[list] = None,
|
||||
) -> ChatCompletion:
|
||||
stream: bool = False,
|
||||
) -> Union[ChatCompletion, AsyncStream[ChatCompletionChunk]]:
|
||||
"""
|
||||
使用 Openai SDK 调用ai获取回复
|
||||
|
||||
@@ -151,34 +130,34 @@ async def make_chat_openai(
|
||||
model=model_name,
|
||||
tools=tools or NOT_GIVEN,
|
||||
timeout=config.marshoai_timeout,
|
||||
stream=stream,
|
||||
**config.marshoai_model_args,
|
||||
)
|
||||
|
||||
|
||||
@from_cache("praises")
|
||||
def get_praises():
|
||||
global praises_json
|
||||
if praises_json is None:
|
||||
praises_file = store.get_plugin_data_file(
|
||||
"praises.json"
|
||||
) # 夸赞名单文件使用localstore存储
|
||||
if not praises_file.exists():
|
||||
with open(praises_file, "w", encoding="utf-8") as f:
|
||||
json.dump(_praises_init_data, f, ensure_ascii=False, indent=4)
|
||||
with open(praises_file, "r", encoding="utf-8") as f:
|
||||
data = json.load(f)
|
||||
praises_json = data
|
||||
praises_file = store.get_plugin_data_file(
|
||||
"praises.json"
|
||||
) # 夸赞名单文件使用localstore存储
|
||||
if not praises_file.exists():
|
||||
with open(praises_file, "w", encoding="utf-8") as f:
|
||||
json.dump(_praises_init_data, f, ensure_ascii=False, indent=4)
|
||||
with open(praises_file, "r", encoding="utf-8") as f:
|
||||
data = json.load(f)
|
||||
praises_json = data
|
||||
return praises_json
|
||||
|
||||
|
||||
@update_to_cache("praises")
|
||||
async def refresh_praises_json():
|
||||
global praises_json
|
||||
praises_file = store.get_plugin_data_file("praises.json")
|
||||
if not praises_file.exists():
|
||||
with open(praises_file, "w", encoding="utf-8") as f:
|
||||
json.dump(_praises_init_data, f, ensure_ascii=False, indent=4) # 异步?
|
||||
async with aiofiles.open(praises_file, "r", encoding="utf-8") as f:
|
||||
data = json.loads(await f.read())
|
||||
praises_json = data
|
||||
return data
|
||||
|
||||
|
||||
def build_praises() -> str:
|
||||
@@ -210,22 +189,21 @@ async def load_context_from_json(name: str, path: str) -> list:
|
||||
return []
|
||||
|
||||
|
||||
@from_cache("nickname")
|
||||
async def get_nicknames():
|
||||
"""获取nickname_json, 优先来源于全局变量"""
|
||||
global nickname_json
|
||||
if nickname_json is None:
|
||||
filename = store.get_plugin_data_file("nickname.json")
|
||||
# noinspection PyBroadException
|
||||
try:
|
||||
async with aiofiles.open(filename, "r", encoding="utf-8") as f:
|
||||
nickname_json = json.loads(await f.read())
|
||||
except Exception:
|
||||
nickname_json = {}
|
||||
"""获取nickname_json, 优先来源于缓存"""
|
||||
filename = store.get_plugin_data_file("nickname.json")
|
||||
# noinspection PyBroadException
|
||||
try:
|
||||
async with aiofiles.open(filename, "r", encoding="utf-8") as f:
|
||||
nickname_json = json.loads(await f.read())
|
||||
except (json.JSONDecodeError, FileNotFoundError):
|
||||
nickname_json = {}
|
||||
return nickname_json
|
||||
|
||||
|
||||
@update_to_cache("nickname")
|
||||
async def set_nickname(user_id: str, name: str):
|
||||
global nickname_json
|
||||
filename = store.get_plugin_data_file("nickname.json")
|
||||
if not filename.exists():
|
||||
data = {}
|
||||
@@ -237,19 +215,25 @@ async def set_nickname(user_id: str, name: str):
|
||||
del data[user_id]
|
||||
with open(filename, "w", encoding="utf-8") as f:
|
||||
json.dump(data, f, ensure_ascii=False, indent=4)
|
||||
nickname_json = data
|
||||
return data
|
||||
|
||||
|
||||
async def get_nickname_by_user_id(user_id: str):
|
||||
nickname_json = await get_nicknames()
|
||||
return nickname_json.get(user_id, "")
|
||||
|
||||
|
||||
@update_to_cache("nickname")
|
||||
async def refresh_nickname_json():
|
||||
"""强制刷新nickname_json, 刷新全局变量"""
|
||||
global nickname_json
|
||||
"""强制刷新nickname_json"""
|
||||
# noinspection PyBroadException
|
||||
try:
|
||||
async with aiofiles.open(
|
||||
store.get_plugin_data_file("nickname.json"), "r", encoding="utf-8"
|
||||
) as f:
|
||||
nickname_json = json.loads(await f.read())
|
||||
except Exception:
|
||||
return nickname_json
|
||||
except (json.JSONDecodeError, FileNotFoundError):
|
||||
logger.error("刷新 nickname_json 表错误:无法载入 nickname.json 文件")
|
||||
|
||||
|
||||
|
||||
@@ -81,4 +81,4 @@ test = [
|
||||
]
|
||||
|
||||
[tool.ruff.lint]
|
||||
ignore = ["E402"]
|
||||
ignore = ["E402", "F405"]
|
||||
|
||||
Reference in New Issue
Block a user