Compare commits

..

No commits in common. "main" and "v1.0.7" have entirely different histories.
main ... v1.0.7

30 changed files with 439 additions and 756 deletions

View File

@ -1,18 +1,26 @@
name: Deploy VitePress site to Liteyuki PaaS
# 构建 VitePress 站点并将其部署到 GitHub Pages 的示例工作流程
#
name: Deploy VitePress site to Pages
on: ["push", "pull_request_target"]
on:
# 在针对 `main` 分支的推送上运行。如果你
# 使用 `master` 分支作为默认分支,请将其更改为 `master`
push:
branches: [main]
# 允许你从 Actions 选项卡手动运行此工作流程
workflow_dispatch:
# 设置 GITHUB_TOKEN 的权限,以允许部署到 GitHub Pages
permissions:
contents: write
statuses: write
# 只允许同时进行一次部署,跳过正在运行和最新队列之间的运行队列
# 但是,不要取消正在进行的运行,因为我们希望允许这些生产部署完成
concurrency:
group: pages
cancel-in-progress: false
env:
MELI_SITE: f31e3b17-c4ea-4d9d-bdce-9417d67fd30e
jobs:
# 构建工作
build:
@ -22,10 +30,12 @@ jobs:
uses: actions/checkout@v4
with:
fetch-depth: 0 # 如果未启用 lastUpdated则不需要
# - uses: pnpm/action-setup@v3 # 如果使用 pnpm请取消注释
# - uses: oven-sh/setup-bun@v1 # 如果使用 Bun请取消注释
- name: Setup Python
uses: actions/setup-python@v2
with:
python-version: "3.11"
python-version: '3.11'
- name: Setup API markdown
run: |-
@ -49,13 +59,9 @@ jobs:
run: |-
pnpm run docs:build
- name: "发布"
run: |
npx -p "@getmeli/cli" meli upload docs/.vitepress/dist \
--url "https://dash.apage.dev" \
--site "$MELI_SITE" \
--token "$MELI_TOKEN" \
--release "$GITHUB_SHA"
env:
MELI_TOKEN: ${{ secrets.MELI_TOKEN }}
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
- name: 部署文档
uses: JamesIves/github-pages-deploy-action@v4
with:
# 这是文档部署到的分支名称
branch: docs
folder: docs/.vitepress/dist

View File

@ -15,7 +15,7 @@ repos:
args: [--config=./pyproject.toml]
- repo: https://github.com/PyCQA/isort
rev: 6.0.1
rev: 6.0.0
hooks:
- id: isort
args: ["--profile", "black"]

2
CNAME
View File

@ -1 +1 @@
marshoai-docs.pages.liteyuki.icu
marsho.liteyuki.icu

View File

@ -1,6 +1,6 @@
<!--suppress LongLine -->
<div align="center">
<a href="https://marshoai-docs.pages.liteyuki.icu"><img src="https://marshoai-docs.pages.liteyuki.icu/marsho-full.svg" width="800" height="430" alt="MarshoLogo"></a>
<a href="https://marsho.liteyuki.icu"><img src="https://marsho.liteyuki.icu/marsho-full.svg" width="800" height="430" alt="MarshoLogo"></a>
<br>
</div>
@ -10,7 +10,7 @@
_✨ 使用 OpenAI 标准格式 API 的聊天机器人插件 ✨_
[![QQ群](https://img.shields.io/badge/QQ群-1029557452-blue.svg?logo=QQ&style=flat-square)](https://qm.qq.com/q/a13iwP5kAw)
[![QQ群](https://img.shields.io/badge/QQ群-1029557452-blue.svg?logo=QQ)](https://qm.qq.com/q/a13iwP5kAw)
[![NoneBot Registry](https://img.shields.io/endpoint?url=https%3A%2F%2Fnbbdg.lgc2333.top%2Fplugin%2Fnonebot-plugin-marshoai&style=flat-square)](https://registry.nonebot.dev/plugin/nonebot-plugin-marshoai:nonebot_plugin_marshoai)
<a href="https://registry.nonebot.dev/plugin/nonebot-plugin-marshoai:nonebot_plugin_marshoai">
<img src="https://img.shields.io/endpoint?url=https%3A%2F%2Fnbbdg.lgc2333.top%2Fplugin-adapters%2Fnonebot-plugin-marshoai&style=flat-square" alt="Supported Adapters">
@ -22,21 +22,20 @@ _✨ 使用 OpenAI 标准格式 API 的聊天机器人插件 ✨_
<img src="https://img.shields.io/badge/Code%20Style-Black-121110.svg?style=flat-square" alt="codestyle">
</div>
<img width="100%" src="https://starify.komoridevs.icu/api/starify?owner=LiteyukiStudio&repo=nonebot-plugin-marshoai" alt="starify" />
## 📖 介绍
通过调用 OpenAI 标准格式 API(例如 GitHub Models API来实现聊天的插件。
插件内置了猫娘小棉Marshoマルショ的人物设定,可以进行可爱的聊天!
通过调用 OpenAI 标准格式 API(例如 GitHub Models API) 来实现聊天的插件。
插件内置了猫娘小棉(Marsho)的人物设定,可以进行可爱的聊天!
_谁不喜欢回复消息快又可爱的猫娘呢_
**对 OneBot 以外的适配器与非 GitHub Models API 的支持未完全经过验证。**
**对 OneBot 以外的适配器与非 GitHub Models API 的支持未经过完全验证。**
[Melobot 实现](https://github.com/LiteyukiStudio/marshoai-melo)
## 🐱 设定
#### 基本信息
- 名字:小棉Marshoマルショ
- 名字:小棉(Marsho)
- 生日9 月 6 日
#### 喜好
@ -48,7 +47,7 @@ _谁不喜欢回复消息快又可爱的猫娘呢_
## 😼 使用
请查看[使用文档](https://marshoai-docs.pages.liteyuki.icu/start/use.html)
请查看[使用文档](https://marsho.liteyuki.icu/start/use)
## ❤ 鸣谢&版权说明

View File

@ -1,6 +1,6 @@
<!--suppress LongLine -->
<div align="center">
<a href="https://marshoai-docs.pages.liteyuki.icu"><img src="https://marshoai-docs.pages.liteyuki.icu/marsho-full.svg" width="800" height="430" alt="MarshoLogo"></a>
<a href="https://marsho.liteyuki.icu"><img src="https://marsho.liteyuki.icu/marsho-full.svg" width="800" height="430" alt="MarshoLogo"></a>
<br>
</div>
@ -48,7 +48,7 @@ Plugin internally installed the catgirl character of Marsho, is able to have a c
- 🐾 Play! I like play with friends!
## 😼 Usage
Please read [Documentation](https://marshoai-docs.pages.liteyuki.icu/start/use.html)
Please read [Documentation](https://marsho.liteyuki.icu/start/install)
## ❤ Thanks&Copyright
This project uses the following code from other projects:

View File

@ -1,10 +1,10 @@
import { VitePressSidebarOptions } from "vitepress-sidebar/types";
import { VitePressSidebarOptions } from "vitepress-sidebar/types"
export const gitea = {
svg: '<svg t="1725391346807" class="icon" viewBox="0 0 1025 1024" version="1.1" xmlns="http://www.w3.org/2000/svg" p-id="5067" width="256" height="256"><path d="M1004.692673 466.396616l-447.094409-447.073929c-25.743103-25.763582-67.501405-25.763582-93.264987 0l-103.873521 103.873521 78.171378 78.171378c12.533635-6.00058 26.562294-9.359266 41.389666-9.359266 53.02219 0 96.00928 42.98709 96.00928 96.00928 0 14.827372-3.358686 28.856031-9.359266 41.389666l127.97824 127.97824c12.533635-6.00058 26.562294-9.359266 41.389666-9.359266 53.02219 0 96.00928 42.98709 96.00928 96.00928s-42.98709 96.00928-96.00928 96.00928-96.00928-42.98709-96.00928-96.00928c0-14.827372 3.358686-28.856031 9.359266-41.389666l-127.97824-127.97824c-3.051489 1.454065-6.184898 2.744293-9.379746 3.870681l0 266.97461c37.273227 13.188988 63.99936 48.721433 63.99936 90.520695 0 53.02219-42.98709 96.00928-96.00928 96.00928s-96.00928-42.98709-96.00928-96.00928c0-41.799262 26.726133-77.331707 63.99936-90.520695l0-266.97461c-37.273227-13.188988-63.99936-48.721433-63.99936-90.520695 0-14.827372 3.358686-28.856031 9.359266-41.389666l-78.171378-78.171378-295.892081 295.871601c-25.743103 25.784062-25.743103 67.542365 0 93.285467l447.114889 447.073929c25.743103 25.743103 67.480925 25.743103 93.264987 0l445.00547-445.00547c25.763582-25.763582 25.763582-67.542365 0-93.285467z" fill="#a2d8f4" p-id="5068"></path></svg>',
};
svg: '<svg t="1725391346807" class="icon" viewBox="0 0 1025 1024" version="1.1" xmlns="http://www.w3.org/2000/svg" p-id="5067" width="256" height="256"><path d="M1004.692673 466.396616l-447.094409-447.073929c-25.743103-25.763582-67.501405-25.763582-93.264987 0l-103.873521 103.873521 78.171378 78.171378c12.533635-6.00058 26.562294-9.359266 41.389666-9.359266 53.02219 0 96.00928 42.98709 96.00928 96.00928 0 14.827372-3.358686 28.856031-9.359266 41.389666l127.97824 127.97824c12.533635-6.00058 26.562294-9.359266 41.389666-9.359266 53.02219 0 96.00928 42.98709 96.00928 96.00928s-42.98709 96.00928-96.00928 96.00928-96.00928-42.98709-96.00928-96.00928c0-14.827372 3.358686-28.856031 9.359266-41.389666l-127.97824-127.97824c-3.051489 1.454065-6.184898 2.744293-9.379746 3.870681l0 266.97461c37.273227 13.188988 63.99936 48.721433 63.99936 90.520695 0 53.02219-42.98709 96.00928-96.00928 96.00928s-96.00928-42.98709-96.00928-96.00928c0-41.799262 26.726133-77.331707 63.99936-90.520695l0-266.97461c-37.273227-13.188988-63.99936-48.721433-63.99936-90.520695 0-14.827372 3.358686-28.856031 9.359266-41.389666l-78.171378-78.171378-295.892081 295.871601c-25.743103 25.784062-25.743103 67.542365 0 93.285467l447.114889 447.073929c25.743103 25.743103 67.480925 25.743103 93.264987 0l445.00547-445.00547c25.763582-25.763582 25.763582-67.542365 0-93.285467z" fill="#a2d8f4" p-id="5068"></path></svg>'
}
export const defaultLang = "zh";
export const defaultLang = 'zh'
const commonSidebarOptions: VitePressSidebarOptions = {
collapsed: true,
@ -13,15 +13,15 @@ const commonSidebarOptions: VitePressSidebarOptions = {
useFolderTitleFromIndexFile: false,
useFolderLinkFromIndexFile: true,
useTitleFromFileHeading: true,
rootGroupText: "MARSHOAI",
rootGroupText: 'MARSHOAI',
includeFolderIndexFile: true,
sortMenusByFrontmatterOrder: true,
};
}
export function generateSidebarConfig(): VitePressSidebarOptions[] {
let sections = ["dev", "start"];
let languages = ["zh", "en"];
let ret: VitePressSidebarOptions[] = [];
let sections = ["dev", "start"]
let languages = ['zh', 'en']
let ret: VitePressSidebarOptions[] = []
for (let language of languages) {
for (let section of sections) {
if (language === defaultLang) {
@ -29,59 +29,53 @@ export function generateSidebarConfig(): VitePressSidebarOptions[] {
basePath: `/${section}/`,
scanStartPath: `docs/${language}/${section}`,
resolvePath: `/${section}/`,
...commonSidebarOptions,
});
...commonSidebarOptions
})
} else {
ret.push({
basePath: `/${language}/${section}/`,
scanStartPath: `docs/${language}/${section}`,
resolvePath: `/${language}/${section}/`,
...commonSidebarOptions,
});
...commonSidebarOptions
})
}
}
}
return ret;
return ret
}
export const ThemeConfig = {
getEditLink: (
editPageText: string
): { pattern: (params: { filePath: string }) => string; text: string } => {
getEditLink: (editPageText: string): { pattern: (params: { filePath: string; }) => string; text: string; } => {
return {
pattern: ({ filePath }: { filePath: string }): string => {
pattern: ({filePath}: { filePath: string; }): string => {
if (!filePath) {
throw new Error("filePath is undefined");
}
const regex = /^(dev\/api|[^\/]+\/dev\/api)/;
if (regex.test(filePath)) {
filePath = filePath
.replace(regex, "")
.replace("index.md", "__init__.py")
.replace(".md", ".py");
const fileName = filePath.split("/").pop();
const parentFolder = filePath.split("/").slice(-2, -1)[0];
if (
fileName &&
parentFolder &&
fileName.split(".")[0] === parentFolder
) {
filePath =
filePath.split("/").slice(0, -1).join("/") + "/__init__.py";
filePath = filePath.replace(regex, '')
.replace('index.md', '__init__.py')
.replace('.md', '.py');
const fileName = filePath.split('/').pop();
const parentFolder = filePath.split('/').slice(-2, -1)[0];
if (fileName && parentFolder && fileName.split('.')[0] === parentFolder) {
filePath = filePath.split('/').slice(0, -1).join('/') + '/__init__.py';
}
return `https://github.com/LiteyukiStudio/nonebot-plugin-marshoai/tree/main/nonebot_plugin_marshoai/${filePath}`;
} else {
return `https://github.com/LiteyukiStudio/nonebot-plugin-marshoai/tree/main/docs/${filePath}`;
}
},
text: editPageText,
text: editPageText
};
},
getOutLine: (label: string): { label: string; level: [number, number] } => {
getOutLine: (label: string): { label: string; level: [number, number]; } => {
return {
label: label,
level: [2, 6],
level: [2, 6]
};
},
};
copyright: 'Copyright (C) 2020-2024 LiteyukiStudio. All Rights Reserved'
}

View File

@ -23,7 +23,7 @@ export const en = defineConfig({
lightModeSwitchTitle: 'Light',
darkModeSwitchTitle: 'Dark',
footer: {
message: "The document is being improved. Suggestions are welcome.<br>Webpage is deployed at <a href='https://meli.liteyuki.icu' target='_blank'>Liteyuki Meli</a> and accelerated by <a href='https://cdn.liteyuki.icu' target='_blank'>Liteyukiflare</a>.",
message: "The document is being improved. Suggestions are welcome.",
copyright: '© 2024 <a href="https://liteyuki.icu" target="_blank">Liteyuki Studio</a>',
}
},

View File

@ -8,13 +8,12 @@ import { generateSidebar } from 'vitepress-sidebar'
// https://vitepress.dev/reference/site-config
export default defineConfig({
head: [
["script", { src: "https://cdn.liteyuki.icu/js/liteyuki_footer.js" }],
['link', { rel: 'icon', type: 'image/x-icon', href: '/favicon.ico' }],
],
rewrites: {
[`${defaultLang}/:rest*`]: ":rest*",
},
cleanUrls: false,
cleanUrls: true,
themeConfig: {
// https://vitepress.dev/reference/default-theme-config
logo: {

View File

@ -23,7 +23,7 @@ export const ja = defineConfig({
lightModeSwitchTitle: 'ライト',
darkModeSwitchTitle: 'ダーク',
footer: {
message: "ドキュメントは改善中です。ご意見をお待ちしております。<br>ウェブサイトは <a href='https://meli.liteyuki.icu' target='_blank'>Liteyuki Meli</a> によってデプロイされ、<a href='https://cdn.liteyuki.icu' target='_blank'>Liteyukiflare</a> によって加速されています。",
message: "ドキュメントは改善中です。ご意見をお待ちしております。",
copyright: '© 2024 <a href="https://liteyuki.icu" target="_blank">Liteyuki Studio</a>',
}
},

View File

@ -23,7 +23,7 @@ export const zh = defineConfig({
lightModeSwitchTitle: '轻色模式',
darkModeSwitchTitle: '深色模式',
footer: {
message: "文档完善中,欢迎提出建议或帮助我们完善。<br>网站部署在 <a href='https://meli.liteyuki.icu' target='_blank'>Liteyuki Meli</a> 由 <a href='https://cdn.liteyuki.icu' target='_blank'>Liteyukiflare</a> 提供加速服务。",
message: "文档完善中,欢迎提出建议或帮助我们完善。",
copyright: '© 2024 <a href="https://liteyuki.icu" target="_blank">Liteyuki Studio</a>',
}
},

View File

@ -65,7 +65,7 @@ When nonebot linked to OneBot v11 adapter, can recieve double click and response
MarshoTools is a feature added in `v0.5.0`, support loading external function library to provide Function Call for Marsho.
## 🧩 Marsho Plugin
Marsho Plugin is a feature added in `v1.0.0`, replacing the old MarshoTools feature. [Documentation](https://marshoai-docs.pages.liteyuki.icu/dev/extension)
Marsho Plugin is a feature added in `v1.0.0`, replacing the old MarshoTools feature. [Documentation](https://marsho.liteyuki.icu/dev/extension)
## 👍 Praise list
@ -148,4 +148,3 @@ Add options in the `.env` file from the diagram below in nonebot2 project.
| MARSHOAI_SINGLE_LATEX_PARSE | `bool` | `false`| Render single-line equation or not |
| MARSHOAI_FIX_TOOLCALLS | `bool` | `true` | Fix tool calls or not |
| MARSHOAI_SEND_THINKING | `bool` | `true` | Send thinking chain or not |
| MARSHOAI_STREAM | `bool` | `false`| 是否通过流式方式请求 API **开启此项后暂无法使用函数调用,无法在 Bot 用户侧聊天界面呈现出流式效果** |

View File

@ -68,7 +68,7 @@ GitHub Models API 的限制较多,不建议使用,建议通过修改`MARSHOA
## 🧩 小棉插件
小棉插件是`v1.0.0`的新增功能,替代旧的小棉工具功能。[使用文档](https://marshoai-docs.pages.liteyuki.icu/dev/extension)
小棉插件是`v1.0.0`的新增功能,替代旧的小棉工具功能。[使用文档](https://marsho.liteyuki.icu/dev/extension)
## 👍 夸赞名单
@ -149,8 +149,6 @@ GitHub Models API 的限制较多,不建议使用,建议通过修改`MARSHOA
| MARSHOAI_SINGLE_LATEX_PARSE | `bool` | `false` | 单行公式是否渲染(当消息富文本解析启用时可用)(如果单行也渲……只能说不好看) |
| MARSHOAI_FIX_TOOLCALLS | `bool` | `true` | 是否修复工具调用(部分模型须关闭,使用 vLLM 部署的模型时须关闭) |
| MARSHOAI_SEND_THINKING | `bool` | `true` | 是否发送思维链(部分模型不支持) |
| MARSHOAI_STREAM | `bool` | `false`| 是否通过流式方式请求 API **开启此项后暂无法使用函数调用,无法在 Bot 用户侧聊天界面呈现出流式效果** |
#### 开发及调试选项

View File

@ -26,19 +26,17 @@ from nonebot.plugin import require
require("nonebot_plugin_alconna")
require("nonebot_plugin_localstore")
require("nonebot_plugin_argot")
import nonebot_plugin_localstore as store # type: ignore
from nonebot import get_driver, logger # type: ignore
from .config import config
# from .hunyuan import *
from .dev import *
from .marsho import *
from .metadata import metadata
# from .hunyuan import *
__author__ = "Asankilp"
__plugin_meta__ = metadata

View File

@ -1,39 +0,0 @@
from ..models import Cache
cache = Cache()
def from_cache(key):
"""
当缓存中有数据时直接返回缓存中的数据否则执行函数并将结果存入缓存
"""
def decorator(func):
async def wrapper(*args, **kwargs):
cached = cache.get(key)
if cached:
return cached
else:
result = await func(*args, **kwargs)
cache.set(key, result)
return result
return wrapper
return decorator
def update_to_cache(key):
"""
执行函数并将结果存入缓存
"""
def decorator(func):
async def wrapper(*args, **kwargs):
result = await func(*args, **kwargs)
cache.set(key, result)
return result
return wrapper
return decorator

View File

@ -1,3 +1,4 @@
import shutil
from io import StringIO
from pathlib import Path
@ -32,7 +33,6 @@ class ConfigModel(BaseModel):
marshoai_enable_sysasuser_prompt: bool = False
marshoai_additional_prompt: str = ""
marshoai_poke_suffix: str = "揉了揉你的猫耳"
marshoai_stream: bool = False
marshoai_enable_richtext_parse: bool = True
"""
是否启用自动消息富文本解析 即若包含图片链接则发送图片若包含LaTeX公式则发送公式图
@ -81,15 +81,15 @@ destination_folder = Path("config/marshoai/")
destination_file = destination_folder / "config.yaml"
def dump_config_to_yaml(cfg: ConfigModel):
return yaml_.dump(cfg.model_dump(), allow_unicode=True, default_flow_style=False)
def dump_config_to_yaml(config: ConfigModel):
return yaml_.dump(config.model_dump(), allow_unicode=True, default_flow_style=False)
def write_default_config(dest_file):
def write_default_config(destination_file):
"""
写入默认配置
"""
with open(dest_file, "w", encoding="utf-8") as f:
with open(destination_file, "w", encoding="utf-8") as f:
with StringIO(dump_config_to_yaml(ConfigModel())) as f2:
f.write(f2.read())
@ -110,17 +110,17 @@ def check_yaml_is_changed():
return True
def merge_configs(existing_cfg, new_cfg):
def merge_configs(old_config, new_config):
"""
合并配置文件
"""
for key, value in new_cfg.items():
if key in existing_cfg:
for key, value in new_config.items():
if key in old_config:
continue
else:
logger.info(f"新增配置项: {key} = {value}")
existing_cfg[key] = value
return existing_cfg
old_config[key] = value
return old_config
config: ConfigModel = get_plugin_config(ConfigModel)

View File

@ -37,7 +37,7 @@ OPENAI_NEW_MODELS: list = [
INTRODUCTION: str = f"""MarshoAI-NoneBot by LiteyukiStudio
你好喵~我是一只可爱的猫娘AI名叫小棉~🐾
我的主页在这里哦~
https://marshoai-docs.pages.liteyuki.icu
https://marsho.liteyuki.icu
使用 {config.marshoai_default_name}.status命令获取状态信息
使用{config.marshoai_default_name}.help命令获取使用说明"""

View File

@ -1,16 +1,15 @@
import os
from pathlib import Path
from nonebot import get_driver, logger, on_command, require
from nonebot import get_driver, logger, require
from nonebot.adapters import Bot, Event
from nonebot.matcher import Matcher
from nonebot.typing import T_State
from nonebot_plugin_argot import add_argot, get_message_id
from nonebot_plugin_marshoai.plugin.load import reload_plugin
from .config import config
from .instances import context
from .marsho import context
from .plugin.func_call.models import SessionContext
require("nonebot_plugin_alconna")
@ -49,21 +48,6 @@ function_call = on_alconna(
permission=SUPERUSER,
)
argot_test = on_command("argot", permission=SUPERUSER)
@argot_test.handle()
async def _():
await argot_test.send(
"aa",
argot={
"name": "test",
"command": "test",
"segment": f"{os.getcwd()}",
"expired_at": 1000,
},
)
@function_call.assign("list")
async def list_functions():

View File

@ -1,300 +0,0 @@
import json
from datetime import timedelta
from typing import Optional, Tuple, Union
from azure.ai.inference.models import (
CompletionsFinishReason,
ImageContentItem,
ImageUrl,
TextContentItem,
ToolMessage,
UserMessage,
)
from nonebot.adapters import Bot, Event
from nonebot.log import logger
from nonebot.matcher import (
Matcher,
current_bot,
current_event,
current_matcher,
)
from nonebot_plugin_alconna.uniseg import (
Text,
UniMessage,
UniMsg,
get_target,
)
from nonebot_plugin_argot import Argot # type: ignore
from openai import AsyncOpenAI, AsyncStream
from openai.types.chat import ChatCompletion, ChatCompletionChunk, ChatCompletionMessage
from .config import config
from .constants import SUPPORT_IMAGE_MODELS
from .instances import target_list
from .models import MarshoContext
from .plugin.func_call.caller import get_function_calls
from .plugin.func_call.models import SessionContext
from .util import (
extract_content_and_think,
get_image_b64,
get_nickname_by_user_id,
get_prompt,
make_chat_openai,
parse_richtext,
)
from .utils.processor import process_chat_stream, process_completion_to_details
class MarshoHandler:
def __init__(
self,
client: AsyncOpenAI,
context: MarshoContext,
):
self.client = client
self.context = context
self.bot: Bot = current_bot.get()
self.event: Event = current_event.get()
# self.state: T_State = current_handler.get().state
self.matcher: Matcher = current_matcher.get()
self.message_id: str = UniMessage.get_message_id(self.event)
self.target = get_target(self.event)
async def process_user_input(
self, user_input: UniMsg, model_name: str
) -> Union[str, list]:
"""
处理用户输入为可输入 API 的格式并添加昵称提示
"""
is_support_image_model = (
model_name.lower()
in SUPPORT_IMAGE_MODELS + config.marshoai_additional_image_models
)
usermsg = [] if is_support_image_model else ""
user_nickname = await get_nickname_by_user_id(self.event.get_user_id())
if user_nickname:
nickname_prompt = f"\n此消息的说话者为: {user_nickname}"
else:
nickname_prompt = ""
for i in user_input: # type: ignore
if i.type == "text":
if is_support_image_model:
usermsg += [TextContentItem(text=i.data["text"] + nickname_prompt).as_dict()] # type: ignore
else:
usermsg += str(i.data["text"] + nickname_prompt) # type: ignore
elif i.type == "image":
if is_support_image_model:
usermsg.append( # type: ignore
ImageContentItem(
image_url=ImageUrl( # type: ignore
url=str(await get_image_b64(i.data["url"])) # type: ignore
) # type: ignore
).as_dict() # type: ignore
) # type: ignore
logger.info(f"输入图片 {i.data['url']}")
elif config.marshoai_enable_support_image_tip:
await UniMessage(
"*此模型不支持图片处理或管理员未启用此模型的图片支持。图片将被忽略。"
).send()
return usermsg # type: ignore
async def handle_single_chat(
self,
user_message: Union[str, list],
model_name: str,
tools_list: list | None,
tool_message: Optional[list] = None,
stream: bool = False,
) -> Union[ChatCompletion, AsyncStream[ChatCompletionChunk]]:
"""
处理单条聊天
"""
context_msg = await get_prompt(model_name) + (
self.context.build(self.target.id, self.target.private)
)
response = await make_chat_openai(
client=self.client,
msg=context_msg + [UserMessage(content=user_message).as_dict()] + (tool_message if tool_message else []), # type: ignore
model_name=model_name,
tools=tools_list if tools_list else None,
stream=stream,
)
return response
async def handle_function_call(
self,
completion: Union[ChatCompletion],
user_message: Union[str, list],
model_name: str,
tools_list: list | None = None,
):
# function call
# 需要获取额外信息,调用函数工具
tool_msg = []
if isinstance(completion, ChatCompletion):
choice = completion.choices[0]
else:
raise ValueError("Unexpected completion type")
# await UniMessage(str(response)).send()
tool_calls = choice.message.tool_calls
# try:
# if tool_calls[0]["function"]["name"].startswith("$"):
# choice.message.tool_calls[0][
# "type"
# ] = "builtin_function" # 兼容 moonshot AI 内置函数的临时方案
# except:
# pass
tool_msg.append(choice.message)
for tool_call in tool_calls: # type: ignore
try:
function_args = json.loads(tool_call.function.arguments)
except json.JSONDecodeError:
function_args = json.loads(
tool_call.function.arguments.replace("'", '"')
)
# 删除args的placeholder参数
if "placeholder" in function_args:
del function_args["placeholder"]
logger.info(
f"调用函数 {tool_call.function.name.replace('-', '.')}\n参数:"
+ "\n".join([f"{k}={v}" for k, v in function_args.items()])
)
await UniMessage(
f"调用函数 {tool_call.function.name.replace('-', '.')}\n参数:"
+ "\n".join([f"{k}={v}" for k, v in function_args.items()])
).send()
if caller := get_function_calls().get(tool_call.function.name):
logger.debug(f"调用插件函数 {caller.full_name}")
# 权限检查,规则检查 TODO
# 实现依赖注入检查函数参数及参数注解类型对Event类型的参数进行注入
func_return = await caller.with_ctx(
SessionContext(
bot=self.bot,
event=self.event,
matcher=self.matcher,
state=None,
)
).call(**function_args)
else:
logger.error(f"未找到函数 {tool_call.function.name.replace('-', '.')}")
func_return = f"未找到函数 {tool_call.function.name.replace('-', '.')}"
tool_msg.append(
ToolMessage(tool_call_id=tool_call.id, content=func_return).as_dict() # type: ignore
)
# tool_msg[0]["tool_calls"][0]["type"] = "builtin_function"
# await UniMessage(str(tool_msg)).send()
return await self.handle_common_chat(
user_message=user_message,
model_name=model_name,
tools_list=tools_list,
tool_message=tool_msg,
)
async def handle_common_chat(
self,
user_message: Union[str, list],
model_name: str,
tools_list: list | None = None,
stream: bool = False,
tool_message: Optional[list] = None,
) -> Optional[Tuple[UserMessage, ChatCompletionMessage]]:
"""
处理一般聊天
"""
global target_list
if stream:
response = await self.handle_stream_request(
user_message=user_message,
model_name=model_name,
tools_list=tools_list,
tools_message=tool_message,
)
else:
response = await self.handle_single_chat( # type: ignore
user_message=user_message,
model_name=model_name,
tools_list=tools_list,
tool_message=tool_message,
)
choice = response.choices[0] # type: ignore
# Sprint(choice)
# 当tool_calls非空时将finish_reason设置为TOOL_CALLS
if choice.message.tool_calls is not None and config.marshoai_fix_toolcalls:
choice.finish_reason = "tool_calls"
logger.info(f"完成原因:{choice.finish_reason}")
if choice.finish_reason == CompletionsFinishReason.STOPPED:
##### DeepSeek-R1 兼容部分 #####
choice_msg_content, choice_msg_thinking, choice_msg_after = (
extract_content_and_think(choice.message)
)
if choice_msg_thinking and config.marshoai_send_thinking:
await UniMessage("思维链:\n" + choice_msg_thinking).send()
##### 兼容部分结束 #####
if [self.target.id, self.target.private] not in target_list:
target_list.append([self.target.id, self.target.private])
# 对话成功发送消息
send_message = UniMessage()
if config.marshoai_enable_richtext_parse:
send_message = await parse_richtext(str(choice_msg_content))
else:
send_message = UniMessage(str(choice_msg_content))
send_message.append(
Argot(
"detail",
Text(await process_completion_to_details(response)),
command="detail",
expired_at=timedelta(minutes=5),
)
)
# send_message.append(
# Argot(
# "debug",
# Text(str(response)),
# command=f"debug",
# expired_at=timedelta(minutes=5),
# )
# )
await send_message.send(reply_to=True)
return UserMessage(content=user_message), choice_msg_after
elif choice.finish_reason == CompletionsFinishReason.CONTENT_FILTERED:
# 对话失败,消息过滤
await UniMessage("*已被内容过滤器过滤。请调整聊天内容后重试。").send(
reply_to=True
)
return None
elif choice.finish_reason == CompletionsFinishReason.TOOL_CALLS:
return await self.handle_function_call(
response, user_message, model_name, tools_list
)
else:
await UniMessage(f"意外的完成原因:{choice.finish_reason}").send()
return None
async def handle_stream_request(
self,
user_message: Union[str, list],
model_name: str,
tools_list: list | None = None,
tools_message: Optional[list] = None,
) -> ChatCompletion:
"""
处理流式请求
"""
response = await self.handle_single_chat(
user_message=user_message,
model_name=model_name,
tools_list=None, # TODO:让流式调用支持工具调用
tool_message=tools_message,
stream=True,
)
if isinstance(response, AsyncStream):
return await process_chat_stream(response)
else:
raise TypeError("Unexpected response type for stream request")

View File

@ -6,7 +6,7 @@ import nonebot_plugin_localstore as store
from nonebot import logger
from .config import config
from .instances import context, driver, target_list, tools
from .instances import *
from .plugin import load_plugin, load_plugins
from .util import get_backup_context, save_context_to_json

View File

@ -1,4 +1,6 @@
# Marsho 的类实例以及全局变量
from azure.ai.inference.aio import ChatCompletionsClient
from azure.core.credentials import AzureKeyCredential
from nonebot import get_driver
from openai import AsyncOpenAI

View File

@ -2,10 +2,15 @@ import contextlib
import traceback
from typing import Optional
import openai
from arclet.alconna import Alconna, AllParam, Args
from azure.ai.inference.models import (
AssistantMessage,
CompletionsFinishReason,
ImageContentItem,
ImageUrl,
TextContentItem,
ToolMessage,
UserMessage,
)
from nonebot import logger, on_command, on_message
@ -13,27 +18,16 @@ from nonebot.adapters import Bot, Event, Message
from nonebot.matcher import Matcher
from nonebot.params import CommandArg
from nonebot.permission import SUPERUSER
from nonebot.rule import to_me
from nonebot.rule import Rule, to_me
from nonebot.typing import T_State
from nonebot_plugin_alconna import (
Emoji,
MsgTarget,
UniMessage,
UniMsg,
message_reaction,
on_alconna,
)
from nonebot_plugin_argot.extension import ArgotExtension # type: ignore
from nonebot_plugin_alconna import MsgTarget, UniMessage, UniMsg, on_alconna
from .config import config
from .constants import INTRODUCTION, SUPPORT_IMAGE_MODELS
from .handler import MarshoHandler
from .hooks import *
from .instances import client, context, model_name, target_list, tools
from .instances import *
from .metadata import metadata
from .plugin.func_call.caller import get_function_calls
from .plugin.func_call.models import SessionContext
from .util import *
from .utils.processor import process_chat_stream
async def at_enable():
@ -41,20 +35,20 @@ async def at_enable():
changemodel_cmd = on_command(
"changemodel", permission=SUPERUSER, priority=96, block=True
"changemodel", permission=SUPERUSER, priority=10, block=True
)
# setprompt_cmd = on_command("prompt",permission=SUPERUSER)
praises_cmd = on_command("praises", permission=SUPERUSER, priority=96, block=True)
add_usermsg_cmd = on_command("usermsg", permission=SUPERUSER, priority=96, block=True)
praises_cmd = on_command("praises", permission=SUPERUSER, priority=10, block=True)
add_usermsg_cmd = on_command("usermsg", permission=SUPERUSER, priority=10, block=True)
add_assistantmsg_cmd = on_command(
"assistantmsg", permission=SUPERUSER, priority=96, block=True
"assistantmsg", permission=SUPERUSER, priority=10, block=True
)
contexts_cmd = on_command("contexts", permission=SUPERUSER, priority=96, block=True)
contexts_cmd = on_command("contexts", permission=SUPERUSER, priority=10, block=True)
save_context_cmd = on_command(
"savecontext", permission=SUPERUSER, priority=96, block=True
"savecontext", permission=SUPERUSER, priority=10, block=True
)
load_context_cmd = on_command(
"loadcontext", permission=SUPERUSER, priority=96, block=True
"loadcontext", permission=SUPERUSER, priority=10, block=True
)
marsho_cmd = on_alconna(
Alconna(
@ -62,43 +56,42 @@ marsho_cmd = on_alconna(
Args["text?", AllParam],
),
aliases=tuple(config.marshoai_aliases),
priority=96,
priority=10,
block=True,
extensions=[ArgotExtension()],
)
resetmem_cmd = on_alconna(
Alconna(
config.marshoai_default_name + ".reset",
),
priority=96,
priority=10,
block=True,
)
marsho_help_cmd = on_alconna(
Alconna(
config.marshoai_default_name + ".help",
),
priority=96,
priority=10,
block=True,
)
marsho_status_cmd = on_alconna(
Alconna(
config.marshoai_default_name + ".status",
),
priority=96,
priority=10,
block=True,
)
marsho_at = on_message(rule=to_me() & at_enable, priority=97)
marsho_at = on_message(rule=to_me() & at_enable, priority=11)
nickname_cmd = on_alconna(
Alconna(
"nickname",
Args["name?", str],
),
priority=96,
priority=10,
block=True,
)
refresh_data_cmd = on_command(
"refresh_data", permission=SUPERUSER, priority=96, block=True
"refresh_data", permission=SUPERUSER, priority=10, block=True
)
@ -236,18 +229,17 @@ async def marsho(
if not text:
# 发送说明
# await UniMessage(metadata.usage + "\n当前使用的模型" + model_name).send()
await message_reaction(Emoji("38"))
await marsho_cmd.finish(INTRODUCTION)
backup_context = await get_backup_context(target.id, target.private)
if backup_context:
context.set_context(
backup_context, target.id, target.private
) # 加载历史记录
logger.info(f"已恢复会话 {target.id} 的上下文备份~")
handler = MarshoHandler(client, context)
try:
user_nickname = await get_nickname_by_user_id(event.get_user_id())
if not user_nickname:
user_id = event.get_user_id()
nicknames = await get_nicknames()
user_nickname = nicknames.get(user_id, "")
if user_nickname != "":
nickname_prompt = (
f"\n*此消息的说话者id为:{user_id},名字为:{user_nickname}*"
)
else:
nickname_prompt = ""
# 用户名无法获取,暂时注释
# user_nickname = event.sender.nickname # 未设置昵称时获取用户名
# nickname_prompt = f"\n*此消息的说话者:{user_nickname}"
@ -261,24 +253,189 @@ async def marsho(
"※你未设置自己的昵称。推荐使用「nickname [昵称]」命令设置昵称来获得个性化(可能)回答。"
).send()
usermsg = await handler.process_user_input(text, model_name)
is_support_image_model = (
model_name.lower()
in SUPPORT_IMAGE_MODELS + config.marshoai_additional_image_models
)
is_openai_new_model = model_name.lower() in OPENAI_NEW_MODELS
usermsg = [] if is_support_image_model else ""
for i in text: # type: ignore
if i.type == "text":
if is_support_image_model:
usermsg += [TextContentItem(text=i.data["text"] + nickname_prompt).as_dict()] # type: ignore
else:
usermsg += str(i.data["text"] + nickname_prompt) # type: ignore
elif i.type == "image":
if is_support_image_model:
usermsg.append( # type: ignore
ImageContentItem(
image_url=ImageUrl( # type: ignore
url=str(await get_image_b64(i.data["url"])) # type: ignore
) # type: ignore
).as_dict() # type: ignore
) # type: ignore
logger.info(f"输入图片 {i.data['url']}")
elif config.marshoai_enable_support_image_tip:
await UniMessage(
"*此模型不支持图片处理或管理员未启用此模型的图片支持。图片将被忽略。"
).send()
backup_context = await get_backup_context(target.id, target.private)
if backup_context:
context.set_context(
backup_context, target.id, target.private
) # 加载历史记录
logger.info(f"已恢复会话 {target.id} 的上下文备份~")
context_msg = get_prompt(model_name) + context.build(target.id, target.private)
tools_lists = tools.tools_list + list(
map(lambda v: v.data(), get_function_calls().values())
)
logger.info(f"正在获取回答,模型:{model_name}")
await message_reaction(Emoji("66"))
# logger.info(f"上下文:{context_msg}")
response = await handler.handle_common_chat(
usermsg, model_name, tools_lists, config.marshoai_stream
response = await make_chat_openai(
client=client,
model_name=model_name,
msg=context_msg + [UserMessage(content=usermsg).as_dict()], # type: ignore
tools=tools_lists if tools_lists else None, # TODO 临时追加函数,后期优化
)
# await UniMessage(str(response)).send()
if response is not None:
context_user, context_assistant = response
context.append(context_user.as_dict(), target.id, target.private)
context.append(context_assistant.to_dict(), target.id, target.private)
choice = response.choices[0]
# Sprint(choice)
# 当tool_calls非空时将finish_reason设置为TOOL_CALLS
if choice.message.tool_calls != None and config.marshoai_fix_toolcalls:
choice.finish_reason = "tool_calls"
logger.info(f"完成原因:{choice.finish_reason}")
if choice.finish_reason == CompletionsFinishReason.STOPPED:
# 当对话成功时将dict的上下文添加到上下文类中
context.append(
UserMessage(content=usermsg).as_dict(), target.id, target.private # type: ignore
)
##### DeepSeek-R1 兼容部分 #####
choice_msg_content, choice_msg_thinking, choice_msg_after = (
extract_content_and_think(choice.message)
)
if choice_msg_thinking and config.marshoai_send_thinking:
await UniMessage("思维链:\n" + choice_msg_thinking).send()
##### 兼容部分结束 #####
context.append(choice_msg_after.to_dict(), target.id, target.private)
if [target.id, target.private] not in target_list:
target_list.append([target.id, target.private])
# 对话成功发送消息
if config.marshoai_enable_richtext_parse:
await (await parse_richtext(str(choice_msg_content))).send(
reply_to=True
)
else:
await UniMessage(str(choice_msg_content)).send(reply_to=True)
elif choice.finish_reason == CompletionsFinishReason.CONTENT_FILTERED:
# 对话失败,消息过滤
await UniMessage("*已被内容过滤器过滤。请调整聊天内容后重试。").send(
reply_to=True
)
return
elif choice.finish_reason == CompletionsFinishReason.TOOL_CALLS:
# function call
# 需要获取额外信息,调用函数工具
tool_msg = []
while choice.message.tool_calls != None:
# await UniMessage(str(response)).send()
tool_calls = choice.message.tool_calls
# try:
# if tool_calls[0]["function"]["name"].startswith("$"):
# choice.message.tool_calls[0][
# "type"
# ] = "builtin_function" # 兼容 moonshot AI 内置函数的临时方案
# except:
# pass
tool_msg.append(choice.message)
for tool_call in tool_calls:
try:
function_args = json.loads(tool_call.function.arguments)
except json.JSONDecodeError:
function_args = json.loads(
tool_call.function.arguments.replace("'", '"')
)
# 删除args的placeholder参数
if "placeholder" in function_args:
del function_args["placeholder"]
logger.info(
f"调用函数 {tool_call.function.name.replace('-', '.')}\n参数:"
+ "\n".join([f"{k}={v}" for k, v in function_args.items()])
)
await UniMessage(
f"调用函数 {tool_call.function.name.replace('-', '.')}\n参数:"
+ "\n".join([f"{k}={v}" for k, v in function_args.items()])
).send()
# TODO 临时追加插件函数,若工具中没有则调用插件函数
if tools.has_function(tool_call.function.name):
logger.debug(f"调用工具函数 {tool_call.function.name}")
func_return = await tools.call(
tool_call.function.name, function_args
) # 获取返回值
else:
if caller := get_function_calls().get(tool_call.function.name):
logger.debug(f"调用插件函数 {caller.full_name}")
# 权限检查,规则检查 TODO
# 实现依赖注入检查函数参数及参数注解类型对Event类型的参数进行注入
func_return = await caller.with_ctx(
SessionContext(
bot=bot,
event=event,
state=state,
matcher=matcher,
)
).call(**function_args)
else:
logger.error(
f"未找到函数 {tool_call.function.name.replace('-', '.')}"
)
func_return = f"未找到函数 {tool_call.function.name.replace('-', '.')}"
tool_msg.append(
ToolMessage(tool_call_id=tool_call.id, content=func_return).as_dict() # type: ignore
)
# tool_msg[0]["tool_calls"][0]["type"] = "builtin_function"
# await UniMessage(str(tool_msg)).send()
request_msg = context_msg + [UserMessage(content=usermsg).as_dict()] + tool_msg # type: ignore
response = await make_chat_openai(
client=client,
model_name=model_name,
msg=request_msg, # type: ignore
tools=(
tools_lists if tools_lists else None
), # TODO 临时追加函数,后期优化
)
choice = response.choices[0]
# 当tool_calls非空时将finish_reason设置为TOOL_CALLS
if choice.message.tool_calls != None:
choice.finish_reason = CompletionsFinishReason.TOOL_CALLS
if choice.finish_reason == CompletionsFinishReason.STOPPED:
# 对话成功 添加上下文
context.append(
UserMessage(content=usermsg).as_dict(), target.id, target.private # type: ignore
)
# context.append(tool_msg, target.id, target.private)
choice_msg_dict = choice.message.to_dict()
if "reasoning_content" in choice_msg_dict:
del choice_msg_dict["reasoning_content"]
context.append(choice_msg_dict, target.id, target.private)
# 发送消息
if config.marshoai_enable_richtext_parse:
await (await parse_richtext(str(choice.message.content))).send(
reply_to=True
)
else:
await UniMessage(str(choice.message.content)).send(reply_to=True)
else:
await marsho_cmd.finish(f"意外的完成原因:{choice.finish_reason}")
else:
await marsho_cmd.finish(f"意外的完成原因:{choice.finish_reason}")
except Exception as e:
await UniMessage(str(e) + suggest_solution(str(e))).send()
traceback.print_exc()
@ -293,25 +450,23 @@ with contextlib.suppress(ImportError): # 优化先不做()
@poke_notify.handle()
async def poke(event: Event):
user_nickname = await get_nickname_by_user_id(event.get_user_id())
usermsg = await get_prompt(model_name) + [
UserMessage(content=f"*{user_nickname}{config.marshoai_poke_suffix}"),
]
user_id = event.get_user_id()
nicknames = await get_nicknames()
user_nickname = nicknames.get(user_id, "")
try:
if config.marshoai_poke_suffix != "":
logger.info(f"收到戳一戳,用户昵称:{user_nickname}")
pre_response = await make_chat_openai(
logger.info(f"收到戳一戳,用户昵称:{user_nickname}用户ID{user_id}")
response = await make_chat_openai(
client=client,
model_name=model_name,
msg=usermsg,
stream=config.marshoai_stream,
msg=get_prompt(model_name)
+ [
UserMessage(
content=f"*{user_nickname}{config.marshoai_poke_suffix}"
),
],
)
if isinstance(pre_response, AsyncStream):
response = await process_chat_stream(pre_response)
else:
response = pre_response
choice = response.choices[0] # type: ignore
choice = response.choices[0]
if choice.finish_reason == CompletionsFinishReason.STOPPED:
content = extract_content_and_think(choice.message)[0]
await UniMessage(" " + str(content)).send(at_sender=True)

View File

@ -5,7 +5,7 @@ from .constants import USAGE
metadata = PluginMetadata(
name="Marsho AI 插件",
description="接入 Azure API 或其他 API 的 AI 聊天插件,支持图片处理,外部函数调用,兼容包括 DeepSeek-R1 QwQ-32B 在内的多个模型",
description="接入 Azure API 或其他 API 的 AI 聊天插件,支持图片处理,外部函数调用,兼容包括 DeepSeek-R1 在内的多个模型",
usage=USAGE,
type="application",
config=ConfigModel,

View File

@ -7,28 +7,9 @@ import sys
import traceback
from nonebot import logger
from typing_extensions import deprecated
from .config import config
class Cache:
"""
缓存类
"""
def __init__(self):
self.cache = {}
def get(self, key):
if key in self.cache:
return self.cache[key]
else:
self.cache[key] = None
return None
def set(self, key, value):
self.cache[key] = value
from .util import *
class MarshoContext:
@ -74,7 +55,6 @@ class MarshoContext:
return self._get_target_dict(is_private).setdefault(target_id, [])
@deprecated("小棉工具已弃用,无法正常调用")
class MarshoTools:
"""
Marsho 的工具类

View File

@ -70,8 +70,8 @@ class Caller:
):
return False, "告诉用户 Permission Denied 权限不足"
# if self.ctx.state is None:
# return False, "State is None"
if self.ctx.state is None:
return False, "State is None"
if self._rule and not await self._rule(
self.ctx.bot, self.ctx.event, self.ctx.state
):
@ -115,10 +115,6 @@ class Caller:
# 检查函数签名,确定依赖注入参数
sig = inspect.signature(func)
for name, param in sig.parameters.items():
# if param.annotation == T_State:
# self.di.state = name
# continue # 防止后续判断T_State子类时报错
if issubclass(param.annotation, Event) or isinstance(
param.annotation, Event
):
@ -137,6 +133,9 @@ class Caller:
):
self.di.matcher = name
if param.annotation == T_State:
self.di.state = name
# 检查默认值情况
for name, param in sig.parameters.items():
if param.default is not inspect.Parameter.empty:

View File

@ -19,7 +19,7 @@ class SessionContext(BaseModel):
bot: Bot
event: Event
matcher: Matcher
state: T_State | None
state: T_State
caller: Any = None
class Config:
@ -30,5 +30,5 @@ class SessionContextDepends(BaseModel):
bot: str | None = None
event: str | None = None
matcher: str | None = None
# state: str | None = None
state: str | None = None
caller: str | None = None

View File

@ -16,7 +16,7 @@ marsho_memory_cmd = on_alconna(
Subcommand("view", alias={"v"}),
Subcommand("reset", alias={"r"}),
),
priority=96,
priority=10,
block=True,
)

View File

@ -2,9 +2,8 @@ import base64
import json
import mimetypes
import re
import ssl
import uuid
from typing import Any, Dict, List, Optional, Union
from typing import Any, Dict, List, Optional
import aiofiles # type: ignore
import httpx
@ -16,19 +15,18 @@ from nonebot.log import logger
from nonebot_plugin_alconna import Image as ImageMsg
from nonebot_plugin_alconna import Text as TextMsg
from nonebot_plugin_alconna import UniMessage
from openai import AsyncOpenAI, AsyncStream, NotGiven
from openai.types.chat import ChatCompletion, ChatCompletionChunk, ChatCompletionMessage
from zhDateTime import DateTime # type: ignore
from openai import AsyncOpenAI, NotGiven
from openai.types.chat import ChatCompletion, ChatCompletionMessage
from zhDateTime import DateTime
from ._types import DeveloperMessage
from .cache.decos import *
from .config import config
from .constants import CODE_BLOCK_PATTERN, IMG_LATEX_PATTERN, OPENAI_NEW_MODELS
from .constants import *
from .deal_latex import ConvertLatex
# nickname_json = None # 记录昵称
# praises_json = None # 记录夸赞名单
loaded_target_list: List[str] = [] # 记录已恢复备份的上下文的列表
nickname_json = None # 记录昵称
praises_json = None # 记录夸赞名单
loaded_target_list = [] # 记录已恢复备份的上下文的列表
NOT_GIVEN = NotGiven()
@ -59,8 +57,6 @@ _praises_init_data = {
"""
初始夸赞名单之数据
"""
_ssl_context = ssl.create_default_context()
_ssl_context.set_ciphers("DEFAULT")
async def get_image_raw_and_type(
@ -77,7 +73,7 @@ async def get_image_raw_and_type(
tuple[bytes, str]: 图片二进制数据, 图片MIME格式
"""
async with httpx.AsyncClient(verify=_ssl_context) as client:
async with httpx.AsyncClient() as client:
response = await client.get(url, headers=_browser_headers, timeout=timeout)
if response.status_code == 200:
# 获取图片数据
@ -101,7 +97,9 @@ async def get_image_b64(url: str, timeout: int = 10) -> Optional[str]:
return: 图片base64编码
"""
if data_type := await get_image_raw_and_type(url, timeout):
if data_type := await get_image_raw_and_type(
url.replace("https://", "http://"), timeout
):
# image_format = content_type.split("/")[1] if content_type else "jpeg"
base64_image = base64.b64encode(data_type[0]).decode("utf-8")
data_url = "data:{};base64,{}".format(data_type[1], base64_image)
@ -110,13 +108,35 @@ async def get_image_b64(url: str, timeout: int = 10) -> Optional[str]:
return None
async def make_chat(
client: ChatCompletionsClient,
msg: list,
model_name: str,
tools: Optional[list] = None,
):
"""
调用ai获取回复
参数:
client: 用于与AI模型进行通信
msg: 消息内容
model_name: 指定AI模型名
tools: 工具列表
"""
return await client.complete(
messages=msg,
model=model_name,
tools=tools,
**config.marshoai_model_args,
)
async def make_chat_openai(
client: AsyncOpenAI,
msg: list,
model_name: str,
tools: Optional[list] = None,
stream: bool = False,
) -> Union[ChatCompletion, AsyncStream[ChatCompletionChunk]]:
) -> ChatCompletion:
"""
使用 Openai SDK 调用ai获取回复
@ -131,38 +151,38 @@ async def make_chat_openai(
model=model_name,
tools=tools or NOT_GIVEN,
timeout=config.marshoai_timeout,
stream=stream,
**config.marshoai_model_args,
)
@from_cache("praises")
async def get_praises():
def get_praises():
global praises_json
if praises_json is None:
praises_file = store.get_plugin_data_file(
"praises.json"
) # 夸赞名单文件使用localstore存储
if not praises_file.exists():
async with aiofiles.open(praises_file, "w", encoding="utf-8") as f:
await f.write(json.dumps(_praises_init_data, ensure_ascii=False, indent=4))
async with aiofiles.open(praises_file, "r", encoding="utf-8") as f:
data = json.loads(await f.read())
with open(praises_file, "w", encoding="utf-8") as f:
json.dump(_praises_init_data, f, ensure_ascii=False, indent=4)
with open(praises_file, "r", encoding="utf-8") as f:
data = json.load(f)
praises_json = data
return praises_json
@update_to_cache("praises")
async def refresh_praises_json():
global praises_json
praises_file = store.get_plugin_data_file("praises.json")
if not praises_file.exists():
with open(praises_file, "w", encoding="utf-8") as f:
json.dump(_praises_init_data, f, ensure_ascii=False, indent=4) # 异步?
async with aiofiles.open(praises_file, "r", encoding="utf-8") as f:
data = json.loads(await f.read())
return data
praises_json = data
async def build_praises() -> str:
praises = await get_praises()
def build_praises() -> str:
praises = get_praises()
result = ["你喜欢以下几个人物,他们有各自的优点:"]
for item in praises["like"]:
result.append(f"名字:{item['name']},优点:{item['advantages']}")
@ -190,21 +210,22 @@ async def load_context_from_json(name: str, path: str) -> list:
return []
@from_cache("nickname")
async def get_nicknames():
"""获取nickname_json, 优先来源于缓存"""
"""获取nickname_json, 优先来源于全局变量"""
global nickname_json
if nickname_json is None:
filename = store.get_plugin_data_file("nickname.json")
# noinspection PyBroadException
try:
async with aiofiles.open(filename, "r", encoding="utf-8") as f:
nickname_json = json.loads(await f.read())
except (json.JSONDecodeError, FileNotFoundError):
except Exception:
nickname_json = {}
return nickname_json
@update_to_cache("nickname")
async def set_nickname(user_id: str, name: str):
global nickname_json
filename = store.get_plugin_data_file("nickname.json")
if not filename.exists():
data = {}
@ -214,35 +235,29 @@ async def set_nickname(user_id: str, name: str):
data[user_id] = name
if name == "" and user_id in data:
del data[user_id]
async with aiofiles.open(filename, "w", encoding="utf-8") as f:
await f.write(json.dumps(data, ensure_ascii=False, indent=4))
return data
with open(filename, "w", encoding="utf-8") as f:
json.dump(data, f, ensure_ascii=False, indent=4)
nickname_json = data
async def get_nickname_by_user_id(user_id: str):
nickname_json = await get_nicknames()
return nickname_json.get(user_id, "")
@update_to_cache("nickname")
async def refresh_nickname_json():
"""强制刷新nickname_json"""
"""强制刷新nickname_json, 刷新全局变量"""
global nickname_json
# noinspection PyBroadException
try:
async with aiofiles.open(
store.get_plugin_data_file("nickname.json"), "r", encoding="utf-8"
) as f:
nickname_json = json.loads(await f.read())
return nickname_json
except (json.JSONDecodeError, FileNotFoundError):
except Exception:
logger.error("刷新 nickname_json 表错误:无法载入 nickname.json 文件")
async def get_prompt(model: str) -> List[Dict[str, Any]]:
def get_prompt(model: str) -> List[Dict[str, Any]]:
"""获取系统提示词"""
prompts = config.marshoai_additional_prompt
if config.marshoai_enable_praises:
praises_prompt = await build_praises()
praises_prompt = build_praises()
prompts += praises_prompt
if config.marshoai_enable_time_prompt:

View File

@ -1,87 +0,0 @@
from nonebot.log import logger
from openai import AsyncStream
from openai.types.chat import ChatCompletion, ChatCompletionChunk, ChatCompletionMessage
from openai.types.chat.chat_completion import Choice
async def process_chat_stream(
stream: AsyncStream[ChatCompletionChunk],
) -> ChatCompletion:
reasoning_contents = ""
answer_contents = ""
last_chunk = None
is_first_token_appeared = False
is_answering = False
async for chunk in stream:
last_chunk = chunk
# print(chunk)
if not is_first_token_appeared:
logger.info(f"{chunk.id}: 第一个 token 已出现")
is_first_token_appeared = True
if not chunk.choices:
logger.info("Usage:", chunk.usage)
else:
delta = chunk.choices[0].delta
if (
hasattr(delta, "reasoning_content")
and delta.reasoning_content is not None
):
reasoning_contents += delta.reasoning_content
else:
if not is_answering:
logger.info(
f"{chunk.id}: 思维链已输出完毕或无 reasoning_content 字段输出"
)
is_answering = True
if delta.content is not None:
answer_contents += delta.content
# print(last_chunk)
# 创建新的 ChatCompletion 对象
if last_chunk and last_chunk.choices:
message = ChatCompletionMessage(
content=answer_contents,
role="assistant",
tool_calls=last_chunk.choices[0].delta.tool_calls, # type: ignore
)
if reasoning_contents != "":
setattr(message, "reasoning_content", reasoning_contents)
choice = Choice(
finish_reason=last_chunk.choices[0].finish_reason, # type: ignore
index=last_chunk.choices[0].index,
message=message,
)
return ChatCompletion(
id=last_chunk.id,
choices=[choice],
created=last_chunk.created,
model=last_chunk.model,
system_fingerprint=last_chunk.system_fingerprint,
object="chat.completion",
usage=last_chunk.usage,
)
else:
return ChatCompletion(
id="",
choices=[],
created=0,
model="",
system_fingerprint="",
object="chat.completion",
usage=None,
)
async def process_completion_to_details(completion: ChatCompletion) -> str:
usage_text = ""
usage = completion.usage
if usage is None:
usage_text = ""
else:
usage_text = str(usage)
details_text = f"""=========消息详情=========
模型: {completion.model}
消息 ID: {completion.id}
用量信息: {usage_text}"""
# print(details_text)
return details_text

32
pdm.lock generated
View File

@ -5,7 +5,7 @@
groups = ["default", "dev", "test"]
strategy = ["inherit_metadata"]
lock_version = "4.5.0"
content_hash = "sha256:6aa043fb1d2d4d384e0d0c698c02a27f22e099828d2973a4baef05c5316f4ee0"
content_hash = "sha256:d7ab3d9ca825de512d4f87ec846f7fddcf3d5796a7c9562e60c8c7d39c058817"
[[metadata.targets]]
requires_python = "~=3.10"
@ -1485,7 +1485,7 @@ files = [
[[package]]
name = "nonebot-plugin-alconna"
version = "0.57.0"
version = "0.54.1"
requires_python = ">=3.9"
summary = "Alconna Adapter for Nonebot"
groups = ["default"]
@ -1499,8 +1499,8 @@ dependencies = [
"tarina<0.7,>=0.6.8",
]
files = [
{file = "nonebot_plugin_alconna-0.57.0-py3-none-any.whl", hash = "sha256:6c4bcce1a9aa176244b4c011b19b1cea00269c4c6794cd4e90d8dd7990ec3ec9"},
{file = "nonebot_plugin_alconna-0.57.0.tar.gz", hash = "sha256:7a9a4bf373f3f6836611dbde1a0917b84441a534dd6f2b20dae3ba6fff142858"},
{file = "nonebot_plugin_alconna-0.54.1-py3-none-any.whl", hash = "sha256:4edb4b081cd64ce37717c7a92d31aadd2cf287a5a0adc2ac86ed82d9bcad5048"},
{file = "nonebot_plugin_alconna-0.54.1.tar.gz", hash = "sha256:66fae03120b8eff25bb0027d65f149e399aa6f73c7585ebdd388d1904cecdeee"},
]
[[package]]
@ -1519,27 +1519,9 @@ files = [
{file = "nonebot_plugin_apscheduler-0.5.0.tar.gz", hash = "sha256:6c0230e99765f275dc83d6639ff33bd6f71203fa10cd1b8a204b0f95530cda86"},
]
[[package]]
name = "nonebot-plugin-argot"
version = "0.1.7"
requires_python = ">=3.10"
summary = "NoneBot 暗语"
groups = ["default"]
dependencies = [
"aiofiles>=24.1.0",
"nonebot-plugin-alconna>=0.51.1",
"nonebot-plugin-apscheduler>=0.5.0",
"nonebot-plugin-localstore>=0.7.4",
"nonebot2>=2.3.2",
]
files = [
{file = "nonebot_plugin_argot-0.1.7-py3-none-any.whl", hash = "sha256:1af939a60967e27aff6f7ce97150d26cba8f1ef0cf216b44372cc0d8e5937204"},
{file = "nonebot_plugin_argot-0.1.7.tar.gz", hash = "sha256:f76c2139c9af1e2de6efdc487b728fbad84737d272bf1f600d085bbe6ed79094"},
]
[[package]]
name = "nonebot-plugin-localstore"
version = "0.7.4"
version = "0.7.3"
requires_python = "<4.0,>=3.9"
summary = "Local Storage Support for NoneBot2"
groups = ["default"]
@ -1550,8 +1532,8 @@ dependencies = [
"typing-extensions<5.0.0,>=4.0.0",
]
files = [
{file = "nonebot_plugin_localstore-0.7.4-py3-none-any.whl", hash = "sha256:3b08030878eadcdd8b9ce3d079da0dc2d0e41dc91f0b2d8cf7fa862a27de9090"},
{file = "nonebot_plugin_localstore-0.7.4.tar.gz", hash = "sha256:85ddc13814bfcd484ab311306823651390020bf44f4fb4733b343a58e72723ce"},
{file = "nonebot_plugin_localstore-0.7.3-py3-none-any.whl", hash = "sha256:1bc239b4b5320df0dc08eada7c4f8ba4cb92d4dc3134bf4646ab5e297bd7e575"},
{file = "nonebot_plugin_localstore-0.7.3.tar.gz", hash = "sha256:1aff10e2dacfc5bc9ce239fd34849f8d7172a118135dbc5aeba1c97605d9959d"},
]
[[package]]

View File

@ -10,7 +10,7 @@ authors = [
]
dependencies = [
"nonebot2>=2.4.0",
"nonebot-plugin-alconna>=0.57.1",
"nonebot-plugin-alconna>=0.48.0",
"nonebot-plugin-localstore>=0.7.1",
"zhDatetime>=2.0.0",
"aiohttp>=3.9",
@ -28,14 +28,13 @@ dependencies = [
"azure-ai-inference>=1.0.0b6",
"watchdog>=6.0.0",
"nonebot-plugin-apscheduler>=0.5.0",
"openai>=1.58.1",
"nonebot-plugin-argot>=0.1.7"
"openai>=1.58.1"
]
license = { text = "MIT, Mulan PSL v2" }
[project.urls]
Homepage = "https://marshoai-docs.pages.liteyuki.icu/"
Homepage = "https://marsho.liteyuki.icu/"
[tool.nonebot]
@ -82,4 +81,4 @@ test = [
]
[tool.ruff.lint]
ignore = ["E402", "F405"]
ignore = ["E402"]