配置两个 AI 功能都使用默认模型
This commit is contained in:
@ -7,7 +7,7 @@ class THQwen(TextHandler):
|
|||||||
name = "qwen"
|
name = "qwen"
|
||||||
|
|
||||||
async def handle(self, env: TextHandlerEnvironment, istream: str | None, args: list[str]) -> TextHandleResult:
|
async def handle(self, env: TextHandlerEnvironment, istream: str | None, args: list[str]) -> TextHandleResult:
|
||||||
llm = get_llm("qwen3-max")
|
llm = get_llm()
|
||||||
messages = []
|
messages = []
|
||||||
|
|
||||||
if istream is not None:
|
if istream is not None:
|
||||||
|
|||||||
@ -91,7 +91,7 @@ async def ask_ai(expression: str, now: datetime.datetime | None = None) -> tuple
|
|||||||
logger.info(f"提醒功能:消息被阿里绿网拦截 message={expression}")
|
logger.info(f"提醒功能:消息被阿里绿网拦截 message={expression}")
|
||||||
return None, ""
|
return None, ""
|
||||||
|
|
||||||
llm = get_llm("qwen3-max")
|
llm = get_llm()
|
||||||
message = await llm.chat([
|
message = await llm.chat([
|
||||||
{ "role": "system", "content": prompt },
|
{ "role": "system", "content": prompt },
|
||||||
{ "role": "user", "content": expression },
|
{ "role": "user", "content": expression },
|
||||||
|
|||||||
Reference in New Issue
Block a user