Compare commits
18 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| 40be5ce335 | |||
| 8e6131473d | |||
|
78bda5fc0a
|
|||
|
97658a6c56
|
|||
|
3fedc685a9
|
|||
|
d1a3e44c45
|
|||
|
f637778173
|
|||
|
145bfedf67
|
|||
|
61b9d733a5
|
|||
| ae59c20e2f | |||
| 0b7d21aeb0 | |||
|
d6ede3e6cd
|
|||
|
07ace8e6e9
|
|||
|
6f08c22b5b
|
|||
|
3e5c1941c8
|
|||
| f6e7dfcd93 | |||
| 1233677eea | |||
| 00bdb90e3c |
6
.gitignore
vendored
6
.gitignore
vendored
@ -1,5 +1,11 @@
|
||||
# 基本的数据文件,以及环境用文件
|
||||
/.env
|
||||
/data
|
||||
/pyrightconfig.json
|
||||
/pyrightconfig.toml
|
||||
|
||||
# 缓存文件
|
||||
__pycache__
|
||||
|
||||
# 可能会偶然生成的 diff 文件
|
||||
/*.diff
|
||||
|
||||
187
QWEN.md
Normal file
187
QWEN.md
Normal file
@ -0,0 +1,187 @@
|
||||
# Konabot Project Context
|
||||
|
||||
## Project Overview
|
||||
|
||||
Konabot is a multi-platform chatbot built using the NoneBot2 framework, primarily used within MTTU (likely an organization or community). The bot supports multiple adapters including Discord, QQ (via Onebot), Minecraft, and Console interfaces.
|
||||
|
||||
### Key Features
|
||||
- Multi-platform support (Discord, QQ, Minecraft, Console)
|
||||
- Rich plugin ecosystem with over 20 built-in plugins
|
||||
- Asynchronous database system with connection pooling (SQLite-based)
|
||||
- Advanced image processing capabilities
|
||||
- Integration with external services like Bilibili analysis
|
||||
- Support for Large Language Models (LLM)
|
||||
- Web rendering capabilities for advanced image generation
|
||||
|
||||
## Technology Stack
|
||||
|
||||
- **Framework**: NoneBot2
|
||||
- **Language**: Python 3.12+
|
||||
- **Dependency Management**: Poetry
|
||||
- **Database**: SQLite with aiosqlite for async operations
|
||||
- **Build System**: Just (task runner)
|
||||
- **Containerization**: Docker
|
||||
- **CI/CD**: Drone CI
|
||||
- **Testing**: Pytest
|
||||
|
||||
## Project Structure
|
||||
|
||||
```
|
||||
konabot/
|
||||
├── bot.py # Main entry point
|
||||
├── pyproject.toml # Project dependencies and metadata
|
||||
├── justfile # Task definitions
|
||||
├── Dockerfile # Container build definition
|
||||
├── .drone.yml # CI/CD pipeline configuration
|
||||
├── konabot/ # Main source code
|
||||
│ ├── common/ # Shared utilities and modules
|
||||
│ │ ├── database/ # Async database manager with connection pooling
|
||||
│ │ ├── llm/ # Large Language Model integration
|
||||
│ │ ├── web_render/ # Web-based image rendering
|
||||
│ │ └── ... # Other utilities
|
||||
│ ├── plugins/ # Plugin modules (core functionality)
|
||||
│ │ ├── air_conditioner/
|
||||
│ │ ├── bilibili_fetch/
|
||||
│ │ ├── gen_qrcode/
|
||||
│ │ ├── hanzi/
|
||||
│ │ ├── idiomgame/
|
||||
│ │ ├── image_process/
|
||||
│ │ ├── roll_dice/
|
||||
│ │ ├── weather/
|
||||
│ │ └── ... (20+ plugins)
|
||||
│ └── test/
|
||||
├── tests/ # Test suite
|
||||
├── scripts/ # Utility scripts
|
||||
├── docs/ # Documentation
|
||||
├── assets/ # Static assets
|
||||
└── data/ # Runtime data storage
|
||||
```
|
||||
|
||||
## Development Environment Setup
|
||||
|
||||
### Prerequisites
|
||||
- Python 3.12+
|
||||
- Git
|
||||
- Poetry (installed via pipx)
|
||||
|
||||
### Installation Steps
|
||||
1. Clone the repository:
|
||||
```bash
|
||||
git clone https://gitea.service.jazzwhom.top/Passthem/konabot.git
|
||||
cd konabot
|
||||
```
|
||||
2. Install dependencies:
|
||||
```bash
|
||||
poetry install
|
||||
```
|
||||
3. Configure environment:
|
||||
- Copy `.env.example` to `.env`
|
||||
- Modify settings as needed for your platform adapters
|
||||
|
||||
### Platform Adapters Configuration
|
||||
- **Discord**: Set `ENABLE_DISCORD=true` and configure bot token
|
||||
- **QQ (Onebot)**: Set `ENABLE_QQ=true` and configure connection
|
||||
- **Console**: Enabled by default, disable with `ENABLE_CONSOLE=false`
|
||||
- **Minecraft**: Set `ENABLE_MINECRAFT=true`
|
||||
|
||||
## Building and Running
|
||||
|
||||
### Development
|
||||
- Auto-reload development mode:
|
||||
```bash
|
||||
poetry run just watch
|
||||
```
|
||||
- Manual start:
|
||||
```bash
|
||||
poetry run python bot.py
|
||||
```
|
||||
|
||||
### Production
|
||||
- Docker container build and run:
|
||||
```bash
|
||||
docker build -t konabot .
|
||||
docker run konabot
|
||||
```
|
||||
|
||||
## Testing
|
||||
|
||||
Run the test suite with:
|
||||
```bash
|
||||
poetry run pytest
|
||||
```
|
||||
|
||||
Tests are located in the `tests/` directory and focus primarily on core functionality like the database manager.
|
||||
|
||||
## Database System
|
||||
|
||||
The project implements a custom asynchronous database manager (`konabot/common/database/__init__.py`) with these features:
|
||||
- Connection pooling for performance
|
||||
- Parameterized queries for security
|
||||
- SQL file execution support
|
||||
- Support for both string and Path objects for file paths
|
||||
- Automatic resource management
|
||||
|
||||
Example usage:
|
||||
```python
|
||||
from konabot.common.database import DatabaseManager
|
||||
|
||||
db = DatabaseManager()
|
||||
results = await db.query("SELECT * FROM users WHERE age > ?", (18,))
|
||||
await db.execute("INSERT INTO users (name, email) VALUES (?, ?)", ("John", "john@example.com"))
|
||||
```
|
||||
|
||||
## Plugin Architecture
|
||||
|
||||
Plugins are organized in `konabot/plugins/` and follow the NoneBot2 plugin structure. Each plugin typically consists of:
|
||||
- `__init__.py`: Main plugin logic using Alconna command parser
|
||||
- Supporting modules for specific functionality
|
||||
|
||||
Popular plugins include:
|
||||
- `roll_dice`: Dice rolling with image generation
|
||||
- `weather`: Weather radar image fetching
|
||||
- `bilibili_fetch`: Bilibili video analysis
|
||||
- `image_process`: Image manipulation tools
|
||||
- `markdown`: Markdown rendering
|
||||
|
||||
## CI/CD Pipeline
|
||||
|
||||
Drone CI is configured with two pipelines:
|
||||
1. **Nightly builds**: Triggered on pushes to master branch
|
||||
2. **Release builds**: Triggered on git tags
|
||||
|
||||
Both pipelines:
|
||||
- Build Docker images
|
||||
- Test plugin loading
|
||||
- Verify Playwright functionality
|
||||
- Send notifications via ntfy
|
||||
|
||||
## Development Conventions
|
||||
|
||||
- Use Poetry for dependency management
|
||||
- Follow NoneBot2 plugin development patterns
|
||||
- Write async code for database operations
|
||||
- Use Alconna for command parsing
|
||||
- Organize SQL queries in separate files when complex
|
||||
- Write tests for core functionality
|
||||
- Document features in the `docs/` directory
|
||||
|
||||
## Common Development Tasks
|
||||
|
||||
1. **Add a new plugin**:
|
||||
- Create a new directory in `konabot/plugins/`
|
||||
- Implement functionality in `__init__.py`
|
||||
- Use Alconna for command definition
|
||||
|
||||
2. **Database operations**:
|
||||
- Use the `DatabaseManager` class
|
||||
- Always parameterize queries
|
||||
- Store complex SQL in separate `.sql` files
|
||||
|
||||
3. **Image processing**:
|
||||
- Leverage existing utilities in `image_process` plugin
|
||||
- Use Pillow and Skia-Python for advanced graphics
|
||||
|
||||
4. **Testing**:
|
||||
- Add tests to the `tests/` directory
|
||||
- Use pytest with async support
|
||||
- Mock external services when needed
|
||||
90
konabot/common/apis/ali_content_safety.py
Normal file
90
konabot/common/apis/ali_content_safety.py
Normal file
@ -0,0 +1,90 @@
|
||||
import asyncio
|
||||
import json
|
||||
|
||||
from alibabacloud_green20220302.client import Client as AlibabaGreenClient
|
||||
from alibabacloud_green20220302.models import TextModerationPlusRequest
|
||||
from alibabacloud_tea_openapi.models import Config as AlibabaTeaConfig
|
||||
from loguru import logger
|
||||
from pydantic import BaseModel
|
||||
|
||||
import nonebot
|
||||
|
||||
|
||||
class AlibabaGreenPluginConfig(BaseModel):
|
||||
module_aligreen_enable: bool = False
|
||||
module_aligreen_access_key_id: str = ""
|
||||
module_aligreen_access_key_secret: str = ""
|
||||
module_aligreen_region_id: str = "cn-shenzhen"
|
||||
module_aligreen_endpoint: str = "green-cip.cn-shenzhen.aliyuncs.com"
|
||||
module_aligreen_service: str = "llm_query_moderation"
|
||||
|
||||
|
||||
class AlibabaGreen:
|
||||
_client: AlibabaGreenClient | None = None
|
||||
_config: AlibabaGreenPluginConfig | None = None
|
||||
|
||||
@staticmethod
|
||||
def get_client() -> AlibabaGreenClient:
|
||||
assert AlibabaGreen._client is not None
|
||||
return AlibabaGreen._client
|
||||
|
||||
@staticmethod
|
||||
def get_config() -> AlibabaGreenPluginConfig:
|
||||
assert AlibabaGreen._config is not None
|
||||
return AlibabaGreen._config
|
||||
|
||||
@staticmethod
|
||||
def init():
|
||||
config = nonebot.get_plugin_config(AlibabaGreenPluginConfig)
|
||||
AlibabaGreen._config = config
|
||||
if not config.module_aligreen_enable:
|
||||
logger.info("该环境未启用阿里内容审查,跳过初始化")
|
||||
return
|
||||
AlibabaGreen._client = AlibabaGreenClient(AlibabaTeaConfig(
|
||||
access_key_id=config.module_aligreen_access_key_id,
|
||||
access_key_secret=config.module_aligreen_access_key_secret,
|
||||
connect_timeout=10000,
|
||||
read_timeout=3000,
|
||||
region_id=config.module_aligreen_region_id,
|
||||
endpoint=config.module_aligreen_endpoint,
|
||||
))
|
||||
|
||||
@staticmethod
|
||||
def _detect_sync(content: str) -> bool:
|
||||
if not AlibabaGreen.get_config().module_aligreen_enable:
|
||||
logger.debug("该环境未启用阿里内容审查,直接跳过")
|
||||
return True
|
||||
|
||||
client = AlibabaGreen.get_client()
|
||||
try:
|
||||
response = client.text_moderation_plus(TextModerationPlusRequest(
|
||||
service=AlibabaGreen.get_config().module_aligreen_service,
|
||||
service_parameters=json.dumps({
|
||||
"content": content,
|
||||
}),
|
||||
))
|
||||
if response.status_code == 200:
|
||||
result = response.body
|
||||
logger.info(f"检测违规内容 API 调用成功:{result}")
|
||||
risk_level: str = result.data.risk_level or "none"
|
||||
if risk_level == "high":
|
||||
return False
|
||||
return True
|
||||
logger.error(f"检测违规内容 API 调用失败:{response}")
|
||||
return True
|
||||
except Exception as e:
|
||||
logger.error("检测违规内容 API 调用失败")
|
||||
logger.exception(e)
|
||||
return True
|
||||
|
||||
@staticmethod
|
||||
async def detect(content: str) -> bool:
|
||||
return await asyncio.to_thread(AlibabaGreen._detect_sync, content)
|
||||
|
||||
|
||||
driver = nonebot.get_driver()
|
||||
|
||||
@driver.on_startup
|
||||
async def _():
|
||||
AlibabaGreen.init()
|
||||
|
||||
@ -59,6 +59,9 @@ def get_llm(llm_model: str | None = None):
|
||||
if llm_model is None:
|
||||
llm_model = llm_config.default_llm
|
||||
if llm_model not in llm_config.llms:
|
||||
raise NotImplementedError("LLM 未配置,该功能无法使用")
|
||||
if llm_config.default_llm in llm_config.llms:
|
||||
logger.warning(f"[LLM] 需求的 LLM 不存在,回退到默认模型 REQUIRED={llm_model}")
|
||||
return llm_config.llms[llm_config.default_llm]
|
||||
raise NotImplementedError("[LLM] LLM 未配置,该功能无法使用")
|
||||
return llm_config.llms[llm_model]
|
||||
|
||||
|
||||
3
konabot/common/ptimeparse/README.md
Normal file
3
konabot/common/ptimeparse/README.md
Normal file
@ -0,0 +1,3 @@
|
||||
# 已废弃
|
||||
|
||||
坏枪用简单的 LLM + 提示词工程,完成了这 200 块的 `qwen3-coder-plus` 都搞不定的 nb 功能
|
||||
@ -1,653 +1,58 @@
|
||||
import re
|
||||
import datetime
|
||||
from typing import Tuple, Optional, Dict, Any
|
||||
"""
|
||||
Professional time parsing module for Chinese and English time expressions.
|
||||
|
||||
from .err import MultipleSpecificationException, TokenUnhandledException
|
||||
This module provides a robust parser for natural language time expressions,
|
||||
supporting both Chinese and English formats with proper whitespace handling.
|
||||
"""
|
||||
|
||||
import datetime
|
||||
from typing import Optional
|
||||
|
||||
from .expression import TimeExpression
|
||||
|
||||
|
||||
def parse(text: str, now: Optional[datetime.datetime] = None) -> datetime.datetime:
|
||||
"""
|
||||
Parse a time expression and return a datetime object.
|
||||
|
||||
Args:
|
||||
text: The time expression to parse
|
||||
now: The reference time (defaults to current time)
|
||||
|
||||
Returns:
|
||||
A datetime object representing the parsed time
|
||||
|
||||
Raises:
|
||||
TokenUnhandledException: If the input cannot be parsed
|
||||
"""
|
||||
return TimeExpression.parse(text, now)
|
||||
|
||||
|
||||
class Parser:
|
||||
"""
|
||||
Parser for time expressions with backward compatibility.
|
||||
|
||||
Maintains the original interface:
|
||||
>>> parser = Parser()
|
||||
>>> result = parser.parse("10分钟后")
|
||||
"""
|
||||
|
||||
def __init__(self, now: Optional[datetime.datetime] = None):
|
||||
self.now = now or datetime.datetime.now()
|
||||
|
||||
def digest_chinese_number(self, text: str) -> Tuple[str, int]:
|
||||
if not text:
|
||||
return text, 0
|
||||
# Handle "两" at start
|
||||
if text.startswith("两"):
|
||||
next_char = text[1] if len(text) > 1 else ''
|
||||
if not next_char or next_char in "十百千万亿":
|
||||
return text[1:], 2
|
||||
s = "零一二三四五六七八九"
|
||||
digits = {c: i for i, c in enumerate(s)}
|
||||
i = 0
|
||||
while i < len(text) and text[i] in s + "十百千万亿":
|
||||
i += 1
|
||||
if i == 0:
|
||||
return text, 0
|
||||
num_str = text[:i]
|
||||
rest = text[i:]
|
||||
|
||||
def parse(s):
|
||||
if not s:
|
||||
return 0
|
||||
if s == "零":
|
||||
return 0
|
||||
if "亿" in s:
|
||||
a, b = s.split("亿", 1)
|
||||
return parse(a) * 100000000 + parse(b)
|
||||
if "万" in s:
|
||||
a, b = s.split("万", 1)
|
||||
return parse(a) * 10000 + parse(b)
|
||||
n = 0
|
||||
t = 0
|
||||
for c in s:
|
||||
if c == "零":
|
||||
continue
|
||||
if c in digits:
|
||||
t = digits[c]
|
||||
elif c == "十":
|
||||
if t == 0:
|
||||
t = 1
|
||||
n += t * 10
|
||||
t = 0
|
||||
elif c == "百":
|
||||
if t == 0:
|
||||
t = 1
|
||||
n += t * 100
|
||||
t = 0
|
||||
elif c == "千":
|
||||
if t == 0:
|
||||
t = 1
|
||||
n += t * 1000
|
||||
t = 0
|
||||
n += t
|
||||
return n
|
||||
|
||||
return rest, parse(num_str)
|
||||
|
||||
|
||||
def parse(self, text: str) -> datetime.datetime:
|
||||
text = text.strip()
|
||||
if not text:
|
||||
raise TokenUnhandledException("Empty input")
|
||||
|
||||
ctx = {
|
||||
"date": None,
|
||||
"time": None,
|
||||
"relative_delta": None,
|
||||
"am_pm": None,
|
||||
"period_word": None,
|
||||
"has_time": False,
|
||||
"has_date": False,
|
||||
"ambiguous_hour": False,
|
||||
"is_24hour": False,
|
||||
"has_relative_date": False,
|
||||
}
|
||||
|
||||
rest = self._parse_all(text, ctx)
|
||||
if rest.strip():
|
||||
raise TokenUnhandledException(f"Unparsed tokens: {rest.strip()}")
|
||||
|
||||
return self._apply_context(ctx)
|
||||
|
||||
def _parse_all(self, text: str, ctx: Dict[str, Any]) -> str:
|
||||
rest = text.lstrip()
|
||||
while True:
|
||||
for parser in [
|
||||
self._parse_absolute_date,
|
||||
self._parse_relative_date,
|
||||
self._parse_relative_time,
|
||||
self._parse_period,
|
||||
self._parse_time,
|
||||
]:
|
||||
new_rest = parser(rest, ctx)
|
||||
if new_rest != rest:
|
||||
rest = new_rest.lstrip()
|
||||
break
|
||||
else:
|
||||
break
|
||||
return rest
|
||||
|
||||
def _add_delta(self, ctx, delta):
|
||||
if ctx["relative_delta"] is None:
|
||||
ctx["relative_delta"] = delta
|
||||
else:
|
||||
ctx["relative_delta"] += delta
|
||||
|
||||
def _parse_absolute_date(self, text: str, ctx: Dict[str, Any]) -> str:
|
||||
text = text.lstrip()
|
||||
m = re.match(r"^(\d{4})-(\d{1,2})-(\d{1,2})T(\d{1,2}):(\d{2})", text)
|
||||
if m:
|
||||
y, mth, d, h, minute = map(int, m.groups())
|
||||
ctx["date"] = datetime.date(y, mth, d)
|
||||
ctx["time"] = datetime.time(h, minute)
|
||||
ctx["has_date"] = True
|
||||
ctx["has_time"] = True
|
||||
ctx["is_24hour"] = True
|
||||
return text[m.end():]
|
||||
m = re.match(r"^(\d{4})-(\d{1,2})-(\d{1,2})", text)
|
||||
if m:
|
||||
y, mth, d = map(int, m.groups())
|
||||
ctx["date"] = datetime.date(y, mth, d)
|
||||
ctx["has_date"] = True
|
||||
return text[m.end():]
|
||||
m = re.match(r"^(\d{4})/(\d{1,2})/(\d{1,2})", text)
|
||||
if m:
|
||||
y, mth, d = map(int, m.groups())
|
||||
ctx["date"] = datetime.date(y, mth, d)
|
||||
ctx["has_date"] = True
|
||||
return text[m.end():]
|
||||
m = re.match(r"^(\d{4})年(\d{1,2})月(\d{1,2})[日号]", text)
|
||||
if m:
|
||||
y, mth, d = map(int, m.groups())
|
||||
ctx["date"] = datetime.date(y, mth, d)
|
||||
ctx["has_date"] = True
|
||||
return text[m.end():]
|
||||
m = re.match(r"^(\d{1,2})月(\d{1,2})[日号]", text)
|
||||
if m:
|
||||
mth, d = map(int, m.groups())
|
||||
ctx["date"] = datetime.date(self.now.year, mth, d)
|
||||
ctx["has_date"] = True
|
||||
return text[m.end():]
|
||||
m = re.match(r"^(.{1,3})月(.{1,3})[日号]", text)
|
||||
if m:
|
||||
m_str, d_str = m.groups()
|
||||
_, mth = self.digest_chinese_number(m_str)
|
||||
_, d = self.digest_chinese_number(d_str)
|
||||
if mth == 0:
|
||||
mth = 1
|
||||
if d == 0:
|
||||
d = 1
|
||||
ctx["date"] = datetime.date(self.now.year, mth, d)
|
||||
ctx["has_date"] = True
|
||||
return text[m.end():]
|
||||
return text
|
||||
|
||||
def _parse_relative_date(self, text: str, ctx: Dict[str, Any]) -> str:
|
||||
text = text.lstrip()
|
||||
"""
|
||||
Parse a time expression and return a datetime object.
|
||||
This maintains backward compatibility with the original interface.
|
||||
|
||||
# Handle "今天", "今晚", "今早", etc.
|
||||
today_variants = [
|
||||
("今晚上", "PM"),
|
||||
("今晚", "PM"),
|
||||
("今早", "AM"),
|
||||
("今天早上", "AM"),
|
||||
("今天早晨", "AM"),
|
||||
("今天上午", "AM"),
|
||||
("今天下午", "PM"),
|
||||
("今天晚上", "PM"),
|
||||
("今天", None),
|
||||
]
|
||||
for variant, period in today_variants:
|
||||
if text.startswith(variant):
|
||||
self._add_delta(ctx, datetime.timedelta(days=0))
|
||||
ctx["has_relative_date"] = True
|
||||
rest = text[len(variant):]
|
||||
if period is not None and ctx["am_pm"] is None:
|
||||
ctx["am_pm"] = period
|
||||
ctx["period_word"] = variant
|
||||
return rest
|
||||
Args:
|
||||
text: The time expression to parse
|
||||
|
||||
Returns:
|
||||
A datetime object representing the parsed time
|
||||
|
||||
Raises:
|
||||
TokenUnhandledException: If the input cannot be parsed
|
||||
"""
|
||||
return TimeExpression.parse(text, self.now)
|
||||
|
||||
mapping = {
|
||||
"明天": 1,
|
||||
"后天": 2,
|
||||
"大后天": 3,
|
||||
"昨天": -1,
|
||||
"前天": -2,
|
||||
"大前天": -3,
|
||||
}
|
||||
for word, days in mapping.items():
|
||||
if text.startswith(word):
|
||||
self._add_delta(ctx, datetime.timedelta(days=days))
|
||||
ctx["has_relative_date"] = True
|
||||
return text[len(word):]
|
||||
m = re.match(r"^(\d+|[零一二三四五六七八九十两]+)天(后|前|以后|之后)", text)
|
||||
if m:
|
||||
num_str, direction = m.groups()
|
||||
if num_str.isdigit():
|
||||
n = int(num_str)
|
||||
else:
|
||||
_, n = self.digest_chinese_number(num_str)
|
||||
days = n if direction in ("后", "以后", "之后") else -n
|
||||
self._add_delta(ctx, datetime.timedelta(days=days))
|
||||
ctx["has_relative_date"] = True
|
||||
return text[m.end():]
|
||||
m = re.match(r"^(本|上|下)周([一二三四五六日])", text)
|
||||
if m:
|
||||
scope, day = m.groups()
|
||||
weekday_map = {"一": 0, "二": 1, "三": 2, "四": 3, "五": 4, "六": 5, "日": 6}
|
||||
target = weekday_map[day]
|
||||
current = self.now.weekday()
|
||||
if scope == "本":
|
||||
delta = target - current
|
||||
elif scope == "上":
|
||||
delta = target - current - 7
|
||||
else:
|
||||
delta = target - current + 7
|
||||
self._add_delta(ctx, datetime.timedelta(days=delta))
|
||||
ctx["has_relative_date"] = True
|
||||
return text[m.end():]
|
||||
return text
|
||||
|
||||
def _parse_period(self, text: str, ctx: Dict[str, Any]) -> str:
|
||||
text = text.lstrip()
|
||||
period_mapping = {
|
||||
"上午": "AM",
|
||||
"早晨": "AM",
|
||||
"早上": "AM",
|
||||
"早": "AM",
|
||||
"中午": "PM",
|
||||
"下午": "PM",
|
||||
"晚上": "PM",
|
||||
"晚": "PM",
|
||||
"凌晨": "AM",
|
||||
}
|
||||
for word, tag in period_mapping.items():
|
||||
if text.startswith(word):
|
||||
if ctx["am_pm"] is not None:
|
||||
raise MultipleSpecificationException("Multiple periods")
|
||||
ctx["am_pm"] = tag
|
||||
ctx["period_word"] = word
|
||||
return text[len(word):]
|
||||
return text
|
||||
|
||||
def _parse_time(self, text: str, ctx: Dict[str, Any]) -> str:
|
||||
if ctx["has_time"]:
|
||||
return text
|
||||
text = text.lstrip()
|
||||
|
||||
# 1. H:MM pattern
|
||||
m = re.match(r"^(\d{1,2}):(\d{2})", text)
|
||||
if m:
|
||||
h, minute = int(m.group(1)), int(m.group(2))
|
||||
if 0 <= h <= 23 and 0 <= minute <= 59:
|
||||
ctx["time"] = datetime.time(h, minute)
|
||||
ctx["has_time"] = True
|
||||
ctx["ambiguous_hour"] = 1 <= h <= 12
|
||||
ctx["is_24hour"] = h > 12 or h == 0
|
||||
return text[m.end():]
|
||||
|
||||
# 2. Parse hour part
|
||||
hour = None
|
||||
rest_after_hour = text
|
||||
is_24hour_format = False
|
||||
|
||||
# Try Chinese number + 点/时
|
||||
temp_rest, num = self.digest_chinese_number(text)
|
||||
if num >= 0:
|
||||
temp_rest_stripped = temp_rest.lstrip()
|
||||
if temp_rest_stripped.startswith("点"):
|
||||
hour = num
|
||||
is_24hour_format = False
|
||||
rest_after_hour = temp_rest_stripped[1:]
|
||||
elif temp_rest_stripped.startswith("时"):
|
||||
hour = num
|
||||
is_24hour_format = True
|
||||
rest_after_hour = temp_rest_stripped[1:]
|
||||
|
||||
if hour is None:
|
||||
m = re.match(r"^(\d{1,2})\s*([点时])", text)
|
||||
if m:
|
||||
hour = int(m.group(1))
|
||||
is_24hour_format = m.group(2) == "时"
|
||||
rest_after_hour = text[m.end():]
|
||||
|
||||
if hour is None:
|
||||
if ctx.get("am_pm") is not None:
|
||||
temp_rest, num = self.digest_chinese_number(text)
|
||||
if 0 <= num <= 23:
|
||||
hour = num
|
||||
is_24hour_format = False
|
||||
rest_after_hour = temp_rest.lstrip()
|
||||
else:
|
||||
m = re.match(r"^(\d{1,2})", text)
|
||||
if m:
|
||||
h_val = int(m.group(1))
|
||||
if 0 <= h_val <= 23:
|
||||
hour = h_val
|
||||
is_24hour_format = False
|
||||
rest_after_hour = text[m.end():].lstrip()
|
||||
|
||||
if hour is None:
|
||||
return text
|
||||
|
||||
if not (0 <= hour <= 23):
|
||||
return text
|
||||
|
||||
# Parse minutes
|
||||
rest = rest_after_hour.lstrip()
|
||||
minute = 0
|
||||
minute_spec_count = 0
|
||||
|
||||
if rest.startswith("钟"):
|
||||
rest = rest[1:].lstrip()
|
||||
|
||||
has_zheng = False
|
||||
if rest.startswith("整"):
|
||||
has_zheng = True
|
||||
rest = rest[1:].lstrip()
|
||||
|
||||
if rest.startswith("半"):
|
||||
minute = 30
|
||||
minute_spec_count += 1
|
||||
rest = rest[1:].lstrip()
|
||||
if rest.startswith("钟"):
|
||||
rest = rest[1:].lstrip()
|
||||
if rest.startswith("整"):
|
||||
rest = rest[1:].lstrip()
|
||||
|
||||
if rest.startswith("一刻"):
|
||||
minute = 15
|
||||
minute_spec_count += 1
|
||||
rest = rest[2:].lstrip()
|
||||
if rest.startswith("钟"):
|
||||
rest = rest[1:].lstrip()
|
||||
|
||||
if rest.startswith("过一刻"):
|
||||
minute = 15
|
||||
minute_spec_count += 1
|
||||
rest = rest[3:].lstrip()
|
||||
if rest.startswith("钟"):
|
||||
rest = rest[1:].lstrip()
|
||||
|
||||
m = re.match(r"^(\d+|[零一二三四五六七八九十]+)分", rest)
|
||||
if m:
|
||||
minute_spec_count += 1
|
||||
m_str = m.group(1)
|
||||
if m_str.isdigit():
|
||||
minute = int(m_str)
|
||||
else:
|
||||
_, minute = self.digest_chinese_number(m_str)
|
||||
rest = rest[m.end():].lstrip()
|
||||
|
||||
if minute_spec_count == 0:
|
||||
temp_rest, num = self.digest_chinese_number(rest)
|
||||
if num > 0 and num <= 59:
|
||||
minute = num
|
||||
minute_spec_count += 1
|
||||
rest = temp_rest.lstrip()
|
||||
else:
|
||||
m = re.match(r"^(\d{1,2})", rest)
|
||||
if m:
|
||||
m_val = int(m.group(1))
|
||||
if 0 <= m_val <= 59:
|
||||
minute = m_val
|
||||
minute_spec_count += 1
|
||||
rest = rest[m.end():].lstrip()
|
||||
|
||||
if has_zheng and minute_spec_count == 0:
|
||||
minute_spec_count = 1
|
||||
|
||||
if minute_spec_count > 1:
|
||||
raise MultipleSpecificationException("Multiple minute specifications")
|
||||
|
||||
if not (0 <= minute <= 59):
|
||||
return text
|
||||
|
||||
# Hours 13-23 are always 24-hour, even with "点"
|
||||
if hour >= 13:
|
||||
is_24hour_format = True
|
||||
|
||||
ctx["time"] = datetime.time(hour, minute)
|
||||
ctx["has_time"] = True
|
||||
ctx["ambiguous_hour"] = 1 <= hour <= 12 and not is_24hour_format
|
||||
ctx["is_24hour"] = is_24hour_format
|
||||
|
||||
return rest
|
||||
|
||||
def _parse_relative_time(self, text: str, ctx: Dict[str, Any]) -> str:
|
||||
text = text.lstrip()
|
||||
|
||||
# 半小时
|
||||
m = re.match(r"^(半)(?:个)?小时?(后|前|以后|之后)", text)
|
||||
if m:
|
||||
direction = m.group(2)
|
||||
hours = 0.5
|
||||
delta = datetime.timedelta(
|
||||
hours=hours if direction in ("后", "以后", "之后") else -hours
|
||||
)
|
||||
self._add_delta(ctx, delta)
|
||||
return text[m.end():]
|
||||
|
||||
# X个半
|
||||
m = re.match(r"^([0-9零一二三四五六七八九十两]+)个半(?:小时?)?(后|前|以后|之后)", text)
|
||||
if m:
|
||||
num_str, direction = m.groups()
|
||||
if num_str.isdigit():
|
||||
base_hours = int(num_str)
|
||||
else:
|
||||
_, base_hours = self.digest_chinese_number(num_str)
|
||||
if base_hours == 0 and num_str != "零":
|
||||
return text
|
||||
if base_hours <= 0:
|
||||
return text
|
||||
hours = base_hours + 0.5
|
||||
delta = datetime.timedelta(
|
||||
hours=hours if direction in ("后", "以后", "之后") else -hours
|
||||
)
|
||||
self._add_delta(ctx, delta)
|
||||
return text[m.end():]
|
||||
|
||||
# 一个半
|
||||
m = re.match(r"^(一个半)小时?(后|前|以后|之后)", text)
|
||||
if m:
|
||||
direction = m.group(2)
|
||||
hours = 1.5
|
||||
delta = datetime.timedelta(
|
||||
hours=hours if direction in ("后", "以后", "之后") else -hours
|
||||
)
|
||||
self._add_delta(ctx, delta)
|
||||
return text[m.end():]
|
||||
|
||||
# X小时
|
||||
m = re.match(r"^([0-9零一二三四五六七八九十两]+)(?:个)?小时?(后|前|以后|之后)", text)
|
||||
if m:
|
||||
num_str, direction = m.groups()
|
||||
if num_str.isdigit():
|
||||
hours = int(num_str)
|
||||
else:
|
||||
_, hours = self.digest_chinese_number(num_str)
|
||||
if hours == 0 and num_str != "零":
|
||||
return text
|
||||
if hours <= 0:
|
||||
return text
|
||||
delta = datetime.timedelta(
|
||||
hours=hours if direction in ("后", "以后", "之后") else -hours
|
||||
)
|
||||
self._add_delta(ctx, delta)
|
||||
return text[m.end():]
|
||||
|
||||
m = re.match(r"^([0-9零一二三四五六七八九十两]+)(?:个)?小时(后|前)", text)
|
||||
if m:
|
||||
num_str, direction = m.groups()
|
||||
if num_str.isdigit():
|
||||
hours = int(num_str)
|
||||
else:
|
||||
_, hours = self.digest_chinese_number(num_str)
|
||||
if hours == 0 and num_str != "零":
|
||||
return text
|
||||
if hours <= 0:
|
||||
return text
|
||||
delta = datetime.timedelta(
|
||||
hours=hours if direction == "后" else -hours
|
||||
)
|
||||
self._add_delta(ctx, delta)
|
||||
return text[m.end():]
|
||||
|
||||
# X分钟
|
||||
m = re.match(r"^([0-9零一二三四五六七八九十两]+)分钟?(后|前|以后|之后)", text)
|
||||
if m:
|
||||
num_str, direction = m.groups()
|
||||
if num_str.isdigit():
|
||||
minutes = int(num_str)
|
||||
else:
|
||||
_, minutes = self.digest_chinese_number(num_str)
|
||||
if minutes == 0 and num_str != "零":
|
||||
return text
|
||||
if minutes <= 0:
|
||||
return text
|
||||
delta = datetime.timedelta(
|
||||
minutes=minutes if direction in ("后", "以后", "之后") else -minutes
|
||||
)
|
||||
self._add_delta(ctx, delta)
|
||||
return text[m.end():]
|
||||
|
||||
m = re.match(r"^([0-9零一二三四五六七八九十两]+)分(后|前|以后|之后)", text)
|
||||
if m:
|
||||
num_str, direction = m.groups()
|
||||
if num_str.isdigit():
|
||||
minutes = int(num_str)
|
||||
else:
|
||||
_, minutes = self.digest_chinese_number(num_str)
|
||||
if minutes == 0 and num_str != "零":
|
||||
return text
|
||||
if minutes <= 0:
|
||||
return text
|
||||
delta = datetime.timedelta(
|
||||
minutes=minutes if direction in ("后", "以后", "之后") else -minutes
|
||||
)
|
||||
self._add_delta(ctx, delta)
|
||||
return text[m.end():]
|
||||
|
||||
m = re.match(r"^([0-9零一二三四五六七八九十两]+)分钟?(后|前)", text)
|
||||
if m:
|
||||
num_str, direction = m.groups()
|
||||
if num_str.isdigit():
|
||||
minutes = int(num_str)
|
||||
else:
|
||||
_, minutes = self.digest_chinese_number(num_str)
|
||||
if minutes == 0 and num_str != "零":
|
||||
return text
|
||||
if minutes <= 0:
|
||||
return text
|
||||
delta = datetime.timedelta(
|
||||
minutes=minutes if direction == "后" else -minutes
|
||||
)
|
||||
self._add_delta(ctx, delta)
|
||||
return text[m.end():]
|
||||
|
||||
m = re.match(r"^([0-9零一二三四五六七八九十两]+)分(后|前)", text)
|
||||
if m:
|
||||
num_str, direction = m.groups()
|
||||
if num_str.isdigit():
|
||||
minutes = int(num_str)
|
||||
else:
|
||||
_, minutes = self.digest_chinese_number(num_str)
|
||||
if minutes == 0 and num_str != "零":
|
||||
return text
|
||||
if minutes <= 0:
|
||||
return text
|
||||
delta = datetime.timedelta(
|
||||
minutes=minutes if direction == "后" else -minutes
|
||||
)
|
||||
self._add_delta(ctx, delta)
|
||||
return text[m.end():]
|
||||
|
||||
# === 秒级支持 ===
|
||||
m = re.match(r"^([0-9零一二三四五六七八九十两]+)秒(后|前|以后|之后)", text)
|
||||
if m:
|
||||
num_str, direction = m.groups()
|
||||
if num_str.isdigit():
|
||||
seconds = int(num_str)
|
||||
else:
|
||||
_, seconds = self.digest_chinese_number(num_str)
|
||||
if seconds == 0 and num_str != "零":
|
||||
return text
|
||||
if seconds <= 0:
|
||||
return text
|
||||
delta = datetime.timedelta(
|
||||
seconds=seconds if direction in ("后", "以后", "之后") else -seconds
|
||||
)
|
||||
self._add_delta(ctx, delta)
|
||||
return text[m.end():]
|
||||
|
||||
m = re.match(r"^([0-9零一二三四五六七八九十两]+)秒(后|前)", text)
|
||||
if m:
|
||||
num_str, direction = m.groups()
|
||||
if num_str.isdigit():
|
||||
seconds = int(num_str)
|
||||
else:
|
||||
_, seconds = self.digest_chinese_number(num_str)
|
||||
if seconds == 0 and num_str != "零":
|
||||
return text
|
||||
if seconds <= 0:
|
||||
return text
|
||||
delta = datetime.timedelta(
|
||||
seconds=seconds if direction == "后" else -seconds
|
||||
)
|
||||
self._add_delta(ctx, delta)
|
||||
return text[m.end():]
|
||||
|
||||
return text
|
||||
|
||||
def _apply_context(self, ctx: Dict[str, Any]) -> datetime.datetime:
|
||||
result = self.now
|
||||
has_date = ctx["has_date"]
|
||||
has_time = ctx["has_time"]
|
||||
has_delta = ctx["relative_delta"] is not None
|
||||
has_relative_date = ctx["has_relative_date"]
|
||||
|
||||
if has_delta:
|
||||
result = result + ctx["relative_delta"]
|
||||
|
||||
if has_date:
|
||||
result = result.replace(
|
||||
year=ctx["date"].year,
|
||||
month=ctx["date"].month,
|
||||
day=ctx["date"].day,
|
||||
)
|
||||
|
||||
if has_time:
|
||||
h = ctx["time"].hour
|
||||
m = ctx["time"].minute
|
||||
|
||||
if ctx["is_24hour"]:
|
||||
# "10 时" → 10:00, no conversion
|
||||
pass
|
||||
|
||||
elif ctx["am_pm"] == "AM":
|
||||
if h == 12:
|
||||
h = 0
|
||||
|
||||
elif ctx["am_pm"] == "PM":
|
||||
if h == 12:
|
||||
if ctx.get("period_word") in ("晚上", "晚"):
|
||||
h = 0
|
||||
result += datetime.timedelta(days=1)
|
||||
else:
|
||||
h = 12
|
||||
elif 1 <= h <= 11:
|
||||
h += 12
|
||||
|
||||
else:
|
||||
# No period and not 24-hour (i.e., "点" format)
|
||||
if ctx["has_relative_date"]:
|
||||
# "明天五点" → 05:00 AM
|
||||
if h == 12:
|
||||
h = 0
|
||||
# keep h as AM hour (1-11 unchanged)
|
||||
else:
|
||||
# Infer from current time
|
||||
am_hour = 0 if h == 12 else h
|
||||
candidate_am = result.replace(hour=am_hour, minute=m, second=0, microsecond=0)
|
||||
if candidate_am < self.now:
|
||||
# AM time is in the past, so use PM
|
||||
if h == 12:
|
||||
h = 12
|
||||
else:
|
||||
h += 12
|
||||
# else: keep as AM (h unchanged)
|
||||
|
||||
if h > 23:
|
||||
h = h % 24
|
||||
|
||||
result = result.replace(hour=h, minute=m, second=0, microsecond=0)
|
||||
|
||||
else:
|
||||
if has_date or (has_relative_date and not has_time):
|
||||
result = result.replace(hour=0, minute=0, second=0, microsecond=0)
|
||||
|
||||
return result
|
||||
|
||||
|
||||
def parse(text: str) -> datetime.datetime:
|
||||
return Parser().parse(text)
|
||||
|
||||
133
konabot/common/ptimeparse/chinese_number.py
Normal file
133
konabot/common/ptimeparse/chinese_number.py
Normal file
@ -0,0 +1,133 @@
|
||||
"""
|
||||
Chinese number parser for the time expression parser.
|
||||
"""
|
||||
|
||||
import re
|
||||
from typing import Tuple
|
||||
|
||||
|
||||
class ChineseNumberParser:
|
||||
"""Parser for Chinese numbers."""
|
||||
|
||||
def __init__(self):
|
||||
self.digits = {"零": 0, "一": 1, "二": 2, "三": 3, "四": 4,
|
||||
"五": 5, "六": 6, "七": 7, "八": 8, "九": 9}
|
||||
self.units = {"十": 10, "百": 100, "千": 1000, "万": 10000, "亿": 100000000}
|
||||
|
||||
def digest(self, text: str) -> Tuple[str, int]:
|
||||
"""
|
||||
Parse a Chinese number from the beginning of text and return the rest and the parsed number.
|
||||
|
||||
Args:
|
||||
text: Text that may start with a Chinese number
|
||||
|
||||
Returns:
|
||||
Tuple of (remaining_text, parsed_number)
|
||||
"""
|
||||
if not text:
|
||||
return text, 0
|
||||
|
||||
# Handle "两" at start
|
||||
if text.startswith("两"):
|
||||
# Check if "两" is followed by a time unit
|
||||
# Look ahead to see if we have a valid pattern like "两小时", "两分钟", etc.
|
||||
if len(text) >= 2:
|
||||
# Check for time units that start with the second character
|
||||
time_units = ["小时", "分钟", "秒"]
|
||||
for unit in time_units:
|
||||
if text[1:].startswith(unit):
|
||||
# Return the text starting from the time unit, not after it
|
||||
# The parser will handle the time unit in the next step
|
||||
return text[1:], 2
|
||||
# Check for single character time units
|
||||
next_char = text[1]
|
||||
if next_char in "时分秒":
|
||||
return text[1:], 2
|
||||
# Check for Chinese number units
|
||||
if next_char in "十百千万亿":
|
||||
# This will be handled by the normal parsing below
|
||||
pass
|
||||
# If "两" is at the end of string, treat it as standalone
|
||||
elif len(text) == 1:
|
||||
return "", 2
|
||||
# Also accept "两" followed by whitespace and then time units
|
||||
elif next_char.isspace():
|
||||
# Check if after whitespace we have time units
|
||||
rest_after_space = text[2:].lstrip()
|
||||
for unit in time_units:
|
||||
if rest_after_space.startswith(unit):
|
||||
# Return the text starting from the time unit
|
||||
space_len = len(text[2:]) - len(rest_after_space)
|
||||
return text[2+space_len:], 2
|
||||
# Check single character time units after whitespace
|
||||
if rest_after_space and rest_after_space[0] in "时分秒":
|
||||
return text[2:], 2
|
||||
else:
|
||||
# Just "两" by itself
|
||||
return "", 2
|
||||
|
||||
s = "零一二三四五六七八九"
|
||||
i = 0
|
||||
while i < len(text) and text[i] in s + "十百千万亿":
|
||||
i += 1
|
||||
if i == 0:
|
||||
return text, 0
|
||||
num_str = text[:i]
|
||||
rest = text[i:]
|
||||
|
||||
return rest, self.parse(num_str)
|
||||
|
||||
def parse(self, text: str) -> int:
|
||||
"""
|
||||
Parse a Chinese number string and return its integer value.
|
||||
|
||||
Args:
|
||||
text: Chinese number string
|
||||
|
||||
Returns:
|
||||
Integer value of the Chinese number
|
||||
"""
|
||||
if not text:
|
||||
return 0
|
||||
if text == "零":
|
||||
return 0
|
||||
if text == "两":
|
||||
return 2
|
||||
|
||||
# Handle special case for "十"
|
||||
if text == "十":
|
||||
return 10
|
||||
|
||||
# Handle numbers with "亿"
|
||||
if "亿" in text:
|
||||
parts = text.split("亿", 1)
|
||||
a, b = parts[0], parts[1]
|
||||
return self.parse(a) * 100000000 + self.parse(b)
|
||||
|
||||
# Handle numbers with "万"
|
||||
if "万" in text:
|
||||
parts = text.split("万", 1)
|
||||
a, b = parts[0], parts[1]
|
||||
return self.parse(a) * 10000 + self.parse(b)
|
||||
|
||||
# Handle remaining numbers
|
||||
result = 0
|
||||
temp = 0
|
||||
|
||||
for char in text:
|
||||
if char == "零":
|
||||
continue
|
||||
elif char == "两":
|
||||
temp = 2
|
||||
elif char in self.digits:
|
||||
temp = self.digits[char]
|
||||
elif char in self.units:
|
||||
unit = self.units[char]
|
||||
if unit == 10 and temp == 0:
|
||||
# Special case for numbers like "十三"
|
||||
temp = 1
|
||||
result += temp * unit
|
||||
temp = 0
|
||||
|
||||
result += temp
|
||||
return result
|
||||
63
konabot/common/ptimeparse/expression.py
Normal file
63
konabot/common/ptimeparse/expression.py
Normal file
@ -0,0 +1,63 @@
|
||||
"""
|
||||
Main time expression parser class that integrates all components.
|
||||
"""
|
||||
|
||||
import datetime
|
||||
from typing import Optional
|
||||
|
||||
from .lexer import Lexer
|
||||
from .parser import Parser
|
||||
from .semantic import SemanticAnalyzer
|
||||
from .ptime_ast import TimeExpressionNode
|
||||
from .err import TokenUnhandledException
|
||||
|
||||
|
||||
class TimeExpression:
|
||||
"""Main class for parsing time expressions."""
|
||||
|
||||
def __init__(self, text: str, now: Optional[datetime.datetime] = None):
|
||||
self.text = text.strip()
|
||||
self.now = now or datetime.datetime.now()
|
||||
|
||||
if not self.text:
|
||||
raise TokenUnhandledException("Empty input")
|
||||
|
||||
# Initialize components
|
||||
self.lexer = Lexer(self.text, self.now)
|
||||
self.parser = Parser(self.text, self.now)
|
||||
self.semantic_analyzer = SemanticAnalyzer(self.now)
|
||||
|
||||
# Parse the expression
|
||||
self.ast = self._parse()
|
||||
|
||||
def _parse(self) -> TimeExpressionNode:
|
||||
"""Parse the time expression and return the AST."""
|
||||
try:
|
||||
return self.parser.parse()
|
||||
except Exception as e:
|
||||
raise TokenUnhandledException(f"Failed to parse '{self.text}': {str(e)}")
|
||||
|
||||
def evaluate(self) -> datetime.datetime:
|
||||
"""Evaluate the time expression and return the datetime."""
|
||||
try:
|
||||
return self.semantic_analyzer.evaluate(self.ast)
|
||||
except Exception as e:
|
||||
raise TokenUnhandledException(f"Failed to evaluate '{self.text}': {str(e)}")
|
||||
|
||||
@classmethod
|
||||
def parse(cls, text: str, now: Optional[datetime.datetime] = None) -> datetime.datetime:
|
||||
"""
|
||||
Parse a time expression and return a datetime object.
|
||||
|
||||
Args:
|
||||
text: The time expression to parse
|
||||
now: The reference time (defaults to current time)
|
||||
|
||||
Returns:
|
||||
A datetime object representing the parsed time
|
||||
|
||||
Raises:
|
||||
TokenUnhandledException: If the input cannot be parsed
|
||||
"""
|
||||
expression = cls(text, now)
|
||||
return expression.evaluate()
|
||||
225
konabot/common/ptimeparse/lexer.py
Normal file
225
konabot/common/ptimeparse/lexer.py
Normal file
@ -0,0 +1,225 @@
|
||||
"""
|
||||
Lexical analyzer for time expressions.
|
||||
"""
|
||||
|
||||
import re
|
||||
from typing import Iterator, Optional
|
||||
import datetime
|
||||
|
||||
from .ptime_token import Token, TokenType
|
||||
from .chinese_number import ChineseNumberParser
|
||||
|
||||
|
||||
class Lexer:
|
||||
"""Lexical analyzer for time expressions."""
|
||||
|
||||
def __init__(self, text: str, now: Optional[datetime.datetime] = None):
|
||||
self.text = text
|
||||
self.pos = 0
|
||||
self.current_char = self.text[self.pos] if self.text else None
|
||||
self.now = now or datetime.datetime.now()
|
||||
self.chinese_parser = ChineseNumberParser()
|
||||
|
||||
# Define token patterns
|
||||
self.token_patterns = [
|
||||
# Whitespace
|
||||
(r'^\s+', TokenType.WHITESPACE),
|
||||
|
||||
# Time separators
|
||||
(r'^:', TokenType.TIME_SEPARATOR),
|
||||
(r'^点', TokenType.TIME_SEPARATOR),
|
||||
(r'^时', TokenType.TIME_SEPARATOR),
|
||||
(r'^分', TokenType.TIME_SEPARATOR),
|
||||
(r'^秒', TokenType.TIME_SEPARATOR),
|
||||
|
||||
# Special time markers
|
||||
(r'^半', TokenType.HALF),
|
||||
(r'^一刻', TokenType.QUARTER),
|
||||
(r'^整', TokenType.ZHENG),
|
||||
(r'^钟', TokenType.ZHONG),
|
||||
|
||||
# Period indicators (must come before relative time patterns to avoid conflicts)
|
||||
(r'^(上午|早晨|早上|清晨|早(?!\d))', TokenType.PERIOD_AM),
|
||||
(r'^(中午|下午|晚上|晚(?!\d)|凌晨|午夜)', TokenType.PERIOD_PM),
|
||||
|
||||
# Week scope (more specific patterns first)
|
||||
(r'^本周', TokenType.WEEK_SCOPE_CURRENT),
|
||||
(r'^上周', TokenType.WEEK_SCOPE_LAST),
|
||||
(r'^下周', TokenType.WEEK_SCOPE_NEXT),
|
||||
|
||||
# Relative directions
|
||||
(r'^(后|以后|之后)', TokenType.RELATIVE_DIRECTION_FORWARD),
|
||||
(r'^(前|以前|之前)', TokenType.RELATIVE_DIRECTION_BACKWARD),
|
||||
|
||||
# Extended relative time
|
||||
(r'^明年', TokenType.RELATIVE_NEXT),
|
||||
(r'^去年', TokenType.RELATIVE_LAST),
|
||||
(r'^今年', TokenType.RELATIVE_THIS),
|
||||
(r'^下(?![午年月周])', TokenType.RELATIVE_NEXT),
|
||||
(r'^(上|去)(?![午年月周])', TokenType.RELATIVE_LAST),
|
||||
(r'^这', TokenType.RELATIVE_THIS),
|
||||
(r'^本(?![周月年])', TokenType.RELATIVE_THIS), # Match "本" but not "本周", "本月", "本年"
|
||||
|
||||
# Week scope (fallback for standalone terms)
|
||||
(r'^本', TokenType.WEEK_SCOPE_CURRENT),
|
||||
(r'^上', TokenType.WEEK_SCOPE_LAST),
|
||||
(r'^下(?![午年月周])', TokenType.WEEK_SCOPE_NEXT),
|
||||
|
||||
# Week days (order matters - longer patterns first)
|
||||
(r'^周一', TokenType.WEEKDAY_MONDAY),
|
||||
(r'^周二', TokenType.WEEKDAY_TUESDAY),
|
||||
(r'^周三', TokenType.WEEKDAY_WEDNESDAY),
|
||||
(r'^周四', TokenType.WEEKDAY_THURSDAY),
|
||||
(r'^周五', TokenType.WEEKDAY_FRIDAY),
|
||||
(r'^周六', TokenType.WEEKDAY_SATURDAY),
|
||||
(r'^周日', TokenType.WEEKDAY_SUNDAY),
|
||||
# Single character weekdays should be matched after numbers
|
||||
# (r'^一', TokenType.WEEKDAY_MONDAY),
|
||||
# (r'^二', TokenType.WEEKDAY_TUESDAY),
|
||||
# (r'^三', TokenType.WEEKDAY_WEDNESDAY),
|
||||
# (r'^四', TokenType.WEEKDAY_THURSDAY),
|
||||
# (r'^五', TokenType.WEEKDAY_FRIDAY),
|
||||
# (r'^六', TokenType.WEEKDAY_SATURDAY),
|
||||
# (r'^日', TokenType.WEEKDAY_SUNDAY),
|
||||
|
||||
# Student-friendly time expressions
|
||||
(r'^早(?=\d)', TokenType.EARLY_MORNING),
|
||||
(r'^晚(?=\d)', TokenType.LATE_NIGHT),
|
||||
|
||||
# Relative today variants
|
||||
(r'^今晚上', TokenType.RELATIVE_TODAY),
|
||||
(r'^今晚', TokenType.RELATIVE_TODAY),
|
||||
(r'^今早', TokenType.RELATIVE_TODAY),
|
||||
(r'^今天早上', TokenType.RELATIVE_TODAY),
|
||||
(r'^今天早晨', TokenType.RELATIVE_TODAY),
|
||||
(r'^今天上午', TokenType.RELATIVE_TODAY),
|
||||
(r'^今天下午', TokenType.RELATIVE_TODAY),
|
||||
(r'^今天晚上', TokenType.RELATIVE_TODAY),
|
||||
(r'^今天', TokenType.RELATIVE_TODAY),
|
||||
|
||||
# Relative days
|
||||
(r'^明天', TokenType.RELATIVE_TOMORROW),
|
||||
(r'^后天', TokenType.RELATIVE_DAY_AFTER_TOMORROW),
|
||||
(r'^大后天', TokenType.RELATIVE_THREE_DAYS_AFTER_TOMORROW),
|
||||
(r'^昨天', TokenType.RELATIVE_YESTERDAY),
|
||||
(r'^前天', TokenType.RELATIVE_DAY_BEFORE_YESTERDAY),
|
||||
(r'^大前天', TokenType.RELATIVE_THREE_DAYS_BEFORE_YESTERDAY),
|
||||
|
||||
# Digits
|
||||
(r'^\d+', TokenType.INTEGER),
|
||||
|
||||
# Time units (must come after date separators to avoid conflicts)
|
||||
(r'^年(?![月日号])', TokenType.YEAR),
|
||||
(r'^月(?![日号])', TokenType.MONTH),
|
||||
(r'^[日号](?![月年])', TokenType.DAY),
|
||||
(r'^天', TokenType.DAY),
|
||||
(r'^周', TokenType.WEEK),
|
||||
(r'^小时', TokenType.HOUR),
|
||||
(r'^分钟', TokenType.MINUTE),
|
||||
(r'^秒', TokenType.SECOND),
|
||||
|
||||
# Date separators (fallback patterns)
|
||||
(r'^年', TokenType.DATE_SEPARATOR),
|
||||
(r'^月', TokenType.DATE_SEPARATOR),
|
||||
(r'^[日号]', TokenType.DATE_SEPARATOR),
|
||||
(r'^[-/]', TokenType.DATE_SEPARATOR),
|
||||
]
|
||||
|
||||
def advance(self):
|
||||
"""Advance the position pointer and set the current character."""
|
||||
self.pos += 1
|
||||
if self.pos >= len(self.text):
|
||||
self.current_char = None
|
||||
else:
|
||||
self.current_char = self.text[self.pos]
|
||||
|
||||
def skip_whitespace(self):
|
||||
"""Skip whitespace characters."""
|
||||
while self.current_char is not None and self.current_char.isspace():
|
||||
self.advance()
|
||||
|
||||
def integer(self) -> int:
|
||||
"""Parse an integer from the input."""
|
||||
result = ''
|
||||
while self.current_char is not None and self.current_char.isdigit():
|
||||
result += self.current_char
|
||||
self.advance()
|
||||
return int(result)
|
||||
|
||||
def chinese_number(self) -> int:
|
||||
"""Parse a Chinese number from the input."""
|
||||
# Find the longest prefix that can be parsed as a Chinese number
|
||||
for i in range(len(self.text) - self.pos, 0, -1):
|
||||
prefix = self.text[self.pos:self.pos + i]
|
||||
try:
|
||||
# Use digest to get both the remaining text and the parsed value
|
||||
remaining, value = self.chinese_parser.digest(prefix)
|
||||
# Check if we actually consumed part of the prefix
|
||||
consumed_length = len(prefix) - len(remaining)
|
||||
if consumed_length > 0:
|
||||
# Advance position by the length of the consumed text
|
||||
for _ in range(consumed_length):
|
||||
self.advance()
|
||||
return value
|
||||
except ValueError:
|
||||
continue
|
||||
# If no Chinese number found, just return 0
|
||||
return 0
|
||||
|
||||
def get_next_token(self) -> Token:
|
||||
"""Lexical analyzer that breaks the sentence into tokens."""
|
||||
while self.current_char is not None:
|
||||
# Skip whitespace
|
||||
if self.current_char.isspace():
|
||||
self.skip_whitespace()
|
||||
continue
|
||||
|
||||
# Try to match each pattern
|
||||
text_remaining = self.text[self.pos:]
|
||||
for pattern, token_type in self.token_patterns:
|
||||
match = re.match(pattern, text_remaining)
|
||||
if match:
|
||||
value = match.group(0)
|
||||
position = self.pos
|
||||
|
||||
# Advance position
|
||||
for _ in range(len(value)):
|
||||
self.advance()
|
||||
|
||||
# Special handling for some tokens
|
||||
if token_type == TokenType.INTEGER:
|
||||
value = int(value)
|
||||
elif token_type == TokenType.RELATIVE_TODAY and value in [
|
||||
"今早上", "今天早上", "今天早晨", "今天上午"
|
||||
]:
|
||||
token_type = TokenType.PERIOD_AM
|
||||
elif token_type == TokenType.RELATIVE_TODAY and value in [
|
||||
"今晚上", "今天下午", "今天晚上"
|
||||
]:
|
||||
token_type = TokenType.PERIOD_PM
|
||||
|
||||
return Token(token_type, value, position)
|
||||
|
||||
# Try to parse Chinese numbers
|
||||
chinese_start_pos = self.pos
|
||||
try:
|
||||
chinese_value = self.chinese_number()
|
||||
if chinese_value > 0:
|
||||
# We successfully parsed a Chinese number
|
||||
return Token(TokenType.CHINESE_NUMBER, chinese_value, chinese_start_pos)
|
||||
except ValueError:
|
||||
pass
|
||||
|
||||
# If no pattern matches, skip the character and continue
|
||||
self.advance()
|
||||
|
||||
# End of file
|
||||
return Token(TokenType.EOF, None, self.pos)
|
||||
|
||||
def tokenize(self) -> Iterator[Token]:
|
||||
"""Generate all tokens from the input."""
|
||||
while True:
|
||||
token = self.get_next_token()
|
||||
yield token
|
||||
if token.type == TokenType.EOF:
|
||||
break
|
||||
846
konabot/common/ptimeparse/parser.py
Normal file
846
konabot/common/ptimeparse/parser.py
Normal file
@ -0,0 +1,846 @@
|
||||
"""
|
||||
Parser for time expressions that builds an Abstract Syntax Tree (AST).
|
||||
"""
|
||||
|
||||
from typing import Iterator, Optional, List
|
||||
import datetime
|
||||
|
||||
from .ptime_token import Token, TokenType
|
||||
from .ptime_ast import (
|
||||
ASTNode, NumberNode, DateNode, TimeNode,
|
||||
RelativeDateNode, RelativeTimeNode, WeekdayNode, TimeExpressionNode
|
||||
)
|
||||
from .lexer import Lexer
|
||||
|
||||
|
||||
class ParserError(Exception):
|
||||
"""Exception raised for parser errors."""
|
||||
pass
|
||||
|
||||
|
||||
class Parser:
|
||||
"""Parser for time expressions that builds an AST."""
|
||||
|
||||
def __init__(self, text: str, now: Optional[datetime.datetime] = None):
|
||||
self.lexer = Lexer(text, now)
|
||||
self.tokens: List[Token] = list(self.lexer.tokenize())
|
||||
self.pos = 0
|
||||
self.now = now or datetime.datetime.now()
|
||||
|
||||
@property
|
||||
def current_token(self) -> Token:
|
||||
"""Get the current token."""
|
||||
if self.pos < len(self.tokens):
|
||||
return self.tokens[self.pos]
|
||||
return Token(TokenType.EOF, None, len(self.tokens))
|
||||
|
||||
def eat(self, token_type: TokenType) -> Token:
|
||||
"""Consume a token of the expected type."""
|
||||
if self.current_token.type == token_type:
|
||||
token = self.current_token
|
||||
self.pos += 1
|
||||
return token
|
||||
else:
|
||||
raise ParserError(
|
||||
f"Expected token {token_type}, got {self.current_token.type} "
|
||||
f"at position {self.current_token.position}"
|
||||
)
|
||||
|
||||
def peek(self, offset: int = 1) -> Token:
|
||||
"""Look ahead at the next token without consuming it."""
|
||||
next_pos = self.pos + offset
|
||||
if next_pos < len(self.tokens):
|
||||
return self.tokens[next_pos]
|
||||
return Token(TokenType.EOF, None, len(self.tokens))
|
||||
|
||||
def parse_number(self) -> NumberNode:
|
||||
"""Parse a number (integer or Chinese number)."""
|
||||
token = self.current_token
|
||||
if token.type == TokenType.INTEGER:
|
||||
self.eat(TokenType.INTEGER)
|
||||
return NumberNode(value=token.value)
|
||||
elif token.type == TokenType.CHINESE_NUMBER:
|
||||
self.eat(TokenType.CHINESE_NUMBER)
|
||||
return NumberNode(value=token.value)
|
||||
else:
|
||||
raise ParserError(
|
||||
f"Expected number, got {token.type} at position {token.position}"
|
||||
)
|
||||
|
||||
def parse_date(self) -> DateNode:
|
||||
"""Parse a date specification."""
|
||||
year_node = None
|
||||
month_node = None
|
||||
day_node = None
|
||||
|
||||
# Try YYYY-MM-DD or YYYY/MM/DD format
|
||||
if (self.current_token.type == TokenType.INTEGER and
|
||||
self.peek().type == TokenType.DATE_SEPARATOR and
|
||||
self.peek().value in ['-', '/'] and
|
||||
self.peek(2).type == TokenType.INTEGER and
|
||||
self.peek(3).type == TokenType.DATE_SEPARATOR and
|
||||
self.peek(3).value in ['-', '/'] and
|
||||
self.peek(4).type == TokenType.INTEGER):
|
||||
|
||||
year_token = self.current_token
|
||||
self.eat(TokenType.INTEGER)
|
||||
separator1 = self.eat(TokenType.DATE_SEPARATOR).value
|
||||
|
||||
month_token = self.current_token
|
||||
self.eat(TokenType.INTEGER)
|
||||
|
||||
separator2 = self.eat(TokenType.DATE_SEPARATOR).value
|
||||
|
||||
day_token = self.current_token
|
||||
self.eat(TokenType.INTEGER)
|
||||
|
||||
year_node = NumberNode(value=year_token.value)
|
||||
month_node = NumberNode(value=month_token.value)
|
||||
day_node = NumberNode(value=day_token.value)
|
||||
|
||||
return DateNode(year=year_node, month=month_node, day=day_node)
|
||||
|
||||
# Try YYYY年MM月DD[日号] format
|
||||
if (self.current_token.type == TokenType.INTEGER and
|
||||
self.peek().type in [TokenType.DATE_SEPARATOR, TokenType.YEAR] and
|
||||
self.peek(2).type == TokenType.INTEGER and
|
||||
self.peek(3).type in [TokenType.DATE_SEPARATOR, TokenType.MONTH] and
|
||||
self.peek(4).type == TokenType.INTEGER):
|
||||
|
||||
year_token = self.current_token
|
||||
self.eat(TokenType.INTEGER)
|
||||
self.eat(self.current_token.type) # 年 (could be DATE_SEPARATOR or YEAR)
|
||||
|
||||
month_token = self.current_token
|
||||
self.eat(TokenType.INTEGER)
|
||||
self.eat(self.current_token.type) # 月 (could be DATE_SEPARATOR or MONTH)
|
||||
|
||||
day_token = self.current_token
|
||||
self.eat(TokenType.INTEGER)
|
||||
# Optional 日 or 号
|
||||
if self.current_token.type in [TokenType.DATE_SEPARATOR, TokenType.DAY]:
|
||||
self.eat(self.current_token.type)
|
||||
|
||||
year_node = NumberNode(value=year_token.value)
|
||||
month_node = NumberNode(value=month_token.value)
|
||||
day_node = NumberNode(value=day_token.value)
|
||||
|
||||
return DateNode(year=year_node, month=month_node, day=day_node)
|
||||
|
||||
# Try MM月DD[日号] format (without year)
|
||||
if (self.current_token.type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER] and
|
||||
self.peek().type in [TokenType.DATE_SEPARATOR, TokenType.MONTH] and
|
||||
self.peek().value == '月' and
|
||||
self.peek(2).type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER]):
|
||||
|
||||
month_token = self.current_token
|
||||
self.eat(month_token.type)
|
||||
self.eat(self.current_token.type) # 月 (could be DATE_SEPARATOR or MONTH)
|
||||
|
||||
day_token = self.current_token
|
||||
self.eat(day_token.type)
|
||||
# Optional 日 or 号
|
||||
if self.current_token.type in [TokenType.DATE_SEPARATOR, TokenType.DAY]:
|
||||
self.eat(self.current_token.type)
|
||||
|
||||
month_node = NumberNode(value=month_token.value)
|
||||
day_node = NumberNode(value=day_token.value)
|
||||
|
||||
return DateNode(year=None, month=month_node, day=day_node)
|
||||
|
||||
# Try Chinese MM月DD[日号] format
|
||||
if (self.current_token.type == TokenType.CHINESE_NUMBER and
|
||||
self.peek().type == TokenType.DATE_SEPARATOR and
|
||||
self.peek().value == '月' and
|
||||
self.peek(2).type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER]):
|
||||
|
||||
month_token = self.current_token
|
||||
self.eat(TokenType.CHINESE_NUMBER)
|
||||
self.eat(TokenType.DATE_SEPARATOR) # 月
|
||||
|
||||
day_token = self.current_token
|
||||
self.eat(day_token.type)
|
||||
# Optional 日 or 号
|
||||
if self.current_token.type == TokenType.DATE_SEPARATOR:
|
||||
self.eat(TokenType.DATE_SEPARATOR)
|
||||
|
||||
month_node = NumberNode(value=month_token.value)
|
||||
day_node = NumberNode(value=day_token.value)
|
||||
|
||||
return DateNode(year=None, month=month_node, day=day_node)
|
||||
|
||||
raise ParserError(
|
||||
f"Unable to parse date at position {self.current_token.position}"
|
||||
)
|
||||
|
||||
def parse_time(self) -> TimeNode:
|
||||
"""Parse a time specification."""
|
||||
hour_node = None
|
||||
minute_node = None
|
||||
second_node = None
|
||||
is_24hour = False
|
||||
period = None
|
||||
|
||||
# Try HH:MM format
|
||||
if (self.current_token.type == TokenType.INTEGER and
|
||||
self.peek().type == TokenType.TIME_SEPARATOR and
|
||||
self.peek().value == ':'):
|
||||
|
||||
hour_token = self.current_token
|
||||
self.eat(TokenType.INTEGER)
|
||||
self.eat(TokenType.TIME_SEPARATOR) # :
|
||||
|
||||
minute_token = self.current_token
|
||||
self.eat(TokenType.INTEGER)
|
||||
|
||||
hour_node = NumberNode(value=hour_token.value)
|
||||
minute_node = NumberNode(value=minute_token.value)
|
||||
is_24hour = True # HH:MM is always interpreted as 24-hour
|
||||
|
||||
# Optional :SS
|
||||
if (self.current_token.type == TokenType.TIME_SEPARATOR and
|
||||
self.peek().type == TokenType.INTEGER):
|
||||
|
||||
self.eat(TokenType.TIME_SEPARATOR) # :
|
||||
second_token = self.current_token
|
||||
self.eat(TokenType.INTEGER)
|
||||
second_node = NumberNode(value=second_token.value)
|
||||
|
||||
return TimeNode(
|
||||
hour=hour_node,
|
||||
minute=minute_node,
|
||||
second=second_node,
|
||||
is_24hour=is_24hour,
|
||||
period=period
|
||||
)
|
||||
|
||||
# Try Chinese time format (X点X分)
|
||||
# First check for period indicators
|
||||
period = None
|
||||
if self.current_token.type in [TokenType.PERIOD_AM, TokenType.PERIOD_PM]:
|
||||
if self.current_token.type == TokenType.PERIOD_AM:
|
||||
period = "AM"
|
||||
else:
|
||||
period = "PM"
|
||||
self.eat(self.current_token.type)
|
||||
|
||||
if self.current_token.type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER, TokenType.EARLY_MORNING, TokenType.LATE_NIGHT]:
|
||||
if self.current_token.type == TokenType.EARLY_MORNING:
|
||||
self.eat(TokenType.EARLY_MORNING)
|
||||
is_24hour = True
|
||||
period = "AM"
|
||||
|
||||
# Expect a number next
|
||||
if self.current_token.type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER]:
|
||||
hour_token = self.current_token
|
||||
self.eat(hour_token.type)
|
||||
hour_node = NumberNode(value=hour_token.value)
|
||||
|
||||
# "早八" should be interpreted as 08:00
|
||||
# If hour is greater than 12, treat as 24-hour
|
||||
if hour_node.value > 12:
|
||||
is_24hour = True
|
||||
period = None
|
||||
else:
|
||||
raise ParserError(
|
||||
f"Expected number after '早', got {self.current_token.type} "
|
||||
f"at position {self.current_token.position}"
|
||||
)
|
||||
elif self.current_token.type == TokenType.LATE_NIGHT:
|
||||
self.eat(TokenType.LATE_NIGHT)
|
||||
is_24hour = True
|
||||
period = "PM"
|
||||
|
||||
# Expect a number next
|
||||
if self.current_token.type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER]:
|
||||
hour_token = self.current_token
|
||||
self.eat(hour_token.type)
|
||||
hour_node = NumberNode(value=hour_token.value)
|
||||
|
||||
# "晚十" should be interpreted as 22:00
|
||||
# Adjust hour to 24-hour format
|
||||
if hour_node.value <= 12:
|
||||
hour_node.value += 12
|
||||
is_24hour = True
|
||||
period = None
|
||||
else:
|
||||
raise ParserError(
|
||||
f"Expected number after '晚', got {self.current_token.type} "
|
||||
f"at position {self.current_token.position}"
|
||||
)
|
||||
else:
|
||||
# Regular time parsing
|
||||
hour_token = self.current_token
|
||||
self.eat(hour_token.type)
|
||||
|
||||
# Check for 点 or 时
|
||||
if self.current_token.type == TokenType.TIME_SEPARATOR:
|
||||
separator = self.current_token.value
|
||||
self.eat(TokenType.TIME_SEPARATOR)
|
||||
|
||||
if separator == '点':
|
||||
is_24hour = False
|
||||
elif separator == '时':
|
||||
is_24hour = True
|
||||
|
||||
hour_node = NumberNode(value=hour_token.value)
|
||||
|
||||
# Optional minutes
|
||||
if self.current_token.type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER]:
|
||||
minute_token = self.current_token
|
||||
self.eat(minute_token.type)
|
||||
|
||||
# Optional 分
|
||||
if self.current_token.type == TokenType.TIME_SEPARATOR and \
|
||||
self.current_token.value == '分':
|
||||
self.eat(TokenType.TIME_SEPARATOR)
|
||||
|
||||
minute_node = NumberNode(value=minute_token.value)
|
||||
|
||||
# Handle special markers
|
||||
if self.current_token.type == TokenType.HALF:
|
||||
self.eat(TokenType.HALF)
|
||||
minute_node = NumberNode(value=30)
|
||||
elif self.current_token.type == TokenType.QUARTER:
|
||||
self.eat(TokenType.QUARTER)
|
||||
minute_node = NumberNode(value=15)
|
||||
elif self.current_token.type == TokenType.ZHENG:
|
||||
self.eat(TokenType.ZHENG)
|
||||
if minute_node is None:
|
||||
minute_node = NumberNode(value=0)
|
||||
|
||||
# Optional 钟
|
||||
if self.current_token.type == TokenType.ZHONG:
|
||||
self.eat(TokenType.ZHONG)
|
||||
else:
|
||||
# If no separator, treat as hour-only time (like "三点")
|
||||
hour_node = NumberNode(value=hour_token.value)
|
||||
is_24hour = False
|
||||
|
||||
return TimeNode(
|
||||
hour=hour_node,
|
||||
minute=minute_node,
|
||||
second=second_node,
|
||||
is_24hour=is_24hour,
|
||||
period=period
|
||||
)
|
||||
|
||||
raise ParserError(
|
||||
f"Unable to parse time at position {self.current_token.position}"
|
||||
)
|
||||
|
||||
def parse_relative_date(self) -> RelativeDateNode:
|
||||
"""Parse a relative date specification."""
|
||||
years = 0
|
||||
months = 0
|
||||
weeks = 0
|
||||
days = 0
|
||||
|
||||
# Handle today variants
|
||||
if self.current_token.type == TokenType.RELATIVE_TODAY:
|
||||
self.eat(TokenType.RELATIVE_TODAY)
|
||||
days = 0
|
||||
elif self.current_token.type == TokenType.RELATIVE_TOMORROW:
|
||||
self.eat(TokenType.RELATIVE_TOMORROW)
|
||||
days = 1
|
||||
elif self.current_token.type == TokenType.RELATIVE_DAY_AFTER_TOMORROW:
|
||||
self.eat(TokenType.RELATIVE_DAY_AFTER_TOMORROW)
|
||||
days = 2
|
||||
elif self.current_token.type == TokenType.RELATIVE_THREE_DAYS_AFTER_TOMORROW:
|
||||
self.eat(TokenType.RELATIVE_THREE_DAYS_AFTER_TOMORROW)
|
||||
days = 3
|
||||
elif self.current_token.type == TokenType.RELATIVE_YESTERDAY:
|
||||
self.eat(TokenType.RELATIVE_YESTERDAY)
|
||||
days = -1
|
||||
elif self.current_token.type == TokenType.RELATIVE_DAY_BEFORE_YESTERDAY:
|
||||
self.eat(TokenType.RELATIVE_DAY_BEFORE_YESTERDAY)
|
||||
days = -2
|
||||
elif self.current_token.type == TokenType.RELATIVE_THREE_DAYS_BEFORE_YESTERDAY:
|
||||
self.eat(TokenType.RELATIVE_THREE_DAYS_BEFORE_YESTERDAY)
|
||||
days = -3
|
||||
else:
|
||||
# Check if this looks like an absolute date pattern before processing
|
||||
# Look ahead to see if this matches absolute date patterns
|
||||
is_likely_absolute_date = False
|
||||
|
||||
# Check for MM月DD[日号] patterns (like "6月20日")
|
||||
if (self.pos + 2 < len(self.tokens) and
|
||||
self.tokens[self.pos].type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER] and
|
||||
self.tokens[self.pos + 1].type in [TokenType.DATE_SEPARATOR, TokenType.MONTH] and
|
||||
self.tokens[self.pos + 1].value == '月' and
|
||||
self.tokens[self.pos + 2].type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER]):
|
||||
is_likely_absolute_date = True
|
||||
|
||||
if is_likely_absolute_date:
|
||||
# This looks like an absolute date, skip relative date parsing
|
||||
raise ParserError("Looks like absolute date format")
|
||||
|
||||
# Try to parse extended relative time expressions
|
||||
# Handle patterns like "明年", "去年", "下个月", "上个月", etc.
|
||||
original_pos = self.pos
|
||||
try:
|
||||
# Check for "今年", "明年", "去年"
|
||||
if self.current_token.type == TokenType.RELATIVE_THIS and self.peek().type == TokenType.YEAR:
|
||||
self.eat(TokenType.RELATIVE_THIS)
|
||||
self.eat(TokenType.YEAR)
|
||||
years = 0 # Current year
|
||||
elif self.current_token.type == TokenType.RELATIVE_NEXT and self.peek().type == TokenType.YEAR:
|
||||
self.eat(TokenType.RELATIVE_NEXT)
|
||||
self.eat(TokenType.YEAR)
|
||||
years = 1 # Next year
|
||||
elif self.current_token.type == TokenType.RELATIVE_LAST and self.peek().type == TokenType.YEAR:
|
||||
self.eat(TokenType.RELATIVE_LAST)
|
||||
self.eat(TokenType.YEAR)
|
||||
years = -1 # Last year
|
||||
elif self.current_token.type == TokenType.RELATIVE_NEXT and self.current_token.value == "明年":
|
||||
self.eat(TokenType.RELATIVE_NEXT)
|
||||
years = 1 # Next year
|
||||
# Check if there's a month after "明年"
|
||||
if (self.current_token.type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER] and
|
||||
self.peek().type == TokenType.MONTH):
|
||||
# Parse the month
|
||||
month_node = self.parse_number()
|
||||
self.eat(TokenType.MONTH) # Eat the "月" token
|
||||
# Store the month in the months field as a special marker
|
||||
# We'll handle this in semantic analysis
|
||||
months = month_node.value - 100 # Use negative offset to indicate absolute month
|
||||
elif self.current_token.type == TokenType.RELATIVE_LAST and self.current_token.value == "去年":
|
||||
self.eat(TokenType.RELATIVE_LAST)
|
||||
years = -1 # Last year
|
||||
elif self.current_token.type == TokenType.RELATIVE_THIS and self.current_token.value == "今年":
|
||||
self.eat(TokenType.RELATIVE_THIS)
|
||||
years = 0 # Current year
|
||||
|
||||
# Check for "这个月", "下个月", "上个月"
|
||||
elif self.current_token.type == TokenType.RELATIVE_THIS and self.peek().type == TokenType.MONTH:
|
||||
self.eat(TokenType.RELATIVE_THIS)
|
||||
self.eat(TokenType.MONTH)
|
||||
months = 0 # Current month
|
||||
elif self.current_token.type == TokenType.RELATIVE_NEXT and self.peek().type == TokenType.MONTH:
|
||||
self.eat(TokenType.RELATIVE_NEXT)
|
||||
self.eat(TokenType.MONTH)
|
||||
months = 1 # Next month
|
||||
|
||||
# Handle patterns like "下个月五号"
|
||||
if (self.current_token.type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER] and
|
||||
self.peek().type == TokenType.DAY):
|
||||
# Parse the day
|
||||
day_node = self.parse_number()
|
||||
self.eat(TokenType.DAY) # Eat the "号" token
|
||||
# Instead of adding days to the current date, we should set a specific day in the target month
|
||||
# We'll handle this in semantic analysis by setting a flag or special value
|
||||
days = 0 # Reset days - we'll handle the day differently
|
||||
# Use a special marker to indicate we want a specific day in the target month
|
||||
# For now, we'll just store the target day in the weeks field as a temporary solution
|
||||
weeks = day_node.value # This is a hack - we'll fix this in semantic analysis
|
||||
elif self.current_token.type == TokenType.RELATIVE_LAST and self.peek().type == TokenType.MONTH:
|
||||
self.eat(TokenType.RELATIVE_LAST)
|
||||
self.eat(TokenType.MONTH)
|
||||
months = -1 # Last month
|
||||
|
||||
# Check for "下周", "上周"
|
||||
elif self.current_token.type == TokenType.RELATIVE_NEXT and self.peek().type == TokenType.WEEK:
|
||||
self.eat(TokenType.RELATIVE_NEXT)
|
||||
self.eat(TokenType.WEEK)
|
||||
weeks = 1 # Next week
|
||||
elif self.current_token.type == TokenType.RELATIVE_LAST and self.peek().type == TokenType.WEEK:
|
||||
self.eat(TokenType.RELATIVE_LAST)
|
||||
self.eat(TokenType.WEEK)
|
||||
weeks = -1 # Last week
|
||||
|
||||
# Handle more complex patterns like "X年后", "X个月后", etc.
|
||||
elif self.current_token.type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER]:
|
||||
# Check if this is likely an absolute date format (e.g., "2025年11月21日")
|
||||
# If the next token after the number is a date separator or date unit,
|
||||
# and the number looks like a year (4 digits) or the pattern continues,
|
||||
# it might be an absolute date. In that case, skip relative date parsing.
|
||||
|
||||
# Look ahead to see if this matches absolute date patterns
|
||||
lookahead_pos = self.pos
|
||||
is_likely_absolute_date = False
|
||||
|
||||
# Check for YYYY-MM-DD or YYYY/MM/DD patterns
|
||||
if (lookahead_pos + 4 < len(self.tokens) and
|
||||
self.tokens[lookahead_pos].type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER] and
|
||||
self.tokens[lookahead_pos + 1].type in [TokenType.DATE_SEPARATOR, TokenType.YEAR] and
|
||||
self.tokens[lookahead_pos + 1].value in ['-', '/', '年'] and
|
||||
self.tokens[lookahead_pos + 2].type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER] and
|
||||
self.tokens[lookahead_pos + 3].type in [TokenType.DATE_SEPARATOR, TokenType.MONTH] and
|
||||
self.tokens[lookahead_pos + 3].value in ['-', '/', '月']):
|
||||
is_likely_absolute_date = True
|
||||
|
||||
# Check for YYYY年MM月DD patterns
|
||||
if (lookahead_pos + 4 < len(self.tokens) and
|
||||
self.tokens[lookahead_pos].type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER] and
|
||||
self.tokens[lookahead_pos + 1].type in [TokenType.DATE_SEPARATOR, TokenType.YEAR] and
|
||||
self.tokens[lookahead_pos + 1].value == '年' and
|
||||
self.tokens[lookahead_pos + 2].type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER] and
|
||||
self.tokens[lookahead_pos + 3].type in [TokenType.DATE_SEPARATOR, TokenType.MONTH] and
|
||||
self.tokens[lookahead_pos + 3].value == '月'):
|
||||
is_likely_absolute_date = True
|
||||
|
||||
# Check for MM月DD[日号] patterns (like "6月20日")
|
||||
if (self.pos + 2 < len(self.tokens) and
|
||||
self.tokens[self.pos].type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER] and
|
||||
self.tokens[self.pos + 1].type in [TokenType.DATE_SEPARATOR, TokenType.MONTH] and
|
||||
self.tokens[self.pos + 1].value == '月' and
|
||||
self.tokens[self.pos + 2].type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER]):
|
||||
is_likely_absolute_date = True
|
||||
|
||||
if is_likely_absolute_date:
|
||||
# This looks like an absolute date, skip relative date parsing
|
||||
raise ParserError("Looks like absolute date format")
|
||||
|
||||
print(f"DEBUG: Parsing complex relative date pattern")
|
||||
# Parse the number
|
||||
number_node = self.parse_number()
|
||||
number_value = number_node.value
|
||||
print(f"DEBUG: Parsed number: {number_value}")
|
||||
|
||||
# Check the unit
|
||||
if self.current_token.type == TokenType.YEAR:
|
||||
self.eat(TokenType.YEAR)
|
||||
years = number_value
|
||||
print(f"DEBUG: Set years to {years}")
|
||||
elif self.current_token.type == TokenType.MONTH:
|
||||
self.eat(TokenType.MONTH)
|
||||
months = number_value
|
||||
print(f"DEBUG: Set months to {months}")
|
||||
elif self.current_token.type == TokenType.WEEK:
|
||||
self.eat(TokenType.WEEK)
|
||||
weeks = number_value
|
||||
print(f"DEBUG: Set weeks to {weeks}")
|
||||
elif self.current_token.type == TokenType.DAY:
|
||||
self.eat(TokenType.DAY)
|
||||
days = number_value
|
||||
print(f"DEBUG: Set days to {days}")
|
||||
else:
|
||||
print(f"DEBUG: Unexpected token type: {self.current_token.type}")
|
||||
raise ParserError(
|
||||
f"Expected time unit, got {self.current_token.type} "
|
||||
f"at position {self.current_token.position}"
|
||||
)
|
||||
|
||||
# Check direction (前/后)
|
||||
if self.current_token.type == TokenType.RELATIVE_DIRECTION_FORWARD:
|
||||
self.eat(TokenType.RELATIVE_DIRECTION_FORWARD)
|
||||
print(f"DEBUG: Forward direction, values are already positive")
|
||||
# Values are already positive
|
||||
elif self.current_token.type == TokenType.RELATIVE_DIRECTION_BACKWARD:
|
||||
self.eat(TokenType.RELATIVE_DIRECTION_BACKWARD)
|
||||
print(f"DEBUG: Backward direction, negating values")
|
||||
years = -years
|
||||
months = -months
|
||||
weeks = -weeks
|
||||
days = -days
|
||||
|
||||
except ParserError:
|
||||
# Reset position if parsing failed
|
||||
self.pos = original_pos
|
||||
raise ParserError(
|
||||
f"Expected relative date, got {self.current_token.type} "
|
||||
f"at position {self.current_token.position}"
|
||||
)
|
||||
|
||||
return RelativeDateNode(years=years, months=months, weeks=weeks, days=days)
|
||||
|
||||
def parse_weekday(self) -> WeekdayNode:
|
||||
"""Parse a weekday specification."""
|
||||
# Parse week scope (本, 上, 下)
|
||||
scope = "current"
|
||||
if self.current_token.type == TokenType.WEEK_SCOPE_CURRENT:
|
||||
self.eat(TokenType.WEEK_SCOPE_CURRENT)
|
||||
scope = "current"
|
||||
elif self.current_token.type == TokenType.WEEK_SCOPE_LAST:
|
||||
self.eat(TokenType.WEEK_SCOPE_LAST)
|
||||
scope = "last"
|
||||
elif self.current_token.type == TokenType.WEEK_SCOPE_NEXT:
|
||||
self.eat(TokenType.WEEK_SCOPE_NEXT)
|
||||
scope = "next"
|
||||
|
||||
# Parse weekday
|
||||
weekday_map = {
|
||||
TokenType.WEEKDAY_MONDAY: 0,
|
||||
TokenType.WEEKDAY_TUESDAY: 1,
|
||||
TokenType.WEEKDAY_WEDNESDAY: 2,
|
||||
TokenType.WEEKDAY_THURSDAY: 3,
|
||||
TokenType.WEEKDAY_FRIDAY: 4,
|
||||
TokenType.WEEKDAY_SATURDAY: 5,
|
||||
TokenType.WEEKDAY_SUNDAY: 6,
|
||||
# Handle Chinese numbers (1=Monday, 2=Tuesday, etc.)
|
||||
TokenType.CHINESE_NUMBER: lambda x: x - 1 if 1 <= x <= 7 else None,
|
||||
}
|
||||
|
||||
if self.current_token.type in weekday_map:
|
||||
if self.current_token.type == TokenType.CHINESE_NUMBER:
|
||||
# Handle numeric weekday (1=Monday, 2=Tuesday, etc.)
|
||||
weekday_num = self.current_token.value
|
||||
if 1 <= weekday_num <= 7:
|
||||
weekday = weekday_num - 1 # Convert to 0-based index
|
||||
self.eat(TokenType.CHINESE_NUMBER)
|
||||
return WeekdayNode(weekday=weekday, scope=scope)
|
||||
else:
|
||||
raise ParserError(
|
||||
f"Invalid weekday number: {weekday_num} "
|
||||
f"at position {self.current_token.position}"
|
||||
)
|
||||
else:
|
||||
weekday = weekday_map[self.current_token.type]
|
||||
self.eat(self.current_token.type)
|
||||
return WeekdayNode(weekday=weekday, scope=scope)
|
||||
|
||||
raise ParserError(
|
||||
f"Expected weekday, got {self.current_token.type} "
|
||||
f"at position {self.current_token.position}"
|
||||
)
|
||||
|
||||
def parse_relative_time(self) -> RelativeTimeNode:
|
||||
"""Parse a relative time specification."""
|
||||
hours = 0.0
|
||||
minutes = 0.0
|
||||
seconds = 0.0
|
||||
|
||||
def parse_relative_time(self) -> RelativeTimeNode:
|
||||
"""Parse a relative time specification."""
|
||||
hours = 0.0
|
||||
minutes = 0.0
|
||||
seconds = 0.0
|
||||
|
||||
# Parse sequences of relative time expressions
|
||||
while self.current_token.type in [
|
||||
TokenType.INTEGER, TokenType.CHINESE_NUMBER,
|
||||
TokenType.HALF, TokenType.QUARTER
|
||||
] or (self.current_token.type == TokenType.RELATIVE_DIRECTION_FORWARD or
|
||||
self.current_token.type == TokenType.RELATIVE_DIRECTION_BACKWARD):
|
||||
|
||||
# Handle 半小时
|
||||
if (self.current_token.type == TokenType.HALF):
|
||||
self.eat(TokenType.HALF)
|
||||
# Optional 个
|
||||
if (self.current_token.type == TokenType.INTEGER and
|
||||
self.current_token.value == "个"):
|
||||
self.eat(TokenType.INTEGER)
|
||||
# Optional 小时
|
||||
if self.current_token.type == TokenType.HOUR:
|
||||
self.eat(TokenType.HOUR)
|
||||
hours += 0.5
|
||||
# Check for direction
|
||||
if self.current_token.type == TokenType.RELATIVE_DIRECTION_FORWARD:
|
||||
self.eat(TokenType.RELATIVE_DIRECTION_FORWARD)
|
||||
elif self.current_token.type == TokenType.RELATIVE_DIRECTION_BACKWARD:
|
||||
self.eat(TokenType.RELATIVE_DIRECTION_BACKWARD)
|
||||
hours = -hours
|
||||
continue
|
||||
|
||||
# Handle 一刻钟 (15 minutes)
|
||||
if self.current_token.type == TokenType.QUARTER:
|
||||
self.eat(TokenType.QUARTER)
|
||||
# Optional 钟
|
||||
if self.current_token.type == TokenType.ZHONG:
|
||||
self.eat(TokenType.ZHONG)
|
||||
minutes += 15
|
||||
# Check for direction
|
||||
if self.current_token.type == TokenType.RELATIVE_DIRECTION_FORWARD:
|
||||
self.eat(TokenType.RELATIVE_DIRECTION_FORWARD)
|
||||
elif self.current_token.type == TokenType.RELATIVE_DIRECTION_BACKWARD:
|
||||
self.eat(TokenType.RELATIVE_DIRECTION_BACKWARD)
|
||||
minutes = -minutes
|
||||
continue
|
||||
|
||||
# Parse number if we have one
|
||||
if self.current_token.type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER]:
|
||||
number_node = self.parse_number()
|
||||
number_value = number_node.value
|
||||
|
||||
# Determine unit and direction
|
||||
unit = None
|
||||
direction = 1 # Forward by default
|
||||
|
||||
# Check for unit
|
||||
if self.current_token.type == TokenType.HOUR:
|
||||
self.eat(TokenType.HOUR)
|
||||
# Optional 个
|
||||
if (self.current_token.type == TokenType.INTEGER and
|
||||
self.current_token.value == "个"):
|
||||
self.eat(TokenType.INTEGER)
|
||||
unit = "hour"
|
||||
elif self.current_token.type == TokenType.MINUTE:
|
||||
self.eat(TokenType.MINUTE)
|
||||
unit = "minute"
|
||||
elif self.current_token.type == TokenType.SECOND:
|
||||
self.eat(TokenType.SECOND)
|
||||
unit = "second"
|
||||
elif self.current_token.type == TokenType.TIME_SEPARATOR:
|
||||
# Handle "X点", "X分", "X秒" format
|
||||
sep_value = self.current_token.value
|
||||
self.eat(TokenType.TIME_SEPARATOR)
|
||||
if sep_value == "点":
|
||||
unit = "hour"
|
||||
# Optional 钟
|
||||
if self.current_token.type == TokenType.ZHONG:
|
||||
self.eat(TokenType.ZHONG)
|
||||
# If we have "X点" without a direction, this is likely an absolute time
|
||||
# Check if there's a direction after
|
||||
if not (self.current_token.type == TokenType.RELATIVE_DIRECTION_FORWARD or
|
||||
self.current_token.type == TokenType.RELATIVE_DIRECTION_BACKWARD):
|
||||
# This is probably an absolute time, not relative time
|
||||
# Push back the number and break
|
||||
break
|
||||
elif sep_value == "分":
|
||||
unit = "minute"
|
||||
# Optional 钟
|
||||
if self.current_token.type == TokenType.ZHONG:
|
||||
self.eat(TokenType.ZHONG)
|
||||
elif sep_value == "秒":
|
||||
unit = "second"
|
||||
else:
|
||||
# If no unit specified, but we have a number followed by a direction,
|
||||
# assume it's hours
|
||||
if (self.current_token.type == TokenType.RELATIVE_DIRECTION_FORWARD or
|
||||
self.current_token.type == TokenType.RELATIVE_DIRECTION_BACKWARD):
|
||||
unit = "hour"
|
||||
else:
|
||||
# If no unit and no direction, this might not be a relative time expression
|
||||
# Push the number back and break
|
||||
# We can't easily push back, so let's break
|
||||
break
|
||||
|
||||
# Check for direction (后/前)
|
||||
if self.current_token.type == TokenType.RELATIVE_DIRECTION_FORWARD:
|
||||
self.eat(TokenType.RELATIVE_DIRECTION_FORWARD)
|
||||
direction = 1
|
||||
elif self.current_token.type == TokenType.RELATIVE_DIRECTION_BACKWARD:
|
||||
self.eat(TokenType.RELATIVE_DIRECTION_BACKWARD)
|
||||
direction = -1
|
||||
|
||||
# Apply the value based on unit
|
||||
if unit == "hour":
|
||||
hours += number_value * direction
|
||||
elif unit == "minute":
|
||||
minutes += number_value * direction
|
||||
elif unit == "second":
|
||||
seconds += number_value * direction
|
||||
continue
|
||||
|
||||
# If we still haven't handled the current token, break
|
||||
break
|
||||
|
||||
return RelativeTimeNode(hours=hours, minutes=minutes, seconds=seconds)
|
||||
|
||||
def parse_time_expression(self) -> TimeExpressionNode:
|
||||
"""Parse a complete time expression."""
|
||||
date_node = None
|
||||
time_node = None
|
||||
relative_date_node = None
|
||||
relative_time_node = None
|
||||
weekday_node = None
|
||||
|
||||
# Parse different parts of the expression
|
||||
while self.current_token.type != TokenType.EOF:
|
||||
# Try to parse date first (absolute dates should take precedence)
|
||||
if self.current_token.type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER]:
|
||||
if date_node is None:
|
||||
original_pos = self.pos
|
||||
try:
|
||||
date_node = self.parse_date()
|
||||
continue
|
||||
except ParserError:
|
||||
# Reset position if parsing failed
|
||||
self.pos = original_pos
|
||||
pass
|
||||
|
||||
# Try to parse relative date
|
||||
if self.current_token.type in [
|
||||
TokenType.RELATIVE_TODAY, TokenType.RELATIVE_TOMORROW,
|
||||
TokenType.RELATIVE_DAY_AFTER_TOMORROW, TokenType.RELATIVE_THREE_DAYS_AFTER_TOMORROW,
|
||||
TokenType.RELATIVE_YESTERDAY, TokenType.RELATIVE_DAY_BEFORE_YESTERDAY,
|
||||
TokenType.RELATIVE_THREE_DAYS_BEFORE_YESTERDAY,
|
||||
TokenType.INTEGER, TokenType.CHINESE_NUMBER, # For patterns like "X年后", "X个月后", etc.
|
||||
TokenType.RELATIVE_NEXT, TokenType.RELATIVE_LAST, TokenType.RELATIVE_THIS
|
||||
]:
|
||||
if relative_date_node is None:
|
||||
original_pos = self.pos
|
||||
try:
|
||||
relative_date_node = self.parse_relative_date()
|
||||
continue
|
||||
except ParserError:
|
||||
# Reset position if parsing failed
|
||||
self.pos = original_pos
|
||||
pass
|
||||
|
||||
# Try to parse relative time first (since it can have numbers)
|
||||
if self.current_token.type in [
|
||||
TokenType.INTEGER, TokenType.CHINESE_NUMBER,
|
||||
TokenType.HALF, TokenType.QUARTER,
|
||||
TokenType.RELATIVE_DIRECTION_FORWARD, TokenType.RELATIVE_DIRECTION_BACKWARD
|
||||
]:
|
||||
if relative_time_node is None:
|
||||
original_pos = self.pos
|
||||
try:
|
||||
relative_time_node = self.parse_relative_time()
|
||||
# Only continue if we actually parsed some relative time
|
||||
if relative_time_node.hours != 0 or relative_time_node.minutes != 0 or relative_time_node.seconds != 0:
|
||||
continue
|
||||
else:
|
||||
# If we didn't parse any relative time, reset position
|
||||
self.pos = original_pos
|
||||
except ParserError:
|
||||
# Reset position if parsing failed
|
||||
self.pos = original_pos
|
||||
pass
|
||||
|
||||
# Try to parse time
|
||||
if self.current_token.type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER, TokenType.TIME_SEPARATOR, TokenType.PERIOD_AM, TokenType.PERIOD_PM]:
|
||||
if time_node is None:
|
||||
original_pos = self.pos
|
||||
try:
|
||||
time_node = self.parse_time()
|
||||
continue
|
||||
except ParserError:
|
||||
# Reset position if parsing failed
|
||||
self.pos = original_pos
|
||||
pass
|
||||
|
||||
# Try to parse time
|
||||
if self.current_token.type in [TokenType.INTEGER, TokenType.CHINESE_NUMBER, TokenType.TIME_SEPARATOR, TokenType.PERIOD_AM, TokenType.PERIOD_PM]:
|
||||
if time_node is None:
|
||||
original_pos = self.pos
|
||||
try:
|
||||
time_node = self.parse_time()
|
||||
continue
|
||||
except ParserError:
|
||||
# Reset position if parsing failed
|
||||
self.pos = original_pos
|
||||
pass
|
||||
|
||||
# Try to parse weekday
|
||||
if self.current_token.type in [
|
||||
TokenType.WEEK_SCOPE_CURRENT, TokenType.WEEK_SCOPE_LAST, TokenType.WEEK_SCOPE_NEXT,
|
||||
TokenType.WEEKDAY_MONDAY, TokenType.WEEKDAY_TUESDAY, TokenType.WEEKDAY_WEDNESDAY,
|
||||
TokenType.WEEKDAY_THURSDAY, TokenType.WEEKDAY_FRIDAY, TokenType.WEEKDAY_SATURDAY,
|
||||
TokenType.WEEKDAY_SUNDAY
|
||||
]:
|
||||
if weekday_node is None:
|
||||
original_pos = self.pos
|
||||
try:
|
||||
weekday_node = self.parse_weekday()
|
||||
continue
|
||||
except ParserError:
|
||||
# Reset position if parsing failed
|
||||
self.pos = original_pos
|
||||
pass
|
||||
|
||||
# If we get here and couldn't parse anything, skip the token
|
||||
self.pos += 1
|
||||
|
||||
return TimeExpressionNode(
|
||||
date=date_node,
|
||||
time=time_node,
|
||||
relative_date=relative_date_node,
|
||||
relative_time=relative_time_node,
|
||||
weekday=weekday_node
|
||||
)
|
||||
|
||||
def parse(self) -> TimeExpressionNode:
|
||||
"""Parse the complete time expression and return the AST."""
|
||||
return self.parse_time_expression()
|
||||
71
konabot/common/ptimeparse/ptime_ast.py
Normal file
71
konabot/common/ptimeparse/ptime_ast.py
Normal file
@ -0,0 +1,71 @@
|
||||
"""
|
||||
Abstract Syntax Tree (AST) nodes for the time expression parser.
|
||||
"""
|
||||
|
||||
from abc import ABC
|
||||
from typing import Optional
|
||||
from dataclasses import dataclass
|
||||
|
||||
|
||||
@dataclass
|
||||
class ASTNode(ABC):
|
||||
"""Base class for all AST nodes."""
|
||||
pass
|
||||
|
||||
|
||||
@dataclass
|
||||
class NumberNode(ASTNode):
|
||||
"""Represents a numeric value."""
|
||||
value: int
|
||||
|
||||
|
||||
@dataclass
|
||||
class DateNode(ASTNode):
|
||||
"""Represents a date specification."""
|
||||
year: Optional[ASTNode]
|
||||
month: Optional[ASTNode]
|
||||
day: Optional[ASTNode]
|
||||
|
||||
|
||||
@dataclass
|
||||
class TimeNode(ASTNode):
|
||||
"""Represents a time specification."""
|
||||
hour: Optional[ASTNode]
|
||||
minute: Optional[ASTNode]
|
||||
second: Optional[ASTNode]
|
||||
is_24hour: bool = False
|
||||
period: Optional[str] = None # AM or PM
|
||||
|
||||
|
||||
@dataclass
|
||||
class RelativeDateNode(ASTNode):
|
||||
"""Represents a relative date specification."""
|
||||
years: int = 0
|
||||
months: int = 0
|
||||
weeks: int = 0
|
||||
days: int = 0
|
||||
|
||||
|
||||
@dataclass
|
||||
class RelativeTimeNode(ASTNode):
|
||||
"""Represents a relative time specification."""
|
||||
hours: float = 0.0
|
||||
minutes: float = 0.0
|
||||
seconds: float = 0.0
|
||||
|
||||
|
||||
@dataclass
|
||||
class WeekdayNode(ASTNode):
|
||||
"""Represents a weekday specification."""
|
||||
weekday: int # 0=Monday, 6=Sunday
|
||||
scope: str # current, last, next
|
||||
|
||||
|
||||
@dataclass
|
||||
class TimeExpressionNode(ASTNode):
|
||||
"""Represents a complete time expression."""
|
||||
date: Optional[DateNode] = None
|
||||
time: Optional[TimeNode] = None
|
||||
relative_date: Optional[RelativeDateNode] = None
|
||||
relative_time: Optional[RelativeTimeNode] = None
|
||||
weekday: Optional[WeekdayNode] = None
|
||||
95
konabot/common/ptimeparse/ptime_token.py
Normal file
95
konabot/common/ptimeparse/ptime_token.py
Normal file
@ -0,0 +1,95 @@
|
||||
"""
|
||||
Token definitions for the time parser.
|
||||
"""
|
||||
|
||||
from enum import Enum
|
||||
from typing import Union
|
||||
from dataclasses import dataclass
|
||||
|
||||
|
||||
class TokenType(Enum):
|
||||
"""Types of tokens recognized by the lexer."""
|
||||
|
||||
# Numbers
|
||||
INTEGER = "INTEGER"
|
||||
CHINESE_NUMBER = "CHINESE_NUMBER"
|
||||
|
||||
# Time units
|
||||
YEAR = "YEAR"
|
||||
MONTH = "MONTH"
|
||||
DAY = "DAY"
|
||||
WEEK = "WEEK"
|
||||
HOUR = "HOUR"
|
||||
MINUTE = "MINUTE"
|
||||
SECOND = "SECOND"
|
||||
|
||||
# Date separators
|
||||
DATE_SEPARATOR = "DATE_SEPARATOR" # -, /, 年, 月, 日, 号
|
||||
|
||||
# Time separators
|
||||
TIME_SEPARATOR = "TIME_SEPARATOR" # :, 点, 时, 分, 秒
|
||||
|
||||
# Period indicators
|
||||
PERIOD_AM = "PERIOD_AM" # 上午, 早上, 早晨, etc.
|
||||
PERIOD_PM = "PERIOD_PM" # 下午, 晚上, 中午, etc.
|
||||
|
||||
# Relative time
|
||||
RELATIVE_TODAY = "RELATIVE_TODAY" # 今天, 今晚, 今早, etc.
|
||||
RELATIVE_TOMORROW = "RELATIVE_TOMORROW" # 明天
|
||||
RELATIVE_DAY_AFTER_TOMORROW = "RELATIVE_DAY_AFTER_TOMORROW" # 后天
|
||||
RELATIVE_THREE_DAYS_AFTER_TOMORROW = "RELATIVE_THREE_DAYS_AFTER_TOMORROW" # 大后天
|
||||
RELATIVE_YESTERDAY = "RELATIVE_YESTERDAY" # 昨天
|
||||
RELATIVE_DAY_BEFORE_YESTERDAY = "RELATIVE_DAY_BEFORE_YESTERDAY" # 前天
|
||||
RELATIVE_THREE_DAYS_BEFORE_YESTERDAY = "RELATIVE_THREE_DAYS_BEFORE_YESTERDAY" # 大前天
|
||||
RELATIVE_DIRECTION_FORWARD = "RELATIVE_DIRECTION_FORWARD" # 后, 以后, 之后
|
||||
RELATIVE_DIRECTION_BACKWARD = "RELATIVE_DIRECTION_BACKWARD" # 前, 以前, 之前
|
||||
|
||||
# Extended relative time
|
||||
RELATIVE_NEXT = "RELATIVE_NEXT" # 下
|
||||
RELATIVE_LAST = "RELATIVE_LAST" # 上, 去
|
||||
RELATIVE_THIS = "RELATIVE_THIS" # 这, 本
|
||||
|
||||
# Week days
|
||||
WEEKDAY_MONDAY = "WEEKDAY_MONDAY"
|
||||
WEEKDAY_TUESDAY = "WEEKDAY_TUESDAY"
|
||||
WEEKDAY_WEDNESDAY = "WEEKDAY_WEDNESDAY"
|
||||
WEEKDAY_THURSDAY = "WEEKDAY_THURSDAY"
|
||||
WEEKDAY_FRIDAY = "WEEKDAY_FRIDAY"
|
||||
WEEKDAY_SATURDAY = "WEEKDAY_SATURDAY"
|
||||
WEEKDAY_SUNDAY = "WEEKDAY_SUNDAY"
|
||||
|
||||
# Week scope
|
||||
WEEK_SCOPE_CURRENT = "WEEK_SCOPE_CURRENT" # 本
|
||||
WEEK_SCOPE_LAST = "WEEK_SCOPE_LAST" # 上
|
||||
WEEK_SCOPE_NEXT = "WEEK_SCOPE_NEXT" # 下
|
||||
|
||||
# Special time markers
|
||||
HALF = "HALF" # 半
|
||||
QUARTER = "QUARTER" # 一刻
|
||||
ZHENG = "ZHENG" # 整
|
||||
ZHONG = "ZHONG" # 钟
|
||||
|
||||
# Student-friendly time expressions
|
||||
EARLY_MORNING = "EARLY_MORNING" # 早X
|
||||
LATE_NIGHT = "LATE_NIGHT" # 晚X
|
||||
|
||||
# Whitespace
|
||||
WHITESPACE = "WHITESPACE"
|
||||
|
||||
# End of input
|
||||
EOF = "EOF"
|
||||
|
||||
|
||||
@dataclass
|
||||
class Token:
|
||||
"""Represents a single token from the lexer."""
|
||||
|
||||
type: TokenType
|
||||
value: Union[str, int]
|
||||
position: int
|
||||
|
||||
def __str__(self):
|
||||
return f"Token({self.type.value}, {repr(self.value)}, {self.position})"
|
||||
|
||||
def __repr__(self):
|
||||
return self.__str__()
|
||||
369
konabot/common/ptimeparse/semantic.py
Normal file
369
konabot/common/ptimeparse/semantic.py
Normal file
@ -0,0 +1,369 @@
|
||||
"""
|
||||
Semantic analyzer for time expressions that evaluates the AST and produces datetime objects.
|
||||
"""
|
||||
|
||||
import datetime
|
||||
import calendar
|
||||
from typing import Optional
|
||||
|
||||
from .ptime_ast import (
|
||||
TimeExpressionNode, DateNode, TimeNode,
|
||||
RelativeDateNode, RelativeTimeNode, WeekdayNode, NumberNode
|
||||
)
|
||||
from .err import TokenUnhandledException
|
||||
|
||||
|
||||
class SemanticAnalyzer:
|
||||
"""Semantic analyzer that evaluates time expression ASTs."""
|
||||
|
||||
def __init__(self, now: Optional[datetime.datetime] = None):
|
||||
self.now = now or datetime.datetime.now()
|
||||
|
||||
def evaluate_number(self, node: NumberNode) -> int:
|
||||
"""Evaluate a number node."""
|
||||
return node.value
|
||||
|
||||
def evaluate_date(self, node: DateNode) -> datetime.date:
|
||||
"""Evaluate a date node."""
|
||||
year = self.now.year
|
||||
month = 1
|
||||
day = 1
|
||||
|
||||
if node.year is not None:
|
||||
year = self.evaluate_number(node.year)
|
||||
if node.month is not None:
|
||||
month = self.evaluate_number(node.month)
|
||||
if node.day is not None:
|
||||
day = self.evaluate_number(node.day)
|
||||
|
||||
return datetime.date(year, month, day)
|
||||
|
||||
def evaluate_time(self, node: TimeNode) -> datetime.time:
|
||||
"""Evaluate a time node."""
|
||||
hour = 0
|
||||
minute = 0
|
||||
second = 0
|
||||
|
||||
if node.hour is not None:
|
||||
hour = self.evaluate_number(node.hour)
|
||||
if node.minute is not None:
|
||||
minute = self.evaluate_number(node.minute)
|
||||
if node.second is not None:
|
||||
second = self.evaluate_number(node.second)
|
||||
|
||||
# Handle 24-hour vs 12-hour format
|
||||
if not node.is_24hour and node.period is not None:
|
||||
if node.period == "AM":
|
||||
if hour == 12:
|
||||
hour = 0
|
||||
elif node.period == "PM":
|
||||
if hour != 12 and hour <= 12:
|
||||
hour += 12
|
||||
|
||||
# Validate time values
|
||||
if not (0 <= hour <= 23):
|
||||
raise TokenUnhandledException(f"Invalid hour: {hour}")
|
||||
if not (0 <= minute <= 59):
|
||||
raise TokenUnhandledException(f"Invalid minute: {minute}")
|
||||
if not (0 <= second <= 59):
|
||||
raise TokenUnhandledException(f"Invalid second: {second}")
|
||||
|
||||
return datetime.time(hour, minute, second)
|
||||
|
||||
def evaluate_relative_date(self, node: RelativeDateNode) -> datetime.timedelta:
|
||||
"""Evaluate a relative date node."""
|
||||
# Start with current time
|
||||
result = self.now
|
||||
|
||||
# Special case: If weeks contains a target day (hacky way to pass target day info)
|
||||
# This is for patterns like "下个月五号"
|
||||
if node.weeks > 0 and node.weeks <= 31: # Valid day range
|
||||
target_day = node.weeks
|
||||
|
||||
# Calculate the target month
|
||||
if node.months != 0:
|
||||
# Handle month arithmetic carefully
|
||||
total_months = result.month + node.months - 1
|
||||
new_year = result.year + total_months // 12
|
||||
new_month = total_months % 12 + 1
|
||||
|
||||
# Handle day overflow (e.g., Jan 31 + 1 month = Feb 28/29)
|
||||
max_day_in_target_month = calendar.monthrange(new_year, new_month)[1]
|
||||
target_day = min(target_day, max_day_in_target_month)
|
||||
|
||||
try:
|
||||
result = result.replace(year=new_year, month=new_month, day=target_day)
|
||||
except ValueError:
|
||||
# Handle edge cases
|
||||
result = result.replace(year=new_year, month=new_month, day=max_day_in_target_month)
|
||||
|
||||
# Return the difference between the new date and the original date
|
||||
return result - self.now
|
||||
|
||||
# Apply years
|
||||
if node.years != 0:
|
||||
# Handle year arithmetic carefully due to leap years
|
||||
new_year = result.year + node.years
|
||||
try:
|
||||
result = result.replace(year=new_year)
|
||||
except ValueError:
|
||||
# Handle leap year edge case (Feb 29 -> Feb 28)
|
||||
result = result.replace(year=new_year, month=2, day=28)
|
||||
|
||||
# Apply months
|
||||
if node.months != 0:
|
||||
# Check if this is a special marker for absolute month (negative offset)
|
||||
if node.months < 0:
|
||||
# This is an absolute month specification (e.g., from "明年五月")
|
||||
absolute_month = node.months + 100
|
||||
if 1 <= absolute_month <= 12:
|
||||
result = result.replace(year=result.year, month=absolute_month, day=result.day)
|
||||
else:
|
||||
# Handle month arithmetic carefully
|
||||
total_months = result.month + node.months - 1
|
||||
new_year = result.year + total_months // 12
|
||||
new_month = total_months % 12 + 1
|
||||
|
||||
# Handle day overflow (e.g., Jan 31 + 1 month = Feb 28/29)
|
||||
new_day = min(result.day, calendar.monthrange(new_year, new_month)[1])
|
||||
|
||||
result = result.replace(year=new_year, month=new_month, day=new_day)
|
||||
|
||||
# Apply weeks and days
|
||||
if node.weeks != 0 or node.days != 0:
|
||||
delta_days = node.weeks * 7 + node.days
|
||||
result = result + datetime.timedelta(days=delta_days)
|
||||
|
||||
return result - self.now
|
||||
|
||||
def evaluate_relative_time(self, node: RelativeTimeNode) -> datetime.timedelta:
|
||||
"""Evaluate a relative time node."""
|
||||
# Convert all values to seconds for precise calculation
|
||||
total_seconds = (
|
||||
node.hours * 3600 +
|
||||
node.minutes * 60 +
|
||||
node.seconds
|
||||
)
|
||||
|
||||
return datetime.timedelta(seconds=total_seconds)
|
||||
|
||||
def evaluate_weekday(self, node: WeekdayNode) -> datetime.timedelta:
|
||||
"""Evaluate a weekday node."""
|
||||
current_weekday = self.now.weekday() # 0=Monday, 6=Sunday
|
||||
target_weekday = node.weekday
|
||||
|
||||
if node.scope == "current":
|
||||
delta = target_weekday - current_weekday
|
||||
elif node.scope == "last":
|
||||
delta = target_weekday - current_weekday - 7
|
||||
elif node.scope == "next":
|
||||
delta = target_weekday - current_weekday + 7
|
||||
else:
|
||||
delta = target_weekday - current_weekday
|
||||
|
||||
return datetime.timedelta(days=delta)
|
||||
|
||||
def infer_smart_time(self, hour: int, minute: int = 0, second: int = 0, base_time: Optional[datetime.datetime] = None) -> datetime.datetime:
|
||||
"""
|
||||
Smart time inference based on current time.
|
||||
|
||||
For example:
|
||||
- If now is 14:30 and user says "3点", interpret as 15:00
|
||||
- If now is 14:30 and user says "1点", interpret as next day 01:00
|
||||
- If now is 8:00 and user says "3点", interpret as 15:00
|
||||
- If now is 8:00 and user says "9点", interpret as 09:00
|
||||
"""
|
||||
# Use base_time if provided, otherwise use self.now
|
||||
now = base_time if base_time is not None else self.now
|
||||
|
||||
# Handle 24-hour format directly (13-23)
|
||||
if 13 <= hour <= 23:
|
||||
candidate = now.replace(hour=hour, minute=minute, second=second, microsecond=0)
|
||||
if candidate <= now:
|
||||
candidate += datetime.timedelta(days=1)
|
||||
return candidate
|
||||
|
||||
# Handle 12 (noon/midnight)
|
||||
if hour == 12:
|
||||
# For 12 specifically, we need to be more careful
|
||||
# Try noon first
|
||||
noon_candidate = now.replace(hour=12, minute=minute, second=second, microsecond=0)
|
||||
midnight_candidate = now.replace(hour=0, minute=minute, second=second, microsecond=0)
|
||||
|
||||
# Special case: If it's afternoon or evening, "十二点" likely means next day midnight
|
||||
if now.hour >= 12:
|
||||
result = midnight_candidate + datetime.timedelta(days=1)
|
||||
return result
|
||||
|
||||
# If noon is in the future and closer than midnight, use it
|
||||
if noon_candidate > now and (midnight_candidate <= now or noon_candidate < midnight_candidate):
|
||||
return noon_candidate
|
||||
# If midnight is in the future, use it
|
||||
elif midnight_candidate > now:
|
||||
return midnight_candidate
|
||||
# Both are in the past, use the closer one
|
||||
elif noon_candidate > midnight_candidate:
|
||||
return noon_candidate
|
||||
# Otherwise use midnight next day
|
||||
else:
|
||||
result = midnight_candidate + datetime.timedelta(days=1)
|
||||
return result
|
||||
|
||||
# Handle 1-11 (12-hour format)
|
||||
if 1 <= hour <= 11:
|
||||
# Calculate 12-hour format candidates
|
||||
pm_hour = hour + 12
|
||||
pm_candidate = now.replace(hour=pm_hour, minute=minute, second=second, microsecond=0)
|
||||
am_candidate = now.replace(hour=hour, minute=minute, second=second, microsecond=0)
|
||||
|
||||
# Special case: If it's afternoon (12:00-18:00) and the hour is 1-6,
|
||||
# user might mean either PM today or AM tomorrow.
|
||||
# But if PM is in the future, that's more likely what they mean.
|
||||
if 12 <= now.hour <= 18 and 1 <= hour <= 6:
|
||||
if pm_candidate > now:
|
||||
return pm_candidate
|
||||
else:
|
||||
# PM is in the past, so use AM tomorrow
|
||||
result = am_candidate + datetime.timedelta(days=1)
|
||||
return result
|
||||
|
||||
# Special case: If it's late evening (after 22:00) and user specifies early morning hours (1-5),
|
||||
# user likely means next day early morning
|
||||
if now.hour >= 22 and 1 <= hour <= 5:
|
||||
result = am_candidate + datetime.timedelta(days=1)
|
||||
return result
|
||||
|
||||
# Special case: In the morning (0-12:00)
|
||||
if now.hour < 12:
|
||||
# In the morning, for hours 1-11, generally prefer AM interpretation
|
||||
# unless it's a very early hour that's much earlier than current time
|
||||
# Only push to next day for very early hours (1-2) that are significantly earlier
|
||||
if hour <= 2 and hour < now.hour and now.hour - hour >= 6:
|
||||
# Very early morning hour that's significantly earlier, use next day
|
||||
result = am_candidate + datetime.timedelta(days=1)
|
||||
return result
|
||||
else:
|
||||
# For morning, generally prefer AM if it's in the future
|
||||
if am_candidate > now:
|
||||
return am_candidate
|
||||
# If PM is in the future, use it
|
||||
elif pm_candidate > now:
|
||||
return pm_candidate
|
||||
# Both are in the past, prefer AM if it's closer
|
||||
elif am_candidate > pm_candidate:
|
||||
return am_candidate
|
||||
# Otherwise use PM next day
|
||||
else:
|
||||
result = pm_candidate + datetime.timedelta(days=1)
|
||||
return result
|
||||
else:
|
||||
# General case: choose the one that's in the future and closer
|
||||
if pm_candidate > now and (am_candidate <= now or pm_candidate < am_candidate):
|
||||
return pm_candidate
|
||||
elif am_candidate > now:
|
||||
return am_candidate
|
||||
# Both are in the past, use the closer one
|
||||
elif pm_candidate > am_candidate:
|
||||
return pm_candidate
|
||||
# Otherwise use AM next day
|
||||
else:
|
||||
result = am_candidate + datetime.timedelta(days=1)
|
||||
return result
|
||||
|
||||
# Handle 0 (midnight)
|
||||
if hour == 0:
|
||||
candidate = now.replace(hour=0, minute=minute, second=second, microsecond=0)
|
||||
if candidate <= now:
|
||||
candidate += datetime.timedelta(days=1)
|
||||
return candidate
|
||||
|
||||
# Default case (should not happen with valid input)
|
||||
candidate = now.replace(hour=hour, minute=minute, second=second, microsecond=0)
|
||||
if candidate <= now:
|
||||
candidate += datetime.timedelta(days=1)
|
||||
return candidate
|
||||
|
||||
def evaluate(self, node: TimeExpressionNode) -> datetime.datetime:
|
||||
"""Evaluate a complete time expression node."""
|
||||
result = self.now
|
||||
|
||||
# Apply relative date (should set time to 00:00:00 for dates)
|
||||
if node.relative_date is not None:
|
||||
delta = self.evaluate_relative_date(node.relative_date)
|
||||
result = result + delta
|
||||
# For relative dates like "今天", "明天", set time to 00:00:00
|
||||
# But only for cases where we're dealing with days, not years/months
|
||||
if (node.date is None and node.time is None and node.weekday is None and
|
||||
node.relative_date.years == 0 and node.relative_date.months == 0):
|
||||
result = result.replace(hour=0, minute=0, second=0, microsecond=0)
|
||||
|
||||
# Apply weekday
|
||||
if node.weekday is not None:
|
||||
delta = self.evaluate_weekday(node.weekday)
|
||||
result = result + delta
|
||||
# For weekdays, set time to 00:00:00
|
||||
if node.date is None and node.time is None:
|
||||
result = result.replace(hour=0, minute=0, second=0, microsecond=0)
|
||||
|
||||
# Apply relative time
|
||||
if node.relative_time is not None:
|
||||
delta = self.evaluate_relative_time(node.relative_time)
|
||||
result = result + delta
|
||||
|
||||
# Apply absolute date
|
||||
if node.date is not None:
|
||||
date = self.evaluate_date(node.date)
|
||||
result = result.replace(year=date.year, month=date.month, day=date.day)
|
||||
# For absolute dates without time, set time to 00:00:00
|
||||
if node.time is None:
|
||||
result = result.replace(hour=0, minute=0, second=0, microsecond=0)
|
||||
|
||||
# Apply time
|
||||
if node.time is not None:
|
||||
time = self.evaluate_time(node.time)
|
||||
|
||||
# Handle explicit period or student-friendly expressions
|
||||
if node.time.is_24hour or node.time.period is not None:
|
||||
# Handle explicit period
|
||||
if not node.time.is_24hour and node.time.period is not None:
|
||||
hour = time.hour
|
||||
minute = time.minute
|
||||
second = time.second
|
||||
|
||||
if node.time.period == "AM":
|
||||
if hour == 12:
|
||||
hour = 0
|
||||
elif node.time.period == "PM":
|
||||
# Special case: "晚上十二点" should be interpreted as next day 00:00
|
||||
if hour == 12 and minute == 0 and second == 0:
|
||||
# Move to next day at 00:00:00
|
||||
result = result.replace(hour=0, minute=0, second=0, microsecond=0) + datetime.timedelta(days=1)
|
||||
# Skip the general replacement since we've already handled it
|
||||
skip_general_replacement = True
|
||||
else:
|
||||
# For other PM times, convert to 24-hour format
|
||||
if hour != 12 and hour <= 12:
|
||||
hour += 12
|
||||
|
||||
# Validate hour
|
||||
if not (0 <= hour <= 23):
|
||||
raise TokenUnhandledException(f"Invalid hour: {hour}")
|
||||
|
||||
# Only do general replacement if we haven't handled it specially
|
||||
if not locals().get('skip_general_replacement', False):
|
||||
result = result.replace(hour=hour, minute=minute, second=second, microsecond=0)
|
||||
else:
|
||||
# Already in 24-hour format
|
||||
result = result.replace(hour=time.hour, minute=time.minute, second=time.second, microsecond=0)
|
||||
else:
|
||||
# Use smart time inference for regular times
|
||||
# But if we have an explicit date, treat the time as 24-hour format
|
||||
if node.date is not None or node.relative_date is not None:
|
||||
# For explicit dates, treat time as 24-hour format
|
||||
result = result.replace(hour=time.hour, minute=time.minute or 0, second=time.second or 0, microsecond=0)
|
||||
else:
|
||||
# Use smart time inference for regular times
|
||||
smart_time = self.infer_smart_time(time.hour, time.minute, time.second, base_time=result)
|
||||
result = smart_time
|
||||
|
||||
return result
|
||||
51
konabot/docs/user/fx.txt
Normal file
51
konabot/docs/user/fx.txt
Normal file
@ -0,0 +1,51 @@
|
||||
## 指令介绍
|
||||
|
||||
`fx` - 用于对图片应用各种滤镜效果的指令
|
||||
|
||||
## 格式
|
||||
|
||||
```
|
||||
fx [滤镜名称] <参数1> <参数2> ...
|
||||
```
|
||||
|
||||
## 示例
|
||||
|
||||
- `fx 模糊`
|
||||
- `fx 阈值 150`
|
||||
- `fx 缩放 2.0`
|
||||
- `fx 色彩 1.8`
|
||||
- `fx 色键 rgb(0,255,0) 50`
|
||||
|
||||
## 可用滤镜列表
|
||||
|
||||
### 基础滤镜
|
||||
* ```fx 模糊 <半径=10>```
|
||||
* ```fx 马赛克 <像素大小=10>```
|
||||
* ```fx 轮廓```
|
||||
* ```fx 锐化```
|
||||
* ```fx 边缘增强```
|
||||
* ```fx 浮雕```
|
||||
* ```fx 查找边缘```
|
||||
* ```fx 平滑```
|
||||
|
||||
### 色彩处理滤镜
|
||||
* ```fx 反色```
|
||||
* ```fx 黑白```
|
||||
* ```fx 阈值 <阈值=128>```
|
||||
* ```fx 对比度 <因子=1.5>```
|
||||
* ```fx 亮度 <因子=1.5>```
|
||||
* ```fx 色彩 <因子=1.5>```
|
||||
* ```fx 色调 <颜色="rgb(255,0,0)">```
|
||||
|
||||
### 几何变换滤镜
|
||||
* ```fx 缩放 <比例=1.5>```
|
||||
* ```fx 波纹 <振幅=5> <波长=20>```
|
||||
|
||||
### 特殊效果滤镜
|
||||
* ```fx 色键 <目标颜色="rgb(255,0,0)"> <容差=60>```
|
||||
|
||||
## 颜色名称支持
|
||||
- **基本颜色**:红、绿、蓝、黄、紫、黑、白、橙、粉、灰、青、靛、棕
|
||||
- **修饰词**:浅、深、亮、暗(可组合使用,如`浅红`、`深蓝`)
|
||||
- **RGB格式**:`rgb(255,0,0)`、`rgb(0,255,0)`、`(255,0,0)` 等
|
||||
- **HEX格式**:`#66ccff`等
|
||||
24
konabot/docs/user/tqszm.txt
Normal file
24
konabot/docs/user/tqszm.txt
Normal file
@ -0,0 +1,24 @@
|
||||
# tqszm
|
||||
|
||||
引用一条消息,让此方帮你提取首字母。
|
||||
|
||||
例子:
|
||||
|
||||
```
|
||||
John: 11-28 16:50:37
|
||||
谁来总结一下今天的工作?
|
||||
|
||||
Jack: 11-28 16:50:55
|
||||
[引用John的消息] tqszm
|
||||
|
||||
此方Bot: 11-28 16:50:56
|
||||
slzjyxjtdgz?
|
||||
```
|
||||
|
||||
或者,你也可以直接以正常指令的方式调用:
|
||||
|
||||
```
|
||||
提取首字母 中山大学
|
||||
> zsdx
|
||||
```
|
||||
|
||||
@ -117,22 +117,18 @@ async def _(target: DepLongTaskTarget, temp: Optional[Union[int, float]] = 1):
|
||||
await send_ac_image(evt, ac)
|
||||
return
|
||||
await ac.update_ac(temperature_delta=temp)
|
||||
if ac.temperature > 40:
|
||||
# 根据温度随机出是否爆炸,40度开始,呈指数增长
|
||||
possibility = -math.e ** ((40-ac.temperature) / 50) + 1
|
||||
if random.random() < possibility:
|
||||
# 打开爆炸图片
|
||||
with open(ASSETS_PATH / "img" / "other" / "boom.jpg", "rb") as f:
|
||||
output = BytesIO()
|
||||
# 爆炸抖动
|
||||
frames = wiggle_transform(np.array(Image.open(f)), intensity=5)
|
||||
pil_frames = [Image.fromarray(frame) for frame in frames]
|
||||
pil_frames[0].save(output, format="GIF", save_all=True, append_images=pil_frames[1:], loop=0, duration=35, disposal=2)
|
||||
output.seek(0)
|
||||
await evt.send(await UniMessage().image(raw=output).export())
|
||||
await ac.broke_ac(CrashType.BURNT)
|
||||
await evt.send("太热啦,空调炸了!")
|
||||
return
|
||||
if ac.burnt:
|
||||
# 打开爆炸图片
|
||||
with open(ASSETS_PATH / "img" / "other" / "boom.jpg", "rb") as f:
|
||||
output = BytesIO()
|
||||
# 爆炸抖动
|
||||
frames = wiggle_transform(np.array(Image.open(f)), intensity=5)
|
||||
pil_frames = [Image.fromarray(frame) for frame in frames]
|
||||
pil_frames[0].save(output, format="GIF", save_all=True, append_images=pil_frames[1:], loop=0, duration=35, disposal=2)
|
||||
output.seek(0)
|
||||
await evt.send(await UniMessage().image(raw=output).export())
|
||||
await evt.send("太热啦,空调炸了!")
|
||||
return
|
||||
await send_ac_image(evt, ac)
|
||||
|
||||
evt = on_alconna(Alconna(
|
||||
@ -152,11 +148,6 @@ async def _(target: DepLongTaskTarget, temp: Optional[Union[int, float]] = 1):
|
||||
await send_ac_image(evt, ac)
|
||||
return
|
||||
await ac.update_ac(temperature_delta=-temp)
|
||||
if ac.temperature < 0:
|
||||
# 根据温度随机出是否冻结,0度开始,呈指数增长
|
||||
possibility = -math.e ** (ac.temperature / 50) + 1
|
||||
if random.random() < possibility:
|
||||
await ac.broke_ac(CrashType.FROZEN)
|
||||
await send_ac_image(evt, ac)
|
||||
|
||||
evt = on_alconna(Alconna(
|
||||
@ -202,3 +193,36 @@ async def _(target: DepLongTaskTarget):
|
||||
params=params
|
||||
)
|
||||
await evt.send(await UniMessage().image(raw=image).export())
|
||||
|
||||
evt = on_alconna(Alconna(
|
||||
"空调最高峰",
|
||||
), use_cmd_start=True, use_cmd_sep=False, skip_for_unmatch=True)
|
||||
|
||||
@evt.handle()
|
||||
async def _(target: DepLongTaskTarget):
|
||||
result = await db_manager.query_by_sql_file(
|
||||
ROOT_PATH / "sql" / "query_peak.sql"
|
||||
)
|
||||
if len(result) == 0:
|
||||
await evt.send("没有空调记录!")
|
||||
return
|
||||
max_temp = result[0].get("max")
|
||||
min_temp = result[0].get("min")
|
||||
his_max = result[0].get("his_max")
|
||||
his_min = result[0].get("his_min")
|
||||
# 再从内存里的空调池中获取最高温度和最低温度
|
||||
for ac in AirConditioner.InstancesPool.values():
|
||||
if ac.on and not ac.burnt and not ac.frozen:
|
||||
if max_temp is None or min_temp is None:
|
||||
max_temp = ac.temperature
|
||||
min_temp = ac.temperature
|
||||
max_temp = max(max_temp, ac.temperature)
|
||||
min_temp = min(min_temp, ac.temperature)
|
||||
if max_temp is None or min_temp is None:
|
||||
await evt.send(f"目前全部空调都被炸掉了!")
|
||||
else:
|
||||
await evt.send(f"全球在线空调最高温度为 {'%.1f' % max_temp}°C,最低温度为 {'%.1f' % min_temp}°C!")
|
||||
if his_max is None or his_min is None:
|
||||
pass
|
||||
else:
|
||||
await evt.send(f"历史最高温度为 {'%.1f' % his_max}°C,最低温度为 {'%.1f' % his_min}°C!\n(要进入历史记录,温度需至少保持 5 分钟)")
|
||||
@ -1,14 +1,22 @@
|
||||
import asyncio
|
||||
from enum import Enum
|
||||
from io import BytesIO
|
||||
import math
|
||||
from pathlib import Path
|
||||
import random
|
||||
import signal
|
||||
import time
|
||||
|
||||
import cv2
|
||||
import numpy as np
|
||||
from PIL import Image, ImageDraw, ImageFont
|
||||
|
||||
from nonebot import logger
|
||||
|
||||
from konabot.common.database import DatabaseManager
|
||||
from konabot.common.path import ASSETS_PATH, FONTS_PATH
|
||||
from konabot.common.path import DATA_PATH
|
||||
import nonebot
|
||||
import json
|
||||
|
||||
ROOT_PATH = Path(__file__).resolve().parent
|
||||
@ -20,9 +28,99 @@ class CrashType(Enum):
|
||||
BURNT = 0
|
||||
FROZEN = 1
|
||||
|
||||
driver = nonebot.get_driver()
|
||||
|
||||
@driver.on_startup
|
||||
async def register_startup_hook():
|
||||
await ac_manager.start_auto_save()
|
||||
|
||||
@driver.on_shutdown
|
||||
async def register_shutdown_hook():
|
||||
"""注册关闭时需要执行的函数"""
|
||||
# 停止自动保存任务
|
||||
if ac_manager:
|
||||
await ac_manager.stop_auto_save()
|
||||
|
||||
class AirConditionerManager:
|
||||
def __init__(self, save_interval: int = 300): # 默认5分钟保存一次
|
||||
self.save_interval = save_interval
|
||||
self._save_task = None
|
||||
self._running = False
|
||||
|
||||
async def start_auto_save(self):
|
||||
"""启动自动保存任务"""
|
||||
self._running = True
|
||||
self._save_task = asyncio.create_task(self._auto_save_loop())
|
||||
|
||||
logger.info(f"自动保存任务已启动,间隔: {self.save_interval}秒")
|
||||
|
||||
async def stop_auto_save(self):
|
||||
"""停止自动保存任务"""
|
||||
if self._save_task:
|
||||
self._running = False
|
||||
self._save_task.cancel()
|
||||
try:
|
||||
await self._save_task
|
||||
except asyncio.CancelledError:
|
||||
pass
|
||||
logger.info("自动保存任务已停止")
|
||||
else:
|
||||
logger.warning("没有正在运行的自动保存任务")
|
||||
|
||||
async def _auto_save_loop(self):
|
||||
"""自动保存循环"""
|
||||
while self._running:
|
||||
try:
|
||||
await asyncio.sleep(self.save_interval)
|
||||
await self.save_all_instances()
|
||||
except asyncio.CancelledError:
|
||||
break
|
||||
except Exception as e:
|
||||
logger.error(f"定时保存失败: {e}")
|
||||
|
||||
async def save_all_instances(self):
|
||||
save_time = time.time()
|
||||
to_remove = []
|
||||
"""保存所有实例到数据库"""
|
||||
for ac_id, ac_instance in AirConditioner.InstancesPool.items():
|
||||
try:
|
||||
await db_manager.execute_by_sql_file(
|
||||
ROOT_PATH / "sql" / "update_ac.sql",
|
||||
[(ac_instance.on, ac_instance.temperature,
|
||||
ac_instance.burnt, ac_instance.frozen, ac_id),(ac_id,)]
|
||||
)
|
||||
if(save_time - ac_instance.instance_get_time >= 300): # 5 分钟
|
||||
to_remove.append(ac_id)
|
||||
except Exception as e:
|
||||
logger.error(f"保存空调 {ac_id} 失败: {e}")
|
||||
|
||||
logger.info(f"定时保存完成,共保存 {len(AirConditioner.InstancesPool)} 个空调实例")
|
||||
|
||||
# 删除时间过长实例
|
||||
for ac_id in to_remove:
|
||||
del AirConditioner.InstancesPool[ac_id]
|
||||
|
||||
logger.info(f"清理长期不活跃的空调实例完成,目前池内共有 {len(AirConditioner.InstancesPool)} 个实例")
|
||||
|
||||
ac_manager = AirConditionerManager(save_interval=300) # 5分钟
|
||||
|
||||
class AirConditioner:
|
||||
InstancesPool: dict[str, 'AirConditioner'] = {}
|
||||
|
||||
@classmethod
|
||||
async def refresh_ac(cls, id: str):
|
||||
cls.InstancesPool[id].instance_get_time = time.time()
|
||||
|
||||
@classmethod
|
||||
async def storage_ac(cls, id: str, ac: 'AirConditioner'):
|
||||
cls.InstancesPool[id] = ac
|
||||
|
||||
@classmethod
|
||||
async def get_ac(cls, id: str) -> 'AirConditioner':
|
||||
if(id in cls.InstancesPool):
|
||||
await cls.refresh_ac(id)
|
||||
return cls.InstancesPool[id]
|
||||
# 如果没有,那么从数据库重新实例化一个 AC 出来
|
||||
result = await db_manager.query_by_sql_file(ROOT_PATH / "sql" / "query_ac.sql", (id,))
|
||||
if len(result) == 0:
|
||||
ac = await cls.create_ac(id)
|
||||
@ -33,6 +131,7 @@ class AirConditioner:
|
||||
ac.temperature = float(ac_data["temperature"])
|
||||
ac.burnt = bool(ac_data["burnt"])
|
||||
ac.frozen = bool(ac_data["frozen"])
|
||||
await cls.storage_ac(id, ac)
|
||||
return ac
|
||||
|
||||
@classmethod
|
||||
@ -40,23 +139,43 @@ class AirConditioner:
|
||||
ac = AirConditioner(id)
|
||||
await db_manager.execute_by_sql_file(
|
||||
ROOT_PATH / "sql" / "insert_ac.sql",
|
||||
(id, ac.on, ac.temperature, ac.burnt, ac.frozen)
|
||||
[(id, ac.on, ac.temperature, ac.burnt, ac.frozen),(id,)]
|
||||
)
|
||||
await cls.storage_ac(id, ac)
|
||||
return ac
|
||||
|
||||
async def change_ac_temp(self, temperature_delta: float) -> None:
|
||||
'''
|
||||
改变空调的温度
|
||||
:param temperature_delta: float 温度变化量
|
||||
'''
|
||||
changed_temp = self.temperature + temperature_delta
|
||||
random_poss = random.random()
|
||||
if temperature_delta < 0 and changed_temp < 0:
|
||||
# 根据温度随机出是否冻结,0度开始,呈指数增长
|
||||
possibility = -math.e ** (changed_temp / 50) + 1
|
||||
if random_poss < possibility:
|
||||
await self.broke_ac(CrashType.FROZEN)
|
||||
elif temperature_delta > 0 and changed_temp > 40:
|
||||
# 根据温度随机出是否烧坏,40度开始,呈指数增长
|
||||
possibility = -math.e ** ((40-changed_temp) / 50) + 1
|
||||
if random_poss < possibility:
|
||||
await self.broke_ac(CrashType.BURNT)
|
||||
self.temperature = changed_temp
|
||||
|
||||
async def update_ac(self, state: bool = None, temperature_delta: float = None, burnt: bool = None, frozen: bool = None) -> 'AirConditioner':
|
||||
if state is not None:
|
||||
self.on = state
|
||||
if temperature_delta is not None:
|
||||
self.temperature += temperature_delta
|
||||
await self.change_ac_temp(temperature_delta)
|
||||
if burnt is not None:
|
||||
self.burnt = burnt
|
||||
if frozen is not None:
|
||||
self.frozen = frozen
|
||||
await db_manager.execute_by_sql_file(
|
||||
ROOT_PATH / "sql" / "update_ac.sql",
|
||||
(self.on, self.temperature, self.burnt, self.frozen, self.id)
|
||||
)
|
||||
# await db_manager.execute_by_sql_file(
|
||||
# ROOT_PATH / "sql" / "update_ac.sql",
|
||||
# (self.on, self.temperature, self.burnt, self.frozen, self.id)
|
||||
# )
|
||||
return self
|
||||
|
||||
async def change_ac(self) -> 'AirConditioner':
|
||||
@ -64,10 +183,10 @@ class AirConditioner:
|
||||
self.temperature = 24
|
||||
self.burnt = False
|
||||
self.frozen = False
|
||||
await db_manager.execute_by_sql_file(
|
||||
ROOT_PATH / "sql" / "update_ac.sql",
|
||||
(self.on, self.temperature, self.burnt, self.frozen, self.id)
|
||||
)
|
||||
# await db_manager.execute_by_sql_file(
|
||||
# ROOT_PATH / "sql" / "update_ac.sql",
|
||||
# (self.on, self.temperature, self.burnt, self.frozen, self.id)
|
||||
# )
|
||||
return self
|
||||
|
||||
def __init__(self, id: str) -> None:
|
||||
@ -77,6 +196,8 @@ class AirConditioner:
|
||||
self.burnt = False
|
||||
self.frozen = False
|
||||
|
||||
self.instance_get_time = time.time()
|
||||
|
||||
async def broke_ac(self, crash_type: CrashType):
|
||||
'''
|
||||
让空调坏掉
|
||||
|
||||
@ -7,6 +7,17 @@ CREATE TABLE IF NOT EXISTS air_conditioner (
|
||||
frozen BOOLEAN NOT NULL
|
||||
);
|
||||
|
||||
CREATE TABLE IF NOT EXISTS air_conditioner_log (
|
||||
log_id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||
log_time DATETIME NOT NULL DEFAULT CURRENT_TIMESTAMP,
|
||||
id VARCHAR(128),
|
||||
"on" BOOLEAN NOT NULL,
|
||||
temperature REAL NOT NULL,
|
||||
burnt BOOLEAN NOT NULL,
|
||||
frozen BOOLEAN NOT NULL,
|
||||
FOREIGN KEY (id) REFERENCES air_conditioner(id)
|
||||
);
|
||||
|
||||
CREATE TABLE IF NOT EXISTS air_conditioner_crash_log (
|
||||
id VARCHAR(128) NOT NULL,
|
||||
crash_type INT NOT NULL,
|
||||
|
||||
@ -1,3 +1,8 @@
|
||||
-- 插入一台新空调
|
||||
INSERT INTO air_conditioner (id, "on", temperature, burnt, frozen)
|
||||
VALUES (?, ?, ?, ?, ?);
|
||||
VALUES (?, ?, ?, ?, ?);
|
||||
-- 使用返回的数据插入日志
|
||||
INSERT INTO air_conditioner_log (id, "on", temperature, burnt, frozen)
|
||||
SELECT id, "on", temperature, burnt, frozen
|
||||
FROM air_conditioner
|
||||
WHERE id = ?;
|
||||
@ -1,4 +1,4 @@
|
||||
-- 查询空调状态,如果没有就插入一条新的记录
|
||||
-- 查询空调状态
|
||||
SELECT *
|
||||
FROM air_conditioner
|
||||
WHERE id = ?;
|
||||
13
konabot/plugins/air_conditioner/sql/query_peak.sql
Normal file
13
konabot/plugins/air_conditioner/sql/query_peak.sql
Normal file
@ -0,0 +1,13 @@
|
||||
-- 查询目前所有空调中的最高温度与最低温度与历史最高低温
|
||||
SELECT
|
||||
(SELECT MAX(temperature) FROM air_conditioner
|
||||
WHERE "on" = TRUE AND NOT frozen AND NOT burnt) AS max,
|
||||
|
||||
(SELECT MIN(temperature) FROM air_conditioner
|
||||
WHERE "on" = TRUE AND NOT frozen AND NOT burnt) AS min,
|
||||
|
||||
(SELECT MAX(temperature) FROM air_conditioner_log
|
||||
WHERE "on" = TRUE AND NOT frozen AND NOT burnt) AS his_max,
|
||||
|
||||
(SELECT MIN(temperature) FROM air_conditioner_log
|
||||
WHERE "on" = TRUE AND NOT frozen AND NOT burnt) AS his_min;
|
||||
@ -1,4 +1,10 @@
|
||||
-- 更新空调状态
|
||||
UPDATE air_conditioner
|
||||
SET "on" = ?, temperature = ?, burnt = ?, frozen = ?
|
||||
WHERE id = ?;
|
||||
|
||||
-- 插入日志记录(从更新后的数据获取)
|
||||
INSERT INTO air_conditioner_log (id, "on", temperature, burnt, frozen)
|
||||
SELECT id, "on", temperature, burnt, frozen
|
||||
FROM air_conditioner
|
||||
WHERE id = ?;
|
||||
@ -6,34 +6,25 @@ from nonebot_plugin_alconna import Reference, Reply, UniMsg
|
||||
from nonebot.adapters import Event
|
||||
|
||||
|
||||
matcher_fix = on_message()
|
||||
|
||||
pattern = (
|
||||
r"^(?:(?:av|cv)\d+|BV[a-zA-Z0-9]{10})|"
|
||||
r"(?:b23\.tv|bili(?:22|23|33|2233)\.cn|\.bilibili\.com|QQ小程序(?:&#93;|]|\])哔哩哔哩).{0,500}"
|
||||
)
|
||||
|
||||
|
||||
@matcher_fix.handle()
|
||||
async def _(msg: UniMsg, event: Event):
|
||||
def _rule(msg: UniMsg):
|
||||
to_search = msg.exclude(Reply, Reference).dump(json=True)
|
||||
to_search2 = msg.exclude(Reply, Reference).extract_plain_text()
|
||||
if not re.search(pattern, to_search) and not re.search(pattern, to_search2):
|
||||
return
|
||||
return False
|
||||
return True
|
||||
|
||||
|
||||
matcher_fix = on_message(rule=_rule)
|
||||
|
||||
@matcher_fix.handle()
|
||||
async def _(event: Event):
|
||||
from nonebot_plugin_analysis_bilibili import handle_analysis
|
||||
|
||||
await handle_analysis(event)
|
||||
|
||||
# b_url: str
|
||||
# b_page: str | None
|
||||
# b_time: str | None
|
||||
#
|
||||
# from nonebot_plugin_analysis_bilibili.analysis_bilibili import extract as bilibili_extract
|
||||
#
|
||||
# b_url, b_page, b_time = bilibili_extract(to_search)
|
||||
# if b_url is None:
|
||||
# return
|
||||
#
|
||||
# await matcher_fix.send(await UniMessage().text(b_url).export())
|
||||
|
||||
|
||||
58
konabot/plugins/fx_process/__init__.py
Normal file
58
konabot/plugins/fx_process/__init__.py
Normal file
@ -0,0 +1,58 @@
|
||||
from io import BytesIO
|
||||
|
||||
from inspect import signature
|
||||
|
||||
from konabot.common.nb.extract_image import DepPILImage
|
||||
from nonebot.adapters import Event as BaseEvent
|
||||
from nonebot import on_message
|
||||
|
||||
from nonebot_plugin_alconna import (
|
||||
UniMessage,
|
||||
UniMsg
|
||||
)
|
||||
|
||||
from konabot.plugins.fx_process.fx_manager import ImageFilterManager
|
||||
|
||||
def is_fx_mentioned(evt: BaseEvent, msg: UniMsg) -> bool:
|
||||
txt = msg.extract_plain_text()
|
||||
if "fx" not in txt[:3]:
|
||||
return False
|
||||
return True
|
||||
|
||||
fx_on = on_message(rule=is_fx_mentioned)
|
||||
|
||||
@fx_on.handle()
|
||||
async def _(msg: UniMsg, event: BaseEvent, img: DepPILImage):
|
||||
args = msg.extract_plain_text().split()
|
||||
if len(args) < 2:
|
||||
return
|
||||
filter_name = args[1]
|
||||
filter_func = ImageFilterManager.get_filter(filter_name)
|
||||
if not filter_func:
|
||||
return
|
||||
# 获取函数最大参数数量
|
||||
sig = signature(filter_func)
|
||||
max_params = len(sig.parameters) - 1 # 减去第一个参数 image
|
||||
# 从 args 提取参数,并转换为适当类型
|
||||
func_args = []
|
||||
for i in range(2, min(len(args), max_params + 2)):
|
||||
# 尝试将参数转换为函数签名中对应的类型
|
||||
param = list(sig.parameters.values())[i - 1]
|
||||
param_type = param.annotation
|
||||
arg_value = args[i]
|
||||
try:
|
||||
if param_type is float:
|
||||
converted_value = float(arg_value)
|
||||
elif param_type is int:
|
||||
converted_value = int(arg_value)
|
||||
else:
|
||||
converted_value = arg_value
|
||||
except Exception:
|
||||
converted_value = arg_value
|
||||
func_args.append(converted_value)
|
||||
# 应用滤镜
|
||||
out_img = filter_func(img, *func_args)
|
||||
output = BytesIO()
|
||||
out_img.save(output, format="PNG")
|
||||
await fx_on.send(await UniMessage().image(raw=output).export())
|
||||
|
||||
50
konabot/plugins/fx_process/color_handle.py
Normal file
50
konabot/plugins/fx_process/color_handle.py
Normal file
@ -0,0 +1,50 @@
|
||||
from typing import Optional
|
||||
from PIL import ImageColor
|
||||
|
||||
class ColorHandle:
|
||||
color_name_map = {
|
||||
"红": (255, 0, 0),
|
||||
"绿": (0, 255, 0),
|
||||
"蓝": (0, 0, 255),
|
||||
"黄": (255, 255, 0),
|
||||
"紫": (128, 0, 128),
|
||||
"黑": (0, 0, 0),
|
||||
"白": (255, 255, 255),
|
||||
"橙": (255, 165, 0),
|
||||
"粉": (255, 192, 203),
|
||||
"灰": (128, 128, 128),
|
||||
"青": (0, 255, 255),
|
||||
"靛": (75, 0, 130),
|
||||
"棕": (165, 42, 42),
|
||||
"浅": (200, 200, 200),
|
||||
"深": (50, 50, 50),
|
||||
"亮": (255, 255, 224),
|
||||
"暗": (47, 79, 79),
|
||||
}
|
||||
|
||||
@staticmethod
|
||||
def set_or_blend_color(ori_color: Optional[tuple], target_color: tuple) -> tuple:
|
||||
# 如果没有指定初始颜色,返回目标颜色
|
||||
if ori_color is None:
|
||||
return target_color
|
||||
# 混合颜色,取平均值
|
||||
blended_color = tuple((o + t) // 2 for o, t in zip(ori_color, target_color))
|
||||
return blended_color
|
||||
|
||||
@staticmethod
|
||||
def parse_color(color_str: str) -> tuple:
|
||||
# 如果是纯括号,则加上前缀 rgb
|
||||
if color_str.startswith('(') and color_str.endswith(')'):
|
||||
color_str = 'rgb' + color_str
|
||||
try:
|
||||
return ImageColor.getrgb(color_str)
|
||||
except ValueError:
|
||||
pass
|
||||
base_color = None
|
||||
color_str = color_str.replace('色', '')
|
||||
for name, rgb in ColorHandle.color_name_map.items():
|
||||
if name in color_str:
|
||||
base_color = ColorHandle.set_or_blend_color(base_color, rgb)
|
||||
if base_color is not None:
|
||||
return base_color
|
||||
return (255, 255, 255) # 默认白色
|
||||
167
konabot/plugins/fx_process/fx_handle.py
Normal file
167
konabot/plugins/fx_process/fx_handle.py
Normal file
@ -0,0 +1,167 @@
|
||||
from PIL import Image, ImageFilter
|
||||
from PIL import ImageEnhance
|
||||
|
||||
from konabot.plugins.fx_process.color_handle import ColorHandle
|
||||
|
||||
import math
|
||||
|
||||
class ImageFilterImplement:
|
||||
@staticmethod
|
||||
def apply_blur(image: Image.Image, radius: float = 10) -> Image.Image:
|
||||
return image.filter(ImageFilter.GaussianBlur(radius))
|
||||
|
||||
# 马赛克
|
||||
@staticmethod
|
||||
def apply_mosaic(image: Image.Image, pixel_size: int = 10) -> Image.Image:
|
||||
if pixel_size <= 0:
|
||||
pixel_size = 1
|
||||
# 缩小图像
|
||||
small_image = image.resize(
|
||||
(image.width // pixel_size, image.height // pixel_size),
|
||||
Image.Resampling.NEAREST
|
||||
)
|
||||
# 放大图像
|
||||
return small_image.resize(image.size, Image.Resampling.NEAREST)
|
||||
|
||||
@staticmethod
|
||||
def apply_contour(image: Image.Image) -> Image.Image:
|
||||
return image.filter(ImageFilter.CONTOUR)
|
||||
|
||||
@staticmethod
|
||||
def apply_sharpen(image: Image.Image) -> Image.Image:
|
||||
return image.filter(ImageFilter.SHARPEN)
|
||||
|
||||
@staticmethod
|
||||
def apply_edge_enhance(image: Image.Image) -> Image.Image:
|
||||
return image.filter(ImageFilter.EDGE_ENHANCE)
|
||||
|
||||
@staticmethod
|
||||
def apply_emboss(image: Image.Image) -> Image.Image:
|
||||
return image.filter(ImageFilter.EMBOSS)
|
||||
|
||||
@staticmethod
|
||||
def apply_find_edges(image: Image.Image) -> Image.Image:
|
||||
return image.filter(ImageFilter.FIND_EDGES)
|
||||
|
||||
@staticmethod
|
||||
def apply_smooth(image: Image.Image) -> Image.Image:
|
||||
return image.filter(ImageFilter.SMOOTH)
|
||||
|
||||
# 反色
|
||||
@staticmethod
|
||||
def apply_invert(image: Image.Image) -> Image.Image:
|
||||
# 确保图像是RGBA模式,保留透明度通道
|
||||
if image.mode != 'RGBA':
|
||||
image = image.convert('RGBA')
|
||||
r, g, b, a = image.split()
|
||||
r = r.point(lambda i: 255 - i)
|
||||
g = g.point(lambda i: 255 - i)
|
||||
b = b.point(lambda i: 255 - i)
|
||||
return Image.merge('RGBA', (r, g, b, a))
|
||||
|
||||
# 黑白灰度
|
||||
@staticmethod
|
||||
def apply_black_white(image: Image.Image) -> Image.Image:
|
||||
# 保留透明度通道
|
||||
if image.mode != 'RGBA':
|
||||
image = image.convert('RGBA')
|
||||
r, g, b, a = image.split()
|
||||
gray = Image.merge('RGB', (r, g, b)).convert('L')
|
||||
return Image.merge('RGBA', (gray, gray, gray, a))
|
||||
|
||||
# 阈值
|
||||
@staticmethod
|
||||
def apply_threshold(image: Image.Image, threshold: int = 128) -> Image.Image:
|
||||
# 保留透明度通道
|
||||
if image.mode != 'RGBA':
|
||||
image = image.convert('RGBA')
|
||||
r, g, b, a = image.split()
|
||||
gray = Image.merge('RGB', (r, g, b)).convert('L')
|
||||
bw = gray.point(lambda x: 255 if x >= threshold else 0, '1')
|
||||
return Image.merge('RGBA', (bw.convert('L'), bw.convert('L'), bw.convert('L'), a))
|
||||
|
||||
# 对比度
|
||||
@staticmethod
|
||||
def apply_contrast(image: Image.Image, factor: float = 1.5) -> Image.Image:
|
||||
enhancer = ImageEnhance.Contrast(image)
|
||||
return enhancer.enhance(factor)
|
||||
|
||||
# 亮度
|
||||
@staticmethod
|
||||
def apply_brightness(image: Image.Image, factor: float = 1.5) -> Image.Image:
|
||||
enhancer = ImageEnhance.Brightness(image)
|
||||
return enhancer.enhance(factor)
|
||||
|
||||
# 色彩
|
||||
@staticmethod
|
||||
def apply_color(image: Image.Image, factor: float = 1.5) -> Image.Image:
|
||||
enhancer = ImageEnhance.Color(image)
|
||||
return enhancer.enhance(factor)
|
||||
|
||||
# 三色调
|
||||
@staticmethod
|
||||
def apply_to_color(image: Image.Image, color: str = 'rgb(255,0,0)') -> Image.Image:
|
||||
if image.mode != 'RGBA':
|
||||
image = image.convert('RGBA')
|
||||
# 转为灰度图
|
||||
gray = image.convert('L')
|
||||
# 获取目标颜色的RGB值
|
||||
rgb_color = ColorHandle.parse_color(color)
|
||||
# 高光默认为白色,阴影默认为黑色
|
||||
highlight = (255, 255, 255)
|
||||
shadow = (0, 0, 0)
|
||||
# 创建新的图像
|
||||
new_image = Image.new('RGBA', image.size)
|
||||
width, height = image.size
|
||||
for x in range(width):
|
||||
for y in range(height):
|
||||
lum = gray.getpixel((x, y))
|
||||
# 计算新颜色
|
||||
new_r = int((rgb_color[0] * lum + shadow[0] * (highlight[0] - lum)) / 255)
|
||||
new_g = int((rgb_color[1] * lum + shadow[1] * (highlight[1] - lum)) / 255)
|
||||
new_b = int((rgb_color[2] * lum + shadow[2] * (highlight[2] - lum)) / 255)
|
||||
a = image.getpixel((x, y))[3] # 保留原图的透明度
|
||||
new_image.putpixel((x, y), (new_r, new_g, new_b, a))
|
||||
return new_image
|
||||
|
||||
# 缩放
|
||||
@staticmethod
|
||||
def apply_resize(image: Image.Image, scale: float = 1.5) -> Image.Image:
|
||||
if scale <= 0:
|
||||
scale = 1.0
|
||||
new_size = (int(image.width * scale), int(image.height * scale))
|
||||
return image.resize(new_size, Image.Resampling.LANCZOS)
|
||||
|
||||
# 波纹
|
||||
@staticmethod
|
||||
def apply_wave(image: Image.Image, amplitude: float = 5, wavelength: float = 20) -> Image.Image:
|
||||
width, height = image.size
|
||||
new_image = Image.new('RGBA', (width, height))
|
||||
for x in range(width):
|
||||
for y in range(height):
|
||||
offset_x = int(amplitude * math.sin(2 * math.pi * y / wavelength))
|
||||
offset_y = int(amplitude * math.cos(2 * math.pi * x / wavelength))
|
||||
new_x = x + offset_x
|
||||
new_y = y + offset_y
|
||||
if 0 <= new_x < width and 0 <= new_y < height:
|
||||
new_image.putpixel((x, y), image.getpixel((new_x, new_y)))
|
||||
else:
|
||||
new_image.putpixel((x, y), (0, 0, 0, 0)) # 透明像素
|
||||
return new_image
|
||||
|
||||
def apply_color_key(image: Image.Image, target_color: str = 'rgb(255,0,0)', tolerance: int = 60) -> Image.Image:
|
||||
if image.mode != 'RGBA':
|
||||
image = image.convert('RGBA')
|
||||
target_rgb = ColorHandle.parse_color(target_color)
|
||||
width, height = image.size
|
||||
new_image = Image.new('RGBA', (width, height))
|
||||
for x in range(width):
|
||||
for y in range(height):
|
||||
r, g, b, a = image.getpixel((x, y))
|
||||
# 计算与目标颜色的距离
|
||||
distance = math.sqrt((r - target_rgb[0]) ** 2 + (g - target_rgb[1]) ** 2 + (b - target_rgb[2]) ** 2)
|
||||
if distance <= tolerance:
|
||||
new_image.putpixel((x, y), (r, g, b, 0)) # 设置为透明
|
||||
else:
|
||||
new_image.putpixel((x, y), (r, g, b, a)) # 保留原像素
|
||||
return new_image
|
||||
28
konabot/plugins/fx_process/fx_manager.py
Normal file
28
konabot/plugins/fx_process/fx_manager.py
Normal file
@ -0,0 +1,28 @@
|
||||
from typing import Optional
|
||||
from konabot.plugins.fx_process.fx_handle import ImageFilterImplement
|
||||
|
||||
class ImageFilterManager:
|
||||
filter_map = {
|
||||
"模糊": ImageFilterImplement.apply_blur,
|
||||
"马赛克": ImageFilterImplement.apply_mosaic,
|
||||
"轮廓": ImageFilterImplement.apply_contour,
|
||||
"锐化": ImageFilterImplement.apply_sharpen,
|
||||
"边缘增强": ImageFilterImplement.apply_edge_enhance,
|
||||
"浮雕": ImageFilterImplement.apply_emboss,
|
||||
"查找边缘": ImageFilterImplement.apply_find_edges,
|
||||
"平滑": ImageFilterImplement.apply_smooth,
|
||||
"反色": ImageFilterImplement.apply_invert,
|
||||
"黑白": ImageFilterImplement.apply_black_white,
|
||||
"阈值": ImageFilterImplement.apply_threshold,
|
||||
"对比度": ImageFilterImplement.apply_contrast,
|
||||
"亮度": ImageFilterImplement.apply_brightness,
|
||||
"色彩": ImageFilterImplement.apply_color,
|
||||
"色调": ImageFilterImplement.apply_to_color,
|
||||
"缩放": ImageFilterImplement.apply_resize,
|
||||
"波纹": ImageFilterImplement.apply_wave,
|
||||
"色键": ImageFilterImplement.apply_color_key,
|
||||
}
|
||||
|
||||
@classmethod
|
||||
def get_filter(cls, name: str) -> Optional[callable]:
|
||||
return cls.filter_map.get(name)
|
||||
@ -117,18 +117,22 @@ class IdiomGameLLM:
|
||||
@classmethod
|
||||
async def storage_idiom(cls, idiom: str):
|
||||
# 将 idiom 存入数据库
|
||||
await db_manager.execute_by_sql_file(
|
||||
ROOT_PATH / "sql" / "insert_custom_word.sql",
|
||||
(idiom,)
|
||||
)
|
||||
# await db_manager.execute_by_sql_file(
|
||||
# ROOT_PATH / "sql" / "insert_custom_word.sql",
|
||||
# (idiom,)
|
||||
# )
|
||||
# 将 idiom 存入本地文件以备后续分析
|
||||
with open(DATA_DIR / "idiom_llm_storage.txt", "a", encoding="utf-8") as f:
|
||||
f.write(idiom + "\n")
|
||||
IdiomGame.append_into_word_list(idiom)
|
||||
|
||||
|
||||
class IdiomGame:
|
||||
# ALL_WORDS = [] # 所有四字词语
|
||||
# ALL_IDIOMS = [] # 所有成语
|
||||
ALL_WORDS = [] # 所有四字词语
|
||||
ALL_IDIOMS = [] # 所有成语
|
||||
INSTANCE_LIST: dict[str, "IdiomGame"] = {} # 群号对应的游戏实例
|
||||
# IDIOM_FIRST_CHAR = {} # 所有成语包括词语的首字字典
|
||||
# AVALIABLE_IDIOM_FIRST_CHAR = {} # 真正有效的成语首字字典
|
||||
IDIOM_FIRST_CHAR = {} # 所有成语包括词语的首字字典
|
||||
AVALIABLE_IDIOM_FIRST_CHAR = {} # 真正有效的成语首字字典
|
||||
|
||||
__inited = False
|
||||
|
||||
@ -153,10 +157,15 @@ class IdiomGame:
|
||||
'''
|
||||
将一个新词加入到词语列表中
|
||||
'''
|
||||
await db_manager.execute_by_sql_file(
|
||||
ROOT_PATH / "sql" / "insert_custom_word.sql",
|
||||
(word,)
|
||||
)
|
||||
if word not in cls.ALL_WORDS:
|
||||
cls.ALL_WORDS.append(word)
|
||||
if word[0] not in cls.IDIOM_FIRST_CHAR:
|
||||
cls.IDIOM_FIRST_CHAR[word[0]] = []
|
||||
cls.IDIOM_FIRST_CHAR[word[0]].append(word)
|
||||
# await db_manager.execute_by_sql_file(
|
||||
# ROOT_PATH / "sql" / "insert_custom_word.sql",
|
||||
# (word,)
|
||||
# )
|
||||
|
||||
def be_able_to_play(self) -> bool:
|
||||
if self.last_play_date != datetime.date.today():
|
||||
@ -169,10 +178,11 @@ class IdiomGame:
|
||||
|
||||
@staticmethod
|
||||
async def random_idiom() -> str:
|
||||
result = await db_manager.query_by_sql_file(
|
||||
ROOT_PATH / "sql" / "random_choose_idiom.sql"
|
||||
)
|
||||
return result[0]["idiom"]
|
||||
# result = await db_manager.query_by_sql_file(
|
||||
# ROOT_PATH / "sql" / "random_choose_idiom.sql"
|
||||
# )
|
||||
# return result[0]["idiom"]
|
||||
return secrets.choice(IdiomGame.ALL_IDIOMS)
|
||||
|
||||
async def choose_start_idiom(self) -> str:
|
||||
"""
|
||||
@ -257,11 +267,12 @@ class IdiomGame:
|
||||
"""
|
||||
判断是否有成语可以接
|
||||
"""
|
||||
result = await db_manager.query_by_sql_file(
|
||||
ROOT_PATH / "sql" / "is_nextable.sql",
|
||||
(last_char,)
|
||||
)
|
||||
return result[0]["DEED"] == 1
|
||||
# result = await db_manager.query_by_sql_file(
|
||||
# ROOT_PATH / "sql" / "is_nextable.sql",
|
||||
# (last_char,)
|
||||
# )
|
||||
# return result[0]["DEED"] == 1
|
||||
return last_char in IdiomGame.AVALIABLE_IDIOM_FIRST_CHAR
|
||||
|
||||
def add_already_idiom(self, idiom: str):
|
||||
if idiom in self.already_idioms:
|
||||
@ -293,12 +304,13 @@ class IdiomGame:
|
||||
state.append(TryVerifyState.WRONG_FIRST_CHAR)
|
||||
return state
|
||||
# 成语是否存在
|
||||
result = await db_manager.query_by_sql_file(
|
||||
ROOT_PATH / "sql" / "query_idiom.sql",
|
||||
(idiom, idiom, idiom)
|
||||
)
|
||||
status_result = result[0]["status"]
|
||||
if status_result == -1:
|
||||
# result = await db_manager.query_by_sql_file(
|
||||
# ROOT_PATH / "sql" / "query_idiom.sql",
|
||||
# (idiom, idiom, idiom)
|
||||
# )
|
||||
# status_result = result[0]["status"]
|
||||
# if status_result == -1:
|
||||
if idiom not in IdiomGame.ALL_IDIOMS and idiom not in IdiomGame.ALL_WORDS:
|
||||
logger.info(f"用户 {user_id} 发送了未知词语 {idiom},正在使用 LLM 进行验证")
|
||||
try:
|
||||
if not await IdiomGameLLM.verify_idiom_with_llm(idiom):
|
||||
@ -320,7 +332,8 @@ class IdiomGame:
|
||||
self.last_idiom = idiom
|
||||
self.last_char = idiom[-1]
|
||||
self.add_score(user_id, 1 * score_k) # 先加 1 分
|
||||
if status_result == 1:
|
||||
# if status_result == 1:
|
||||
if idiom in IdiomGame.ALL_IDIOMS:
|
||||
state.append(TryVerifyState.VERIFIED_AND_REAL)
|
||||
self.add_score(user_id, 4 * score_k) # 再加 4 分
|
||||
self.remain_rounds -= 1
|
||||
@ -357,22 +370,26 @@ class IdiomGame:
|
||||
|
||||
@classmethod
|
||||
async def random_idiom_starting_with(cls, first_char: str) -> Optional[str]:
|
||||
# await cls.init_lexicon()
|
||||
# result = await db_manager.query_by_sql_file(
|
||||
# ROOT_PATH / "sql" / "query_idiom_start_with.sql",
|
||||
# (first_char,)
|
||||
# )
|
||||
# if len(result) == 0:
|
||||
# return None
|
||||
# return result[0]["idiom"]
|
||||
await cls.init_lexicon()
|
||||
result = await db_manager.query_by_sql_file(
|
||||
ROOT_PATH / "sql" / "query_idiom_start_with.sql",
|
||||
(first_char,)
|
||||
)
|
||||
if len(result) == 0:
|
||||
if first_char not in cls.AVALIABLE_IDIOM_FIRST_CHAR:
|
||||
return None
|
||||
return result[0]["idiom"]
|
||||
return secrets.choice(cls.AVALIABLE_IDIOM_FIRST_CHAR[first_char])
|
||||
|
||||
@classmethod
|
||||
async def init_lexicon(cls):
|
||||
if cls.__inited:
|
||||
return
|
||||
await db_manager.execute_by_sql_file(
|
||||
ROOT_PATH / "sql" / "create_table.sql"
|
||||
) # 确保数据库初始化
|
||||
# await db_manager.execute_by_sql_file(
|
||||
# ROOT_PATH / "sql" / "create_table.sql"
|
||||
# ) # 确保数据库初始化
|
||||
cls.__inited = True
|
||||
|
||||
# 成语大表
|
||||
@ -439,10 +456,10 @@ class IdiomGame:
|
||||
ALL_IDIOMS = [idiom["word"] for idiom in ALL_IDIOMS_INFOS] + THUOCL_IDIOMS
|
||||
ALL_IDIOMS = list(set(ALL_IDIOMS)) # 去重
|
||||
# 批量插入数据库
|
||||
await db_manager.execute_many_values_by_sql_file(
|
||||
ROOT_PATH / "sql" / "insert_idiom.sql",
|
||||
[(idiom,) for idiom in ALL_IDIOMS]
|
||||
)
|
||||
# await db_manager.execute_many_values_by_sql_file(
|
||||
# ROOT_PATH / "sql" / "insert_idiom.sql",
|
||||
# [(idiom,) for idiom in ALL_IDIOMS]
|
||||
# )
|
||||
|
||||
|
||||
# 其他四字词语表,仅表示可以有这个词
|
||||
@ -451,29 +468,33 @@ class IdiomGame:
|
||||
+ THUOCL_WORDS
|
||||
+ COMMON_WORDS
|
||||
)
|
||||
|
||||
cls.ALL_WORDS = ALL_WORDS + LOCAL_LLM_WORDS
|
||||
cls.ALL_IDIOMS = ALL_IDIOMS
|
||||
|
||||
# 插入数据库
|
||||
await db_manager.execute_many_values_by_sql_file(
|
||||
ROOT_PATH / "sql" / "insert_word.sql",
|
||||
[(word,) for word in ALL_WORDS]
|
||||
)
|
||||
# await db_manager.execute_many_values_by_sql_file(
|
||||
# ROOT_PATH / "sql" / "insert_word.sql",
|
||||
# [(word,) for word in ALL_WORDS]
|
||||
# )
|
||||
|
||||
# 自定义词语 LOCAL_LLM_WORDS 插入数据库,兼容用
|
||||
await db_manager.execute_many_values_by_sql_file(
|
||||
ROOT_PATH / "sql" / "insert_custom_word.sql",
|
||||
[(word,) for word in LOCAL_LLM_WORDS]
|
||||
)
|
||||
# await db_manager.execute_many_values_by_sql_file(
|
||||
# ROOT_PATH / "sql" / "insert_custom_word.sql",
|
||||
# [(word,) for word in LOCAL_LLM_WORDS]
|
||||
# )
|
||||
|
||||
# # 根据成语大表,划分出成语首字字典
|
||||
# for idiom in cls.ALL_IDIOMS + cls.ALL_WORDS:
|
||||
# if idiom[0] not in cls.IDIOM_FIRST_CHAR:
|
||||
# cls.IDIOM_FIRST_CHAR[idiom[0]] = []
|
||||
# cls.IDIOM_FIRST_CHAR[idiom[0]].append(idiom)
|
||||
# 根据成语大表,划分出成语首字字典
|
||||
for idiom in cls.ALL_IDIOMS + cls.ALL_WORDS:
|
||||
if idiom[0] not in cls.IDIOM_FIRST_CHAR:
|
||||
cls.IDIOM_FIRST_CHAR[idiom[0]] = []
|
||||
cls.IDIOM_FIRST_CHAR[idiom[0]].append(idiom)
|
||||
|
||||
# # 根据真正的成语大表,划分出有效成语首字字典
|
||||
# for idiom in cls.ALL_IDIOMS:
|
||||
# if idiom[0] not in cls.AVALIABLE_IDIOM_FIRST_CHAR:
|
||||
# cls.AVALIABLE_IDIOM_FIRST_CHAR[idiom[0]] = []
|
||||
# cls.AVALIABLE_IDIOM_FIRST_CHAR[idiom[0]].append(idiom)
|
||||
# 根据真正的成语大表,划分出有效成语首字字典
|
||||
for idiom in cls.ALL_IDIOMS:
|
||||
if idiom[0] not in cls.AVALIABLE_IDIOM_FIRST_CHAR:
|
||||
cls.AVALIABLE_IDIOM_FIRST_CHAR[idiom[0]] = []
|
||||
cls.AVALIABLE_IDIOM_FIRST_CHAR[idiom[0]].append(idiom)
|
||||
|
||||
|
||||
evt = on_alconna(
|
||||
|
||||
@ -10,7 +10,9 @@ from nonebot_plugin_alconna import (Alconna, Args, UniMessage, UniMsg,
|
||||
on_alconna)
|
||||
from nonebot_plugin_apscheduler import scheduler
|
||||
|
||||
from konabot.common import username
|
||||
from konabot.common.longtask import DepLongTaskTarget
|
||||
from konabot.common.pager import PagerQuery
|
||||
from konabot.plugins.kona_ph.core.message import (get_daily_report,
|
||||
get_daily_report_v2,
|
||||
get_puzzle_description,
|
||||
@ -76,7 +78,7 @@ async def _(target: DepLongTaskTarget):
|
||||
|
||||
|
||||
cmd_history = on_alconna(Alconna(
|
||||
"历史题目",
|
||||
"re:历史(题目|谜题)",
|
||||
Args["page?", int],
|
||||
Args["index_id?", str],
|
||||
), rule=to_me())
|
||||
@ -118,6 +120,24 @@ async def _(target: DepLongTaskTarget, index_id: str = "", page: int = 1):
|
||||
await target.send_message(msg)
|
||||
|
||||
|
||||
cmd_leadboard = on_alconna(Alconna(
|
||||
"re:此方(解谜|谜题)排行榜",
|
||||
Args["page?", int],
|
||||
))
|
||||
|
||||
@cmd_leadboard.handle()
|
||||
async def _(target: DepLongTaskTarget, page: int = 1):
|
||||
async with puzzle_manager() as manager:
|
||||
result = manager.get_leadboard(PagerQuery(page, 10))
|
||||
await target.send_message(result.to_unimessage(
|
||||
title="此方解谜排行榜",
|
||||
formatter=lambda data: (
|
||||
f"✨ {data[1]} 已完成 | "
|
||||
f"{username.get_username(data[0])}"
|
||||
)
|
||||
))
|
||||
|
||||
|
||||
@scheduler.scheduled_job("cron", hour="8")
|
||||
async def _():
|
||||
async with puzzle_manager() as manager:
|
||||
|
||||
@ -17,7 +17,7 @@ class PuzzleImageManager:
|
||||
21,
|
||||
)
|
||||
img_name = f"{id}{suffix}"
|
||||
(KONAPH_IMAGE_BASE / img_name).write_bytes(data)
|
||||
_ = (KONAPH_IMAGE_BASE / img_name).write_bytes(data)
|
||||
return img_name
|
||||
|
||||
def remove_puzzle_image(self, img_name: str):
|
||||
|
||||
@ -33,7 +33,7 @@ def get_puzzle_description(puzzle: Puzzle, with_answer: bool = False) -> UniMess
|
||||
)
|
||||
|
||||
result = result.text(f"\n\n出题者:{get_username(puzzle.author_id)}")
|
||||
|
||||
|
||||
if with_answer:
|
||||
result = result.text(f"\n\n题目答案:{puzzle.flag}")
|
||||
else:
|
||||
|
||||
@ -1,5 +1,6 @@
|
||||
import asyncio
|
||||
import datetime
|
||||
import functools
|
||||
import random
|
||||
import re
|
||||
from contextlib import asynccontextmanager
|
||||
@ -7,6 +8,7 @@ from contextlib import asynccontextmanager
|
||||
import nanoid
|
||||
from pydantic import BaseModel, Field, ValidationError
|
||||
|
||||
from konabot.common.pager import PagerQuery
|
||||
from konabot.plugins.kona_ph.core.path import KONAPH_DATA_JSON
|
||||
|
||||
|
||||
@ -112,6 +114,10 @@ class PuzzleManager(BaseModel):
|
||||
|
||||
index_id_counter: int = 1
|
||||
submissions: dict[str, dict[str, list[PuzzleSubmission]]] = {}
|
||||
"""
|
||||
类型:{ [raw_id: str]: { [user_id: str]: PuzzleSubmission[] } }
|
||||
"""
|
||||
|
||||
last_checked_date: datetime.date = Field(
|
||||
default_factory=lambda: get_today_date() - datetime.timedelta(days=1)
|
||||
)
|
||||
@ -235,6 +241,20 @@ class PuzzleManager(BaseModel):
|
||||
if p.author_id == user
|
||||
], key=lambda p: p.created_at, reverse=True)
|
||||
|
||||
def get_leadboard(self, pager: PagerQuery):
|
||||
return pager.apply(sorted([
|
||||
(user, sum((
|
||||
len([
|
||||
s for s in sl.get(user, [])
|
||||
if s.success
|
||||
]) for sl in self.submissions.values()
|
||||
)))
|
||||
for user in functools.reduce(
|
||||
lambda x, y: x | y,
|
||||
(set(sl.keys()) for sl in self.submissions.values()),
|
||||
)
|
||||
], key=lambda t: t[1], reverse=True))
|
||||
|
||||
|
||||
lock = asyncio.Lock()
|
||||
|
||||
|
||||
@ -10,7 +10,6 @@ from pydantic import BaseModel
|
||||
from konabot.common.longtask import DepLongTaskTarget
|
||||
from konabot.common.nb.exc import BotExceptionMessage
|
||||
from konabot.common.nb.extract_image import download_image_bytes
|
||||
from konabot.common.nb.qq_broadcast import qq_broadcast
|
||||
from konabot.common.username import get_username
|
||||
from konabot.plugins.kona_ph.core.image import get_image_manager
|
||||
from konabot.plugins.kona_ph.core.message import (get_puzzle_description, get_puzzle_hint_list,
|
||||
@ -19,6 +18,7 @@ from konabot.plugins.kona_ph.core.message import (get_puzzle_description, get_pu
|
||||
from konabot.plugins.kona_ph.core.storage import (Puzzle, PuzzleHint, PuzzleManager,
|
||||
get_today_date,
|
||||
puzzle_manager)
|
||||
from konabot.plugins.poster.service import broadcast
|
||||
|
||||
PUZZLE_PAGE_SIZE = 10
|
||||
|
||||
@ -344,7 +344,7 @@ def create_admin_commands():
|
||||
p = manager.get_today_puzzle(strong=True)
|
||||
if p is None:
|
||||
return await target.send_message("上架失败了orz,可能是没题了")
|
||||
await qq_broadcast(config.plugin_puzzle_playgroup, get_puzzle_description(p))
|
||||
await broadcast("每日谜题", get_puzzle_description(p))
|
||||
return await target.send_message("Ok!")
|
||||
|
||||
@cmd_admin.assign("preview")
|
||||
|
||||
@ -1,9 +1,10 @@
|
||||
from nonebot import on_message
|
||||
from nonebot_plugin_alconna import UniMessage, UniMsg
|
||||
from nonebot_plugin_alconna import UniMessage
|
||||
|
||||
evt = on_message()
|
||||
from konabot.common.nb.match_keyword import match_keyword
|
||||
|
||||
evt = on_message(rule=match_keyword("喵"))
|
||||
|
||||
@evt.handle()
|
||||
async def _(msg: UniMsg):
|
||||
if msg.extract_plain_text() == "喵":
|
||||
await evt.send(await UniMessage().text("喵").export())
|
||||
async def _():
|
||||
await evt.send(await UniMessage().text("喵").export())
|
||||
|
||||
@ -1,3 +1,4 @@
|
||||
import re
|
||||
import aiohttp
|
||||
import asyncio as asynkio
|
||||
from math import ceil
|
||||
@ -13,9 +14,10 @@ from nonebot_plugin_alconna import Alconna, Args, Subcommand, UniMessage, UniMsg
|
||||
from pydantic import BaseModel
|
||||
|
||||
from konabot.common.longtask import DepLongTaskTarget, LongTask, create_longtask, handle_long_task, longtask_data
|
||||
from konabot.common.ptimeparse import Parser
|
||||
from konabot.common.nb.match_keyword import match_keyword
|
||||
from konabot.plugins.simple_notify.ask_llm import ask_ai
|
||||
|
||||
evt = on_message()
|
||||
evt = on_message(rule=match_keyword(re.compile("^.+提醒我.+$")))
|
||||
|
||||
(Path(__file__).parent.parent.parent.parent / "data").mkdir(exist_ok=True)
|
||||
DATA_FILE_PATH = Path(__file__).parent.parent.parent.parent / "data" / "notify.json"
|
||||
@ -75,21 +77,12 @@ async def _(msg: UniMsg, mEvt: Event, target: DepLongTaskTarget):
|
||||
return
|
||||
|
||||
text = msg.extract_plain_text()
|
||||
if "提醒我" not in text:
|
||||
return
|
||||
|
||||
segments = text.split("提醒我", maxsplit=1)
|
||||
if len(segments) != 2:
|
||||
return
|
||||
|
||||
notify_time, notify_text = segments
|
||||
try:
|
||||
target_time = Parser().parse(notify_time)
|
||||
logger.info(f"从 {notify_time} 解析出了时间:{target_time}")
|
||||
except Exception:
|
||||
logger.info(f"无法从 {notify_time} 中解析出时间")
|
||||
return
|
||||
if not notify_text:
|
||||
target_time, notify_text = await ask_ai(text)
|
||||
if target_time is None:
|
||||
return
|
||||
|
||||
await create_longtask(
|
||||
|
||||
158
konabot/plugins/simple_notify/ask_llm.py
Normal file
158
konabot/plugins/simple_notify/ask_llm.py
Normal file
@ -0,0 +1,158 @@
|
||||
import datetime
|
||||
import json
|
||||
import re
|
||||
|
||||
from loguru import logger
|
||||
|
||||
from konabot.common.apis.ali_content_safety import AlibabaGreen
|
||||
from konabot.common.llm import get_llm
|
||||
|
||||
|
||||
SYSTEM_PROMPT = """你是一个专门解析提醒请求的助手。请分析用户输入,识别其中是否包含提醒信息,并输出标准化的JSON格式结果。
|
||||
|
||||
输入格式通常是:"xxxx提醒我yyyy",其中:
|
||||
- xxxx 是用户提供的时间信息
|
||||
- yyyy 是提醒内容。有些时候用户会有一些需求。你可以在合理的范围进行衍生。
|
||||
|
||||
输出要求:
|
||||
- 必须是有效的JSON对象
|
||||
- 包含以下字段:
|
||||
* datetime: 如果是绝对时间,填入ISO 8601格式的日期时间字符串;否则为null
|
||||
* datetime_delta: 如果是相对时间,填入ISO 8601持续时间格式;否则为null
|
||||
* datetime_delta_minus: 如果时间偏移量是负数,则此项为 true,否则为 false
|
||||
* content: 提醒内容的字符串
|
||||
* is_notice: 布尔值,表示这是否是真正的提醒请求
|
||||
|
||||
时间处理规则:
|
||||
- 绝对时间示例:如果 xxxx 输入了非常明确的时间点,如"2024年12月25日" → 转换为具体datetime
|
||||
- 相对时间示例:如果 xxxx 没有输入非常明确的时间点,如"10分钟后"、"2小时后"、"3天后" → 转换为datetime_delta
|
||||
- 如果用户输入了需要计算的时间,你需要计算出正确的结果,如"10分钟后的8分钟前" → 转换为 “PT2M”
|
||||
- zzzz 是系统提供的时间,每句话肯定都有,这不是你判断相对或绝对时间的依据,需严格按照 xxx 来判断
|
||||
- datetime和datetime_delta有且仅有一个不为null
|
||||
|
||||
时间格式要求:
|
||||
- datetime: "YYYY-MM-DDTHH:MM:SS" (ISO 8601)
|
||||
- datetime_delta: "PxYxMxDTxHxMxS" 格式 (如"PT1H30M"表示1小时30分钟,"P3DT4H"表示三天四小时,"P5MT2M"表示五个月两分钟)
|
||||
|
||||
判断标准:
|
||||
- is_notice=true: 明确包含时间+提醒内容的请求
|
||||
- is_notice=false: 闲聊、疑问句、或不符合提醒格式的内容
|
||||
|
||||
示例:
|
||||
用户:"明天下午2点提醒我开会"
|
||||
输出:{"datetime": "2024-01-16T14:00:00", "datetime_delta": null,
|
||||
"datetime_delta_minus": false, "content": "开会", "is_notice": true}
|
||||
|
||||
用户:"5分钟后提醒我关火"
|
||||
输出:{"datetime": null, "datetime_delta": "PT5M", "datetime_delta_minus": false, "content": "关火", "is_notice": true}
|
||||
|
||||
用户:"5分钟前提醒我关火"
|
||||
输出:{"datetime": null, "datetime_delta": "PT5M", "datetime_delta_minus": true, "content": "关火", "is_notice": true}
|
||||
|
||||
用户:"五百年后提醒我关火"
|
||||
输出:{"datetime": null, "datetime_delta": "P500Y", "datetime_delta_minus": false, "content": "关火", "is_notice": true}
|
||||
|
||||
用户:"昨天提醒我关火"
|
||||
输出:{"datetime": null, "datetime_delta": "P1D", "datetime_delta_minus": true, "content": "关火", "is_notice": true}
|
||||
|
||||
用户:"什么是提醒功能?"
|
||||
输出:{"datetime": null, "datetime_delta": null, "datetime_delta_minus": false, "content": "", "is_notice": false}
|
||||
|
||||
用户:"过一会会,用可爱的语气提醒我该睡觉了"
|
||||
输出:{"datetime": null, "datetime_delta": "PT10M", "datetime_delta_minus": false, "content": "呼呼!该睡觉了哦!ヾ(•ω•`)o", "is_notice": true}
|
||||
|
||||
请严格按照上述格式输出JSON,不要添加任何其他文字说明。现在是 DATETIME"""
|
||||
|
||||
pt_pattern = re.compile(
|
||||
r"^P"
|
||||
r"((?P<year>\d+)Y)?"
|
||||
r"((?P<month>\d+)M)?"
|
||||
r"((?P<day>\d+)D)?"
|
||||
r"(T((?P<hour>\d+)H)?"
|
||||
r"((?P<minute>\d+)M)?"
|
||||
r"((?P<second>\d+)S)?)?$"
|
||||
)
|
||||
|
||||
|
||||
def tryint(s: str | None):
|
||||
if s:
|
||||
if re.match(r"^\d+$", s):
|
||||
return int(s)
|
||||
return 0
|
||||
|
||||
|
||||
async def ask_ai(expression: str, now: datetime.datetime | None = None) -> tuple[datetime.datetime | None, str]:
|
||||
if now is None:
|
||||
now = datetime.datetime.now()
|
||||
prompt = SYSTEM_PROMPT.replace("DATETIME", f"{now}, 星期 {now.weekday() + 1}")
|
||||
|
||||
is_safe = await AlibabaGreen.detect(expression)
|
||||
if not is_safe:
|
||||
logger.info(f"提醒功能:消息被阿里绿网拦截 message={expression}")
|
||||
return None, ""
|
||||
|
||||
llm = get_llm("qwen3-max")
|
||||
message = await llm.chat([
|
||||
{ "role": "system", "content": prompt },
|
||||
{ "role": "user", "content": expression },
|
||||
])
|
||||
result = message.content
|
||||
if result is None:
|
||||
return (None, "")
|
||||
|
||||
try:
|
||||
data = json.loads(result)
|
||||
except json.JSONDecodeError:
|
||||
logger.info(f"提醒功能:解析 AI 返回值时出现问题 raw={result}")
|
||||
return (None, "")
|
||||
|
||||
datetime_absolute = data.get("datetime", None)
|
||||
datetime_delta = data.get("datetime_delta", None)
|
||||
is_minus = data.get("datetime_delta_minus", False)
|
||||
content = data.get("content", "")
|
||||
is_notice = data.get("is_notice", False)
|
||||
|
||||
if not is_notice:
|
||||
return (None, "")
|
||||
if datetime_absolute:
|
||||
try:
|
||||
res = datetime.datetime.strptime(datetime_absolute, "%Y-%m-%dT%H:%M:%S"), content
|
||||
logger.info(f"提醒功能:使用绝对时间解析 AI 返回值 raw={result} target={res[0]}")
|
||||
return res
|
||||
except ValueError:
|
||||
pass
|
||||
|
||||
if datetime_delta and (match := pt_pattern.match(datetime_delta)):
|
||||
years = tryint(match.group("year"))
|
||||
months = tryint(match.group("month"))
|
||||
days = tryint(match.group("day"))
|
||||
hours = tryint(match.group("hour"))
|
||||
minutes = tryint(match.group("minute"))
|
||||
seconds = tryint(match.group("second"))
|
||||
|
||||
dt = datetime.timedelta(days=days, hours=hours, minutes=minutes, seconds=seconds)
|
||||
if is_minus:
|
||||
dt = -dt
|
||||
if is_minus:
|
||||
now2 = now.replace(year=now.year - years)
|
||||
m = now2.month
|
||||
if (months - m) >= 0:
|
||||
neg_months = -(m - months - 1)
|
||||
neg_years = (neg_months + 11) // 12
|
||||
target_month = 12 - ((neg_months - 1) % 12)
|
||||
now2 = now2.replace(year=now2.year - neg_years, month=target_month)
|
||||
else:
|
||||
now2 = now2.replace(month=m - months)
|
||||
else:
|
||||
now2 = now.replace(year=now.year + years)
|
||||
m = now2.month
|
||||
now2 = now2.replace(
|
||||
year=now2.year + (m + months - 1) // 12,
|
||||
month=(m + months - 1) % 12 + 1
|
||||
)
|
||||
logger.info(f"提醒功能:使用相对时间解析 AI 返回值 raw={result} target={now2+dt}")
|
||||
return (now2 + dt, content)
|
||||
|
||||
logger.warning(f"提醒功能:解析 AI 返回值时没有找到解析方法 raw={result}")
|
||||
return (None, "")
|
||||
|
||||
71
konabot/plugins/tqszm/__init__.py
Normal file
71
konabot/plugins/tqszm/__init__.py
Normal file
@ -0,0 +1,71 @@
|
||||
"""
|
||||
提取首字母
|
||||
|
||||
...谁需要啊!!!
|
||||
"""
|
||||
|
||||
import functools
|
||||
|
||||
from loguru import logger
|
||||
from nonebot import on_message
|
||||
from nonebot.rule import KeywordsRule, Rule, ToMeRule
|
||||
from nonebot.adapters.onebot.v11.bot import Bot as OBBot
|
||||
from nonebot.adapters.onebot.v11.event import MessageEvent as OBMessageEvent
|
||||
from nonebot.adapters.onebot.v11.message import MessageSegment as OBMessageSegment
|
||||
from nonebot_plugin_alconna import Text, UniMsg
|
||||
from nonebot_plugin_alconna.uniseg.adapters.onebot11.builder import Onebot11MessageBuilder
|
||||
from pypinyin import pinyin, Style as PinyinStyle
|
||||
|
||||
from konabot.common.longtask import DepLongTaskTarget
|
||||
from konabot.common.apis.ali_content_safety import AlibabaGreen
|
||||
|
||||
|
||||
keywords = ("szmtq", "tqszm", "提取首字母", "首字母提取", )
|
||||
|
||||
|
||||
cmd_tqszm = on_message(rule=Rule(ToMeRule(), KeywordsRule(*keywords)))
|
||||
|
||||
@cmd_tqszm.handle()
|
||||
async def _(target: DepLongTaskTarget, msg: UniMsg, evt: OBMessageEvent | None = None, bot: OBBot | None = None):
|
||||
texts = ""
|
||||
command_occurred = False
|
||||
is_reply_mode = False
|
||||
|
||||
if evt is not None and bot is not None:
|
||||
reply = await Onebot11MessageBuilder().extract_reply(evt, bot)
|
||||
if reply is not None:
|
||||
is_reply_mode = True
|
||||
for seg in reply.data.get('msg', []):
|
||||
if isinstance(seg, OBMessageSegment) and seg.type == 'text':
|
||||
segtext: str = seg.data.get('text', '')
|
||||
texts += segtext
|
||||
|
||||
if is_reply_mode:
|
||||
for seg in msg:
|
||||
if isinstance(seg, Text):
|
||||
if all(seg.text.strip() != k for k in keywords):
|
||||
return
|
||||
else:
|
||||
for seg in msg:
|
||||
if isinstance(seg, Text):
|
||||
if command_occurred:
|
||||
texts += seg.text
|
||||
continue
|
||||
if any(seg.text.startswith(w) for w in keywords):
|
||||
command_occurred = True
|
||||
texts += next(
|
||||
seg.text.removeprefix(w) for w in keywords
|
||||
if seg.text.startswith(w)
|
||||
).removeprefix(" ")
|
||||
|
||||
result = pinyin(texts, style=PinyinStyle.FIRST_LETTER, errors=lambda x: x)
|
||||
print(result)
|
||||
result_text = functools.reduce(lambda x, y: x + y, [
|
||||
p[0] for p in result if len(p) > 0
|
||||
], "")
|
||||
|
||||
if not await AlibabaGreen.detect(result_text):
|
||||
logger.info(f"对首字母序列的检测未通过安全测试 RAW={texts} FORMATTED={result_text}")
|
||||
|
||||
await target.send_message(result_text, at=False)
|
||||
|
||||
304
poetry.lock
generated
304
poetry.lock
generated
@ -20,6 +20,23 @@ type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "aiofiles"
|
||||
version = "24.1.0"
|
||||
description = "File support for asyncio."
|
||||
optional = false
|
||||
python-versions = ">=3.8"
|
||||
groups = ["main"]
|
||||
files = [
|
||||
{file = "aiofiles-24.1.0-py3-none-any.whl", hash = "sha256:b4ec55f4195e3eb5d7abd1bf7e061763e864dd4954231fb8539a0ef8bb8260e5"},
|
||||
{file = "aiofiles-24.1.0.tar.gz", hash = "sha256:22a075c9e5a3810f0c2e48f3008c94d68c65d763b9b03857924c99e57355166c"},
|
||||
]
|
||||
|
||||
[package.source]
|
||||
type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "aiohappyeyeballs"
|
||||
version = "2.6.1"
|
||||
@ -233,6 +250,187 @@ type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "alibabacloud-credentials"
|
||||
version = "1.0.3"
|
||||
description = "The alibabacloud credentials module of alibabaCloud Python SDK."
|
||||
optional = false
|
||||
python-versions = ">=3.7"
|
||||
groups = ["main"]
|
||||
files = [
|
||||
{file = "alibabacloud-credentials-1.0.3.tar.gz", hash = "sha256:9d8707e96afc6f348e23f5677ed15a21c2dfce7cfe6669776548ee4c80e1dfaf"},
|
||||
{file = "alibabacloud_credentials-1.0.3-py3-none-any.whl", hash = "sha256:30c8302f204b663c655d97e1c283ee9f9f84a6257d7901b931477d6cf34445a8"},
|
||||
]
|
||||
|
||||
[package.dependencies]
|
||||
aiofiles = ">=22.1.0,<25.0.0"
|
||||
alibabacloud-credentials-api = ">=1.0.0,<2.0.0"
|
||||
alibabacloud-tea = ">=0.4.0"
|
||||
APScheduler = ">=3.10.0,<4.0.0"
|
||||
|
||||
[package.source]
|
||||
type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "alibabacloud-credentials-api"
|
||||
version = "1.0.0"
|
||||
description = "Alibaba Cloud Gateway SPI SDK Library for Python"
|
||||
optional = false
|
||||
python-versions = ">=3.7"
|
||||
groups = ["main"]
|
||||
files = [
|
||||
{file = "alibabacloud-credentials-api-1.0.0.tar.gz", hash = "sha256:8c340038d904f0218d7214a8f4088c31912bfcf279af2cbc7d9be4897a97dd2f"},
|
||||
]
|
||||
|
||||
[package.source]
|
||||
type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "alibabacloud-endpoint-util"
|
||||
version = "0.0.4"
|
||||
description = "The endpoint-util module of alibabaCloud Python SDK."
|
||||
optional = false
|
||||
python-versions = "*"
|
||||
groups = ["main"]
|
||||
files = [
|
||||
{file = "alibabacloud_endpoint_util-0.0.4.tar.gz", hash = "sha256:a593eb8ddd8168d5dc2216cd33111b144f9189fcd6e9ca20e48f358a739bbf90"},
|
||||
]
|
||||
|
||||
[package.source]
|
||||
type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "alibabacloud-gateway-spi"
|
||||
version = "0.0.3"
|
||||
description = "Alibaba Cloud Gateway SPI SDK Library for Python"
|
||||
optional = false
|
||||
python-versions = ">=3.6"
|
||||
groups = ["main"]
|
||||
files = [
|
||||
{file = "alibabacloud_gateway_spi-0.0.3.tar.gz", hash = "sha256:10d1c53a3fc5f87915fbd6b4985b98338a776e9b44a0263f56643c5048223b8b"},
|
||||
]
|
||||
|
||||
[package.dependencies]
|
||||
alibabacloud_credentials = ">=0.3.4"
|
||||
|
||||
[package.source]
|
||||
type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "alibabacloud-green20220302"
|
||||
version = "3.0.1"
|
||||
description = "Alibaba Cloud Green (20220302) SDK Library for Python"
|
||||
optional = false
|
||||
python-versions = ">=3.6"
|
||||
groups = ["main"]
|
||||
files = [
|
||||
{file = "alibabacloud_green20220302-3.0.1-py3-none-any.whl", hash = "sha256:01cd4a402f8de45423cf5c52c59fc696255ef5f291c3e1e10ac4985033e3e1d9"},
|
||||
{file = "alibabacloud_green20220302-3.0.1.tar.gz", hash = "sha256:083ad03e7c553ec127b69ad5b039f42aaac10730784df319ead2f8a0d8ee928a"},
|
||||
]
|
||||
|
||||
[package.dependencies]
|
||||
alibabacloud-endpoint-util = ">=0.0.4,<1.0.0"
|
||||
alibabacloud-openapi-util = ">=0.2.2,<1.0.0"
|
||||
alibabacloud-tea-openapi = ">=0.3.16,<1.0.0"
|
||||
alibabacloud-tea-util = ">=0.3.13,<1.0.0"
|
||||
|
||||
[package.source]
|
||||
type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "alibabacloud-openapi-util"
|
||||
version = "0.2.2"
|
||||
description = "Aliyun Tea OpenApi Library for Python"
|
||||
optional = false
|
||||
python-versions = "*"
|
||||
groups = ["main"]
|
||||
files = [
|
||||
{file = "alibabacloud_openapi_util-0.2.2.tar.gz", hash = "sha256:ebbc3906f554cb4bf8f513e43e8a33e8b6a3d4a0ef13617a0e14c3dda8ef52a8"},
|
||||
]
|
||||
|
||||
[package.dependencies]
|
||||
alibabacloud_tea_util = ">=0.0.2"
|
||||
cryptography = ">=3.0.0"
|
||||
|
||||
[package.source]
|
||||
type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "alibabacloud-tea"
|
||||
version = "0.4.3"
|
||||
description = "The tea module of alibabaCloud Python SDK."
|
||||
optional = false
|
||||
python-versions = ">=3.7"
|
||||
groups = ["main"]
|
||||
files = [
|
||||
{file = "alibabacloud-tea-0.4.3.tar.gz", hash = "sha256:ec8053d0aa8d43ebe1deb632d5c5404339b39ec9a18a0707d57765838418504a"},
|
||||
]
|
||||
|
||||
[package.dependencies]
|
||||
aiohttp = ">=3.7.0,<4.0.0"
|
||||
requests = ">=2.21.0,<3.0.0"
|
||||
|
||||
[package.source]
|
||||
type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "alibabacloud-tea-openapi"
|
||||
version = "0.4.2"
|
||||
description = "Alibaba Cloud openapi SDK Library for Python"
|
||||
optional = false
|
||||
python-versions = ">=3.7"
|
||||
groups = ["main"]
|
||||
files = [
|
||||
{file = "alibabacloud_tea_openapi-0.4.2-py3-none-any.whl", hash = "sha256:c498065a297fd1972ed7709ef935c9ce1f9757f267f68933de6e63853f37366f"},
|
||||
{file = "alibabacloud_tea_openapi-0.4.2.tar.gz", hash = "sha256:0a8d79374ca692469472355a125969c8a22cc5fb08328c75c26663ccf5c8b168"},
|
||||
]
|
||||
|
||||
[package.dependencies]
|
||||
alibabacloud-credentials = ">=1.0.2,<2.0.0"
|
||||
alibabacloud-gateway-spi = ">=0.0.2,<1.0.0"
|
||||
alibabacloud-tea-util = ">=0.3.13,<1.0.0"
|
||||
cryptography = ">=3.0.0,<45.0.0"
|
||||
darabonba-core = ">=1.0.3,<2.0.0"
|
||||
|
||||
[package.source]
|
||||
type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "alibabacloud-tea-util"
|
||||
version = "0.3.14"
|
||||
description = "The tea-util module of alibabaCloud Python SDK."
|
||||
optional = false
|
||||
python-versions = ">=3.6"
|
||||
groups = ["main"]
|
||||
files = [
|
||||
{file = "alibabacloud_tea_util-0.3.14-py3-none-any.whl", hash = "sha256:10d3e5c340d8f7ec69dd27345eb2fc5a1dab07875742525edf07bbe86db93bfe"},
|
||||
{file = "alibabacloud_tea_util-0.3.14.tar.gz", hash = "sha256:708e7c9f64641a3c9e0e566365d2f23675f8d7c2a3e2971d9402ceede0408cdb"},
|
||||
]
|
||||
|
||||
[package.dependencies]
|
||||
alibabacloud-tea = ">=0.3.3"
|
||||
|
||||
[package.source]
|
||||
type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "annotated-doc"
|
||||
version = "0.0.3"
|
||||
@ -982,6 +1180,93 @@ type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "cryptography"
|
||||
version = "44.0.3"
|
||||
description = "cryptography is a package which provides cryptographic recipes and primitives to Python developers."
|
||||
optional = false
|
||||
python-versions = ">=3.7, !=3.9.0, !=3.9.1"
|
||||
groups = ["main"]
|
||||
files = [
|
||||
{file = "cryptography-44.0.3-cp37-abi3-macosx_10_9_universal2.whl", hash = "sha256:962bc30480a08d133e631e8dfd4783ab71cc9e33d5d7c1e192f0b7c06397bb88"},
|
||||
{file = "cryptography-44.0.3-cp37-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4ffc61e8f3bf5b60346d89cd3d37231019c17a081208dfbbd6e1605ba03fa137"},
|
||||
{file = "cryptography-44.0.3-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:58968d331425a6f9eedcee087f77fd3c927c88f55368f43ff7e0a19891f2642c"},
|
||||
{file = "cryptography-44.0.3-cp37-abi3-manylinux_2_28_aarch64.whl", hash = "sha256:e28d62e59a4dbd1d22e747f57d4f00c459af22181f0b2f787ea83f5a876d7c76"},
|
||||
{file = "cryptography-44.0.3-cp37-abi3-manylinux_2_28_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:af653022a0c25ef2e3ffb2c673a50e5a0d02fecc41608f4954176f1933b12359"},
|
||||
{file = "cryptography-44.0.3-cp37-abi3-manylinux_2_28_x86_64.whl", hash = "sha256:157f1f3b8d941c2bd8f3ffee0af9b049c9665c39d3da9db2dc338feca5e98a43"},
|
||||
{file = "cryptography-44.0.3-cp37-abi3-manylinux_2_34_aarch64.whl", hash = "sha256:c6cd67722619e4d55fdb42ead64ed8843d64638e9c07f4011163e46bc512cf01"},
|
||||
{file = "cryptography-44.0.3-cp37-abi3-manylinux_2_34_x86_64.whl", hash = "sha256:b424563394c369a804ecbee9b06dfb34997f19d00b3518e39f83a5642618397d"},
|
||||
{file = "cryptography-44.0.3-cp37-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:c91fc8e8fd78af553f98bc7f2a1d8db977334e4eea302a4bfd75b9461c2d8904"},
|
||||
{file = "cryptography-44.0.3-cp37-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:25cd194c39fa5a0aa4169125ee27d1172097857b27109a45fadc59653ec06f44"},
|
||||
{file = "cryptography-44.0.3-cp37-abi3-win32.whl", hash = "sha256:3be3f649d91cb182c3a6bd336de8b61a0a71965bd13d1a04a0e15b39c3d5809d"},
|
||||
{file = "cryptography-44.0.3-cp37-abi3-win_amd64.whl", hash = "sha256:3883076d5c4cc56dbef0b898a74eb6992fdac29a7b9013870b34efe4ddb39a0d"},
|
||||
{file = "cryptography-44.0.3-cp39-abi3-macosx_10_9_universal2.whl", hash = "sha256:5639c2b16764c6f76eedf722dbad9a0914960d3489c0cc38694ddf9464f1bb2f"},
|
||||
{file = "cryptography-44.0.3-cp39-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f3ffef566ac88f75967d7abd852ed5f182da252d23fac11b4766da3957766759"},
|
||||
{file = "cryptography-44.0.3-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:192ed30fac1728f7587c6f4613c29c584abdc565d7417c13904708db10206645"},
|
||||
{file = "cryptography-44.0.3-cp39-abi3-manylinux_2_28_aarch64.whl", hash = "sha256:7d5fe7195c27c32a64955740b949070f21cba664604291c298518d2e255931d2"},
|
||||
{file = "cryptography-44.0.3-cp39-abi3-manylinux_2_28_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:3f07943aa4d7dad689e3bb1638ddc4944cc5e0921e3c227486daae0e31a05e54"},
|
||||
{file = "cryptography-44.0.3-cp39-abi3-manylinux_2_28_x86_64.whl", hash = "sha256:cb90f60e03d563ca2445099edf605c16ed1d5b15182d21831f58460c48bffb93"},
|
||||
{file = "cryptography-44.0.3-cp39-abi3-manylinux_2_34_aarch64.whl", hash = "sha256:ab0b005721cc0039e885ac3503825661bd9810b15d4f374e473f8c89b7d5460c"},
|
||||
{file = "cryptography-44.0.3-cp39-abi3-manylinux_2_34_x86_64.whl", hash = "sha256:3bb0847e6363c037df8f6ede57d88eaf3410ca2267fb12275370a76f85786a6f"},
|
||||
{file = "cryptography-44.0.3-cp39-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:b0cc66c74c797e1db750aaa842ad5b8b78e14805a9b5d1348dc603612d3e3ff5"},
|
||||
{file = "cryptography-44.0.3-cp39-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:6866df152b581f9429020320e5eb9794c8780e90f7ccb021940d7f50ee00ae0b"},
|
||||
{file = "cryptography-44.0.3-cp39-abi3-win32.whl", hash = "sha256:c138abae3a12a94c75c10499f1cbae81294a6f983b3af066390adee73f433028"},
|
||||
{file = "cryptography-44.0.3-cp39-abi3-win_amd64.whl", hash = "sha256:5d186f32e52e66994dce4f766884bcb9c68b8da62d61d9d215bfe5fb56d21334"},
|
||||
{file = "cryptography-44.0.3-pp310-pypy310_pp73-macosx_10_9_x86_64.whl", hash = "sha256:cad399780053fb383dc067475135e41c9fe7d901a97dd5d9c5dfb5611afc0d7d"},
|
||||
{file = "cryptography-44.0.3-pp310-pypy310_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:21a83f6f35b9cc656d71b5de8d519f566df01e660ac2578805ab245ffd8523f8"},
|
||||
{file = "cryptography-44.0.3-pp310-pypy310_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:fc3c9babc1e1faefd62704bb46a69f359a9819eb0292e40df3fb6e3574715cd4"},
|
||||
{file = "cryptography-44.0.3-pp310-pypy310_pp73-manylinux_2_34_aarch64.whl", hash = "sha256:e909df4053064a97f1e6565153ff8bb389af12c5c8d29c343308760890560aff"},
|
||||
{file = "cryptography-44.0.3-pp310-pypy310_pp73-manylinux_2_34_x86_64.whl", hash = "sha256:dad80b45c22e05b259e33ddd458e9e2ba099c86ccf4e88db7bbab4b747b18d06"},
|
||||
{file = "cryptography-44.0.3-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:479d92908277bed6e1a1c69b277734a7771c2b78633c224445b5c60a9f4bc1d9"},
|
||||
{file = "cryptography-44.0.3-pp311-pypy311_pp73-macosx_10_9_x86_64.whl", hash = "sha256:896530bc9107b226f265effa7ef3f21270f18a2026bc09fed1ebd7b66ddf6375"},
|
||||
{file = "cryptography-44.0.3-pp311-pypy311_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:9b4d4a5dbee05a2c390bf212e78b99434efec37b17a4bff42f50285c5c8c9647"},
|
||||
{file = "cryptography-44.0.3-pp311-pypy311_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:02f55fb4f8b79c1221b0961488eaae21015b69b210e18c386b69de182ebb1259"},
|
||||
{file = "cryptography-44.0.3-pp311-pypy311_pp73-manylinux_2_34_aarch64.whl", hash = "sha256:dd3db61b8fe5be220eee484a17233287d0be6932d056cf5738225b9c05ef4fff"},
|
||||
{file = "cryptography-44.0.3-pp311-pypy311_pp73-manylinux_2_34_x86_64.whl", hash = "sha256:978631ec51a6bbc0b7e58f23b68a8ce9e5f09721940933e9c217068388789fe5"},
|
||||
{file = "cryptography-44.0.3-pp311-pypy311_pp73-win_amd64.whl", hash = "sha256:5d20cc348cca3a8aa7312f42ab953a56e15323800ca3ab0706b8cd452a3a056c"},
|
||||
{file = "cryptography-44.0.3.tar.gz", hash = "sha256:fe19d8bc5536a91a24a8133328880a41831b6c5df54599a8417b62fe015d3053"},
|
||||
]
|
||||
|
||||
[package.dependencies]
|
||||
cffi = {version = ">=1.12", markers = "platform_python_implementation != \"PyPy\""}
|
||||
|
||||
[package.extras]
|
||||
docs = ["sphinx (>=5.3.0)", "sphinx-rtd-theme (>=3.0.0) ; python_version >= \"3.8\""]
|
||||
docstest = ["pyenchant (>=3)", "readme-renderer (>=30.0)", "sphinxcontrib-spelling (>=7.3.1)"]
|
||||
nox = ["nox (>=2024.4.15)", "nox[uv] (>=2024.3.2) ; python_version >= \"3.8\""]
|
||||
pep8test = ["check-sdist ; python_version >= \"3.8\"", "click (>=8.0.1)", "mypy (>=1.4)", "ruff (>=0.3.6)"]
|
||||
sdist = ["build (>=1.0.0)"]
|
||||
ssh = ["bcrypt (>=3.1.5)"]
|
||||
test = ["certifi (>=2024)", "cryptography-vectors (==44.0.3)", "pretend (>=0.7)", "pytest (>=7.4.0)", "pytest-benchmark (>=4.0)", "pytest-cov (>=2.10.1)", "pytest-xdist (>=3.5.0)"]
|
||||
test-randomorder = ["pytest-randomly"]
|
||||
|
||||
[package.source]
|
||||
type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "darabonba-core"
|
||||
version = "1.0.4"
|
||||
description = "The darabonba module of alibabaCloud Python SDK."
|
||||
optional = false
|
||||
python-versions = ">=3.7"
|
||||
groups = ["main"]
|
||||
files = [
|
||||
{file = "darabonba_core-1.0.4-py3-none-any.whl", hash = "sha256:4c3bc1d76d5af1087297b6afde8e960ea2f54f93e725e2df8453f0b4bb27dd24"},
|
||||
{file = "darabonba_core-1.0.4.tar.gz", hash = "sha256:6ede4e9bfd458148bab19ab2331716ae9b5c226ba5f6d221de6f88ee65704137"},
|
||||
]
|
||||
|
||||
[package.dependencies]
|
||||
aiohttp = ">=3.7.0,<4.0.0"
|
||||
alibabacloud-tea = "*"
|
||||
requests = ">=2.21.0,<3.0.0"
|
||||
|
||||
[package.source]
|
||||
type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "distro"
|
||||
version = "1.9.0"
|
||||
@ -3454,6 +3739,23 @@ type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "pypinyin"
|
||||
version = "0.55.0"
|
||||
description = "汉字拼音转换模块/工具."
|
||||
optional = false
|
||||
python-versions = ">=2.6, !=3.0.*, !=3.1.*, !=3.2.*, <4"
|
||||
groups = ["main"]
|
||||
files = [
|
||||
{file = "pypinyin-0.55.0-py2.py3-none-any.whl", hash = "sha256:d53b1e8ad2cdb815fb2cb604ed3123372f5a28c6f447571244aca36fc62a286f"},
|
||||
{file = "pypinyin-0.55.0.tar.gz", hash = "sha256:b5711b3a0c6f76e67408ec6b2e3c4987a3a806b7c528076e7c7b86fcf0eaa66b"},
|
||||
]
|
||||
|
||||
[package.source]
|
||||
type = "legacy"
|
||||
url = "https://pypi.tuna.tsinghua.edu.cn/simple"
|
||||
reference = "mirrors"
|
||||
|
||||
[[package]]
|
||||
name = "pytest"
|
||||
version = "9.0.1"
|
||||
@ -4681,4 +4983,4 @@ reference = "mirrors"
|
||||
[metadata]
|
||||
lock-version = "2.1"
|
||||
python-versions = ">=3.12,<4.0"
|
||||
content-hash = "2c341fdc0d5b29ad3b24516c46e036b2eff4c11e244047d114971039255c2ac4"
|
||||
content-hash = "59498c038a603c90f051d2f360cb9226ec0fc4470942c0a7cf34f832701f0ce7"
|
||||
|
||||
@ -29,6 +29,8 @@ dependencies = [
|
||||
"imageio (>=2.37.2,<3.0.0)",
|
||||
"aiosqlite (>=0.20.0,<1.0.0)",
|
||||
"sqlparse (>=0.5.0,<1.0.0)",
|
||||
"alibabacloud-green20220302 (>=3.0.1,<4.0.0)",
|
||||
"pypinyin (>=0.55.0,<0.56.0)",
|
||||
]
|
||||
|
||||
[tool.poetry]
|
||||
|
||||
@ -12,4 +12,5 @@ def filter(change: Change, path: str) -> bool:
|
||||
return False
|
||||
if Path(path).absolute().is_relative_to((base / ".git").absolute()):
|
||||
return False
|
||||
print(path)
|
||||
return True
|
||||
|
||||
Reference in New Issue
Block a user