Compare commits
5 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| 3ce230adfe | |||
| 4f885554ca | |||
| 7ebcb8add4 | |||
| e18cc82792 | |||
| eb28cd0a0c |
BIN
assets/img/meme/snaur_1_base.png
Executable file
BIN
assets/img/meme/snaur_1_base.png
Executable file
Binary file not shown.
|
After Width: | Height: | Size: 1.1 MiB |
BIN
assets/img/meme/snaur_1_top.png
Executable file
BIN
assets/img/meme/snaur_1_top.png
Executable file
Binary file not shown.
|
After Width: | Height: | Size: 1008 KiB |
@ -45,7 +45,7 @@
|
||||
- 帧数必须为正整数(> 0)。
|
||||
- 若原始帧数 ≤ 指定帧数,则保留全部帧。
|
||||
|
||||
--s <速度>(可选)
|
||||
--speed <速度>(可选)
|
||||
- 调整 gif 图的速度。若为负数,则代表倒放
|
||||
|
||||
使用方式
|
||||
|
||||
39
konabot/plugins/bilibili_fetch/__init__.py
Normal file
39
konabot/plugins/bilibili_fetch/__init__.py
Normal file
@ -0,0 +1,39 @@
|
||||
import re
|
||||
|
||||
from nonebot import on_message
|
||||
from nonebot_plugin_alconna import Reference, Reply, UniMsg
|
||||
|
||||
from nonebot.adapters import Event
|
||||
|
||||
|
||||
matcher_fix = on_message()
|
||||
|
||||
pattern = (
|
||||
r"^(?:(?:av|cv)\d+|BV[a-zA-Z0-9]{10})|"
|
||||
r"(?:b23\.tv|bili(?:22|23|33|2233)\.cn|\.bilibili\.com|QQ小程序(?:&#93;|]|\])哔哩哔哩).{0,500}"
|
||||
)
|
||||
|
||||
|
||||
@matcher_fix.handle()
|
||||
async def _(msg: UniMsg, event: Event):
|
||||
to_search = msg.exclude(Reply, Reference).dump(json=True)
|
||||
to_search2 = msg.exclude(Reply, Reference).extract_plain_text()
|
||||
if not re.search(pattern, to_search) and not re.search(pattern, to_search2):
|
||||
return
|
||||
|
||||
from nonebot_plugin_analysis_bilibili import handle_analysis
|
||||
|
||||
await handle_analysis(event)
|
||||
|
||||
# b_url: str
|
||||
# b_page: str | None
|
||||
# b_time: str | None
|
||||
#
|
||||
# from nonebot_plugin_analysis_bilibili.analysis_bilibili import extract as bilibili_extract
|
||||
#
|
||||
# b_url, b_page, b_time = bilibili_extract(to_search)
|
||||
# if b_url is None:
|
||||
# return
|
||||
#
|
||||
# await matcher_fix.send(await UniMessage().text(b_url).export())
|
||||
|
||||
@ -1,25 +0,0 @@
|
||||
import re
|
||||
from loguru import logger
|
||||
from nonebot import on_message
|
||||
from nonebot_plugin_alconna import Reference, Reply, UniMsg
|
||||
|
||||
from nonebot.adapters import Event
|
||||
|
||||
|
||||
matcher_fix = on_message()
|
||||
|
||||
pattern = (
|
||||
r"^(?:(?:av|cv)\d+|BV[a-zA-Z0-9]{10})|"
|
||||
r"(?:b23\.tv|bili(?:22|23|33|2233)\.cn|\.bilibili\.com|QQ小程序(?:&#93;|]|\])哔哩哔哩).{0,500}"
|
||||
)
|
||||
|
||||
|
||||
@matcher_fix.handle()
|
||||
async def _(msg: UniMsg, event: Event):
|
||||
to_search = msg.exclude(Reply, Reference).dump(json=True)
|
||||
if not re.search(pattern, to_search):
|
||||
return
|
||||
logger.info("检测到有 Bilibili 相关的消息,直接进行一个调用")
|
||||
_module = __import__("nonebot_plugin_analysis_bilibili")
|
||||
await _module.handle_analysis(event)
|
||||
|
||||
@ -2,11 +2,11 @@ from io import BytesIO
|
||||
from typing import Iterable, cast
|
||||
|
||||
from nonebot import on_message
|
||||
from nonebot_plugin_alconna import (Alconna, Args, Field, MultiVar, Text,
|
||||
from nonebot_plugin_alconna import (Alconna, Args, Field, Image, MultiVar, Text,
|
||||
UniMessage, UniMsg, on_alconna)
|
||||
|
||||
from konabot.common.nb.extract_image import extract_image_from_message
|
||||
from konabot.plugins.memepack.drawing.display import draw_cao_display
|
||||
from konabot.common.nb.extract_image import PIL_Image, extract_image_from_message
|
||||
from konabot.plugins.memepack.drawing.display import draw_cao_display, draw_snaur_display
|
||||
from konabot.plugins.memepack.drawing.saying import (draw_cute_ten,
|
||||
draw_geimao, draw_mnk,
|
||||
draw_pt, draw_suan)
|
||||
@ -139,3 +139,17 @@ async def _(msg: UniMsg, evt: Event, bot: Bot):
|
||||
.text(err)
|
||||
.export()
|
||||
)
|
||||
|
||||
|
||||
snaur_display_cmd = on_alconna(Alconna(
|
||||
"卵总展示",
|
||||
Args["image", Image | None],
|
||||
))
|
||||
|
||||
@snaur_display_cmd.handle()
|
||||
async def _(img: PIL_Image):
|
||||
img_processed = await draw_snaur_display(img)
|
||||
img_data = BytesIO()
|
||||
img_processed.save(img_data, "PNG")
|
||||
await snaur_display_cmd.send(await UniMessage().image(raw=img_data).export())
|
||||
|
||||
|
||||
@ -4,10 +4,12 @@ from typing import Any, cast
|
||||
import cv2
|
||||
import numpy as np
|
||||
import PIL.Image
|
||||
import PIL.ImageChops
|
||||
import PIL.ImageEnhance
|
||||
|
||||
from konabot.common.path import ASSETS_PATH
|
||||
|
||||
cao_image = PIL.Image.open(ASSETS_PATH / "img" / "meme" / "caoimg1.png")
|
||||
cao_image = PIL.Image.open(ASSETS_PATH / "img" / "meme" / "caoimg1.png")
|
||||
CAO_QUAD_POINTS = np.float32(cast(Any, [
|
||||
[392, 540],
|
||||
[577, 557],
|
||||
@ -15,6 +17,16 @@ CAO_QUAD_POINTS = np.float32(cast(Any, [
|
||||
[381, 687],
|
||||
]))
|
||||
|
||||
snaur_image_base = PIL.Image.open(ASSETS_PATH / "img" / "meme" / "snaur_1_base.png")
|
||||
snaur_image_top = PIL.Image.open(ASSETS_PATH / "img" / "meme" / "snaur_1_top.png")
|
||||
SNAUR_RATIO = (1 / 2) ** .5
|
||||
SNAUR_QUAD_POINTS = np.float32(cast(Any, [
|
||||
[0, 466 ],
|
||||
[673, 471 ],
|
||||
[640, 1196],
|
||||
[106, 1280],
|
||||
]))
|
||||
|
||||
def _draw_cao_display(image: PIL.Image.Image):
|
||||
src = np.array(image.convert("RGB"))
|
||||
h, w = src.shape[:2]
|
||||
@ -43,3 +55,68 @@ def _draw_cao_display(image: PIL.Image.Image):
|
||||
|
||||
async def draw_cao_display(image: PIL.Image.Image):
|
||||
return await asyncio.to_thread(_draw_cao_display, image)
|
||||
|
||||
|
||||
def _draw_snaur_display(image: PIL.Image.Image):
|
||||
src = np.array(image.convert("RGB"))
|
||||
_h, _w = src.shape[:2]
|
||||
|
||||
if _w / _h < SNAUR_RATIO:
|
||||
_w_target = _w
|
||||
_h_target = int(_w / SNAUR_RATIO)
|
||||
else:
|
||||
_w_target = int(_h * SNAUR_RATIO)
|
||||
_h_target = _h
|
||||
|
||||
x_center = _w / 2
|
||||
y_center = _h / 2
|
||||
|
||||
x1 = int(x_center - _w_target / 2)
|
||||
x2 = int(x_center + _w_target / 2)
|
||||
y1 = int(y_center - _h_target / 2)
|
||||
y2 = int(y_center + _h_target / 2)
|
||||
|
||||
src = src[y1:y2, x1:x2, :]
|
||||
|
||||
h, w = src.shape[:2]
|
||||
src_points = np.float32(cast(Any, [
|
||||
[0, 0],
|
||||
[w, 0],
|
||||
[w, h],
|
||||
[0, h],
|
||||
]))
|
||||
dst_points = SNAUR_QUAD_POINTS
|
||||
M = cv2.getPerspectiveTransform(cast(Any, src_points), cast(Any, dst_points))
|
||||
output_size = snaur_image_top.size
|
||||
output_w, output_h = output_size
|
||||
warped = cv2.warpPerspective(
|
||||
src,
|
||||
M,
|
||||
(output_w, output_h),
|
||||
flags=cv2.INTER_LINEAR,
|
||||
borderMode=cv2.BORDER_CONSTANT,
|
||||
borderValue=(0, 0, 0)
|
||||
)
|
||||
|
||||
result = PIL.Image.fromarray(warped, 'RGB').convert('RGBA')
|
||||
result = PIL.ImageChops.multiply(result, snaur_image_base)
|
||||
|
||||
r, g, b, a = result.split()
|
||||
a = a.point(lambda p: int(p * 0.8))
|
||||
f2 = lambda p: int(p * 0.8 + 0.2)
|
||||
r = r.point(f2)
|
||||
g = g.point(f2)
|
||||
b = b.point(f2)
|
||||
result = PIL.Image.merge('RGBA', (r, g, b, a))
|
||||
result = PIL.Image.alpha_composite(snaur_image_base, result)
|
||||
|
||||
enhancer = PIL.ImageEnhance.Color(result)
|
||||
result = enhancer.enhance(0.85)
|
||||
|
||||
result = PIL.Image.alpha_composite(result, snaur_image_top)
|
||||
return result
|
||||
|
||||
|
||||
async def draw_snaur_display(image: PIL.Image.Image) -> PIL.Image.Image:
|
||||
return await asyncio.to_thread(_draw_snaur_display, image)
|
||||
|
||||
|
||||
Reference in New Issue
Block a user