mirror of
https://github.com/zhenxun-org/zhenxun_bot.git
synced 2025-12-15 14:22:55 +08:00
296 lines
9.3 KiB
Python
296 lines
9.3 KiB
Python
from asyncio import Semaphore
|
||
from collections.abc import Iterable
|
||
from typing import Any, ClassVar
|
||
from typing_extensions import Self
|
||
from urllib.parse import urlparse
|
||
|
||
import nonebot
|
||
from nonebot.utils import is_coroutine_callable
|
||
from tortoise import Tortoise
|
||
from tortoise.backends.base.client import BaseDBAsyncClient
|
||
from tortoise.connection import connections
|
||
from tortoise.models import Model as TortoiseModel
|
||
|
||
from zhenxun.configs.config import BotConfig
|
||
from zhenxun.utils.enum import DbLockType
|
||
from zhenxun.utils.exception import HookPriorityException
|
||
from zhenxun.utils.manager.priority_manager import PriorityLifecycle
|
||
|
||
from .cache import CacheRoot
|
||
from .log import logger
|
||
|
||
SCRIPT_METHOD = []
|
||
MODELS: list[str] = []
|
||
|
||
driver = nonebot.get_driver()
|
||
|
||
CACHE_FLAG = False
|
||
|
||
|
||
@driver.on_bot_connect
|
||
def _():
|
||
global CACHE_FLAG
|
||
CACHE_FLAG = True
|
||
|
||
|
||
class Model(TortoiseModel):
|
||
"""
|
||
自动添加模块
|
||
"""
|
||
|
||
sem_data: ClassVar[dict[str, dict[str, Semaphore]]] = {}
|
||
|
||
def __init_subclass__(cls, **kwargs):
|
||
if cls.__module__ not in MODELS:
|
||
MODELS.append(cls.__module__)
|
||
|
||
if func := getattr(cls, "_run_script", None):
|
||
SCRIPT_METHOD.append((cls.__module__, func))
|
||
if enable_lock := getattr(cls, "enable_lock", []):
|
||
"""创建锁"""
|
||
cls.sem_data[cls.__module__] = {}
|
||
for lock in enable_lock:
|
||
cls.sem_data[cls.__module__][lock] = Semaphore(1)
|
||
|
||
@classmethod
|
||
def get_semaphore(cls, lock_type: DbLockType):
|
||
return cls.sem_data.get(cls.__module__, {}).get(lock_type, None)
|
||
|
||
@classmethod
|
||
def get_cache_type(cls):
|
||
return getattr(cls, "cache_type", None) if CACHE_FLAG else None
|
||
|
||
@classmethod
|
||
async def create(
|
||
cls, using_db: BaseDBAsyncClient | None = None, **kwargs: Any
|
||
) -> Self:
|
||
return await super().create(using_db=using_db, **kwargs)
|
||
|
||
@classmethod
|
||
async def get_or_create(
|
||
cls,
|
||
defaults: dict | None = None,
|
||
using_db: BaseDBAsyncClient | None = None,
|
||
**kwargs: Any,
|
||
) -> tuple[Self, bool]:
|
||
result, is_create = await super().get_or_create(
|
||
defaults=defaults, using_db=using_db, **kwargs
|
||
)
|
||
if is_create and (cache_type := cls.get_cache_type()):
|
||
await CacheRoot.reload(cache_type)
|
||
return (result, is_create)
|
||
|
||
@classmethod
|
||
async def update_or_create(
|
||
cls,
|
||
defaults: dict | None = None,
|
||
using_db: BaseDBAsyncClient | None = None,
|
||
**kwargs: Any,
|
||
) -> tuple[Self, bool]:
|
||
result = await super().update_or_create(
|
||
defaults=defaults, using_db=using_db, **kwargs
|
||
)
|
||
if cache_type := cls.get_cache_type():
|
||
await CacheRoot.reload(cache_type)
|
||
return result
|
||
|
||
@classmethod
|
||
async def bulk_create( # type: ignore
|
||
cls,
|
||
objects: Iterable[Self],
|
||
batch_size: int | None = None,
|
||
ignore_conflicts: bool = False,
|
||
update_fields: Iterable[str] | None = None,
|
||
on_conflict: Iterable[str] | None = None,
|
||
using_db: BaseDBAsyncClient | None = None,
|
||
) -> list[Self]:
|
||
result = await super().bulk_create(
|
||
objects=objects,
|
||
batch_size=batch_size,
|
||
ignore_conflicts=ignore_conflicts,
|
||
update_fields=update_fields,
|
||
on_conflict=on_conflict,
|
||
using_db=using_db,
|
||
)
|
||
if cache_type := cls.get_cache_type():
|
||
await CacheRoot.reload(cache_type)
|
||
return result
|
||
|
||
@classmethod
|
||
async def bulk_update( # type: ignore
|
||
cls,
|
||
objects: Iterable[Self],
|
||
fields: Iterable[str],
|
||
batch_size: int | None = None,
|
||
using_db: BaseDBAsyncClient | None = None,
|
||
) -> int:
|
||
result = await super().bulk_update(
|
||
objects=objects,
|
||
fields=fields,
|
||
batch_size=batch_size,
|
||
using_db=using_db,
|
||
)
|
||
if cache_type := cls.get_cache_type():
|
||
await CacheRoot.reload(cache_type)
|
||
return result
|
||
|
||
async def save(
|
||
self,
|
||
using_db: BaseDBAsyncClient | None = None,
|
||
update_fields: Iterable[str] | None = None,
|
||
force_create: bool = False,
|
||
force_update: bool = False,
|
||
):
|
||
if getattr(self, "id", None) is None:
|
||
sem = self.get_semaphore(DbLockType.CREATE)
|
||
else:
|
||
sem = self.get_semaphore(DbLockType.UPDATE)
|
||
if sem:
|
||
async with sem:
|
||
await super().save(
|
||
using_db=using_db,
|
||
update_fields=update_fields,
|
||
force_create=force_create,
|
||
force_update=force_update,
|
||
)
|
||
else:
|
||
await super().save(
|
||
using_db=using_db,
|
||
update_fields=update_fields,
|
||
force_create=force_create,
|
||
force_update=force_update,
|
||
)
|
||
if CACHE_FLAG and (cache_type := getattr(self, "cache_type", None)):
|
||
await CacheRoot.reload(cache_type)
|
||
|
||
async def delete(self, using_db: BaseDBAsyncClient | None = None):
|
||
await super().delete(using_db=using_db)
|
||
if CACHE_FLAG and (cache_type := getattr(self, "cache_type", None)):
|
||
await CacheRoot.reload(cache_type)
|
||
|
||
|
||
class DbUrlIsNode(HookPriorityException):
|
||
"""
|
||
数据库链接地址为空
|
||
"""
|
||
|
||
pass
|
||
|
||
|
||
class DbConnectError(Exception):
|
||
"""
|
||
数据库连接错误
|
||
"""
|
||
|
||
pass
|
||
|
||
|
||
@PriorityLifecycle.on_startup(priority=1)
|
||
async def init():
|
||
if not BotConfig.db_url:
|
||
error = f"""
|
||
**********************************************************************
|
||
🌟 **************************** 配置为空 ************************* 🌟
|
||
🚀 请打开 WebUi 进行基础配置 🚀
|
||
🌐 配置地址:http://{driver.config.host}:{driver.config.port}/#/configure 🌐
|
||
***********************************************************************
|
||
***********************************************************************
|
||
"""
|
||
raise DbUrlIsNode("\n" + error.strip())
|
||
try:
|
||
db_url = BotConfig.db_url
|
||
url_scheme = db_url.split(":", 1)[0]
|
||
config = None
|
||
|
||
if url_scheme in ("postgres", "postgresql"):
|
||
# 解析 db_url
|
||
url = urlparse(db_url)
|
||
credentials = {
|
||
"host": url.hostname,
|
||
"port": url.port or 5432,
|
||
"user": url.username,
|
||
"password": url.password,
|
||
"database": url.path.lstrip("/"),
|
||
"minsize": 1,
|
||
"maxsize": 50,
|
||
}
|
||
config = {
|
||
"connections": {
|
||
"default": {
|
||
"engine": "tortoise.backends.asyncpg",
|
||
"credentials": credentials,
|
||
}
|
||
},
|
||
"apps": {
|
||
"models": {
|
||
"models": MODELS,
|
||
"default_connection": "default",
|
||
}
|
||
},
|
||
}
|
||
elif url_scheme in ("mysql", "mysql+aiomysql"):
|
||
url = urlparse(db_url)
|
||
credentials = {
|
||
"host": url.hostname,
|
||
"port": url.port or 3306,
|
||
"user": url.username,
|
||
"password": url.password,
|
||
"database": url.path.lstrip("/"),
|
||
"minsize": 1,
|
||
"maxsize": 50,
|
||
}
|
||
config = {
|
||
"connections": {
|
||
"default": {
|
||
"engine": "tortoise.backends.mysql",
|
||
"credentials": credentials,
|
||
}
|
||
},
|
||
"apps": {
|
||
"models": {
|
||
"models": MODELS,
|
||
"default_connection": "default",
|
||
}
|
||
},
|
||
}
|
||
else:
|
||
# sqlite 或其它,直接用 db_url
|
||
await Tortoise.init(
|
||
db_url=db_url,
|
||
modules={"models": MODELS},
|
||
timezone="Asia/Shanghai",
|
||
)
|
||
|
||
if config:
|
||
await Tortoise.init(config=config)
|
||
|
||
if SCRIPT_METHOD:
|
||
db = Tortoise.get_connection("default")
|
||
logger.debug(
|
||
"即将运行SCRIPT_METHOD方法, 合计 "
|
||
f"<u><y>{len(SCRIPT_METHOD)}</y></u> 个..."
|
||
)
|
||
sql_list = []
|
||
for module, func in SCRIPT_METHOD:
|
||
try:
|
||
sql = await func() if is_coroutine_callable(func) else func()
|
||
if sql:
|
||
sql_list += sql
|
||
except Exception as e:
|
||
logger.debug(f"{module} 执行SCRIPT_METHOD方法出错...", e=e)
|
||
for sql in sql_list:
|
||
logger.debug(f"执行SQL: {sql}")
|
||
try:
|
||
await db.execute_query_dict(sql)
|
||
except Exception as e:
|
||
logger.debug(f"执行SQL: {sql} 错误...", e=e)
|
||
if sql_list:
|
||
logger.debug("SCRIPT_METHOD方法执行完毕!")
|
||
await Tortoise.generate_schemas()
|
||
logger.info("Database loaded successfully!")
|
||
except Exception as e:
|
||
raise DbConnectError(f"数据库连接错误... e:{e}") from e
|
||
|
||
|
||
async def disconnect():
|
||
await connections.close_all() |