zhenxun_bot/plugins/pixiv/data_source.py
2021-05-20 19:23:32 +08:00

85 lines
3.2 KiB
Python

import aiohttp
import aiofiles
from configs.path_config import IMAGE_PATH
from util.utils import get_local_proxy
from util.user_agent import get_user_agent
from bs4 import BeautifulSoup
import feedparser
from util.init_result import image
from asyncio.exceptions import TimeoutError
import platform
if platform.system() == 'Windows':
import asyncio
asyncio.set_event_loop_policy(asyncio.WindowsSelectorEventLoopPolicy())
async def get_pixiv_urls(mode: str, num: int = 5, date: str = '') -> 'list, list, int':
url = 'https://rsshub.app/pixiv/ranking/{}'
iurl = url.format(mode)
if date:
iurl += f'/{date}'
return await parser_data(iurl, num)
async def download_pixiv_imgs(urls: list, user_id: int) -> str:
result = ''
index = 0
for img in urls:
async with aiohttp.ClientSession(headers=get_user_agent()) as session:
for _ in range(3):
async with session.get(img, proxy=get_local_proxy(), timeout=3) as response:
async with aiofiles.open(IMAGE_PATH + f'temp/{user_id}_{index}_pixiv.jpg', 'wb') as f:
try:
await f.write(await response.read())
result += image(f'{user_id}_{index}_pixiv.jpg', 'temp')
index += 1
break
except TimeoutError:
# result += '\n这张图下载失败了..\n'
pass
else:
result += '\n这张图下载失败了..\n'
return result
async def search_pixiv_urls(keyword: str, num: int, order: str, r18: int) -> 'list, list':
url = 'https://rsshub.app/pixiv/search/{}/{}/{}'.format(keyword, order, r18)
return await parser_data(url, num)
async def parser_data(url: str, num: int) -> 'list, list, int':
text_list = []
urls = []
async with aiohttp.ClientSession() as session:
for _ in range(3):
try:
async with session.get(url, proxy=get_local_proxy(), timeout=4) as response:
data = feedparser.parse(await response.text())['entries']
break
except TimeoutError:
pass
else:
return ['网络不太好,也许过一会就好了'], [], 998
try:
if len(data) == 0:
return ['没有搜索到喔'], [], 997
if num > len(data):
num = len(data)
data = data[:num]
for data in data:
soup = BeautifulSoup(data['summary'], 'lxml')
title = "标题:" + data['title']
pl = soup.find_all('p')
author = pl[0].text.split('-')[0].strip()
imgs = []
text_list.append(f'{title}\n{author}\n')
for p in pl[1:]:
imgs.append(p.find('img').get('src'))
urls.append(imgs)
except ValueError:
return ['是网站坏了啊,也许过一会就好了'], [], 999
return text_list, urls, 200
# asyncio.get_event_loop().run_until_complete(get_pixiv_urls('day'))