diff --git a/misskaty/helper/subscene_helper.py b/misskaty/helper/subscene_helper.py index e8e15e41..1c058ba3 100644 --- a/misskaty/helper/subscene_helper.py +++ b/misskaty/helper/subscene_helper.py @@ -1,9 +1,9 @@ -import cfscrape +import cloudscraper from bs4 import BeautifulSoup async def down_page(url): - f = cfscrape.create_scraper() + f = cloudscraper.create_scraper() resp = f.get(url).text soup = BeautifulSoup(resp, "lxml") maindiv = soup.body.find("div", class_="subtitle").find("div", class_="top left") diff --git a/misskaty/plugins/dev.py b/misskaty/plugins/dev.py index 263a6dae..0574515f 100644 --- a/misskaty/plugins/dev.py +++ b/misskaty/plugins/dev.py @@ -7,7 +7,7 @@ import html import pickle import json import traceback -import cfscrape +import cloudscraper import aiohttp from datetime import datetime from shutil import disk_usage @@ -212,7 +212,7 @@ async def cmd_eval(self: Client, ctx: Message, strings) -> Optional[str]: "re": re, "os": os, "asyncio": asyncio, - "cfscrape": cfscrape, + "cloudscraper": cloudscraper, "json": json, "aiohttp": aiohttp, "print": _print, diff --git a/misskaty/plugins/subscene_dl.py b/misskaty/plugins/subscene_dl.py index 23725a32..298650eb 100644 --- a/misskaty/plugins/subscene_dl.py +++ b/misskaty/plugins/subscene_dl.py @@ -2,7 +2,7 @@ import asyncio import logging import os -import cfscrape +import cloudscraper from bs4 import BeautifulSoup from pykeyboard import InlineButton, InlineKeyboard from pyrogram import Client, filters @@ -24,7 +24,7 @@ SUB_DL_DICT = {} async def getTitleSub(msg, kueri, CurrentPage, user): if not SUB_TITLE_DICT.get(msg.id): sdata = [] - scraper = cfscrape.create_scraper() + scraper = cloudscraper.create_scraper() param = {"query": kueri} r = scraper.post("https://subscene.com/subtitles/searchbytitle", data=param).text soup = BeautifulSoup(r, "lxml") @@ -61,7 +61,7 @@ async def getTitleSub(msg, kueri, CurrentPage, user): async def getListSub(msg, link, CurrentPage, user): if not SUB_DL_DICT.get(msg.id): sdata = [] - scraper = cfscrape.create_scraper() + scraper = cloudscraper.create_scraper() kuki = {"LanguageFilter": "13,44,50"} # Only filter language English, Malay, Indonesian r = scraper.get(link, cookies=kuki).text soup = BeautifulSoup(r, "lxml") @@ -194,7 +194,7 @@ async def dlsub_callback(self: Client, callback_query: CallbackQuery): await callback_query.answer("Invalid callback data, please send CMD again..") await asyncio.sleep(3) return await callback_query.message.delete_msg() - scraper = cfscrape.create_scraper() + scraper = cloudscraper.create_scraper() res = await down_page(link) dl = scraper.get(res.get("download_url")) f = open(f"{title}.zip", mode="wb").write(dl.content) diff --git a/misskaty/plugins/web_scraper.py b/misskaty/plugins/web_scraper.py index d5010cfa..25a6cf3c 100644 --- a/misskaty/plugins/web_scraper.py +++ b/misskaty/plugins/web_scraper.py @@ -6,7 +6,7 @@ """ import re import logging -import cfscrape +import cloudscraper from bs4 import BeautifulSoup from pykeyboard import InlineKeyboard, InlineButton from pyrogram import filters, Client @@ -384,7 +384,7 @@ async def getDataGomov(msg, kueri, CurrentPage, user, strings): # getData samehada async def getSame(msg, query, current_page, strings): if not SCRAP_DICT.get(msg.id): - cfse = cfscrape.CloudflareScraper() + cfse = cloudscraper.create_scraper() try: if query: data = cfse.get(f"{web['samehadaku']}/?s={query}", headers=headers) diff --git a/requirements.txt b/requirements.txt index f3c64338..ab29b51f 100644 --- a/requirements.txt +++ b/requirements.txt @@ -26,5 +26,5 @@ deep-translator telethon pyrate_limiter cachetools -cfscrape +cloudscraper openai \ No newline at end of file