""" * @author yasir * @created 2022-12-01 09:12:27 * @projectName MissKatyPyro * Copyright @YasirPedia All rights reserved """ import contextlib import logging import re import sys import traceback import cloudscraper import httpx from bs4 import BeautifulSoup from cachetools import TTLCache from pykeyboard import InlineButton, InlineKeyboard from pyrogram.errors import QueryIdInvalid from pyrogram.types import Message from database import dbname from misskaty import app from misskaty.helper import Cache, Kusonime, fetch, post_to_telegraph, use_chat_lang __MODULE__ = "WebScraper" __HELP__ = """ /melongmovie [query ] - Scrape website data from MelongMovie Web. /lk21 [query ] - Scrape website data from LayarKaca21. /pahe [query ] - Scrape website data from Pahe.li. /terbit21 [query ] - Scrape website data from Terbit21. /savefilm21 [query ] - Scrape website data from Savefilm21. /movieku [query ] - Scrape website data from Movieku.cc /kusonime [query ] - Scrape website data from Kusonime /lendrive [query ] - Scrape website data from Lendrive /klikxxi [query ] - Scrape website data from Klikxxi aka GoMov. /samehadaku [query ] - Scrape website data from Samehadaku. /nodrakor [query ] - Scrape website data from NoDrakor """ LOGGER = logging.getLogger("MissKaty") SCRAP_DICT = Cache(filename="scraper_cache.db", path="cache", in_memory=False) data_kuso = Cache(filename="kuso_cache.db", path="cache", in_memory=False) savedict = TTLCache(maxsize=1000, ttl=3600) webdb = dbname["web"] web = { "yasirapi": "https://yasirapi.eu.org", "yasirapi_v2": "https://v2.yasirapi.eu.org", "pahe": "pahe.ink", "savefilm21": "https://sf6.savefilm21digital.com", "melongmovie": "https://melongmovie.store", "terbit21": "https://terbit21.cc", "lk21": "https://tv12.lk21official.my/", "gomov": "https://klikxxi.com", "movieku": "https://movieku.cloud", "kusonime": "https://kusonime.com", "lendrive": "https://lendrive.web.id", "samehadaku": "https://samehadaku.help", "oplovers": "https://oploverz.red", "nodrakor": "https://no-drakor.xyz", } def split_arr(arr, size: 5): arrs = [] while len(arr) > size: pice = arr[:size] arrs.append(pice) arr = arr[size:] arrs.append(arr) return arrs # Terbit21 GetData async def getDataTerbit21(msg, kueri, CurrentPage, strings): if not SCRAP_DICT.get(msg.id): with contextlib.redirect_stdout(sys.stderr): try: if kueri: terbitjson = await fetch.get( f"{web['yasirapi']}/terbit21?q={kueri}" ) else: terbitjson = await fetch.get(f"{web['yasirapi']}/terbit21") terbitjson.raise_for_status() except httpx.HTTPError as exc: await msg.edit_msg( f"ERROR: Failed to fetch data from {exc.request.url} - {exc}" ) return None, None res = terbitjson.json() if not res.get("result"): await msg.edit_msg(strings("no_result"), del_in=5) return None, None SCRAP_DICT.add(msg.id, [split_arr(res["result"], 6), kueri], timeout=1800) index = int(CurrentPage - 1) PageLen = len(SCRAP_DICT[msg.id][0]) if kueri: TerbitRes = strings("header_with_query").format(web="Terbit21", kueri=kueri) else: TerbitRes = strings("header_no_query").format(web="Terbit21", cmd="terbit21") for c, i in enumerate(SCRAP_DICT[msg.id][0][index], start=1): TerbitRes += f"{index*6+c}. {i['judul']}\n{strings('cat_text')}: {i['kategori']}\n" TerbitRes += ( "\n" if re.search(r"Complete|Ongoing", i["kategori"]) else f"{strings('dl_text')}\n\n" ) return TerbitRes, PageLen # LK21 GetData async def getDatalk21(msg, kueri, CurrentPage, strings): if not SCRAP_DICT.get(msg.id): with contextlib.redirect_stdout(sys.stderr): try: if kueri: lk21json = await fetch.get(f"{web['yasirapi_v2']}/lk21?q={kueri}") else: lk21json = await fetch.get(f"{web['yasirapi_v2']}/lk21") lk21json.raise_for_status() except httpx.HTTPError as exc: await msg.edit_msg( f"ERROR: Failed to fetch data from {exc.request.url} - {exc}" ) return None, None res = lk21json.json() if not res.get("result"): await msg.edit_msg(strings("no_result"), del_in=5) return None, None SCRAP_DICT.add(msg.id, [split_arr(res["result"], 6), kueri], timeout=1800) index = int(CurrentPage - 1) PageLen = len(SCRAP_DICT[msg.id][0]) if kueri: lkResult = strings("header_with_query").format(web="Layarkaca21", kueri=kueri) else: lkResult = strings("header_no_query").format(web="Layarkaca21", cmd="lk21") for c, i in enumerate(SCRAP_DICT[msg.id][0][index], start=1): lkResult += f"{index*6+c}. {i['judul']}\n{strings('cat_text')}: {i['kategori']}\n" lkResult += ( "\n" if re.search(r"Complete|Ongoing", i["kategori"]) else f"{strings('dl_text')}\n\n" ) return lkResult, PageLen # Pahe GetData async def getDataPahe(msg, kueri, CurrentPage, strings): if not SCRAP_DICT.get(msg.id): with contextlib.redirect_stdout(sys.stderr): try: if kueri: pahejson = await fetch.get( f"{web['yasirapi']}/pahe?q={kueri}&domain={web['pahe']}" ) else: pahejson = await fetch.get( f"{web['yasirapi']}/pahe?domain={web['pahe']}" ) pahejson.raise_for_status() except httpx.HTTPError as exc: await msg.edit_msg( f"ERROR: Failed to fetch data from {exc.request.url} - {exc}" ) return None, None res = pahejson.json() if not res.get("result"): await msg.edit_msg(strings("no_result"), del_in=5) return None, None SCRAP_DICT.add(msg.id, [split_arr(res["result"], 6), kueri], timeout=1800) index = int(CurrentPage - 1) PageLen = len(SCRAP_DICT[msg.id][0]) paheResult = ( strings("header_with_query").format(web="Pahe", kueri=kueri) if kueri else strings("header_no_query").format(web="Pahe", cmd="pahe") ) for c, i in enumerate(SCRAP_DICT[msg.id][0][index], start=1): paheResult += f"{index*6+c}. {i['judul']}\n\n" return paheResult, PageLen # Kusonime GetData async def getDataKuso(msg, kueri, CurrentPage, user, strings): if not SCRAP_DICT.get(msg.id): kusodata = [] with contextlib.redirect_stdout(sys.stderr): try: data = await fetch.get( f"{web['kusonime']}/?s={kueri}", follow_redirects=True ) data.raise_for_status() except httpx.HTTPError as exc: await msg.edit_msg( f"ERROR: Failed to fetch data from {exc.request.url} - {exc}", disable_web_page_preview=True, ) return None, 0, None, None res = BeautifulSoup(data, "lxml").find_all("h2", {"class": "episodeye"}) for i in res: ress = i.find_all("a")[0] title = ress.text link = ress["href"] kusodata.append({"title": title, "link": link}) if not kusodata: await msg.edit_msg(strings("no_result"), del_in=5) return None, 0, None, None SCRAP_DICT.add(msg.id, [split_arr(kusodata, 10), kueri], timeout=1800) index = int(CurrentPage - 1) PageLen = len(SCRAP_DICT[msg.id][0]) extractbtn1 = [] extractbtn2 = [] kusoResult = ( strings("header_no_query").format(web="Kusonime", cmd="kusonime") if kueri == "" else strings("header_with_query").format(web="Kusonime", kueri=kueri) ) for c, i in enumerate(SCRAP_DICT[msg.id][0][index], start=1): kusoResult += f"{index*6+c}. {i['title']}\n{i['link']}\n\n" if c < 6: extractbtn1.append( InlineButton( index * 6 + c, f"kusoextract#{CurrentPage}#{c}#{user}#{msg.id}" ) ) else: extractbtn2.append( InlineButton( index * 6 + c, f"kusoextract#{CurrentPage}#{c}#{user}#{msg.id}" ) ) return kusoResult, PageLen, extractbtn1, extractbtn2 # Movieku GetData async def getDataMovieku(msg, kueri, CurrentPage, user, strings): if not SCRAP_DICT.get(msg.id): moviekudata = [] with contextlib.redirect_stdout(sys.stderr): try: data = await fetch.get( f"{web['movieku']}/?s={kueri}", follow_redirects=True ) data.raise_for_status() except httpx.HTTPError as exc: await msg.edit_msg( f"ERROR: Failed to fetch data from {exc.request.url} - {exc}" ) return None, 0, None r = BeautifulSoup(data, "lxml") res = r.find_all(class_="bx") for i in res: judul = i.find_all("a")[0]["title"] link = i.find_all("a")[0]["href"] typ = i.find(class_="overlay").text typee = typ.strip() if typ.strip() != "" else "~" moviekudata.append({"judul": judul, "link": link, "type": typee}) if not moviekudata: await msg.edit_msg(strings("no_result"), del_in=5) return None, 0, None SCRAP_DICT.add(msg.id, [split_arr(moviekudata, 6), kueri], timeout=1800) index = int(CurrentPage - 1) PageLen = len(SCRAP_DICT[msg.id][0]) extractbtn = [] moviekuResult = ( strings("header_no_query").format(web="Movieku", cmd="movieku") if kueri == "" else strings("header_with_query").format(web="Movieku", kueri=kueri) ) for c, i in enumerate(SCRAP_DICT[msg.id][0][index], start=1): moviekuResult += f"{index*6+c}. {i['judul']}\n{strings('quality')}/Status: {i['type']}\n\n" extractbtn.append( InlineButton( index * 6 + c, f"moviekuextract#{CurrentPage}#{c}#{user}#{msg.id}" ) ) return moviekuResult, PageLen, extractbtn # NoDrakor GetData async def getDataNodrakor(msg, kueri, CurrentPage, user, strings): if not SCRAP_DICT.get(msg.id): nodrakordata = [] with contextlib.redirect_stdout(sys.stderr): try: data = await fetch.get( f"{web['nodrakor']}/?s={kueri}", follow_redirects=True, ) data.raise_for_status() except httpx.HTTPError as exc: await msg.edit_msg( f"HTTP Exception for {exc.request.url} - {exc}", disable_web_page_preview=True, ) return None, 0, None text = BeautifulSoup(data, "lxml") entry = text.find_all(class_="entry-header") if entry[0].text.strip() == "Nothing Found": if not kueri: await msg.edit_msg(strings("no_result"), del_in=5) else: await msg.edit_msg( strings("no_result_w_query").format(kueri=kueri), del_in=5 ) return None, 0, None for i in entry: genre = i.find(class_="gmr-movie-on") genre = f"{genre.text}" if genre else "N/A" judul = i.find(class_="entry-title").find("a").text link = i.find(class_="entry-title").find("a").get("href") nodrakordata.append({"judul": judul, "link": link, "genre": genre}) SCRAP_DICT.add(msg.id, [split_arr(nodrakordata, 6), kueri], timeout=1800) index = int(CurrentPage - 1) PageLen = len(SCRAP_DICT[msg.id][0]) extractbtn = [] nodrakorResult = ( strings("header_no_query").format(web="NoDrakor", cmd="nodrakor") if kueri == "" else strings("header_with_query").format(web="NoDrakor", kueri=kueri) ) for c, i in enumerate(SCRAP_DICT[msg.id][0][index], start=1): nodrakorResult += f"{index*6+c}. {i['judul']}\nGenre: {i['genre']}\n\n" extractbtn.append( InlineButton( index * 6 + c, f"nodrakorextract#{CurrentPage}#{c}#{user}#{msg.id}" ) ) return nodrakorResult, PageLen, extractbtn # Savefilm21 GetData async def getDataSavefilm21(msg, kueri, CurrentPage, user, strings): if not SCRAP_DICT.get(msg.id): sfdata = [] with contextlib.redirect_stdout(sys.stderr): try: data = await fetch.get( f"{web['savefilm21']}/?s={kueri}", follow_redirects=True, ) data.raise_for_status() except httpx.HTTPError as exc: await msg.edit_msg( f"HTTP Exception for {exc.request.url} - {exc}", disable_web_page_preview=True, ) return None, 0, None text = BeautifulSoup(data, "lxml") entry = text.find_all(class_="entry-header") if "Tidak Ditemukan" in entry[0].text: if not kueri: await msg.edit_msg(strings("no_result"), del_in=5) else: await msg.edit_msg( strings("no_result_w_query").format(kueri=kueri), del_in=5 ) return None, 0, None for i in entry: genre = i.find(class_="gmr-movie-on").text genre = f"{genre}" if genre != "" else "N/A" judul = i.find(class_="entry-title").find("a").text link = i.find(class_="entry-title").find("a").get("href") sfdata.append({"judul": judul, "link": link, "genre": genre}) SCRAP_DICT.add(msg.id, [split_arr(sfdata, 6), kueri], timeout=1800) index = int(CurrentPage - 1) PageLen = len(SCRAP_DICT[msg.id][0]) extractbtn = [] sfResult = ( strings("header_no_query").format(web="Savefilm21", cmd="savefilm21") if kueri == "" else strings("header_with_query").format(web="Savefilm21", kueri=kueri) ) for c, i in enumerate(SCRAP_DICT[msg.id][0][index], start=1): sfResult += f"{index*6+c}. {i['judul']}\nGenre: {i['genre']}\n\n" extractbtn.append( InlineButton( index * 6 + c, f"sf21extract#{CurrentPage}#{c}#{user}#{msg.id}" ) ) return sfResult, PageLen, extractbtn # Lendrive GetData async def getDataLendrive(msg, kueri, CurrentPage, user, strings): if not SCRAP_DICT.get(msg.id): with contextlib.redirect_stdout(sys.stderr): try: if kueri: data = await fetch.get( f"{web['lendrive']}/?s={kueri}", follow_redirects=True, ) else: data = await fetch.get(web["lendrive"], follow_redirects=True) data.raise_for_status() except httpx.HTTPError as exc: await msg.edit_msg( f"ERROR: Failed to fetch data from {exc.request.url} - {exc}", disable_web_page_preview=True, ) return None, 0, None res = BeautifulSoup(data, "lxml") lenddata = [] for o in res.find_all(class_="bsx"): title = o.find("a")["title"] link = o.find("a")["href"] status = ( o.find(class_="epx").text if o.find(class_="epx") else "Not Provided by BOT" ) kualitas = ( o.find(class_="typez TV").text if o.find(class_="typez TV") else o.find(class_="typez BD") ) lenddata.append( { "judul": title, "link": link, "quality": kualitas or "N/A", "status": status, } ) if not lenddata: await msg.edit_msg(strings("no_result"), del_in=5) return None, 0, None savedict[msg.id] = [split_arr(lenddata, 6), kueri] index = int(CurrentPage - 1) PageLen = len(savedict[msg.id][0]) extractbtn = [] lenddataResult = ( strings("header_no_query").format(web="Lendrive", cmd="lendrive") if kueri == "" else strings("header_with_query").format(web="Lendrive", kueri=kueri) ) for c, i in enumerate(savedict[msg.id][0][index], start=1): lenddataResult += f"{index*6+c}. {i['judul']}\n{strings('quality')}: {i['quality']}\nStatus: {i['status']}\n\n" extractbtn.append( InlineButton( index * 6 + c, f"lendriveextract#{CurrentPage}#{c}#{user}#{msg.id}" ) ) return lenddataResult, PageLen, extractbtn # MelongMovie GetData async def getDataMelong(msg, kueri, CurrentPage, user, strings): if not SCRAP_DICT.get(msg.id): with contextlib.redirect_stdout(sys.stderr): try: data = await fetch.get( f"{web['melongmovie']}/?s={kueri}", follow_redirects=True, ) data.raise_for_status() except httpx.HTTPError as exc: await msg.edit_msg( f"HTTP Exception for {exc.request.url} - {exc}", disable_web_page_preview=True, ) return None, 0, None bs4 = BeautifulSoup(data, "lxml") melongdata = [] for res in bs4.select(".box"): dd = res.select("a") url = dd[0]["href"] title = dd[0]["title"] try: quality = dd[0].find(class_="quality").text except: quality = "N/A" melongdata.append({"judul": title, "link": url, "quality": quality}) if not melongdata: await msg.edit_msg(strings("no_result"), del_in=5) return None, 0, None SCRAP_DICT.add(msg.id, [split_arr(melongdata, 6), kueri], timeout=1800) index = int(CurrentPage - 1) PageLen = len(SCRAP_DICT[msg.id][0]) extractbtn = [] melongResult = ( strings("header_no_query").format(web="Melongmovie", cmd="melongmovie") if kueri == "" else strings("header_with_query").format(web="Melongmovie", kueri=kueri) ) for c, i in enumerate(SCRAP_DICT[msg.id][0][index], start=1): melongResult += f"{index*6+c}. {i['judul']}\n{strings('quality')}: {i['quality']}\n\n" extractbtn.append( InlineButton( index * 6 + c, f"melongextract#{CurrentPage}#{c}#{user}#{msg.id}" ) ) return melongResult, PageLen, extractbtn # GoMov GetData async def getDataGomov(msg, kueri, CurrentPage, user, strings): if not SCRAP_DICT.get(msg.id): with contextlib.redirect_stdout(sys.stderr): try: gomovv = await fetch.get( f"{web['gomov']}/?s={kueri}", follow_redirects=True ) gomovv.raise_for_status() except httpx.HTTPError as exc: await msg.edit_msg( f"ERROR: Failed to fetch data from {exc.request.url} - {exc}", disable_web_page_preview=True, ) return None, 0, None text = BeautifulSoup(gomovv, "lxml") entry = text.find_all(class_="entry-header") if entry[0].text.strip() == "Nothing Found": if not kueri: await msg.edit_msg(strings("no_result"), del_in=5) else: await msg.edit_msg( strings("no_result_w_query").format(kueri=kueri), del_in=5 ) return None, 0, None else: data = [] for i in entry: genre = i.find(class_="gmr-movie-on") genre = f"{genre.text}" if genre else "N/A" judul = i.find(class_="entry-title").find("a").text link = i.find(class_="entry-title").find("a").get("href") data.append({"judul": judul, "link": link, "genre": genre}) SCRAP_DICT.add(msg.id, [split_arr(data, 6), kueri], timeout=1800) index = int(CurrentPage - 1) PageLen = len(SCRAP_DICT[msg.id][0]) extractbtn = [] gomovResult = ( strings("header_with_query").format(web="GoMov", kueri=kueri) if kueri else strings("header_no_query").format(web="GoMov", cmd="gomov") ) for c, i in enumerate(SCRAP_DICT[msg.id][0][index], start=1): gomovResult += f"{index*6+c}. {i['judul']}\nGenre: {i['genre']}\n\n" if not re.search(r"Series", i["genre"]): extractbtn.append( InlineButton( index * 6 + c, f"gomovextract#{CurrentPage}#{c}#{user}#{msg.id}" ) ) gomovResult += strings("unsupport_dl_btn") return gomovResult, PageLen, extractbtn # getData samehada async def getSame(msg, query, current_page, strings): if not SCRAP_DICT.get(msg.id): cfse = cloudscraper.create_scraper() if query: data = cfse.get(f"{web['samehadaku']}/?s={query}") else: data = cfse.get(web["samehadaku"]) if data.status_code != 200: await msg.edit_msg(strings("err_getweb").format(err=data.status_code)) return None, None res = BeautifulSoup(data.text, "lxml").find_all(class_="animposx") sdata = [] for i in res: url = i.find("a")["href"] title = i.find("a")["title"] sta = ( i.find(class_="type TV").text if i.find(class_="type TV") else "Ongoing" ) rate = i.find(class_="score") sdata.append({"url": url, "title": title, "sta": sta, "rate": rate}) if not sdata: await msg.edit_msg(strings("no_result"), del_in=5) return None, None savedict[msg.id] = [split_arr(sdata, 10), query] index = int(current_page - 1) PageLen = len(savedict[msg.id][0]) sameresult = "".join( f"{index * 6 + c}. {i['title']}\nStatus: {i['sta']}\nRating: {i['rate']}\n\n" for c, i in enumerate(savedict[msg.id][0][index], start=1) ) return sameresult, PageLen # SameHada CMD @app.on_cmd("samehadaku", no_channel=True) @use_chat_lang() async def same_search(_, msg, strings): query = msg.text.split(maxsplit=1)[1] if len(msg.command) > 1 else None bmsg = await msg.reply_msg(strings("get_data"), quote=True) sameres, PageLen = await getSame(bmsg, query, 1, strings) if not sameres: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, 1, "page_same#{number}" + f"#{bmsg.id}#{msg.from_user.id}" ) keyboard.row(InlineButton(strings("cl_btn"), f"close#{msg.from_user.id}")) await bmsg.edit_msg(sameres, disable_web_page_preview=True, reply_markup=keyboard) # Terbit21 CMD @app.on_cmd("terbit21", no_channel=True) @use_chat_lang() async def terbit21_s(_, message, strings): kueri = " ".join(message.command[1:]) if not kueri: kueri = None pesan = await message.reply_msg(strings("get_data"), quote=True) CurrentPage = 1 terbitres, PageLen = await getDataTerbit21(pesan, kueri, CurrentPage, strings) if not terbitres: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_terbit21#{number}" + f"#{pesan.id}#{message.from_user.id}", ) keyboard.row(InlineButton(strings("cl_btn"), f"close#{message.from_user.id}")) await pesan.edit_msg( terbitres, disable_web_page_preview=True, reply_markup=keyboard ) # LK21 CMD @app.on_cmd("lk21", no_channel=True) @use_chat_lang() async def lk21_s(_, message, strings): kueri = " ".join(message.command[1:]) if not kueri: kueri = None pesan = await message.reply_msg(strings("get_data"), quote=True) CurrentPage = 1 lkres, PageLen = await getDatalk21(pesan, kueri, CurrentPage, strings) if not lkres: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_lk21#{number}" + f"#{pesan.id}#{message.from_user.id}", ) keyboard.row(InlineButton(strings("cl_btn"), f"close#{message.from_user.id}")) await pesan.edit_msg(lkres, disable_web_page_preview=True, reply_markup=keyboard) # Pahe CMD @app.on_cmd("pahe", no_channel=True) @use_chat_lang() async def pahe_s(_, message, strings): kueri = " ".join(message.command[1:]) if not kueri: kueri = "" pesan = await message.reply_msg(strings("get_data"), quote=True) CurrentPage = 1 paheres, PageLen = await getDataPahe(pesan, kueri, CurrentPage, strings) if not paheres: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_pahe#{number}" + f"#{pesan.id}#{message.from_user.id}", ) keyboard.row(InlineButton(strings("cl_btn"), f"close#{message.from_user.id}")) await pesan.edit_msg(paheres, disable_web_page_preview=True, reply_markup=keyboard) # Gomov CMD @app.on_cmd(["gomov", "klikxxi"], no_channel=True) @use_chat_lang() async def gomov_s(self, message, strings): kueri = " ".join(message.command[1:]) if not kueri: kueri = "" pesan = await message.reply_msg(strings("get_data"), quote=True) CurrentPage = 1 gomovres, PageLen, btn = await getDataGomov( pesan, kueri, CurrentPage, message.from_user.id, strings ) if not gomovres: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_gomov#{number}" + f"#{pesan.id}#{message.from_user.id}", ) keyboard.row(InlineButton(strings("ex_data"), user_id=self.me.id)) keyboard.row(*btn) keyboard.row(InlineButton(strings("cl_btn"), f"close#{message.from_user.id}")) await pesan.edit_msg(gomovres, disable_web_page_preview=True, reply_markup=keyboard) # MelongMovie CMD @app.on_cmd("melongmovie", no_channel=True) @use_chat_lang() async def melong_s(self, message, strings): kueri = " ".join(message.command[1:]) if not kueri: kueri = "" pesan = await message.reply_msg(strings("get_data"), quote=True) CurrentPage = 1 melongres, PageLen, btn = await getDataMelong( pesan, kueri, CurrentPage, message.from_user.id, strings ) if not melongres: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_melong#{number}" + f"#{pesan.id}#{message.from_user.id}", ) keyboard.row(InlineButton(strings("ex_data"), user_id=self.me.id)) keyboard.row(*btn) keyboard.row(InlineButton(strings("cl_btn"), f"close#{message.from_user.id}")) try: await pesan.edit_msg( melongres, disable_web_page_preview=True, reply_markup=keyboard ) except Exception as err: await pesan.edit_msg( f"ERROR: {err}", disable_web_page_preview=True, reply_markup=keyboard ) # Savefilm21 CMD @app.on_cmd("savefilm21", no_channel=True) @use_chat_lang() async def savefilm_s(self, message, strings): kueri = " ".join(message.command[1:]) if not kueri: kueri = "" pesan = await message.reply_msg(strings("get_data"), quote=True) CurrentPage = 1 savefilmres, PageLen, btn = await getDataSavefilm21( pesan, kueri, CurrentPage, message.from_user.id, strings ) if not savefilmres: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_sf21#{number}" + f"#{pesan.id}#{message.from_user.id}", ) keyboard.row(InlineButton(strings("ex_data"), user_id=self.me.id)) keyboard.row(*btn) keyboard.row(InlineButton(strings("cl_btn"), f"close#{message.from_user.id}")) await pesan.edit_msg( savefilmres, disable_web_page_preview=True, reply_markup=keyboard ) # NoDrakor CMD @app.on_cmd("nodrakor", no_channel=True) @use_chat_lang() async def nodrakor_s(self, message, strings): kueri = " ".join(message.command[1:]) if not kueri: kueri = "" pesan = await message.reply_msg(strings("get_data"), quote=True) CurrentPage = 1 nodrakorres, PageLen, btn = await getDataNodrakor( pesan, kueri, CurrentPage, message.from_user.id, strings ) if not nodrakorres: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_nodrakor#{number}" + f"#{pesan.id}#{message.from_user.id}", ) keyboard.row(InlineButton(strings("ex_data"), user_id=self.me.id)) keyboard.row(*btn) keyboard.row(InlineButton(strings("cl_btn"), f"close#{message.from_user.id}")) await pesan.edit_msg( nodrakorres, disable_web_page_preview=True, reply_markup=keyboard ) # Kusonime CMD @app.on_cmd("kusonime", no_channel=True) @use_chat_lang() async def kusonime_s(self, message, strings): kueri = " ".join(message.command[1:]) if not kueri: kueri = "" pesan = await message.reply_msg(strings("get_data"), quote=True) CurrentPage = 1 kusores, PageLen, btn1, btn2 = await getDataKuso( pesan, kueri, CurrentPage, message.from_user.id, strings ) if not kusores: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_kuso#{number}" + f"#{pesan.id}#{message.from_user.id}", ) keyboard.row(InlineButton(strings("ex_data"), user_id=self.me.id)) keyboard.row(*btn1) if btn2: keyboard.row(*btn2) keyboard.row(InlineButton(strings("cl_btn"), f"close#{message.from_user.id}")) await pesan.edit_msg(kusores, disable_web_page_preview=True, reply_markup=keyboard) # Lendrive CMD @app.on_cmd("lendrive", no_channel=True) @use_chat_lang() async def lendrive_s(self, ctx: Message, strings): kueri = ctx.input if not kueri: kueri = "" pesan = await ctx.reply_msg(strings("get_data"), quote=True) CurrentPage = 1 lendres, PageLen, btn = await getDataLendrive( pesan, kueri, CurrentPage, ctx.from_user.id, strings ) if not lendres: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_lendrive#{number}" + f"#{pesan.id}#{ctx.from_user.id}", ) keyboard.row(InlineButton(strings("ex_data"), user_id=self.me.id)) keyboard.row(*btn) keyboard.row(InlineButton(strings("cl_btn"), f"close#{ctx.from_user.id}")) await pesan.edit_msg(lendres, disable_web_page_preview=True, reply_markup=keyboard) # Movieku CMD @app.on_cmd("movieku", no_channel=True) @use_chat_lang() async def movieku_s(self, ctx: Message, strings): kueri = ctx.input if not kueri: kueri = "" pesan = await ctx.reply_msg(strings("get_data"), quote=True) CurrentPage = 1 moviekures, PageLen, btn = await getDataMovieku(pesan, kueri, CurrentPage, ctx.from_user.id, strings) if not moviekures: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_movieku#{number}" + f"#{pesan.id}#{ctx.from_user.id}", ) keyboard.row(InlineButton(strings("ex_data"), user_id=self.me.id)) keyboard.row(*btn) keyboard.row(InlineButton(strings("cl_btn"), f"close#{ctx.from_user.id}")) await pesan.edit_msg( moviekures, disable_web_page_preview=True, reply_markup=keyboard ) # Savefillm21 Page Callback @app.on_cb("page_sf21#") @use_chat_lang() async def sf21page_callback(self, callback_query, strings): try: if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) message_id = int(callback_query.data.split("#")[2]) CurrentPage = int(callback_query.data.split("#")[1]) kueri = SCRAP_DICT[message_id][1] except (IndexError, ValueError): # Gatau napa err ini return except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb")) except QueryIdInvalid: return try: savefilmres, PageLen, btn = await getDataSavefilm21( callback_query.message, kueri, CurrentPage, callback_query.from_user.id, strings, ) except TypeError: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_sf21#{number}" + f"#{message_id}#{callback_query.from_user.id}", ) keyboard.row(InlineButton(strings("ex_data"), user_id=self.me.id)) keyboard.row(*btn) keyboard.row( InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}") ) await callback_query.message.edit_msg( savefilmres, disable_web_page_preview=True, reply_markup=keyboard ) # NoDrakor Page Callback @app.on_cb("page_nodrakor#") @use_chat_lang() async def nodrakorpage_cb(self, callback_query, strings): try: if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) message_id = int(callback_query.data.split("#")[2]) CurrentPage = int(callback_query.data.split("#")[1]) kueri = SCRAP_DICT[message_id][1] except (IndexError, ValueError): return except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb")) except QueryIdInvalid: return try: nodrakorres, PageLen, btn = await getDataNodrakor( callback_query.message, kueri, CurrentPage, callback_query.from_user.id, strings, ) except TypeError: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_nodrakor#{number}" + f"#{message_id}#{callback_query.from_user.id}", ) keyboard.row(InlineButton(strings("ex_data"), user_id=self.me.id)) keyboard.row(*btn) keyboard.row( InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}") ) await callback_query.message.edit_msg( nodrakorres, disable_web_page_preview=True, reply_markup=keyboard ) # Kuso Page Callback @app.on_cb("page_kuso#") @use_chat_lang() async def kusopage_callback(self, callback_query, strings): try: if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) message_id = int(callback_query.data.split("#")[2]) CurrentPage = int(callback_query.data.split("#")[1]) kueri = SCRAP_DICT[message_id][1] except QueryIdInvalid: return except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb")) try: kusores, PageLen, btn1, btn2 = await getDataKuso( callback_query.message, kueri, CurrentPage, callback_query.from_user.id, strings, ) except TypeError: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_kuso#{number}" + f"#{message_id}#{callback_query.from_user.id}", ) keyboard.row(InlineButton(strings("ex_data"), user_id=self.me.id)) keyboard.row(*btn1) if btn2: keyboard.row(*btn2) keyboard.row( InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}") ) await callback_query.message.edit_msg( kusores, disable_web_page_preview=True, reply_markup=keyboard ) # Lendrive Page Callback @app.on_cb("page_lendrive#") @use_chat_lang() async def lendrivepage_callback(self, callback_query, strings): try: if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) message_id = int(callback_query.data.split("#")[2]) CurrentPage = int(callback_query.data.split("#")[1]) kueri = savedict[message_id][1] except QueryIdInvalid: return except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb")) try: lendres, PageLen, btn = await getDataLendrive( callback_query.message, kueri, CurrentPage, callback_query.from_user.id, strings, ) except TypeError: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_lendrive#{number}" + f"#{message_id}#{callback_query.from_user.id}", ) keyboard.row(InlineButton(strings("ex_data"), user_id=self.me.id)) keyboard.row(*btn) keyboard.row( InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}") ) await callback_query.message.edit_msg( lendres, disable_web_page_preview=True, reply_markup=keyboard ) # Movieku Page Callback @app.on_cb("page_movieku#") @use_chat_lang() async def moviekupage_callback(self, callback_query, strings): try: if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) message_id = int(callback_query.data.split("#")[2]) CurrentPage = int(callback_query.data.split("#")[1]) kueri = SCRAP_DICT[message_id][1] except QueryIdInvalid: return except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb"), True) try: moviekures, PageLen, btn = await getDataMovieku( callback_query.message, kueri, CurrentPage, callback_query.from_user.id, strings ) except TypeError: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_movieku#{number}" + f"#{message_id}#{callback_query.from_user.id}", ) keyboard.row(InlineButton(strings("ex_data"), user_id=self.me.id)) keyboard.row(*btn) keyboard.row( InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}") ) await callback_query.message.edit_msg( moviekures, disable_web_page_preview=True, reply_markup=keyboard ) # Samehada Page Callback @app.on_cb("page_same#") @use_chat_lang() async def samepg(_, query, strings): try: _, current_page, _id, user_id = query.data.split("#") if int(user_id) != query.from_user.id: return await query.answer(strings("unauth"), True) lquery = savedict[int(_id)][1] except QueryIdInvalid: return except KeyError: return await query.message.edit_msg(strings("invalid_cb")) try: sameres, PageLen = await getSame( query.message, lquery, int(current_page), strings ) except TypeError: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, int(current_page), "page_same#{number}" + f"#{_id}#{query.from_user.id}", ) keyboard.row(InlineButton(strings("cl_btn"), f"close#{query.from_user.id}")) await query.message.edit_msg( sameres, disable_web_page_preview=True, reply_markup=keyboard ) # Terbit21 Page Callback @app.on_cb("page_terbit21#") @use_chat_lang() async def terbit21page_callback(_, callback_query, strings): try: if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) message_id = int(callback_query.data.split("#")[2]) CurrentPage = int(callback_query.data.split("#")[1]) kueri = SCRAP_DICT[message_id][1] except QueryIdInvalid: return except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb")) try: terbitres, PageLen = await getDataTerbit21( callback_query.message, kueri, CurrentPage, strings ) except TypeError: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_terbit21#{number}" + f"#{message_id}#{callback_query.from_user.id}", ) keyboard.row( InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}") ) await callback_query.message.edit_msg( terbitres, disable_web_page_preview=True, reply_markup=keyboard ) # Page Callback Melong @app.on_cb("page_melong#") @use_chat_lang() async def melongpage_callback(self, callback_query, strings): try: if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) message_id = int(callback_query.data.split("#")[2]) CurrentPage = int(callback_query.data.split("#")[1]) kueri = SCRAP_DICT[message_id][1] except QueryIdInvalid: return except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb")) try: terbitres, PageLen, btn = await getDataMelong( callback_query.message, kueri, CurrentPage, callback_query.from_user.id, strings, ) except TypeError: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_melong#{number}" + f"#{message_id}#{callback_query.from_user.id}", ) keyboard.row(InlineButton(strings("ex_data"), user_id=self.me.id)) keyboard.row(*btn) keyboard.row( InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}") ) await callback_query.message.edit_msg( terbitres, disable_web_page_preview=True, reply_markup=keyboard ) # Lk21 Page Callback @app.on_cb("page_lk21#") @use_chat_lang() async def lk21page_callback(_, callback_query, strings): try: if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) message_id = int(callback_query.data.split("#")[2]) CurrentPage = int(callback_query.data.split("#")[1]) kueri = SCRAP_DICT[message_id][1] except QueryIdInvalid: return except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb")) try: lkres, PageLen = await getDatalk21( callback_query.message, kueri, CurrentPage, strings ) except TypeError: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_lk21#{number}" + f"#{message_id}#{callback_query.from_user.id}", ) keyboard.row( InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}") ) await callback_query.message.edit_msg( lkres, disable_web_page_preview=True, reply_markup=keyboard ) # Pahe Page Callback @app.on_cb("page_pahe#") @use_chat_lang() async def pahepage_callback(_, callback_query, strings): try: if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) message_id = int(callback_query.data.split("#")[2]) CurrentPage = int(callback_query.data.split("#")[1]) kueri = SCRAP_DICT[message_id][1] except QueryIdInvalid: return except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb")) try: lkres, PageLen = await getDataPahe( callback_query.message, kueri, CurrentPage, strings ) except TypeError: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_pahe#{number}" + f"#{message_id}#{callback_query.from_user.id}", ) keyboard.row( InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}") ) await callback_query.message.edit_msg( lkres, disable_web_page_preview=True, reply_markup=keyboard ) # Gomov Page Callback @app.on_cb("page_gomov#") @use_chat_lang() async def gomovpage_callback(self, callback_query, strings): try: if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) message_id = int(callback_query.data.split("#")[2]) CurrentPage = int(callback_query.data.split("#")[1]) kueri = SCRAP_DICT[message_id][1] except QueryIdInvalid: return except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb")) try: gomovres, PageLen, btn = await getDataGomov( callback_query.message, kueri, CurrentPage, callback_query.from_user.id, strings, ) except TypeError: return keyboard = InlineKeyboard() keyboard.paginate( PageLen, CurrentPage, "page_gomov#{number}" + f"#{message_id}#{callback_query.from_user.id}", ) keyboard.row(InlineButton(strings("ex_data"), user_id=self.me.id)) keyboard.row(*btn) keyboard.row( InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}") ) await callback_query.message.edit_msg( gomovres, disable_web_page_preview=True, reply_markup=keyboard ) ### Scrape DDL Link From Web ### # Kusonime DDL @app.on_cb("kusoextract#") @use_chat_lang() async def kusonime_scrap(client, callback_query, strings): try: if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) idlink = int(callback_query.data.split("#")[2]) message_id = int(callback_query.data.split("#")[4]) CurrentPage = int(callback_query.data.split("#")[1]) link = SCRAP_DICT[message_id][0][CurrentPage - 1][idlink - 1].get("link") except QueryIdInvalid: return except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb")) kuso = Kusonime() keyboard = InlineKeyboard() keyboard.row( InlineButton( strings("back_btn"), f"page_kuso#{CurrentPage}#{message_id}#{callback_query.from_user.id}", ), InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}"), ) try: if init_url := data_kuso.get(link, False): await callback_query.message.edit_msg( init_url.get("ph_url"), reply_markup=keyboard ) tgh = await kuso.telegraph(link, client.me.username) data_kuso[link] = {"ph_url": tgh} return await callback_query.message.edit_msg(tgh, reply_markup=keyboard) except Exception as e: LOGGER.error(f"clases: {e.__class__}, moduleName: {e.__class__.__name__}") if str(e).startswith("ERROR"): return await callback_query.message.edit_msg(e, reply_markup=keyboard) # Savefilm21 DDL @app.on_cb("sf21extract#") @use_chat_lang() async def savefilm21_scrap(_, callback_query, strings): try: if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) idlink = int(callback_query.data.split("#")[2]) message_id = int(callback_query.data.split("#")[4]) CurrentPage = int(callback_query.data.split("#")[1]) link = SCRAP_DICT[message_id][0][CurrentPage - 1][idlink - 1].get("link") except QueryIdInvalid: return except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb")) keyboard = InlineKeyboard() keyboard.row( InlineButton( strings("back_btn"), f"page_sf21#{CurrentPage}#{message_id}#{callback_query.from_user.id}", ), InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}"), ) with contextlib.redirect_stdout(sys.stderr): try: html = await fetch.get(link) html.raise_for_status() soup = BeautifulSoup(html.text, "lxml") res = soup.find_all(class_="button button-shadow") res = "".join(f"{i.text}\n{i['href']}\n\n" for i in res) await callback_query.message.edit_msg( strings("res_scrape").format(link=link, kl=res), reply_markup=keyboard ) except httpx.HTTPError as exc: await callback_query.message.edit_msg( f"HTTP Exception for {exc.request.url} - {exc}", reply_markup=keyboard, ) except Exception as err: await callback_query.message.edit_msg( f"ERROR: {err}", reply_markup=keyboard ) # NoDrakor DDL @app.on_cb("nodrakorextract#") @use_chat_lang() async def nodrakorddl_scrap(_, callback_query, strings): try: if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) idlink = int(callback_query.data.split("#")[2]) message_id = int(callback_query.data.split("#")[4]) CurrentPage = int(callback_query.data.split("#")[1]) link = SCRAP_DICT[message_id][0][CurrentPage - 1][idlink - 1].get("link") except QueryIdInvalid: return except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb")) keyboard = InlineKeyboard() keyboard.row( InlineButton( strings("back_btn"), f"page_nodrakor#{CurrentPage}#{message_id}#{callback_query.from_user.id}", ), InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}"), ) with contextlib.redirect_stdout(sys.stderr): try: html = await fetch.get(link) html.raise_for_status() soup = BeautifulSoup(html.text, "lxml") if "/tv/" in link: result = soup.find( "div", {"entry-content entry-content-single"} ).find_all("p") msg = "".join(str(f"{i}\n") for i in result) link = await post_to_telegraph(False, "MissKaty NoDrakor", msg) return await callback_query.message.edit_msg( strings("res_scrape").format(link=link, kl=link), reply_markup=keyboard, ) res = soup.find_all(class_="button button-shadow") res = "".join(f"{i.text}\n{i['href']}\n\n" for i in res) if len(res) > 3500: link = await post_to_telegraph(False, "MissKaty NoDrakor", res) return await callback_query.message.edit_msg( strings("res_scrape").format(link=link, kl=link), reply_markup=keyboard, ) await callback_query.message.edit_msg( strings("res_scrape").format(link=link, kl=res), reply_markup=keyboard ) except httpx.HTTPError as exc: await callback_query.message.edit_msg( f"HTTP Exception for {exc.request.url} - {exc}", reply_markup=keyboard, ) except Exception as err: await callback_query.message.edit_msg( f"ERROR: {err}", reply_markup=keyboard ) # Scrape DDL Link Melongmovie @app.on_cb("moviekuextract#") @use_chat_lang() async def movieku_scrap(_, callback_query, strings): try: if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) idlink = int(callback_query.data.split("#")[2]) message_id = int(callback_query.data.split("#")[4]) CurrentPage = int(callback_query.data.split("#")[1]) link = SCRAP_DICT[message_id][0][CurrentPage - 1][idlink - 1].get("link") except QueryIdInvalid: return except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb")) keyboard = InlineKeyboard() keyboard.row( InlineButton( strings("back_btn"), f"page_movieku#{CurrentPage}#{message_id}#{callback_query.from_user.id}", ), InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}"), ) with contextlib.redirect_stdout(sys.stderr): try: html = await fetch.get(link) html.raise_for_status() soup = BeautifulSoup(html.text, "lxml") data = {} output = [] total_links = 0 valid_resolutions = {'1080p', '720p', '480p', '360p'} current_title = None for element in soup.find_all(['h3', 'p']): if element.name == 'h3' and 'smokettl' in element.get('class', []): current_title = element.text.strip() if current_title not in data: data[current_title] = [] elif element.name == 'p' and current_title: strong_tag = element.find('strong') if strong_tag: resolution = strong_tag.text.strip() if resolution in valid_resolutions: total_links += 1 links = ', '.join([f'{a.text.strip()}' for a in element.find_all('a')]) data[current_title].append(f"{resolution} {links}") for title, resolutions in data.items(): output.append(title) output.extend(resolutions) output.append('') if total_links > 70: url = await post_to_telegraph(False, link, "
".join(output)) return await callback_query.message.edit_msg(strings("res_scrape").format(link=link, kl=f"Your result is too long, i have pasted your result on Telegraph:\n{url}"), reply_markup=keyboard) if "\n".join(output) == "": output = "\nOpen link in browser, click on episode page and use /movieku_scrap [page link] commands for extract download link" return await callback_query.message.edit_msg(strings("res_scrape").format(link=link, kl=output), reply_markup=keyboard) await callback_query.message.edit_msg(strings("res_scrape").format(link=link, kl="\n".join(output)), reply_markup=keyboard) except httpx.HTTPError as exc: await callback_query.message.edit_msg( f"HTTP Exception for {exc.request.url} - {exc}", reply_markup=keyboard, ) except Exception as err: await callback_query.message.edit_msg( f"ERROR: {err}", reply_markup=keyboard ) # Scrape DDL Link Melongmovie @app.on_cb("melongextract#") @use_chat_lang() async def melong_scrap(_, callback_query, strings): try: if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) idlink = int(callback_query.data.split("#")[2]) message_id = int(callback_query.data.split("#")[4]) CurrentPage = int(callback_query.data.split("#")[1]) link = SCRAP_DICT[message_id][0][CurrentPage - 1][idlink - 1].get("link") except QueryIdInvalid: return except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb")) keyboard = InlineKeyboard() keyboard.row( InlineButton( strings("back_btn"), f"page_melong#{CurrentPage}#{message_id}#{callback_query.from_user.id}", ), InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}"), ) with contextlib.redirect_stdout(sys.stderr): try: html = await fetch.get(link) html.raise_for_status() soup = BeautifulSoup(html.text, "lxml") rep = "" for ep in soup.findAll(text=re.compile(r"(?i)episode\s+\d+|LINK DOWNLOAD")): hardsub = ep.findPrevious("div") softsub = ep.findNext("div") rep += f"{hardsub}\n{softsub}" await callback_query.message.edit_msg( strings("res_scrape").format(link=link, kl=rep), reply_markup=keyboard ) except httpx.HTTPError as exc: await callback_query.message.edit_msg( f"HTTP Exception for {exc.request.url} - {exc}", reply_markup=keyboard, ) except Exception as err: await callback_query.message.edit_msg( f"ERROR: {err}", reply_markup=keyboard ) # Scrape DDL Link Gomov @app.on_cb("gomovextract#") @use_chat_lang() async def gomov_dl(_, callback_query, strings): try: if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) idlink = int(callback_query.data.split("#")[2]) message_id = int(callback_query.data.split("#")[4]) CurrentPage = int(callback_query.data.split("#")[1]) link = SCRAP_DICT[message_id][0][CurrentPage - 1][idlink - 1].get("link") except QueryIdInvalid: return except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb")) keyboard = InlineKeyboard() keyboard.row( InlineButton( strings("back_btn"), f"page_gomov#{CurrentPage}#{message_id}#{callback_query.from_user.id}", ), InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}"), ) with contextlib.redirect_stdout(sys.stderr): try: html = await fetch.get(link) html.raise_for_status() soup = BeautifulSoup(html.text, "lxml") entry = soup.find(class_="gmr-download-wrap clearfix") hasil = soup.find(class_="title-download").text for i in entry.find(class_="list-inline gmr-download-list clearfix"): title = i.find("a").text ddl = i.find("a")["href"] hasil += f"\n{title}\n{ddl}\n" await callback_query.message.edit_msg( strings("res_scrape").format(link=link, kl=hasil), reply_markup=keyboard ) except httpx.HTTPError as exc: await callback_query.message.edit_msg( f"HTTP Exception for {exc.request.url} - {exc}", reply_markup=keyboard, ) except Exception as err: await callback_query.message.edit_msg( f"ERROR: {err}", reply_markup=keyboard ) @app.on_cb("lendriveextract#") @use_chat_lang() async def lendrive_dl(_, callback_query, strings): if callback_query.from_user.id != int(callback_query.data.split("#")[3]): return await callback_query.answer(strings("unauth"), True) idlink = int(callback_query.data.split("#")[2]) message_id = int(callback_query.data.split("#")[4]) CurrentPage = int(callback_query.data.split("#")[1]) try: link = savedict[message_id][0][CurrentPage - 1][idlink - 1].get("link") except KeyError: return await callback_query.message.edit_msg(strings("invalid_cb")) keyboard = InlineKeyboard() keyboard.row( InlineButton( strings("back_btn"), f"page_lendrive#{CurrentPage}#{message_id}#{callback_query.from_user.id}", ), InlineButton(strings("cl_btn"), f"close#{callback_query.from_user.id}"), ) with contextlib.redirect_stdout(sys.stderr): try: hmm = await fetch.get(link) hmm.raise_for_status() q = BeautifulSoup(hmm.text, "lxml") j = q.findAll("div", class_="soraurlx") kl = "" for i in j: if not i.find("a"): continue kl += f"{i.find('strong')}:\n" kl += "".join( f"[ {a.text} ]\n" for a in i.findAll("a") ) await callback_query.message.edit_msg( strings("res_scrape").format(link=link, kl=kl), reply_markup=keyboard ) except httpx.HTTPError as exc: await callback_query.message.edit_msg( f"HTTP Exception for {exc.request.url} - {exc}", reply_markup=keyboard, ) except Exception as err: await callback_query.message.edit_msg( f"ERROR: {err}", reply_markup=keyboard ) # Manual Scrape DDL Movieku.CC incase cannot auto scrape from button @app.on_cmd("movieku_scrap") @use_chat_lang() async def muviku_scrap(_, message, strings): with contextlib.redirect_stdout(sys.stderr): try: link = message.text.split(maxsplit=1)[1] html = await fetch.get(link) html.raise_for_status() soup = BeautifulSoup(html.text, "lxml") data = {} output = [] total_links = 0 valid_resolutions = {'1080p', '720p', '480p', '360p'} current_title = None for element in soup.find_all(['h3', 'p']): if element.name == 'h3' and 'smokettl' in element.get('class', []): current_title = element.text.strip() if current_title not in data: data[current_title] = [] elif element.name == 'p' and current_title: strong_tag = element.find('strong') if strong_tag: resolution = strong_tag.text.strip() if resolution in valid_resolutions: links = ', '.join([f'{a.text.strip()}' for a in element.find_all('a')]) data[current_title].append(f"{resolution} {links}") for title, resolutions in data.items(): output.append(title) output.extend(resolutions) output.append('') for res in resolutions: total_links += res.count('