mirror of
https://github.com/yasirarism/MissKatyPyro.git
synced 2026-01-08 21:04:50 +00:00
Fix more
This commit is contained in:
parent
9c4f3e3c78
commit
c621ea8cae
1 changed files with 1 additions and 257 deletions
|
|
@ -388,259 +388,3 @@ async def savefilm21(_, msg):
|
||||||
await m.delete()
|
await m.delete()
|
||||||
LOGGER.error(e)
|
LOGGER.error(e)
|
||||||
await msg.reply(f"ERROR: {e}", True)
|
await msg.reply(f"ERROR: {e}", True)
|
||||||
|
|
||||||
|
|
||||||
@app.on_message(filters.command(["melongmovie"], COMMAND_HANDLER))
|
|
||||||
@capture_err
|
|
||||||
async def melongmovie(_, msg):
|
|
||||||
SITE = "http://167.99.31.48"
|
|
||||||
try:
|
|
||||||
judul = msg.text.split(" ", 1)[1]
|
|
||||||
except IndexError:
|
|
||||||
judul = None
|
|
||||||
data = []
|
|
||||||
m = await msg.reply("**__⏳ Please wait, scraping data ...__**", True)
|
|
||||||
if judul is not None:
|
|
||||||
try:
|
|
||||||
html = await http.get(f"{SITE}/?s={judul}", headers=headers)
|
|
||||||
bs4 = BeautifulSoup(html.text, "lxml")
|
|
||||||
for res in bs4.select(".box"):
|
|
||||||
dd = res.select("a")
|
|
||||||
url = dd[0]["href"]
|
|
||||||
title = dd[0]["title"]
|
|
||||||
try:
|
|
||||||
quality = dd[0].find(class_="quality").text
|
|
||||||
except:
|
|
||||||
quality = "N/A"
|
|
||||||
data.append({"judul": title, "link": url, "quality": quality})
|
|
||||||
if not data:
|
|
||||||
await m.delete()
|
|
||||||
return await msg.reply("404 Result not FOUND!", True)
|
|
||||||
await m.delete()
|
|
||||||
head = f"<b>#MelongMovie Results For:</b> <code>{judul}</code>\n\n"
|
|
||||||
msgs = ""
|
|
||||||
for c, i in enumerate(data, start=1):
|
|
||||||
msgs += f"<b>{c}. <a href='{i['link']}'>{i['judul']}</a></b>\n<b>Quality:</b> {i['quality']}\n<b>Extract:</b> <code>/{msg.command[0]}_scrap {i['link']}</code>\n\n"
|
|
||||||
if len(head.encode("utf-8") + msgs.encode("utf-8")) >= 4000:
|
|
||||||
await msg.reply(
|
|
||||||
head + msgs,
|
|
||||||
True,
|
|
||||||
disable_web_page_preview=True,
|
|
||||||
reply_markup=InlineKeyboardMarkup(
|
|
||||||
[
|
|
||||||
[
|
|
||||||
InlineKeyboardButton(
|
|
||||||
text="❌ Close",
|
|
||||||
callback_data=f"close#{msg.from_user.id}",
|
|
||||||
)
|
|
||||||
]
|
|
||||||
]
|
|
||||||
),
|
|
||||||
)
|
|
||||||
await asyncio.sleep(2)
|
|
||||||
msgs = ""
|
|
||||||
if msgs != "":
|
|
||||||
await msg.reply(
|
|
||||||
head + msgs,
|
|
||||||
True,
|
|
||||||
disable_web_page_preview=True,
|
|
||||||
reply_markup=InlineKeyboardMarkup(
|
|
||||||
[
|
|
||||||
[
|
|
||||||
InlineKeyboardButton(
|
|
||||||
text="❌ Close",
|
|
||||||
callback_data=f"close#{msg.from_user.id}",
|
|
||||||
)
|
|
||||||
]
|
|
||||||
]
|
|
||||||
),
|
|
||||||
)
|
|
||||||
except Exception as e:
|
|
||||||
await m.delete()
|
|
||||||
LOGGER.error(e)
|
|
||||||
await msg.reply(str(e), True)
|
|
||||||
else:
|
|
||||||
try:
|
|
||||||
html = await http.get(SITE, headers=headers)
|
|
||||||
bs4 = BeautifulSoup(html.text, "lxml")
|
|
||||||
for res in bs4.select(".box"):
|
|
||||||
dd = res.select("a")
|
|
||||||
url = dd[0]["href"]
|
|
||||||
title = dd[0]["title"]
|
|
||||||
try:
|
|
||||||
quality = dd[0].find(class_="quality").text
|
|
||||||
except:
|
|
||||||
quality = "N/A"
|
|
||||||
data.append({"judul": title, "link": url, "quality": quality})
|
|
||||||
if not data:
|
|
||||||
await m.delete()
|
|
||||||
return await msg.reply("404 Result not FOUND!", True)
|
|
||||||
await m.delete()
|
|
||||||
head = f"<b>#MelongMovie Latest:</b>\n🌀 Use /{msg.command[0]} [title] to start search with title.\n\n"
|
|
||||||
msgs = ""
|
|
||||||
for c, i in enumerate(data, start=1):
|
|
||||||
msgs += f"<b>{c}. <a href='{i['link']}'>{i['judul']}</a></b>\n<b>Quality:</b> {i['quality']}\n<b>Extract:</b> <code>/{msg.command[0]}_scrap {i['link']}</code>\n\n"
|
|
||||||
if len(head.encode("utf-8") + msgs.encode("utf-8")) >= 4000:
|
|
||||||
await msg.reply(
|
|
||||||
head + msgs,
|
|
||||||
True,
|
|
||||||
disable_web_page_preview=True,
|
|
||||||
reply_markup=InlineKeyboardMarkup(
|
|
||||||
[
|
|
||||||
[
|
|
||||||
InlineKeyboardButton(
|
|
||||||
text="❌ Close",
|
|
||||||
callback_data=f"close#{msg.from_user.id}",
|
|
||||||
)
|
|
||||||
]
|
|
||||||
]
|
|
||||||
),
|
|
||||||
)
|
|
||||||
await asyncio.sleep(2)
|
|
||||||
msgs = ""
|
|
||||||
if msgs != "":
|
|
||||||
await msg.reply(
|
|
||||||
head + msgs,
|
|
||||||
True,
|
|
||||||
disable_web_page_preview=True,
|
|
||||||
reply_markup=InlineKeyboardMarkup(
|
|
||||||
[
|
|
||||||
[
|
|
||||||
InlineKeyboardButton(
|
|
||||||
text="❌ Close",
|
|
||||||
callback_data=f"close#{msg.from_user.id}",
|
|
||||||
)
|
|
||||||
]
|
|
||||||
]
|
|
||||||
),
|
|
||||||
)
|
|
||||||
except Exception as e:
|
|
||||||
await m.delete()
|
|
||||||
LOGGER.error(e)
|
|
||||||
await msg.reply(str(e), True)
|
|
||||||
|
|
||||||
|
|
||||||
@app.on_message(filters.command(["savefilm21_scrap"], COMMAND_HANDLER))
|
|
||||||
@capture_err
|
|
||||||
async def savefilm21_scrap(_, message):
|
|
||||||
try:
|
|
||||||
link = message.text.split(" ", maxsplit=1)[1]
|
|
||||||
headers = {"User-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582"}
|
|
||||||
|
|
||||||
html = await http.get(link, headers=headers)
|
|
||||||
soup = BeautifulSoup(html.text, "lxml")
|
|
||||||
res = soup.find_all(class_="button button-shadow")
|
|
||||||
res = "".join(f"{i.text}\n{i['href']}\n\n" for i in res)
|
|
||||||
await message.reply(
|
|
||||||
f"<b>Hasil Scrap dari {link}</b>:\n\n{res}",
|
|
||||||
disable_web_page_preview=True,
|
|
||||||
reply_markup=InlineKeyboardMarkup(
|
|
||||||
[
|
|
||||||
[
|
|
||||||
InlineKeyboardButton(
|
|
||||||
text="❌ Close",
|
|
||||||
callback_data=f"close#{message.from_user.id}",
|
|
||||||
)
|
|
||||||
]
|
|
||||||
]
|
|
||||||
),
|
|
||||||
)
|
|
||||||
except IndexError:
|
|
||||||
return await message.reply(f"Gunakan command /{message.command[0]} <b>[link]</b> untuk scrap link download")
|
|
||||||
except Exception as e:
|
|
||||||
await message.reply(f"ERROR: {str(e)}")
|
|
||||||
|
|
||||||
|
|
||||||
@app.on_message(filters.command(["nodrakor_scrap"], COMMAND_HANDLER))
|
|
||||||
@capture_err
|
|
||||||
async def nodrakor_scrap(_, message):
|
|
||||||
try:
|
|
||||||
link = message.text.split(" ", maxsplit=1)[1]
|
|
||||||
headers = {"User-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582"}
|
|
||||||
|
|
||||||
html = await http.get(link, headers=headers)
|
|
||||||
soup = BeautifulSoup(html.text, "lxml")
|
|
||||||
hasil = soup.find_all(class_="gmr-download-wrap clearfix")[0]
|
|
||||||
await message.reply(f"<b>Hasil Scrap dari {link}</b>:\n{hasil}")
|
|
||||||
except IndexError:
|
|
||||||
return await message.reply(f"Gunakan command /{message.command[0]} <b>[link]</b> untuk scrap link download")
|
|
||||||
except Exception as e:
|
|
||||||
await message.reply(f"ERROR: {str(e)}")
|
|
||||||
|
|
||||||
|
|
||||||
# Scrape Link Download Movieku.CC
|
|
||||||
@app.on_message(filters.command(["movieku_scrap"], COMMAND_HANDLER))
|
|
||||||
@capture_err
|
|
||||||
async def muviku_scrap(_, message):
|
|
||||||
try:
|
|
||||||
link = message.text.split(" ", maxsplit=1)[1]
|
|
||||||
headers = {"User-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582"}
|
|
||||||
|
|
||||||
html = await http.get(link, headers=headers)
|
|
||||||
soup = BeautifulSoup(html.text, "lxml")
|
|
||||||
res = soup.find_all(class_="smokeurl")
|
|
||||||
data = []
|
|
||||||
for i in res:
|
|
||||||
for b in range(len(i.find_all("a"))):
|
|
||||||
link = i.find_all("a")[b]["href"]
|
|
||||||
kualitas = i.find_all("a")[b].text
|
|
||||||
# print(f"{kualitas}\n{link
|
|
||||||
data.append({"link": link, "kualitas": kualitas})
|
|
||||||
if not data:
|
|
||||||
return await message.reply("Oops, data film tidak ditemukan.")
|
|
||||||
res = "".join(f"<b>Host: {i['kualitas']}</b>\n{i['link']}\n\n" for i in data)
|
|
||||||
await message.reply(res)
|
|
||||||
except IndexError:
|
|
||||||
return await message.reply(f"Gunakan command /{message.command[0]} <b>[link]</b> untuk scrap link download")
|
|
||||||
except Exception as e:
|
|
||||||
await message.reply(f"ERROR: {str(e)}")
|
|
||||||
|
|
||||||
|
|
||||||
@app.on_message(filters.command(["melongmovie_scrap"], COMMAND_HANDLER))
|
|
||||||
@capture_err
|
|
||||||
async def melong_scrap(_, message):
|
|
||||||
try:
|
|
||||||
link = message.text.split(" ", maxsplit=1)[1]
|
|
||||||
headers = {"User-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582"}
|
|
||||||
|
|
||||||
html = await http.get(link, headers=headers)
|
|
||||||
soup = BeautifulSoup(html.text, "lxml")
|
|
||||||
for ep in soup.findAll(text=re.compile(r"(?i)episode\s+\d+|LINK DOWNLOAD")):
|
|
||||||
hardsub = ep.findPrevious("div")
|
|
||||||
softsub = ep.findNext("div")
|
|
||||||
rep = f"{hardsub}\n{softsub}"
|
|
||||||
await message.reply(rep)
|
|
||||||
except IndexError:
|
|
||||||
await message.reply(f"Gunakan command /{message.command[0]} <b>[link]</b> untuk scrap link download")
|
|
||||||
|
|
||||||
|
|
||||||
@app.on_message(filters.command(["gomov_scrap", "zonafilm_scrap"], COMMAND_HANDLER))
|
|
||||||
@capture_err
|
|
||||||
async def gomov_zonafilm_dl(_, message):
|
|
||||||
try:
|
|
||||||
link = message.text.split(" ", maxsplit=1)[1]
|
|
||||||
headers = {"User-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582"}
|
|
||||||
|
|
||||||
html = await http.get(link, headers=headers)
|
|
||||||
soup = BeautifulSoup(html.text, "lxml")
|
|
||||||
entry = soup.find(class_="gmr-download-wrap clearfix")
|
|
||||||
hasil = soup.find(class_="title-download").text
|
|
||||||
for i in entry.find(class_="list-inline gmr-download-list clearfix"):
|
|
||||||
title = i.find("a").text
|
|
||||||
link = i.find("a")["href"]
|
|
||||||
hasil += f"\n{title}\n{link}\n"
|
|
||||||
await message.reply(
|
|
||||||
hasil,
|
|
||||||
reply_markup=InlineKeyboardMarkup(
|
|
||||||
[
|
|
||||||
[
|
|
||||||
InlineKeyboardButton(
|
|
||||||
text="❌ Close",
|
|
||||||
callback_data=f"close#{message.from_user.id}",
|
|
||||||
)
|
|
||||||
]
|
|
||||||
]
|
|
||||||
),
|
|
||||||
)
|
|
||||||
except IndexError:
|
|
||||||
await message.reply(f"Gunakan command /{message.command[0]} <b>[link]</b> untuk scrap link download")
|
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue