mirror of
https://github.com/yasirarism/MissKatyPyro.git
synced 2026-01-03 02:54:52 +00:00
Fix more
This commit is contained in:
parent
9c4f3e3c78
commit
c621ea8cae
1 changed files with 1 additions and 257 deletions
|
|
@ -387,260 +387,4 @@ async def savefilm21(_, msg):
|
|||
except Exception as e:
|
||||
await m.delete()
|
||||
LOGGER.error(e)
|
||||
await msg.reply(f"ERROR: {e}", True)
|
||||
|
||||
|
||||
@app.on_message(filters.command(["melongmovie"], COMMAND_HANDLER))
|
||||
@capture_err
|
||||
async def melongmovie(_, msg):
|
||||
SITE = "http://167.99.31.48"
|
||||
try:
|
||||
judul = msg.text.split(" ", 1)[1]
|
||||
except IndexError:
|
||||
judul = None
|
||||
data = []
|
||||
m = await msg.reply("**__⏳ Please wait, scraping data ...__**", True)
|
||||
if judul is not None:
|
||||
try:
|
||||
html = await http.get(f"{SITE}/?s={judul}", headers=headers)
|
||||
bs4 = BeautifulSoup(html.text, "lxml")
|
||||
for res in bs4.select(".box"):
|
||||
dd = res.select("a")
|
||||
url = dd[0]["href"]
|
||||
title = dd[0]["title"]
|
||||
try:
|
||||
quality = dd[0].find(class_="quality").text
|
||||
except:
|
||||
quality = "N/A"
|
||||
data.append({"judul": title, "link": url, "quality": quality})
|
||||
if not data:
|
||||
await m.delete()
|
||||
return await msg.reply("404 Result not FOUND!", True)
|
||||
await m.delete()
|
||||
head = f"<b>#MelongMovie Results For:</b> <code>{judul}</code>\n\n"
|
||||
msgs = ""
|
||||
for c, i in enumerate(data, start=1):
|
||||
msgs += f"<b>{c}. <a href='{i['link']}'>{i['judul']}</a></b>\n<b>Quality:</b> {i['quality']}\n<b>Extract:</b> <code>/{msg.command[0]}_scrap {i['link']}</code>\n\n"
|
||||
if len(head.encode("utf-8") + msgs.encode("utf-8")) >= 4000:
|
||||
await msg.reply(
|
||||
head + msgs,
|
||||
True,
|
||||
disable_web_page_preview=True,
|
||||
reply_markup=InlineKeyboardMarkup(
|
||||
[
|
||||
[
|
||||
InlineKeyboardButton(
|
||||
text="❌ Close",
|
||||
callback_data=f"close#{msg.from_user.id}",
|
||||
)
|
||||
]
|
||||
]
|
||||
),
|
||||
)
|
||||
await asyncio.sleep(2)
|
||||
msgs = ""
|
||||
if msgs != "":
|
||||
await msg.reply(
|
||||
head + msgs,
|
||||
True,
|
||||
disable_web_page_preview=True,
|
||||
reply_markup=InlineKeyboardMarkup(
|
||||
[
|
||||
[
|
||||
InlineKeyboardButton(
|
||||
text="❌ Close",
|
||||
callback_data=f"close#{msg.from_user.id}",
|
||||
)
|
||||
]
|
||||
]
|
||||
),
|
||||
)
|
||||
except Exception as e:
|
||||
await m.delete()
|
||||
LOGGER.error(e)
|
||||
await msg.reply(str(e), True)
|
||||
else:
|
||||
try:
|
||||
html = await http.get(SITE, headers=headers)
|
||||
bs4 = BeautifulSoup(html.text, "lxml")
|
||||
for res in bs4.select(".box"):
|
||||
dd = res.select("a")
|
||||
url = dd[0]["href"]
|
||||
title = dd[0]["title"]
|
||||
try:
|
||||
quality = dd[0].find(class_="quality").text
|
||||
except:
|
||||
quality = "N/A"
|
||||
data.append({"judul": title, "link": url, "quality": quality})
|
||||
if not data:
|
||||
await m.delete()
|
||||
return await msg.reply("404 Result not FOUND!", True)
|
||||
await m.delete()
|
||||
head = f"<b>#MelongMovie Latest:</b>\n🌀 Use /{msg.command[0]} [title] to start search with title.\n\n"
|
||||
msgs = ""
|
||||
for c, i in enumerate(data, start=1):
|
||||
msgs += f"<b>{c}. <a href='{i['link']}'>{i['judul']}</a></b>\n<b>Quality:</b> {i['quality']}\n<b>Extract:</b> <code>/{msg.command[0]}_scrap {i['link']}</code>\n\n"
|
||||
if len(head.encode("utf-8") + msgs.encode("utf-8")) >= 4000:
|
||||
await msg.reply(
|
||||
head + msgs,
|
||||
True,
|
||||
disable_web_page_preview=True,
|
||||
reply_markup=InlineKeyboardMarkup(
|
||||
[
|
||||
[
|
||||
InlineKeyboardButton(
|
||||
text="❌ Close",
|
||||
callback_data=f"close#{msg.from_user.id}",
|
||||
)
|
||||
]
|
||||
]
|
||||
),
|
||||
)
|
||||
await asyncio.sleep(2)
|
||||
msgs = ""
|
||||
if msgs != "":
|
||||
await msg.reply(
|
||||
head + msgs,
|
||||
True,
|
||||
disable_web_page_preview=True,
|
||||
reply_markup=InlineKeyboardMarkup(
|
||||
[
|
||||
[
|
||||
InlineKeyboardButton(
|
||||
text="❌ Close",
|
||||
callback_data=f"close#{msg.from_user.id}",
|
||||
)
|
||||
]
|
||||
]
|
||||
),
|
||||
)
|
||||
except Exception as e:
|
||||
await m.delete()
|
||||
LOGGER.error(e)
|
||||
await msg.reply(str(e), True)
|
||||
|
||||
|
||||
@app.on_message(filters.command(["savefilm21_scrap"], COMMAND_HANDLER))
|
||||
@capture_err
|
||||
async def savefilm21_scrap(_, message):
|
||||
try:
|
||||
link = message.text.split(" ", maxsplit=1)[1]
|
||||
headers = {"User-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582"}
|
||||
|
||||
html = await http.get(link, headers=headers)
|
||||
soup = BeautifulSoup(html.text, "lxml")
|
||||
res = soup.find_all(class_="button button-shadow")
|
||||
res = "".join(f"{i.text}\n{i['href']}\n\n" for i in res)
|
||||
await message.reply(
|
||||
f"<b>Hasil Scrap dari {link}</b>:\n\n{res}",
|
||||
disable_web_page_preview=True,
|
||||
reply_markup=InlineKeyboardMarkup(
|
||||
[
|
||||
[
|
||||
InlineKeyboardButton(
|
||||
text="❌ Close",
|
||||
callback_data=f"close#{message.from_user.id}",
|
||||
)
|
||||
]
|
||||
]
|
||||
),
|
||||
)
|
||||
except IndexError:
|
||||
return await message.reply(f"Gunakan command /{message.command[0]} <b>[link]</b> untuk scrap link download")
|
||||
except Exception as e:
|
||||
await message.reply(f"ERROR: {str(e)}")
|
||||
|
||||
|
||||
@app.on_message(filters.command(["nodrakor_scrap"], COMMAND_HANDLER))
|
||||
@capture_err
|
||||
async def nodrakor_scrap(_, message):
|
||||
try:
|
||||
link = message.text.split(" ", maxsplit=1)[1]
|
||||
headers = {"User-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582"}
|
||||
|
||||
html = await http.get(link, headers=headers)
|
||||
soup = BeautifulSoup(html.text, "lxml")
|
||||
hasil = soup.find_all(class_="gmr-download-wrap clearfix")[0]
|
||||
await message.reply(f"<b>Hasil Scrap dari {link}</b>:\n{hasil}")
|
||||
except IndexError:
|
||||
return await message.reply(f"Gunakan command /{message.command[0]} <b>[link]</b> untuk scrap link download")
|
||||
except Exception as e:
|
||||
await message.reply(f"ERROR: {str(e)}")
|
||||
|
||||
|
||||
# Scrape Link Download Movieku.CC
|
||||
@app.on_message(filters.command(["movieku_scrap"], COMMAND_HANDLER))
|
||||
@capture_err
|
||||
async def muviku_scrap(_, message):
|
||||
try:
|
||||
link = message.text.split(" ", maxsplit=1)[1]
|
||||
headers = {"User-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582"}
|
||||
|
||||
html = await http.get(link, headers=headers)
|
||||
soup = BeautifulSoup(html.text, "lxml")
|
||||
res = soup.find_all(class_="smokeurl")
|
||||
data = []
|
||||
for i in res:
|
||||
for b in range(len(i.find_all("a"))):
|
||||
link = i.find_all("a")[b]["href"]
|
||||
kualitas = i.find_all("a")[b].text
|
||||
# print(f"{kualitas}\n{link
|
||||
data.append({"link": link, "kualitas": kualitas})
|
||||
if not data:
|
||||
return await message.reply("Oops, data film tidak ditemukan.")
|
||||
res = "".join(f"<b>Host: {i['kualitas']}</b>\n{i['link']}\n\n" for i in data)
|
||||
await message.reply(res)
|
||||
except IndexError:
|
||||
return await message.reply(f"Gunakan command /{message.command[0]} <b>[link]</b> untuk scrap link download")
|
||||
except Exception as e:
|
||||
await message.reply(f"ERROR: {str(e)}")
|
||||
|
||||
|
||||
@app.on_message(filters.command(["melongmovie_scrap"], COMMAND_HANDLER))
|
||||
@capture_err
|
||||
async def melong_scrap(_, message):
|
||||
try:
|
||||
link = message.text.split(" ", maxsplit=1)[1]
|
||||
headers = {"User-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582"}
|
||||
|
||||
html = await http.get(link, headers=headers)
|
||||
soup = BeautifulSoup(html.text, "lxml")
|
||||
for ep in soup.findAll(text=re.compile(r"(?i)episode\s+\d+|LINK DOWNLOAD")):
|
||||
hardsub = ep.findPrevious("div")
|
||||
softsub = ep.findNext("div")
|
||||
rep = f"{hardsub}\n{softsub}"
|
||||
await message.reply(rep)
|
||||
except IndexError:
|
||||
await message.reply(f"Gunakan command /{message.command[0]} <b>[link]</b> untuk scrap link download")
|
||||
|
||||
|
||||
@app.on_message(filters.command(["gomov_scrap", "zonafilm_scrap"], COMMAND_HANDLER))
|
||||
@capture_err
|
||||
async def gomov_zonafilm_dl(_, message):
|
||||
try:
|
||||
link = message.text.split(" ", maxsplit=1)[1]
|
||||
headers = {"User-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582"}
|
||||
|
||||
html = await http.get(link, headers=headers)
|
||||
soup = BeautifulSoup(html.text, "lxml")
|
||||
entry = soup.find(class_="gmr-download-wrap clearfix")
|
||||
hasil = soup.find(class_="title-download").text
|
||||
for i in entry.find(class_="list-inline gmr-download-list clearfix"):
|
||||
title = i.find("a").text
|
||||
link = i.find("a")["href"]
|
||||
hasil += f"\n{title}\n{link}\n"
|
||||
await message.reply(
|
||||
hasil,
|
||||
reply_markup=InlineKeyboardMarkup(
|
||||
[
|
||||
[
|
||||
InlineKeyboardButton(
|
||||
text="❌ Close",
|
||||
callback_data=f"close#{message.from_user.id}",
|
||||
)
|
||||
]
|
||||
]
|
||||
),
|
||||
)
|
||||
except IndexError:
|
||||
await message.reply(f"Gunakan command /{message.command[0]} <b>[link]</b> untuk scrap link download")
|
||||
await msg.reply(f"ERROR: {e}", True)
|
||||
Loading…
Reference in a new issue