mirror of
https://github.com/yasirarism/MissKatyPyro.git
synced 2026-01-07 20:34:52 +00:00
Update web_scraper.py
This commit is contained in:
parent
cd5ca0eb87
commit
467d851ea5
1 changed files with 4 additions and 4 deletions
|
|
@ -129,7 +129,7 @@ async def getDataPahe(msg, kueri, CurrentPage):
|
||||||
async def getDataKuso(msg, kueri, CurrentPage, user):
|
async def getDataKuso(msg, kueri, CurrentPage, user):
|
||||||
if not SCRAP_DICT.get(msg.id):
|
if not SCRAP_DICT.get(msg.id):
|
||||||
kusodata = []
|
kusodata = []
|
||||||
data = await http.get(f'https://kusonime.com/?s={kueri}', headers=headers, verify=False)
|
data = await http.get(f'https://kusonime.com/?s={kueri}', headers=headers)
|
||||||
res = BeautifulSoup(data.text, "lxml").find_all("h2", {"class": "episodeye"})
|
res = BeautifulSoup(data.text, "lxml").find_all("h2", {"class": "episodeye"})
|
||||||
for i in res:
|
for i in res:
|
||||||
ress = i.find_all("a")[0]
|
ress = i.find_all("a")[0]
|
||||||
|
|
@ -168,7 +168,7 @@ async def getDataKuso(msg, kueri, CurrentPage, user):
|
||||||
async def getDataMovieku(msg, kueri, CurrentPage):
|
async def getDataMovieku(msg, kueri, CurrentPage):
|
||||||
if not SCRAP_DICT.get(msg.id):
|
if not SCRAP_DICT.get(msg.id):
|
||||||
moviekudata = []
|
moviekudata = []
|
||||||
data = await http.get(f'https://107.152.37.223/?s={kueri}', headers=headers, verify=False)
|
data = await http.get(f'https://107.152.37.223/?s={kueri}', headers=headers)
|
||||||
r = BeautifulSoup(data.text, "lxml")
|
r = BeautifulSoup(data.text, "lxml")
|
||||||
res = r.find_all(class_="bx")
|
res = r.find_all(class_="bx")
|
||||||
for i in res:
|
for i in res:
|
||||||
|
|
@ -199,7 +199,7 @@ async def getDataMovieku(msg, kueri, CurrentPage):
|
||||||
async def getDataSavefilm21(msg, kueri, CurrentPage, user):
|
async def getDataSavefilm21(msg, kueri, CurrentPage, user):
|
||||||
if not SCRAP_DICT.get(msg.id):
|
if not SCRAP_DICT.get(msg.id):
|
||||||
sfdata = []
|
sfdata = []
|
||||||
data = await http.get(f'https://185.99.135.215/?s={kueri}', headers=headers, verify=False)
|
data = await http.get(f'https://185.99.135.215/?s={kueri}', headers=headers)
|
||||||
text = BeautifulSoup(data.text, "lxml")
|
text = BeautifulSoup(data.text, "lxml")
|
||||||
entry = text.find_all(class_="entry-header")
|
entry = text.find_all(class_="entry-header")
|
||||||
if "Tidak Ditemukan" in entry[0].text:
|
if "Tidak Ditemukan" in entry[0].text:
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue