Try fix scraper and add sample auto restart docker container to prevent memory leak..

Signed-off-by: Yasir Aris M <git@yasirdev.my.id>
This commit is contained in:
Yasir Aris M 2023-10-15 23:09:46 +07:00
parent 6cc1a2ae5d
commit bb58c41b9b
2 changed files with 8 additions and 2 deletions

View file

@ -10,4 +10,10 @@ services:
command: bash start.sh
environment:
CONFIG_FILE_URL: $CONFIG_FILE_URL
restart: on-failure
restart: on-failure
# restarter:
# image: docker:cli
# volumes: ["/var/run/docker.sock:/var/run/docker.sock"]
# command: ["/bin/sh", "-c", "while true; do sleep 259200; docker restart ${CONTAINER_NAME}; done"]
# restart: ${RESTART}

View file

@ -271,7 +271,7 @@ async def getDataNodrakor(msg, kueri, CurrentPage, user, strings):
return None, 0, None
text = BeautifulSoup(data, "lxml")
entry = text.find_all(class_="entry-header")
if "Nothing Found" in entry[0].text:
if entry[0].text.strip() == "Nothing Found":
if not kueri:
await msg.edit_msg(strings("no_result"), del_in=5)
else: