mirror of
https://gitlab.com/chicken-riders/RcGcDw.git
synced 2025-02-23 00:24:09 +00:00
99 lines
3.7 KiB
Python
99 lines
3.7 KiB
Python
# -*- coding: utf-8 -*-
|
|
|
|
# Recent changes Goat compatible Discord webhook is a project for using a webhook as recent changes page from MediaWiki.
|
|
# Copyright (C) 2020 Frisk
|
|
|
|
# This program is free software: you can redistribute it and/or modify
|
|
# it under the terms of the GNU General Public License as published by
|
|
# the Free Software Foundation, either version 3 of the License, or
|
|
# (at your option) any later version.
|
|
|
|
# This program is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU General Public License for more details.
|
|
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
import logging, gettext, schedule, requests
|
|
from src.configloader import settings
|
|
|
|
from src.discussion_formatters import embed_formatter, compact_formatter
|
|
from src.misc import datafile, messagequeue
|
|
from src.session import session
|
|
|
|
# Initialize translation
|
|
|
|
t = gettext.translation('discussions', localedir='locale', languages=[settings["lang"]])
|
|
_ = t.gettext
|
|
|
|
# Create a custom logger
|
|
|
|
discussion_logger = logging.getLogger("rcgcdw.disc")
|
|
|
|
# Create a variable in datafile if it doesn't exist yet (in files <1.10)
|
|
|
|
if "discussion_id" not in datafile.data:
|
|
datafile.data["discussion_id"] = 0
|
|
datafile.save_datafile()
|
|
|
|
storage = datafile.data
|
|
|
|
fetch_url = "https://services.fandom.com/discussion/{wikiid}/posts?sortDirection=descending&sortKey=creation_date&limit={limit}".format(wikiid=settings["fandom_discussions"]["wiki_id"], limit=settings["fandom_discussions"]["limit"])
|
|
|
|
|
|
def fetch_discussions():
|
|
messagequeue.resend_msgs()
|
|
request = safe_request(fetch_url)
|
|
if request:
|
|
try:
|
|
request_json = request.json()["_embedded"]["doc:posts"]
|
|
request_json.reverse()
|
|
except ValueError:
|
|
discussion_logger.warning("ValueError in fetching discussions")
|
|
return None
|
|
except KeyError:
|
|
discussion_logger.warning("Wiki returned %s" % (request_json.json()))
|
|
return None
|
|
else:
|
|
if request_json:
|
|
for post in request_json:
|
|
if int(post["id"]) > storage["discussion_id"]:
|
|
parse_discussion_post(post)
|
|
if int(post["id"]) > storage["discussion_id"]:
|
|
storage["discussion_id"] = int(post["id"])
|
|
datafile.save_datafile()
|
|
|
|
def parse_discussion_post(post):
|
|
"""Initial post recognition & handling"""
|
|
post_type = post.get("funnel", "TEXT")
|
|
if post_type == "TEXT":
|
|
formatter(post, post_type)
|
|
elif post_type == "POLL":
|
|
formatter(post, post_type)
|
|
else:
|
|
discussion_logger.warning("The type of {} is an unknown discussion post type. Please post an issue on the project page to have it added https://gitlab.com/piotrex43/RcGcDw/-/issues.")
|
|
|
|
|
|
def safe_request(url):
|
|
"""Function to assure safety of request, and do not crash the script on exceptions,"""
|
|
try:
|
|
request = session.get(url, timeout=10, allow_redirects=False, headers={"Accept": "application/hal+json"})
|
|
except requests.exceptions.Timeout:
|
|
discussion_logger.warning("Reached timeout error for request on link {url}".format(url=url))
|
|
return None
|
|
except requests.exceptions.ConnectionError:
|
|
discussion_logger.warning("Reached connection error for request on link {url}".format(url=url))
|
|
return None
|
|
except requests.exceptions.ChunkedEncodingError:
|
|
discussion_logger.warning("Detected faulty response from the web server for request on link {url}".format(url=url))
|
|
return None
|
|
else:
|
|
if 499 < request.status_code < 600:
|
|
return None
|
|
return request
|
|
|
|
formatter = embed_formatter if settings["fandom_discussions"]["appearance"]["mode"] == "embed" else compact_formatter
|
|
|
|
schedule.every(settings["fandom_discussions"]["cooldown"]).seconds.do(fetch_discussions) |