Merge pull request #141 from stijndcl/free-games

Check for free games
pull/142/head
Stijn De Clercq 2022-10-18 21:36:28 +02:00 committed by GitHub
commit 1c249adb46
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
18 changed files with 636 additions and 86 deletions

View File

@ -0,0 +1,30 @@
"""Add free games
Revision ID: 9fb84b4d9f0b
Revises: 11388e39bb90
Create Date: 2022-10-13 19:17:58.032182
"""
import sqlalchemy as sa
from alembic import op
# revision identifiers, used by Alembic.
revision = "9fb84b4d9f0b"
down_revision = "11388e39bb90"
branch_labels = None
depends_on = None
def upgrade() -> None:
# ### commands auto generated by Alembic - please adjust! ###
op.create_table(
"free_games", sa.Column("free_game_id", sa.Integer(), nullable=False), sa.PrimaryKeyConstraint("free_game_id")
)
# ### end Alembic commands ###
def downgrade() -> None:
# ### commands auto generated by Alembic - please adjust! ###
op.drop_table("free_games")
# ### end Alembic commands ###

View File

@ -0,0 +1,20 @@
from sqlalchemy import select
from sqlalchemy.ext.asyncio import AsyncSession
from database.schemas import FreeGame
__all__ = ["add_free_games", "filter_present_games"]
async def add_free_games(session: AsyncSession, game_ids: list[int]):
"""Bulk-add a list of IDs into the database"""
games = [FreeGame(free_game_id=game_id) for game_id in game_ids]
session.add_all(games)
await session.commit()
async def filter_present_games(session: AsyncSession, game_ids: list[int]) -> list[int]:
"""Filter a list of game IDs down to the ones that aren't in the database yet"""
statement = select(FreeGame.free_game_id).where(FreeGame.free_game_id.in_(game_ids))
matches: list[int] = (await session.execute(statement)).scalars().all()
return list(set(game_ids).difference(matches))

View File

@ -33,6 +33,7 @@ __all__ = [
"DadJoke", "DadJoke",
"Deadline", "Deadline",
"EasterEgg", "EasterEgg",
"FreeGame",
"GitHubLink", "GitHubLink",
"Link", "Link",
"MemeTemplate", "MemeTemplate",
@ -174,6 +175,14 @@ class EasterEgg(Base):
startswith: bool = Column(Boolean, nullable=False, server_default="1") startswith: bool = Column(Boolean, nullable=False, server_default="1")
class FreeGame(Base):
"""A temporarily free game"""
__tablename__ = "free_games"
free_game_id: int = Column(Integer, primary_key=True)
class GitHubLink(Base): class GitHubLink(Base):
"""A user's GitHub link""" """A user's GitHub link"""

View File

@ -18,7 +18,8 @@ from didier.data.embeds.schedules import (
get_schedule_for_day, get_schedule_for_day,
parse_schedule_from_content, parse_schedule_from_content,
) )
from didier.data.embeds.ufora.announcements import fetch_ufora_announcements from didier.data.rss_feeds.free_games import fetch_free_games
from didier.data.rss_feeds.ufora import fetch_ufora_announcements
from didier.decorators.tasks import timed_task from didier.decorators.tasks import timed_task
from didier.utils.discord.checks import is_owner from didier.utils.discord.checks import is_owner
from didier.utils.types.datetime import LOCAL_TIMEZONE, tz_aware_now from didier.utils.types.datetime import LOCAL_TIMEZONE, tz_aware_now
@ -48,6 +49,7 @@ class Tasks(commands.Cog):
self._tasks = { self._tasks = {
"birthdays": self.check_birthdays, "birthdays": self.check_birthdays,
"free_games": self.pull_free_games,
"schedules": self.pull_schedules, "schedules": self.pull_schedules,
"reminders": self.reminders, "reminders": self.reminders,
"ufora": self.pull_ufora_announcements, "ufora": self.pull_ufora_announcements,
@ -61,12 +63,17 @@ class Tasks(commands.Cog):
if settings.BIRTHDAY_ANNOUNCEMENT_CHANNEL is not None: if settings.BIRTHDAY_ANNOUNCEMENT_CHANNEL is not None:
self.check_birthdays.start() self.check_birthdays.start()
# Only pull free gmaes if a channel was provided
if settings.FREE_GAMES_CHANNEL is not None:
self.pull_free_games.start()
# Only pull announcements if a token was provided # Only pull announcements if a token was provided
if settings.UFORA_RSS_TOKEN is not None and settings.UFORA_ANNOUNCEMENTS_CHANNEL is not None: if settings.UFORA_RSS_TOKEN is not None and settings.UFORA_ANNOUNCEMENTS_CHANNEL is not None:
self.pull_ufora_announcements.start() self.pull_ufora_announcements.start()
self.remove_old_ufora_announcements.start() self.remove_old_ufora_announcements.start()
# Start other tasks # Start other tasks
self.init_schedules.start()
self.reminders.start() self.reminders.start()
self.reset_wordle_word.start() self.reset_wordle_word.start()
self.pull_schedules.start() self.pull_schedules.start()
@ -128,6 +135,36 @@ class Tasks(commands.Cog):
async def _before_check_birthdays(self): async def _before_check_birthdays(self):
await self.client.wait_until_ready() await self.client.wait_until_ready()
@tasks.loop(count=1)
async def init_schedules(self, **kwargs):
"""Tasks that loads the schedules in memory on startup"""
_ = kwargs
await self.client.load_schedules()
@init_schedules.before_loop
async def _before_init_schedules(self):
await self.client.wait_until_ready()
@tasks.loop(minutes=15)
async def pull_free_games(self, **kwargs):
"""Task that checks for free games occasionally"""
_ = kwargs
# No channel to send the embeds to
if settings.FREE_GAMES_CHANNEL is None:
return
async with self.client.postgres_session as session:
games = await fetch_free_games(self.client.http_session, session)
channel = self.client.get_channel(settings.FREE_GAMES_CHANNEL)
for game in games:
await channel.send(embed=game.to_embed())
@pull_free_games.before_loop
async def _before_free_games(self):
await self.client.wait_until_ready()
@tasks.loop(time=DAILY_RESET_TIME) @tasks.loop(time=DAILY_RESET_TIME)
@timed_task(enums.TaskType.SCHEDULES) @timed_task(enums.TaskType.SCHEDULES)
async def pull_schedules(self, **kwargs): async def pull_schedules(self, **kwargs):
@ -166,6 +203,10 @@ class Tasks(commands.Cog):
# Only replace cached version if all schedules succeeded # Only replace cached version if all schedules succeeded
self.client.schedules = new_schedules self.client.schedules = new_schedules
@pull_schedules.before_loop
async def _before_pull_schedules(self):
await self.client.wait_until_ready()
@tasks.loop(minutes=10) @tasks.loop(minutes=10)
@timed_task(enums.TaskType.UFORA_ANNOUNCEMENTS) @timed_task(enums.TaskType.UFORA_ANNOUNCEMENTS)
async def pull_ufora_announcements(self, **kwargs): async def pull_ufora_announcements(self, **kwargs):

View File

@ -0,0 +1,122 @@
import html
from typing import Optional
import discord
from aiohttp import ClientSession
from overrides import overrides
from pydantic import validator
from didier.data.embeds.base import EmbedPydantic
from didier.data.scrapers.common import GameStorePage
from didier.data.scrapers.steam import get_steam_webpage_info
from didier.utils.discord import colours
__all__ = ["SEPARATOR", "FreeGameEmbed"]
from didier.utils.discord.constants import Limits
from didier.utils.types.string import abbreviate
SEPARATOR = " • Free • "
def _get_store_info(store: str) -> tuple[Optional[str], discord.Colour]:
"""Get the image url for a given store"""
store = store.lower()
if "epic" in store:
return (
"https://cdn2.unrealengine.com/"
"Unreal+Engine%2Feg-logo-filled-1255x1272-0eb9d144a0f981d1cbaaa1eb957de7a3207b31bb.png",
colours.epic_games_white(),
)
if "gog" in store:
return (
"https://upload.wikimedia.org/wikipedia/commons/thumb/2/2e/GOG.com_logo.svg/1679px-GOG.com_logo.svg.png",
colours.gog_purple(),
)
if "steam" in store:
return (
"https://upload.wikimedia.org/wikipedia/commons/thumb/8/83/"
"Steam_icon_logo.svg/2048px-Steam_icon_logo.svg.png",
colours.steam_blue(),
)
return None, discord.Colour.random()
class FreeGameEmbed(EmbedPydantic):
"""Embed for free games"""
dc_identifier: int
link: str
title: str
name: Optional[str] = None
store: Optional[str] = None
store_page: Optional[GameStorePage] = None
@validator("title")
def _clean_title(cls, value: str) -> str:
return html.unescape(value)
async def update(self, http_session: ClientSession):
"""Scrape the store page to fetch some information"""
self.name, self.store = self.title.split(SEPARATOR)
store = (self.store or "").lower()
if "steam" in store:
self.store_page = await get_steam_webpage_info(http_session, self.link)
elif "epic" in store:
self.link = "https://store.epicgames.com/free-games"
if self.store_page is not None and self.store_page.url is not None:
self.link = self.store_page.url
@overrides
def to_embed(self, **kwargs) -> discord.Embed:
embed = discord.Embed()
embed.set_author(name=self.store)
store_image, store_colour = _get_store_info(self.store or "")
if store_image is not None:
embed.set_thumbnail(url=store_image)
# Populate with scraped info
if self.store_page is not None:
embed.title = self.store_page.title
embed.set_image(url=self.store_page.image)
embed.description = abbreviate(self.store_page.description, Limits.EMBED_DESCRIPTION_LENGTH)
if self.store_page.original_price is not None and self.store_page.discounted_price is not None:
if self.store_page.discount_percentage is not None:
discount_pct_str = f" ({self.store_page.discount_percentage})"
else:
discount_pct_str = ""
embed.add_field(
name="Price",
value=f"~~{self.store_page.original_price}~~ **{self.store_page.discounted_price}** "
f"{discount_pct_str}",
inline=False,
)
embed.add_field(name="Open in browser", value=f"[{self.link}]({self.link})")
if self.store_page.xdg_open_url is not None:
embed.add_field(
name="Open in app", value=f"[{self.store_page.xdg_open_url}]({self.store_page.xdg_open_url})"
)
else:
embed.title = self.name
embed.add_field(name="Open in browser", value=f"[{self.link}]({self.link})")
embed.url = self.link
embed.colour = store_colour
return embed

View File

@ -1,18 +1,11 @@
import re
from dataclasses import dataclass, field from dataclasses import dataclass, field
from datetime import datetime from datetime import datetime
from typing import Optional from typing import Optional
from zoneinfo import ZoneInfo from zoneinfo import ZoneInfo
import async_timeout
import discord import discord
import feedparser
from aiohttp import ClientSession
from markdownify import markdownify as md from markdownify import markdownify as md
from sqlalchemy.ext.asyncio import AsyncSession
import settings
from database.crud import ufora_announcements as crud
from database.schemas import UforaCourse from database.schemas import UforaCourse
from didier.data.embeds.base import EmbedBaseModel from didier.data.embeds.base import EmbedBaseModel
from didier.utils.discord.colours import ghent_university_blue from didier.utils.discord.colours import ghent_university_blue
@ -20,8 +13,6 @@ from didier.utils.types.datetime import LOCAL_TIMEZONE, int_to_weekday
from didier.utils.types.string import leading from didier.utils.types.string import leading
__all__ = [ __all__ = [
"fetch_ufora_announcements",
"parse_ids",
"UforaNotification", "UforaNotification",
] ]
@ -107,68 +98,3 @@ class UforaNotification(EmbedBaseModel):
":" ":"
f"{leading('0', str(self.published_dt.second))}" f"{leading('0', str(self.published_dt.second))}"
) )
def parse_ids(url: str) -> Optional[tuple[int, int]]:
"""Parse the notification & course id out of a notification url"""
match = re.search(r"\d+-\d+$", url)
if not match:
return None
spl = match[0].split("-")
return int(spl[0]), int(spl[1])
async def fetch_ufora_announcements(
http_session: ClientSession, database_session: AsyncSession
) -> list[UforaNotification]:
"""Fetch all new announcements"""
notifications: list[UforaNotification] = []
# No token provided, don't fetch announcements
if settings.UFORA_RSS_TOKEN is None:
return notifications
courses = await crud.get_courses_with_announcements(database_session)
for course in courses:
course_announcement_ids = list(map(lambda announcement: announcement.announcement_id, course.announcements))
course_url = (
f"https://ufora.ugent.be/d2l/le/news/rss/{course.course_id}/course?token={settings.UFORA_RSS_TOKEN}"
)
# Get the updated feed
with async_timeout.timeout(10):
async with http_session.get(course_url) as response:
feed = feedparser.parse(await response.text())
# Remove old notifications
fresh_feed: list[dict] = []
for entry in feed["entries"]:
parsed = parse_ids(entry["id"])
if parsed is None:
continue
if parsed[0] not in course_announcement_ids:
fresh_feed.append(entry)
if fresh_feed:
for item in fresh_feed:
# Parse id's out
# Technically this can't happen but Mypy angry
parsed = parse_ids(item["id"])
if parsed is None:
continue
# Create a new notification
notification_id, course_id = parsed
notification = UforaNotification(item, course, notification_id, course_id)
notifications.append(notification)
# Create new db entry
await crud.create_new_announcement(database_session, notification_id, course, notification.published_dt)
return notifications

View File

View File

@ -0,0 +1,51 @@
import logging
from http import HTTPStatus
import feedparser
from aiohttp import ClientSession
from sqlalchemy.ext.asyncio import AsyncSession
from database.crud.free_games import add_free_games, filter_present_games
from didier.data.embeds.free_games import SEPARATOR, FreeGameEmbed
logger = logging.getLogger(__name__)
__all__ = ["fetch_free_games"]
async def fetch_free_games(http_session: ClientSession, database_session: AsyncSession) -> list[FreeGameEmbed]:
"""Get a fresh list of free games"""
url = "https://pepeizqdeals.com/?call_custom_simple_rss=1&csrp_cat=12"
async with http_session.get(url) as response:
if response.status != HTTPStatus.OK:
logger.error("Free games GET-request failed with status code %d." % response.status)
return []
feed = feedparser.parse(await response.text())
games: list[FreeGameEmbed] = []
game_ids: list[int] = []
for entry in feed["entries"]:
# Game isn't free
if SEPARATOR not in entry["title"]:
continue
game = FreeGameEmbed.parse_obj(entry)
games.append(game)
game_ids.append(game.dc_identifier)
# Filter out games that we already know
filtered_ids = await filter_present_games(database_session, game_ids)
# Insert new games into the database
await add_free_games(database_session, filtered_ids)
games = list(filter(lambda x: x.dc_identifier in filtered_ids, games))
# Look up additional info
for game in games:
await game.update(http_session)
return games

View File

@ -0,0 +1,78 @@
import re
from typing import Optional
import async_timeout
import feedparser
from aiohttp import ClientSession
from sqlalchemy.ext.asyncio import AsyncSession
import settings
from database.crud import ufora_announcements as crud
from didier.data.embeds.ufora.announcements import UforaNotification
__all__ = ["parse_ids", "fetch_ufora_announcements"]
def parse_ids(url: str) -> Optional[tuple[int, int]]:
"""Parse the notification & course id out of a notification url"""
match = re.search(r"\d+-\d+$", url)
if not match:
return None
spl = match[0].split("-")
return int(spl[0]), int(spl[1])
async def fetch_ufora_announcements(
http_session: ClientSession, database_session: AsyncSession
) -> list[UforaNotification]:
"""Fetch all new announcements"""
notifications: list[UforaNotification] = []
# No token provided, don't fetch announcements
if settings.UFORA_RSS_TOKEN is None:
return notifications
courses = await crud.get_courses_with_announcements(database_session)
for course in courses:
course_announcement_ids = list(map(lambda announcement: announcement.announcement_id, course.announcements))
course_url = (
f"https://ufora.ugent.be/d2l/le/news/rss/{course.course_id}/course?token={settings.UFORA_RSS_TOKEN}"
)
# Get the updated feed
with async_timeout.timeout(10):
async with http_session.get(course_url) as response:
feed = feedparser.parse(await response.text())
# Remove old notifications
fresh_feed: list[dict] = []
for entry in feed["entries"]:
parsed = parse_ids(entry["id"])
if parsed is None:
continue
if parsed[0] not in course_announcement_ids:
fresh_feed.append(entry)
if fresh_feed:
for item in fresh_feed:
# Parse id's out
# Technically this can't happen but Mypy angry
parsed = parse_ids(item["id"])
if parsed is None:
continue
# Create a new notification
notification_id, course_id = parsed
notification = UforaNotification(item, course, notification_id, course_id)
notifications.append(notification)
# Create new db entry
await crud.create_new_announcement(database_session, notification_id, course, notification.published_dt)
return notifications

View File

@ -0,0 +1,58 @@
from dataclasses import dataclass
from typing import Optional, cast
from bs4 import BeautifulSoup, Tag
__all__ = ["GameStorePage", "parse_open_graph_tags"]
@dataclass
class GameStorePage:
"""Dataclass for information on a game's store page"""
description: str
image: str
title: str
xdg_open_url: Optional[str] = None
url: Optional[str] = None
discount_expiry: Optional[int] = None # TODO
discounted_price: Optional[str] = None
original_price: Optional[str] = None
discount_percentage: Optional[str] = None
def parse_open_graph_tags(soup: BeautifulSoup) -> Optional[GameStorePage]:
"""Parse Open Graph Protocol tags out of a webpage
If any of the required tags were not found, this returns None
"""
head = soup.find("head")
if head is None:
return None
head = cast(Tag, head)
title_tag = head.find("meta", property="og:title")
if title_tag is None:
return None
description_tag = head.find("meta", property="og:description")
if description_tag is None:
return None
image_tag = head.find("meta", property="og:image")
if image_tag is None:
return None
url_tag = head.find("meta", property="og:url")
if url_tag is None:
url = None
else:
url = str(url_tag["content"]) # type: ignore
description = str(description_tag["content"]) # type: ignore
image = str(image_tag["content"]) # type: ignore
title = str(title_tag["content"]) # type: ignore
return GameStorePage(title=title, description=description, url=url, image=image)

View File

@ -74,15 +74,10 @@ def get_search_results(bs: BeautifulSoup) -> list[str]:
async def google_search(http_client: ClientSession, query: str): async def google_search(http_client: ClientSession, query: str):
"""Get the first 10 Google search results""" """Get the first 10 Google search results"""
headers = {
"User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_4) AppleWebKit/537.36 (KHTML, like Gecko) "
"Chrome/83.0.4103.97 Safari/537.36"
}
query = urlencode({"q": query}) query = urlencode({"q": query})
# Request 20 results in case of duplicates, bad matches, ... # Request 20 results in case of duplicates, bad matches, ...
async with http_client.get(f"https://www.google.com/search?{query}&num=20&hl=en", headers=headers) as response: async with http_client.get(f"https://www.google.com/search?{query}&num=20&hl=en") as response:
# Something went wrong # Something went wrong
if response.status != http.HTTPStatus.OK: if response.status != http.HTTPStatus.OK:
return SearchData(query, response.status) return SearchData(query, response.status)

View File

@ -0,0 +1,123 @@
import re
from dataclasses import dataclass
from http import HTTPStatus
from typing import Optional, cast
from aiohttp import ClientSession
from bs4 import BeautifulSoup, Tag
from didier.data.scrapers.common import GameStorePage, parse_open_graph_tags
__all__ = ["get_steam_webpage_info"]
@dataclass
class _PriceInfo:
# These are strings because they aren't used as floats,
# and this avoids possible rounding errors
original_price: str
discounted_price: str
discount_percentage: Optional[str]
def __post_init__(self):
"""Fix the price formats"""
self.original_price = "" + self.original_price.replace(",--", ",00").removesuffix("")
self.discounted_price = "" + self.discounted_price.replace(",--", ",00").removesuffix("")
if self.discounted_price == "€0,00":
self.discounted_price = "Free"
def _shorten_url(url: str) -> str:
match = re.search(r"https://store.steampowered.com/app/(\d+)/", url)
if match is None or not match.groups():
return url
return f"https://s.team/a/{match.groups()[0]}"
def _parse_xdg_open_url(url: str) -> Optional[str]:
match = re.search(r"/app/(\d+)/", url)
if match is None or not match.groups():
return None
return f"steam://store/{match.groups()[0]}"
def _get_steam_discounts(soup: BeautifulSoup) -> Optional[_PriceInfo]:
discount_wrapper_tag = soup.find("div", class_="discount_block")
if discount_wrapper_tag is None:
return None
discount_wrapper_tag = cast(Tag, discount_wrapper_tag)
# Parsing the original (non-discounted) price
original_price_tag = discount_wrapper_tag.find("div", class_="discount_original_price")
if original_price_tag is None:
return None
original_price_tag = cast(Tag, original_price_tag)
original_price = original_price_tag.text
if original_price is None:
return None
# Parsing the discounted price
discounted_price_tag = discount_wrapper_tag.find("div", class_="discount_final_price")
if discounted_price_tag is None:
return None
discounted_price_tag = cast(Tag, discounted_price_tag)
discounted_price = discounted_price_tag.text
if discounted_price is None:
return None
percentage_tag = discount_wrapper_tag.find("div", class_="discount_pct")
if percentage_tag is None:
percentage = None
else:
percentage = percentage_tag.text
return _PriceInfo(original_price=original_price, discounted_price=discounted_price, discount_percentage=percentage)
def _clean_title(title: str) -> str:
match = re.search(r"Save [\d,]+% on (.*) on Steam", title)
if match is None or not match.groups():
return title
return match.groups()[0]
async def get_steam_webpage_info(http_session: ClientSession, url: str) -> Optional[GameStorePage]:
"""Scrape a Steam page"""
# If not currently on a Steam page, follow a redirect chain until you are
if not url.startswith("https://store.steampowered.com/"):
async with http_session.head(url, allow_redirects=True) as response:
url = str(response.url)
async with http_session.get(url) as response:
if response.status != HTTPStatus.OK:
return None
page = await response.text()
soup = BeautifulSoup(page, "html.parser")
page_tags = parse_open_graph_tags(soup)
if page_tags is None:
return None
if page_tags.url is None:
page_tags.url = url
page_tags.title = _clean_title(page_tags.title)
page_tags.xdg_open_url = _parse_xdg_open_url(page_tags.url)
page_tags.url = _shorten_url(page_tags.url)
price_info = _get_steam_discounts(soup)
if price_info is not None:
page_tags.original_price = price_info.original_price
page_tags.discounted_price = price_info.discounted_price
page_tags.discount_percentage = price_info.discount_percentage
return page_tags

View File

@ -78,9 +78,6 @@ class Didier(commands.Bot):
# Create directories that are ignored on GitHub # Create directories that are ignored on GitHub
self._create_ignored_directories() self._create_ignored_directories()
# Load schedules
await self.load_schedules()
# Load the Wordle dictionary # Load the Wordle dictionary
self._load_wordle_words() self._load_wordle_words()
@ -90,7 +87,12 @@ class Didier(commands.Bot):
await self.database_caches.initialize_caches(session) await self.database_caches.initialize_caches(session)
# Create aiohttp session # Create aiohttp session
self.http_session = ClientSession() self.http_session = ClientSession(
headers={
"User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) "
"Chrome/105.0.0.0 Safari/537.36"
}
)
# Load extensions # Load extensions
await self._load_initial_extensions() await self._load_initial_extensions()

View File

@ -1,15 +1,21 @@
import discord import discord
__all__ = [ __all__ = [
"epic_games_white",
"error_red", "error_red",
"github_white", "github_white",
"ghent_university_blue", "ghent_university_blue",
"ghent_university_yellow", "ghent_university_yellow",
"google_blue", "google_blue",
"steam_blue",
"urban_dictionary_green", "urban_dictionary_green",
] ]
def epic_games_white() -> discord.Colour:
return discord.Colour.from_rgb(255, 255, 255)
def error_red() -> discord.Colour: def error_red() -> discord.Colour:
return discord.Colour.red() return discord.Colour.red()
@ -26,9 +32,17 @@ def ghent_university_yellow() -> discord.Colour:
return discord.Colour.from_rgb(255, 210, 0) return discord.Colour.from_rgb(255, 210, 0)
def gog_purple() -> discord.Colour:
return discord.Colour.purple()
def google_blue() -> discord.Colour: def google_blue() -> discord.Colour:
return discord.Colour.from_rgb(66, 133, 244) return discord.Colour.from_rgb(66, 133, 244)
def steam_blue() -> discord.Colour:
return discord.Colour.from_rgb(102, 192, 244)
def urban_dictionary_green() -> discord.Colour: def urban_dictionary_green() -> discord.Colour:
return discord.Colour.from_rgb(220, 255, 0) return discord.Colour.from_rgb(220, 255, 0)

View File

@ -27,6 +27,8 @@ __all__ = [
"DISCORD_TEST_GUILDS", "DISCORD_TEST_GUILDS",
"DISCORD_BOOS_REACT", "DISCORD_BOOS_REACT",
"DISCORD_CUSTOM_COMMAND_PREFIX", "DISCORD_CUSTOM_COMMAND_PREFIX",
"ERRORS_CHANNEL",
"FREE_GAMES_CHANNEL",
"UFORA_ANNOUNCEMENTS_CHANNEL", "UFORA_ANNOUNCEMENTS_CHANNEL",
"UFORA_RSS_TOKEN", "UFORA_RSS_TOKEN",
"IMGFLIP_NAME", "IMGFLIP_NAME",
@ -65,6 +67,7 @@ DISCORD_BOOS_REACT: str = env.str("DISCORD_BOOS_REACT", "<:boos:6296037858402631
DISCORD_CUSTOM_COMMAND_PREFIX: str = env.str("DISCORD_CUSTOM_COMMAND_PREFIX", "?") DISCORD_CUSTOM_COMMAND_PREFIX: str = env.str("DISCORD_CUSTOM_COMMAND_PREFIX", "?")
BIRTHDAY_ANNOUNCEMENT_CHANNEL: Optional[int] = env.int("BIRTHDAY_ANNOUNCEMENT_CHANNEL", None) BIRTHDAY_ANNOUNCEMENT_CHANNEL: Optional[int] = env.int("BIRTHDAY_ANNOUNCEMENT_CHANNEL", None)
ERRORS_CHANNEL: Optional[int] = env.int("ERRORS_CHANNEL", None) ERRORS_CHANNEL: Optional[int] = env.int("ERRORS_CHANNEL", None)
FREE_GAMES_CHANNEL: Optional[int] = env.int("FREE_GAMES_CHANNEL", None)
UFORA_ANNOUNCEMENTS_CHANNEL: Optional[int] = env.int("UFORA_ANNOUNCEMENTS_CHANNEL", None) UFORA_ANNOUNCEMENTS_CHANNEL: Optional[int] = env.int("UFORA_ANNOUNCEMENTS_CHANNEL", None)
"""Discord Role ID's""" """Discord Role ID's"""

View File

@ -1,5 +1,7 @@
import asyncio import asyncio
from typing import AsyncGenerator, Generator import json
import pathlib
from typing import AsyncGenerator, Generator, Union
from unittest.mock import MagicMock from unittest.mock import MagicMock
import pytest import pytest
@ -66,3 +68,22 @@ def mock_client() -> Didier:
mock_client.user = mock_user mock_client.user = mock_user
return mock_client return mock_client
"""Data providers"""
def _provide(name: str) -> Union[dict, str]:
location = pathlib.Path(__file__).parent / "test_data" / name
with open(location, "r") as fp:
if name.endswith(".json"):
return json.load(fp)
return fp.read()
@pytest.fixture
def free_games_response() -> str:
"""Fixture to get an example response from the free games RSS feed"""
return _provide("free_games.rss")

View File

@ -0,0 +1,40 @@
<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:wfw="http://wellformedweb.org/CommentAPI/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:sy="http://purl.org/rss/1.0/modules/syndication/" xmlns:slash="http://purl.org/rss/1.0/modules/slash/" xmlns:media="http://search.yahoo.com/mrss/" xmlns:wp="http://wordpress.org/export/1.2/" xmlns:excerpt="http://wordpress.org/export/1.2/excerpt/" >
<channel>
<title>pepeizq&#039;s deals</title>
<description>Follow the latest deals for PC games from legit stores such as Steam, Humble, Fanatical, Gamesplanet, GOG and more</description>
<link>https://pepeizqdeals.com</link>
<lastBuildDate>Thu, 13 Oct 2022 17:11:24 +0000</lastBuildDate>
<item>
<title><![CDATA[Minion Masters &#8211; Torment • Free • Steam]]></title>
<link><![CDATA[https://pepeizqdeals.com/55623/minion-masters-torment-free-steam/]]></link>
<pubDate>Thu, 13 Oct 2022 18:08:41 +0100</pubDate>
<dc:creator>pepeizq</dc:creator>
<dc:identifier>55623</dc:identifier>
<dc:modified>2022-10-13 18:08:59</dc:modified>
<dc:created unix="1665684521">2022-10-13 18:08:41</dc:created>
<guid isPermaLink="true"><![CDATA[https://pepeizqdeals.com/55623/minion-masters-torment-free-steam/]]></guid><category>12</category>
<description><![CDATA[]]></description><content:encoded><![CDATA[]]></content:encoded><enclosure url="https://pepeizqdeals.com/wp-content/uploads/2022/10/imagenWeb286-19-8-510-en.webp"/><media:content url="https://pepeizqdeals.com/wp-content/uploads/2022/10/imagenWeb286-19-8-510-en.webp" height="150" width="150" type="image/jpeg"/>
</item>
<item>
<title><![CDATA[Darkwood + ToeJam &#038; Earl: Back in the Groove! • Free • Epic Games Store]]></title>
<link><![CDATA[https://pepeizqdeals.com/55616/darkwood-toejam-earl-back-in-the-groove-free-epic-games-store/]]></link>
<pubDate>Thu, 13 Oct 2022 17:03:59 +0100</pubDate>
<dc:creator>pepeizq</dc:creator>
<dc:identifier>55616</dc:identifier>
<dc:modified>2022-10-13 17:04:17</dc:modified>
<dc:created unix="1665680639">2022-10-13 17:03:59</dc:created>
<guid isPermaLink="true"><![CDATA[https://pepeizqdeals.com/55616/darkwood-toejam-earl-back-in-the-groove-free-epic-games-store/]]></guid><category>12</category>
<description><![CDATA[]]></description><content:encoded><![CDATA[]]></content:encoded><enclosure url="https://pepeizqdeals.com/wp-content/uploads/2022/10/imagenWeb286-18-3-139-en.webp"/><media:content url="https://pepeizqdeals.com/wp-content/uploads/2022/10/imagenWeb286-18-3-139-en.webp" height="150" width="150" type="image/jpeg"/>
</item>
<item>
<title><![CDATA[Rebel Inc: Escalation &#8211; Sand &#038; Secrets • Free • Steam]]></title>
<link><![CDATA[https://pepeizqdeals.com/54874/rebel-inc-escalation-sand-secrets-free-steam/]]></link>
<pubDate>Tue, 20 Sep 2022 18:08:52 +0100</pubDate>
<dc:creator>pepeizq</dc:creator>
<dc:identifier>54874</dc:identifier>
<dc:modified>2022-09-20 18:09:03</dc:modified>
<dc:created unix="1663697332">2022-09-20 18:08:52</dc:created>
<guid isPermaLink="true"><![CDATA[https://pepeizqdeals.com/54874/rebel-inc-escalation-sand-secrets-free-steam/]]></guid><category>12</category>
<description><![CDATA[]]></description><content:encoded><![CDATA[]]></content:encoded><enclosure url=""/><media:content url="" height="" width="" type=""/>
</item></channel></rss><!-- end of xml string -->

View File

@ -0,0 +1,17 @@
from sqlalchemy import select
from sqlalchemy.ext.asyncio import AsyncSession
from database.crud import free_games as crud
from database.schemas import FreeGame
async def test_add_games(postgres: AsyncSession):
"""Test adding new games"""
statement = select(FreeGame)
games = (await postgres.execute(statement)).scalars().all()
assert not games
await crud.add_free_games(postgres, [1, 2, 3, 4])
games = (await postgres.execute(statement)).scalars().all()
assert len(games) == 4