155 lines
5 KiB
Python
155 lines
5 KiB
Python
"""
|
|
IGDB
|
|
|
|
use (e.g. "portal-2") as id, which is different from real id in IGDB API
|
|
"""
|
|
|
|
import datetime
|
|
import json
|
|
import logging
|
|
|
|
import requests
|
|
from django.conf import settings
|
|
from igdb.wrapper import IGDBWrapper
|
|
|
|
from catalog.common import *
|
|
from catalog.models import *
|
|
|
|
_logger = logging.getLogger(__name__)
|
|
|
|
|
|
def _igdb_access_token():
|
|
try:
|
|
token = requests.post(
|
|
f"https://id.twitch.tv/oauth2/token?client_id={settings.IGDB_CLIENT_ID}&client_secret={settings.IGDB_CLIENT_SECRET}&grant_type=client_credentials"
|
|
).json()["access_token"]
|
|
except Exception:
|
|
_logger.error("unable to obtain IGDB token")
|
|
token = "<invalid>"
|
|
return token
|
|
|
|
|
|
_wrapper = IGDBWrapper(settings.IGDB_CLIENT_ID, _igdb_access_token())
|
|
|
|
|
|
def search_igdb_by_3p_url(steam_url):
|
|
r = IGDB.api_query("websites", f'fields *, game.*; where url = "{steam_url}";')
|
|
if not r:
|
|
return None
|
|
r = sorted(r, key=lambda w: w["game"]["id"])
|
|
return IGDB(url=r[0]["game"]["url"])
|
|
|
|
|
|
@SiteManager.register
|
|
class IGDB(AbstractSite):
|
|
SITE_NAME = SiteName.IGDB
|
|
ID_TYPE = IdType.IGDB
|
|
URL_PATTERNS = [
|
|
r"\w+://www\.igdb\.com/games/([a-zA-Z0-9\-_]+)",
|
|
r"\w+://m\.igdb\.com/games/([a-zA-Z0-9\-_]+)",
|
|
]
|
|
WIKI_PROPERTY_ID = "?"
|
|
DEFAULT_MODEL = Game
|
|
|
|
@classmethod
|
|
def id_to_url(cls, id_value):
|
|
return "https://www.igdb.com/games/" + id_value
|
|
|
|
@classmethod
|
|
def api_query(cls, p, q):
|
|
key = "igdb:" + p + "/" + q
|
|
if get_mock_mode():
|
|
r = BasicDownloader(key).download().json()
|
|
else:
|
|
r = json.loads(_wrapper.api_request(p, q))
|
|
if settings.DOWNLOADER_SAVEDIR:
|
|
with open(
|
|
settings.DOWNLOADER_SAVEDIR + "/" + get_mock_file(key),
|
|
"w",
|
|
encoding="utf-8",
|
|
) as fp:
|
|
fp.write(json.dumps(r))
|
|
return r
|
|
|
|
def scrape(self):
|
|
fields = "*, cover.url, genres.name, platforms.name, involved_companies.*, involved_companies.company.name"
|
|
r = self.api_query("games", f'fields {fields}; where url = "{self.url}";')[0]
|
|
brief = r["summary"] if "summary" in r else ""
|
|
brief += "\n\n" + r["storyline"] if "storyline" in r else ""
|
|
developer = None
|
|
publisher = None
|
|
release_date = None
|
|
genre = None
|
|
platform = None
|
|
if "involved_companies" in r:
|
|
developer = next(
|
|
iter(
|
|
[
|
|
c["company"]["name"]
|
|
for c in r["involved_companies"]
|
|
if c["developer"]
|
|
]
|
|
),
|
|
None,
|
|
)
|
|
publisher = next(
|
|
iter(
|
|
[
|
|
c["company"]["name"]
|
|
for c in r["involved_companies"]
|
|
if c["publisher"]
|
|
]
|
|
),
|
|
None,
|
|
)
|
|
if "platforms" in r:
|
|
ps = sorted(r["platforms"], key=lambda p: p["id"])
|
|
platform = [(p["name"] if p["id"] != 6 else "Windows") for p in ps]
|
|
if "first_release_date" in r:
|
|
release_date = datetime.datetime.fromtimestamp(
|
|
r["first_release_date"], datetime.timezone.utc
|
|
).strftime("%Y-%m-%d")
|
|
if "genres" in r:
|
|
genre = [g["name"] for g in r["genres"]]
|
|
websites = self.api_query(
|
|
"websites", f'fields *; where game.url = "{self.url}";'
|
|
)
|
|
steam_url = None
|
|
official_site = None
|
|
for website in websites:
|
|
if website["category"] == 1:
|
|
official_site = website["url"]
|
|
elif website["category"] == 13:
|
|
steam_url = website["url"]
|
|
pd = ResourceContent(
|
|
metadata={
|
|
"title": r["name"],
|
|
"other_title": [],
|
|
"developer": [developer],
|
|
"publisher": [publisher],
|
|
"release_date": release_date,
|
|
"genre": genre,
|
|
"platform": platform,
|
|
"brief": brief,
|
|
"official_site": official_site,
|
|
"igdb_id": r["id"],
|
|
"cover_image_url": "https:"
|
|
+ r["cover"]["url"].replace("t_thumb", "t_cover_big")
|
|
if r.get("cover")
|
|
else None,
|
|
}
|
|
)
|
|
if steam_url:
|
|
pd.lookup_ids[IdType.Steam] = SiteManager.get_site_cls_by_id_type(
|
|
IdType.Steam
|
|
).url_to_id(steam_url)
|
|
if pd.metadata["cover_image_url"]:
|
|
imgdl = BasicImageDownloader(pd.metadata["cover_image_url"], self.url)
|
|
try:
|
|
pd.cover_image = imgdl.download().content
|
|
pd.cover_image_extention = imgdl.extention
|
|
except Exception:
|
|
_logger.debug(
|
|
f'failed to download cover for {self.url} from {pd.metadata["cover_image_url"]}'
|
|
)
|
|
return pd
|