You cannot select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

106 lines
3.4 KiB
Python

from __future__ import annotations
import re
from typing import List, Set, TypeAlias
from pony import orm
from ...models import MediaCollection, MediaCollectionLink, MediaElement
from ..generic import (
ChangedReport,
ExtractedDataOnline,
ExtractedDataOffline,
SuitableLevel,
)
from .base import CollectionExtractor
DataType: TypeAlias = List[List[MediaElement]]
class AggregatedCollectionExtractor(CollectionExtractor[DataType]):
__uri_regex = re.compile(r"^aggregated:///(?P<id>\d+(,\d+)*)")
@classmethod
def __get_id(cls, uri: str) -> List[int]:
m = cls.__uri_regex.search(uri)
if not m:
raise Exception(f"Failed to parse aggregated collection uri {uri!r}")
return [int(i) for i in m.group("id").split(",")]
@classmethod
def __get_collections(cls, uri: str) -> List[MediaCollection]:
aggregate_ids = cls.__get_id(uri)
colls = [MediaCollection[i] for i in aggregate_ids]
return colls
def __init__(self) -> None:
super().__init__(
key=".extractor/.aggregated",
long_name="Aggregated Lists",
name="aggregated",
)
def uri_suitable(self, uri: str) -> SuitableLevel:
return SuitableLevel.always_or_no(self.__uri_regex.match(uri) is not None)
def can_extract_offline(self, uri: str) -> bool:
return True
def _cache_expired(self, object: MediaCollection) -> bool:
colls = self.__get_collections(object.primary_uri)
for c in colls:
if c.last_updated is None or object.last_updated <= c.last_updated:
return True
return False
def _extract_offline(self, uri: str) -> ExtractedDataOffline[DataType]:
coll_id = ",".join(str(i) for i in self.__get_id(uri))
return ExtractedDataOffline[DataType](
extractor_name=self.name,
object_key=coll_id,
object_uri=uri,
)
def _extract_online(self, uri: str) -> ExtractedDataOnline[DataType]:
colls = self.__get_collections(uri)
coll_id = ",".join(str(c.id) for c in colls)
return ExtractedDataOnline[DataType](
extractor_name=self.name,
object_key=coll_id,
object_uri=uri,
data=[
[
l.element
for l in orm.select(l for l in c.media_links).order_by(
MediaCollectionLink.sort_key
)
]
for c in colls
],
)
def _update_object_raw(
self,
object: MediaCollection,
data: DataType,
) -> ChangedReport:
if object.title is None or "[aggregated]" not in object.title:
object.title = f"[aggregated] {object.primary_uri}"
object.creator = None
object.set_watch_in_order_auto(True)
all_links: Set[int] = set(
orm.select(link.element.id for link in object.media_links)
)
for season, media_list in enumerate(data):
for episode, media in enumerate(media_list):
all_links.discard(media.id)
object.add_episode(
media=media,
season=season + 1,
episode=episode + 1,
)
orm.delete(link for link in object.media_links if link.element.id in all_links)
object.set_as_only_uri(object.primary_uri)
return ChangedReport.ChangedSome # TODO improve