# -*- coding: utf-8 -*- import io import logging import os import rarfile import re import zipfile from subzero.language import Language from guessit import guessit from requests import Session from six import text_type from subliminal.providers import ParserBeautifulSoup, Provider from subliminal import __short_version__ from subliminal.cache import SHOW_EXPIRATION_TIME, region from subliminal.score import get_equivalent_release_groups from subliminal.subtitle import SUBTITLE_EXTENSIONS, Subtitle, fix_line_ending, guess_matches from subliminal.utils import sanitize, sanitize_release_group from subliminal.video import Episode logger = logging.getLogger(__name__) year_re = re.compile(r'^\((\d{4})\)$') class Subs4SeriesSubtitle(Subtitle): """Subs4Series Subtitle.""" provider_name = 'subs4series' def __init__(self, language, page_link, series, year, version, download_link): super(Subs4SeriesSubtitle, self).__init__(language, page_link=page_link) self.series = series self.year = year self.version = version self.download_link = download_link self.hearing_impaired = None self.encoding = 'windows-1253' @property def id(self): return self.download_link def get_matches(self, video): matches = set() # episode if isinstance(video, Episode): # series name if video.series and sanitize(self.series) in ( sanitize(name) for name in [video.series] + video.alternative_series): matches.add('series') # year if video.original_series and self.year is None or video.year and video.year == self.year: matches.add('year') # release_group if (video.release_group and self.version and any(r in sanitize_release_group(self.version) for r in get_equivalent_release_groups(sanitize_release_group(video.release_group)))): matches.add('release_group') # other properties matches |= guess_matches(video, guessit(self.version, {'type': 'episode'}), partial=True) return matches class Subs4SeriesProvider(Provider): """Subs4Series Provider.""" languages = {Language(l) for l in ['ell', 'eng']} video_types = (Episode,) server_url = 'https://www.subs4series.com' search_url = '/search_report.php?search={}&searchType=1' episode_link = '/tv-series/{show_id}/season-{season:d}/episode-{episode:d}' subtitle_class = Subs4SeriesSubtitle def __init__(self): self.session = None def initialize(self): self.session = Session() self.session.headers['User-Agent'] = 'Subliminal/{}'.format(__short_version__) def terminate(self): self.session.close() def get_show_ids(self, title, year=None): """Get the best matching show id for `series` and `year`. First search in the result of :meth:`_get_show_suggestions`. :param title: show title. :param year: year of the show, if any. :type year: int :return: the show id, if found. :rtype: str """ title_sanitized = sanitize(title).lower() show_ids = self._get_suggestions(title) matched_show_ids = [] for show in show_ids: show_id = None show_title = sanitize(show['title']) # attempt with year if not show_id and year: logger.debug('Getting show id with year') show_id = '/'.join(show['link'].rsplit('/', 2)[1:]) if show_title == '{title} {year:d}'.format( title=title_sanitized, year=year) else None # attempt clean if not show_id: logger.debug('Getting show id') show_id = '/'.join(show['link'].rsplit('/', 2)[1:]) if show_title == title_sanitized else None if show_id: matched_show_ids.append(show_id) return matched_show_ids @region.cache_on_arguments(expiration_time=SHOW_EXPIRATION_TIME, to_str=text_type, should_cache_fn=lambda value: value) def _get_suggestions(self, title): """Search the show or movie id from the `title` and `year`. :param str title: title of the show. :return: the show suggestions found. :rtype: dict """ # make the search logger.info('Searching show ids with %r', title) r = self.session.get(self.server_url + text_type(self.search_url).format(title), headers={'Referer': self.server_url}, timeout=10) r.raise_for_status() if not r.content: logger.debug('No data returned from provider') return {} soup = ParserBeautifulSoup(r.content, ['lxml', 'html.parser']) series = [{'link': l.attrs['value'], 'title': l.text} for l in soup.select('select[name="Mov_sel"] > option[value]')] logger.debug('Found suggestions: %r', series) return series def query(self, show_id, series, season, episode, title): # get the season list of the show logger.info('Getting the subtitle list of show id %s', show_id) if all((show_id, season, episode)): page_link = self.server_url + self.episode_link.format(show_id=show_id, season=season, episode=episode) else: return [] r = self.session.get(page_link, timeout=10) r.raise_for_status() if not r.content: logger.debug('No data returned from provider') return [] soup = ParserBeautifulSoup(r.content, ['lxml', 'html.parser']) year_num = None matches = year_re.match(str(soup.select_one('#dates_header_br > table div').contents[2]).strip()) if matches: year_num = int(matches.group(1)) show_title = str(soup.select_one('#dates_header_br > table u').contents[0]).strip() subtitles = [] # loop over episode rows for subtitle in soup.select('table.table_border div[align="center"] > div'): # read common info version = subtitle.find('b').text download_link = self.server_url + subtitle.find('a')['href'] language = Language.fromalpha2(subtitle.find('img')['src'].split('/')[-1].split('.')[0]) subtitle = self.subtitle_class(language, page_link, show_title, year_num, version, download_link) logger.debug('Found subtitle %r', subtitle) subtitles.append(subtitle) return subtitles def list_subtitles(self, video, languages): # lookup show_id titles = [video.series] + video.alternative_series if isinstance(video, Episode) else [] show_ids = None for title in titles: show_ids = self.get_show_ids(title, video.year) if show_ids and len(show_ids) > 0: break subtitles = [] # query for subtitles with the show_id for show_id in show_ids: subtitles += [s for s in self.query(show_id, video.series, video.season, video.episode, video.title) if s.language in languages] return subtitles def download_subtitle(self, subtitle): if isinstance(subtitle, Subs4SeriesSubtitle): # download the subtitle logger.info('Downloading subtitle %r', subtitle) r = self.session.get(subtitle.download_link, headers={'Referer': subtitle.page_link}, timeout=10) r.raise_for_status() if not r.content: logger.debug('Unable to download subtitle. No data returned from provider') return soup = ParserBeautifulSoup(r.content, ['lxml', 'html.parser']) download_element = soup.select_one('a.style55ws') if not download_element: download_element = soup.select_one('form[method="post"]') target = download_element['action'] if download_element else None else: target = download_element['href'] if not target: logger.debug('Unable to download subtitle. No download link found') return download_url = self.server_url + target r = self.session.get(download_url, headers={'Referer': subtitle.download_link}, timeout=10) r.raise_for_status() if not r.content: logger.debug('Unable to download subtitle. No data returned from provider') return archive = _get_archive(r.content) subtitle_content = _get_subtitle_from_archive(archive) if archive else r.content if subtitle_content: subtitle.content = fix_line_ending(subtitle_content) else: logger.debug('Could not extract subtitle from %r', archive) def _get_archive(content): # open the archive archive_stream = io.BytesIO(content) archive = None if rarfile.is_rarfile(archive_stream): logger.debug('Identified rar archive') archive = rarfile.RarFile(archive_stream) elif zipfile.is_zipfile(archive_stream): logger.debug('Identified zip archive') archive = zipfile.ZipFile(archive_stream) return archive def _get_subtitle_from_archive(archive): for name in archive.namelist(): # discard hidden files if os.path.split(name)[-1].startswith('.'): continue # discard non-subtitle files if not name.lower().endswith(SUBTITLE_EXTENSIONS): continue return archive.read(name) return None