Add defensive programming to site downloaders
This commit is contained in:
@@ -7,7 +7,7 @@ from typing import Optional
|
||||
import bs4
|
||||
from praw.models import Submission
|
||||
|
||||
from bulkredditdownloader.exceptions import NotADownloadableLinkError
|
||||
from bulkredditdownloader.exceptions import SiteDownloaderError
|
||||
from bulkredditdownloader.resource import Resource
|
||||
from bulkredditdownloader.site_authenticator import SiteAuthenticator
|
||||
from bulkredditdownloader.site_downloaders.base_downloader import BaseDownloader
|
||||
@@ -21,8 +21,9 @@ class Erome(BaseDownloader):
|
||||
|
||||
def find_resources(self, authenticator: Optional[SiteAuthenticator] = None) -> list[Resource]:
|
||||
links = self._get_links(self.post.url)
|
||||
|
||||
if not links:
|
||||
raise NotADownloadableLinkError('Erome parser could not find any links')
|
||||
raise SiteDownloaderError('Erome parser could not find any links')
|
||||
|
||||
out = []
|
||||
for link in links:
|
||||
|
||||
Reference in New Issue
Block a user