|
1 | | -import itertools, os |
| 1 | +import os |
2 | 2 | from pathlib import Path |
3 | 3 | from types import ModuleType |
4 | 4 |
|
5 | 5 | import gallery_dl |
6 | | -from gallery_dl.exception import StopExtraction |
7 | | -from gallery_dl.extractor.message import Message |
8 | 6 | from gallery_dl.job import Job |
9 | | -from gallery_dl.util import build_duration_func |
10 | 7 |
|
11 | 8 | from library.mediadb import db_media, db_playlists |
12 | 9 | from library.utils import consts, printing, strings |
@@ -141,70 +138,57 @@ def download(args, m): |
141 | 138 | ) |
142 | 139 |
|
143 | 140 |
|
144 | | -class GeneratorJob(Job): |
145 | | - def __init__(self, *args, **kwargs): |
146 | | - super().__init__(*args, **kwargs) |
147 | | - if hasattr(super(), "_init"): |
148 | | - super()._init() |
149 | | - self.dispatched = False |
| 141 | +class UrlJob(Job): |
| 142 | + resolve = 1 # depth of queue resolution |
| 143 | + |
| 144 | + def __init__(self, url, parent=None, resolve=None): |
| 145 | + super().__init__(url, parent) |
| 146 | + self.results = [] |
150 | 147 | self.visited = set() |
151 | | - self.status = 0 |
152 | | - |
153 | | - def message_generator(self): |
154 | | - extractor = self.extractor |
155 | | - sleep = build_duration_func(extractor.config("sleep-extractor")) |
156 | | - if sleep: |
157 | | - extractor.sleep(sleep(), "extractor") |
158 | | - |
159 | | - try: |
160 | | - for msg in extractor: |
161 | | - self.dispatch(msg) |
162 | | - if self.dispatched: |
163 | | - yield msg |
164 | | - self.dispatched = False |
165 | | - except StopExtraction: |
166 | | - pass |
167 | | - |
168 | | - def run(self): |
169 | | - for msg in self.message_generator(): |
170 | | - ident, url, kwdict = msg |
171 | | - if ident == Message.Url: |
172 | | - yield (msg[1], msg[2]) |
173 | | - |
174 | | - elif ident == Message.Queue: |
175 | | - if url in self.visited: |
176 | | - continue |
177 | | - self.visited.add(url) |
178 | | - |
179 | | - cls = kwdict.get("_extractor") |
180 | | - if cls: |
181 | | - extr = cls.from_url(url) |
182 | | - else: |
183 | | - extr = self.extractor.find(url) |
184 | | - |
185 | | - if extr: |
186 | | - job = self.__class__(extr, self) |
187 | | - yield from job.run() |
188 | | - else: |
189 | | - raise TypeError |
| 148 | + if resolve is not None: |
| 149 | + self.resolve = resolve |
| 150 | + |
| 151 | + if self.resolve > 0: |
| 152 | + self.handle_queue = self.handle_queue_resolve |
190 | 153 |
|
191 | 154 | def handle_url(self, url, kwdict): |
192 | | - self.dispatched = True |
| 155 | + self.results.append((url, kwdict)) |
193 | 156 |
|
194 | 157 | def handle_queue(self, url, kwdict): |
195 | | - self.dispatched = True |
| 158 | + # unresolved queue entry |
| 159 | + self.results.append((url, kwdict)) |
| 160 | + |
| 161 | + def handle_queue_resolve(self, url, kwdict): |
| 162 | + if url in self.visited: |
| 163 | + return |
| 164 | + self.visited.add(url) |
| 165 | + |
| 166 | + cls = kwdict.get("_extractor") |
| 167 | + if cls: |
| 168 | + extr = cls.from_url(url) |
| 169 | + else: |
| 170 | + extr = self.extractor.find(url) |
| 171 | + |
| 172 | + if not extr: |
| 173 | + self.results.append((url, kwdict)) |
| 174 | + return |
| 175 | + |
| 176 | + job = self.__class__(extr, self, self.resolve - 1) |
| 177 | + job.results = self.results # shared accumulator |
| 178 | + job.visited = self.visited # shared visited set |
| 179 | + job.run() |
196 | 180 |
|
197 | 181 |
|
198 | 182 | def get_playlist_metadata(args, playlist_path): |
199 | 183 | gallery_dl = load_module_level_gallery_dl(args) |
200 | 184 |
|
201 | 185 | added_media_count = 0 |
202 | | - job = GeneratorJob(playlist_path) |
203 | | - gen = job.run() |
| 186 | + job = UrlJob(playlist_path) |
| 187 | + job.run() |
| 188 | + |
| 189 | + is_playlist = len(job.results) > 1 |
204 | 190 |
|
205 | | - first_two = list(itertools.islice(gen, 2)) |
206 | | - is_playlist = len(first_two) > 1 |
207 | | - for webpath, info in itertools.chain(first_two, gen): |
| 191 | + for webpath, info in job.results: |
208 | 192 | errors = parse_gdl_job_status(job.status, playlist_path) |
209 | 193 | extractor_key = "gdl_" + job.extractor.category |
210 | 194 |
|
|
0 commit comments