|
24 | 24 | from .pagination import DEFAULT_PAGE_SIZE, paginated_format
|
25 | 25 | from .storages import get_presigned_url, delete_crawl_file_object, get_wacz_logs
|
26 | 26 | from .users import User
|
27 |
| -from .utils import dt_now, ts_now, get_redis_crawl_stats |
| 27 | +from .utils import dt_now, ts_now, get_redis_crawl_stats, parse_jsonl_error_messages |
28 | 28 |
|
29 | 29 |
|
30 | 30 | CRAWL_STATES = (
|
@@ -784,12 +784,14 @@ async def get_errors_from_redis(
|
784 | 784 | try:
|
785 | 785 | redis = await self.get_redis(crawl_id)
|
786 | 786 | errors = await redis.lrange(f"{crawl_id}:e", skip, page_size)
|
787 |
| - total = len(errors) |
788 | 787 | except exceptions.ConnectionError:
|
789 | 788 | # pylint: disable=raise-missing-from
|
790 | 789 | raise HTTPException(status_code=503, detail="redis_connection_error")
|
791 | 790 |
|
792 |
| - return errors, total |
| 791 | + parsed_errors = parse_jsonl_error_messages(errors) |
| 792 | + total = len(parsed_errors) |
| 793 | + |
| 794 | + return parsed_errors, total |
793 | 795 |
|
794 | 796 | async def get_redis(self, crawl_id):
|
795 | 797 | """get redis url for crawl id"""
|
@@ -1240,8 +1242,9 @@ async def get_crawl_errors(
|
1240 | 1242 | skip = (page - 1) * pageSize
|
1241 | 1243 | upper_bound = skip + pageSize - 1
|
1242 | 1244 | errors = crawl.errors[skip:upper_bound]
|
1243 |
| - total = len(errors) |
1244 |
| - return paginated_format(errors, total, page, pageSize) |
| 1245 | + parsed_errors = parse_jsonl_error_messages(errors) |
| 1246 | + total = len(parsed_errors) |
| 1247 | + return paginated_format(parsed_errors, total, page, pageSize) |
1245 | 1248 |
|
1246 | 1249 | errors, total = await ops.get_errors_from_redis(crawl_id, pageSize, page)
|
1247 | 1250 | return paginated_format(errors, total, page, pageSize)
|
|
0 commit comments