-
-
Notifications
You must be signed in to change notification settings - Fork 29
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Critical Errors Viewer #738
Comments
Instead, we should store this list in mongodb, since it'll be of limited size. This way, if the crawl has failed / was canceled, can still see the list of errors. Thinking this error list should be either 50 or 100 lines at most. |
I wasn't sure if this was necessary or if we could just be more careful about what is logged as an error in the crawler and highlight those in the UI (or add a new critical level to the crawler logs, as mentioned) but this is an important point. Until we have live-streaming logs, crawls that fail to produce a WACZ will be pretty opaque. It seems reasonable to me that the initial Logs tab implementation could include both these critical errors and then the full logs from the WACZ if available, and I agree that up to a few dozen JSON lines per crawl seems reasonable to store in redis (short-term) and mongodb (long-term). I'm curious if this will impact how we want to think about live-streaming logs from running crawls, or if that solution may eliminate the need for some of what's scoped in this issue. |
Would it be more helpful to perhaps store a list of the most common errors a workflow is encountering instead of a list of the most recent errors? If a workflow has 50 of error X and 45 of error Y it is presumably a lot more useful to know both of these things instead of 50 error X messages. As for the rest of what is proposed here, I don't see this feature as very useful existing alongside a proper full featured log viewer. Giving users helpful information at a glance and directing them towards the log viewer if they require more details would be great and storing / caching the most common types of errors to display in places like the workflows page in a tooltip or the workflow details overview page would be really great. A less feature-rich log viewer is a reasonable first iteration if need be, but obviously we don't want to expend more time than we need for things that will be reworked later as more features come online. |
From latest discussion:
|
Separate from the full logging system (#631 and #330), we want to quickly be able to show carefully curated critical errors that user should be made aware of while a crawl is running and after. This is a subset of the full, filterable list of logs that will come from every crawler.
The idea would be to:
Store the list in a separateerrors.log
file, which can then be streamed from the WACZ.logger.error
and filter what that's used for now, or add a new categorylogger.critical
that would count the error as added to the critical errors listThe text was updated successfully, but these errors were encountered: