Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Maximum number of warnings? #2308

Open
dginev opened this issue Jan 31, 2024 · 1 comment
Open

Maximum number of warnings? #2308

dginev opened this issue Jan 31, 2024 · 1 comment

Comments

@dginev
Copy link
Collaborator

dginev commented Jan 31, 2024

The current record-holder for largest asset in ar5iv is arXiv:1501.02683, which has a ZIP of 647 MB in size.

To add insult to injury, the bundle contains a single cortex.log file, containing the latexml conversion logging information - and totaling a whopping 7.7 million lines.

As it turns out, those lines are almost entirely the same warning message, emitted in a busy loop until the timeout is reached (45 minutes). It reads:

Warning:expected:<number> Missing number, treated as zero
        at 1501.02683.tex; line 2441 col 0 - line 2441 col 12
        while processing \count@
        Next token is T_CS[\relax] ( == Core::Definition::Primitive[\relax]) more: \ifnum\count@<\value{proofcount}
\advance\count@\@ne\relax\expandafter\iterate\fi\let\iterate\relax
        In Core::Gullet[@0x559bfe9cba78] /dev/shm/hN2Sba2fTg/1501.02683.tex; 
 from line 2441 col 0 to line 2441 col 12

It is very clearly a pathological case of expansion gone wrong. But it suggests an idea for a new guard, similar to the "100 Error" guard that terminates execution. Maybe 10,000 Warnings?

Curious if @brucemiller finds this a good idea, and how many warnings could be considered a "healthy maximum".

@dginev dginev added this to the LaTeXML-0.8.9 milestone Jan 31, 2024
@dginev
Copy link
Collaborator Author

dginev commented Jan 31, 2024

Aside: The reason I stumbled on this issue is that ar5iv had some uptime challenges due to the web service running out of RAM while the site was being fully crawled. The example paper likely contributed to that.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant