Skip to content

Commit 7983d88

Browse files
feat(requirements): update to latest mathy_core/mathy_envs versions (#57)
1 parent d8a4cc0 commit 7983d88

File tree

3 files changed

+14
-28
lines changed

3 files changed

+14
-28
lines changed

libraries/mathy_python/requirements.txt

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5,8 +5,8 @@ srsly
55
pydantic>=1.0.0
66
typer<0.4.0,>=0.3.2
77
wasabi
8-
mathy_core>=0.8.0,<0.9.0
9-
mathy_envs[gym]>=0.9.2,<0.10.0
8+
mathy_core>=0.8.4,<0.9.0
9+
mathy_envs[gym]>=0.10.0,<0.11.0
1010
fragile==0.0.47
1111
tqdm>=4.43.0
1212
# new python feature backports

libraries/website/docs/snippets/cas/tokenizer_manual.ipynb

Lines changed: 6 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -12,21 +12,14 @@
1212
"!pip install mathy --upgrade\n",
1313
"from typing import List\n",
1414
"\n",
15-
"from mathy_core import (\n",
16-
" Token,\n",
17-
" TokenConstant,\n",
18-
" TokenEOF,\n",
19-
" Tokenizer,\n",
20-
" TokenPlus,\n",
21-
" TokenVariable,\n",
22-
")\n",
15+
"from mathy_core import Token, TOKEN_TYPES, Tokenizer\n",
2316
"\n",
2417
"manual_tokens: List[Token] = [\n",
25-
" Token(\"4\", TokenConstant),\n",
26-
" Token(\"x\", TokenVariable),\n",
27-
" Token(\"+\", TokenPlus),\n",
28-
" Token(\"2\", TokenConstant),\n",
29-
" Token(\"\", TokenEOF),\n",
18+
" Token(\"4\", TOKEN_TYPES.Constant),\n",
19+
" Token(\"x\", TOKEN_TYPES.Variable),\n",
20+
" Token(\"+\", TOKEN_TYPES.Plus),\n",
21+
" Token(\"2\", TOKEN_TYPES.Constant),\n",
22+
" Token(\"\", TOKEN_TYPES.EOF),\n",
3023
"]\n",
3124
"auto_tokens: List[Token] = Tokenizer().tokenize(\"4x + 2\")\n",
3225
"\n",

libraries/website/docs/snippets/cas/tokenizer_manual.py

Lines changed: 6 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -1,20 +1,13 @@
11
from typing import List
22

3-
from mathy_core import (
4-
Token,
5-
TokenConstant,
6-
TokenEOF,
7-
Tokenizer,
8-
TokenPlus,
9-
TokenVariable,
10-
)
3+
from mathy_core import Token, TOKEN_TYPES, Tokenizer
114

125
manual_tokens: List[Token] = [
13-
Token("4", TokenConstant),
14-
Token("x", TokenVariable),
15-
Token("+", TokenPlus),
16-
Token("2", TokenConstant),
17-
Token("", TokenEOF),
6+
Token("4", TOKEN_TYPES.Constant),
7+
Token("x", TOKEN_TYPES.Variable),
8+
Token("+", TOKEN_TYPES.Plus),
9+
Token("2", TOKEN_TYPES.Constant),
10+
Token("", TOKEN_TYPES.EOF),
1811
]
1912
auto_tokens: List[Token] = Tokenizer().tokenize("4x + 2")
2013

0 commit comments

Comments
 (0)