Skip to content

Commit 5365a26

Browse files
feat(mathy_alpha_sm): add pretrained model with simplified architecture
- remove time-dimension LSTM initial state handling. I had to tile the LSTM state to get this to work and it felt really hacky. It also expanded the size of the data by quite a bit with duplication. Removing this and going with a single LSTM works just as well or better. - adds a dense layer to the value head to fix layernorm issue with 1 unit output.
1 parent c33ea9c commit 5365a26

File tree

3 files changed

+47
-42
lines changed

3 files changed

+47
-42
lines changed

libraries/mathy_alpha_sm/mathy_alpha_sm/model.config.json

Lines changed: 47 additions & 42 deletions
Original file line numberDiff line numberDiff line change
@@ -6,47 +6,52 @@
66
"author": "Justin DuJardin",
77
"email": "justin@dujardinconsulting.com",
88
"url": "https://mathy.ai",
9-
"mathy_version": ">=0.5.1,<1.0.0",
10-
"units": 512,
11-
"embedding_units": 512,
12-
"lstm_units": 128,
13-
"topics": [
9+
"mathy_version":">=0.6.0,<1.0.0",
10+
"prediction_window_size":7,
11+
"use_lstm":true,
12+
"units":512,
13+
"embedding_units":512,
14+
"lstm_units":128,
15+
"topics":[
1416
"poly",
15-
"binomial",
16-
"complex"
17+
"complex",
18+
"binomial"
1719
],
18-
"difficulty": "easy",
19-
"init_model_from": null,
20-
"train": false,
21-
"verbose": false,
22-
"lr": 0.0003,
23-
"max_eps": 15000,
24-
"summary_interval": 100,
25-
"gamma": 0.99,
26-
"num_workers": 4,
27-
"td_lambda": 0.2,
28-
"print_training": false,
29-
"print_mode": "terminal",
30-
"use_term_order": false,
31-
"use_tree_complexity": false,
32-
"use_grouping_control": true,
33-
"clip_grouping_control": false,
34-
"use_env_features": false,
35-
"use_node_values": true,
36-
"update_gradients_every": 64,
37-
"num_thinking_steps_begin": 3,
38-
"mcts_sims": 10,
39-
"worker_wait": 0.1,
40-
"profile": false,
41-
"action_strategy": "a3c",
42-
"main_worker_use_epsilon": false,
43-
"e_greedy_min": 0.01,
44-
"e_greedy_max": 0.1,
45-
"entropy_loss_scaling": 0.05,
46-
"normalize_entropy_loss": true,
47-
"aux_tasks_weight_scale": 0.1,
48-
"teacher_start_evaluations_at_episode": 50,
49-
"teacher_evaluation_steps": 20,
50-
"teacher_promote_wins": 0.95,
51-
"teacher_demote_wins": 0.6
52-
}
20+
"difficulty":null,
21+
"model_name":"model",
22+
"init_model_from":null,
23+
"train":false,
24+
"verbose":true,
25+
"lr":0.001,
26+
"max_eps":5000,
27+
"summary_interval":100,
28+
"gamma":0.99,
29+
"num_workers":4,
30+
"td_lambda":0.2,
31+
"print_training":false,
32+
"print_model_call_times":false,
33+
"print_mode":"terminal",
34+
"normalization_style":"layer",
35+
"use_term_order":false,
36+
"use_tree_complexity":false,
37+
"use_grouping_control":false,
38+
"clip_grouping_control":true,
39+
"use_env_features":true,
40+
"use_node_values":true,
41+
"update_gradients_every":6,
42+
"num_thinking_steps_begin":3,
43+
"mcts_sims":10,
44+
"worker_wait":0.01,
45+
"profile":false,
46+
"action_strategy":"a3c",
47+
"main_worker_use_epsilon":false,
48+
"e_greedy_min":0.01,
49+
"e_greedy_max":0.3,
50+
"entropy_loss_scaling":0.05,
51+
"normalize_entropy_loss":true,
52+
"aux_tasks_weight_scale":0.1,
53+
"teacher_start_evaluations_at_episode":50,
54+
"teacher_evaluation_steps":5,
55+
"teacher_promote_wins":0.95,
56+
"teacher_demote_wins":0.6
57+
}
2.52 MB
Binary file not shown.
5.03 MB
Binary file not shown.

0 commit comments

Comments
 (0)