diff --git a/examples/TimeSeriesLSTM.ipynb b/examples/TimeSeriesLSTM.ipynb index b8f86925..6fdb3811 100644 --- a/examples/TimeSeriesLSTM.ipynb +++ b/examples/TimeSeriesLSTM.ipynb @@ -20,16 +20,16 @@ "name": "stdout", "output_type": "stream", "text": [ - "--2023-02-11 16:24:56-- https://raw.githubusercontent.com/jbrownlee/Datasets/master/airline-passengers.csv\n", - "Resolving raw.githubusercontent.com (raw.githubusercontent.com)... 2606:50c0:8001::154, 2606:50c0:8002::154, 2606:50c0:8003::154, ...\n", - "Connecting to raw.githubusercontent.com (raw.githubusercontent.com)|2606:50c0:8001::154|:443... connected.\n", + "--2023-02-11 16:28:25-- https://raw.githubusercontent.com/jbrownlee/Datasets/master/airline-passengers.csv\n", + "Resolving raw.githubusercontent.com (raw.githubusercontent.com)... 2606:50c0:8000::154, 2606:50c0:8003::154, 2606:50c0:8002::154, ...\n", + "Connecting to raw.githubusercontent.com (raw.githubusercontent.com)|2606:50c0:8000::154|:443... connected.\n", "HTTP request sent, awaiting response... 200 OK\n", "Length: 2180 (2,1K) [text/plain]\n", - "Saving to: ‘airline-passengers.csv.3’\n", + "Saving to: ‘airline-passengers.csv.4’\n", "\n", "airline-passengers. 100%[===================>] 2,13K --.-KB/s in 0s \n", "\n", - "2023-02-11 16:24:56 (28,2 MB/s) - ‘airline-passengers.csv.3’ saved [2180/2180]\n", + "2023-02-11 16:28:25 (31,9 MB/s) - ‘airline-passengers.csv.4’ saved [2180/2180]\n", "\n" ] } @@ -178,7 +178,7 @@ { "data": { "application/vnd.jupyter.widget-view+json": { - "model_id": "b0c41dfea5bd4cbe91ebf7efcbcf0674", + "model_id": "591aa4bfce744f54b6862a7b843fb1c0", "version_major": 2, "version_minor": 0 }, @@ -204,3808 +204,6 @@ { "cell_type": "code", "execution_count": 10, - "id": "ddd413fe", - "metadata": {}, - "outputs": [ - { - "data": { - "image/svg+xml": [ - "\n", - "\n", - "G\n", - "\n", - "\n", - "\n", - "0\n", - "\n", - "FactNeuron:0:0.04826254826254825 x(4)\n", - "val: 0.05\n", - "grad: 0\n", - "dim: []\n", - "fcn: \n", - "\n", - "\n", - "\n", - "1\n", - "\n", - "FactNeuron:1:0.05405405405405403 x(3)\n", - "val: 0.05\n", - "grad: 0\n", - "dim: []\n", - "fcn: \n", - "\n", - "\n", - "\n", - "2\n", - "\n", - "FactNeuron:2:0.02702702702702703 x(2)\n", - "val: 0.03\n", - "grad: 0\n", - "dim: []\n", - "fcn: \n", - "\n", - "\n", - "\n", - "3\n", - "\n", - "FactNeuron:3:0.015444015444015441 x(1)\n", - "val: 0.02\n", - "grad: 0\n", - "dim: []\n", - "fcn: \n", - "\n", - "\n", - "\n", - "4\n", - "\n", - "FactNeuron:4:[0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0] lstm1_h0\n", - "val: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: \n", - "\n", - "\n", - "\n", - "8\n", - "\n", - "WeightedRuleNeuron:8:lstm1_out__o(T) :- {20, 1} x(T), {20, 20} lstm1_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "8->3\n", - "\n", - "\n", - "4:w4:[20, 1]:[0.61,0.97,-0.44,-0.47,-0.14,-0.94,0.63,-0.86,-0.09,-0.8,0.36,0.42,-0.65,0.19,0.91,0.76,-0.61,-0.6,-0.19,0.43]\n", - "\n", - "\n", - "\n", - "8->4\n", - "\n", - "\n", - "5:w5:[20, 20]:[\n", - "[0.72,0.53,0.15,-0.5,0.14,-0.67,-0.08,-0.83,-0.83,0.2,0.92,0.41,0.24,-0.61,0.99,-0.06,-0.81,-0.67,0.66,0.42],\n", - "[-0.7,-0.76,-0.46,-0.8,0.55,0.47,0.68,0.55,-0.86,-0.61,-0.04,0.39,0.06,0.38,0.09,0.78,0.02,-0.28,0.42,-0.06],\n", - "[0.74,-0.76,-0.53,-0.18,0.39,0.97,0.1,-0.1,0.78,-0.23,-0.44,0.32,0.68,0.21,-0.65,0.05,0.82,-0.84,0.88,0.7],\n", - "[-0.13,-0.75,0.84,0.86,0.82,-0.42,0.52,-0.75,0.55,0.04,0.18,-0.79,0.15,0.9,0.97,0.55,0.21,-0.91,-0.74,-0.77],\n", - "[-0.74,0.57,0.26,0.94,-0.7,-0.2,-0.3,-0.82,-0.65,-0.77,-0.71,0.16,0.94,-0.09,0.62,-0.33,-0.49,-0.09,0.61,0.65],\n", - "[-0.14,-0.15,-0.85,-0.76,0.9,0.57,-0.45,0.93,0.34,0.23,0.95,-0.49,0.1,0.73,0.91,-0.66,-0.28,-0.82,0.1,0.94],\n", - "[0.49,0.94,-0.85,-0.41,0.37,0.94,0.62,0.49,0.67,0.78,0.29,-0.45,-0.42,-0.71,-0.02,-0.9,-0.15,-0.63,0.42,-0.53],\n", - "[-0.67,0.19,-0.06,-0.64,0.06,0.84,0.12,0.33,0.35,0.36,-0.02,0.25,0.33,-0.21,0.59,-0.65,-0.53,-0.97,-0.27,0.32],\n", - "[-0.46,0.27,-0.21,0.33,-0.38,0.65,0.57,-0.73,0.07,0.3,-0.45,0.96,0.51,0.77,-0.05,0.2,0.16,0.35,0.72,-0.2],\n", - "[0.41,-0.02,-0.53,-0.54,-0.21,0.94,0.41,-0.11,-0.04,0.94,-0.7,-0.46,-0.74,0.67,-0.41,0.81,-0.59,0.52,-0.89,0.8],\n", - "[0.21,0.89,-0.28,-0.78,-0.59,-0.95,0.83,0.46,-0.89,-0.61,-0.06,0.49,-0.5,-0.3,0.68,-0.29,0.16,0.1,0.53,0.67],\n", - "[-0.37,0.98,0.6,0.36,0.62,0.49,0.9,0.3,-0.28,0.35,-0.39,0.92,0.26,0.39,-0.84,0.28,-0.38,-0.93,0.26,0.4],\n", - "[-0.71,0.99,0.24,0.4,-0.4,-0.97,-0.43,0.83,-0.76,0.95,0.77,-0.31,0.49,0.96,0.73,-0.15,0.2,-0.97,0.68,0.13],\n", - "[0.87,-0.44,-0,-0.86,-0.06,-0.62,0.65,0.02,-0.77,0.04,-0.14,-0.73,-0.8,-0.05,0.45,-0.51,-0.43,0.69,0.15,-0.28],\n", - "[-0.13,-0.97,0.27,-0.69,-0.61,0.7,-0.95,-0.56,-0.85,-0.64,-0.9,0.75,-0.48,-0.89,0.62,-0.07,-0.51,-0.04,-0.86,0.45],\n", - "[-0.35,-0.88,-0.06,-0.84,0.41,0.05,-0.44,-0.92,0.19,0.44,-0.63,-0.82,-0.97,0.16,-0.67,0.81,-0.8,0.97,0.08,0.9],\n", - "[0.34,-0.09,-0.48,-0.71,-0.05,-0.36,-0.68,-0.18,-0.98,0.4,-0.13,0.31,-0.92,0.58,0.61,-0.82,-0.99,-0.16,0.95,-0.38],\n", - "[-0.32,-0.58,0.36,-0.64,-0.58,-0.64,-0.23,-0.78,0.39,0.97,0.95,0.88,0.06,0.48,-0.99,-0.39,-0.18,0.02,0.96,-0.99],\n", - "[0.35,0.32,0.94,0.32,0.35,-0.47,-0.51,0.25,-0.41,0.99,-0.89,-0.97,-0.3,-0.98,0.23,-0.44,-0.47,0.23,-0.23,0.83],\n", - "[-0.75,0.32,-0.68,-0.13,0.28,0.75,-0.69,0.88,0.42,0.49,-0.49,-0.46,-0.7,0.3,0.49,0.82,-0.68,0.18,-0.11,0.33]\n", - "]\n", - "\n", - "\n", - "\n", - "21\n", - "\n", - "WeightedRuleNeuron:21:lstm1_out__i(T) :- {20, 1} x(T), {20, 20} lstm1_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "21->3\n", - "\n", - "\n", - "0:w0:[20, 1]:[0.75,0.98,0.3,0.15,0.74,0.4,-0.45,0.34,0.4,-0.25,-0.22,-0.14,0.35,-0.51,-0.06,0.93,0.37,-0.03,-0.62,0.86]\n", - "\n", - "\n", - "\n", - "21->4\n", - "\n", - "\n", - "1:w1:[20, 20]:[\n", - "[-0.28,0.14,-0.67,-0.56,0.68,0.25,-0.87,0.41,-0.87,-0.13,-0.83,0.91,-0.52,0.74,-0.95,-0.12,0.16,0.99,-0.73,-0.75],\n", - "[0.46,-0.44,0.42,0.58,-0.59,0.89,0.97,-0.77,-0.81,0.21,-0.76,0.75,0.65,-0.61,-0.45,-0.17,-0.73,-0.73,0.21,-0.91],\n", - "[0.34,0.68,0.31,0.37,-0.27,0.02,-0.48,0.62,0.4,-0.19,0.94,-0.25,-0.4,0.49,-0.19,-0.57,0.26,0.31,-0.17,-0.63],\n", - "[0.26,0.33,0.6,-0.97,-0.92,-0.17,-0.23,0.78,0.24,0.17,0.46,-0.27,-0.84,-0.55,-0.34,-0.19,0.15,0.87,-0.61,-0.96],\n", - "[0.31,0.56,0.81,-0.72,-0.4,-0.89,0.47,-0.66,-0.61,-0.12,-0.57,0.82,0.5,0.12,0.48,-0.31,-0.28,-0.4,0.79,-0.86],\n", - "[-0.15,0.61,-0.65,-0.64,-0.56,0.09,-0.34,0.16,-0.81,-0.2,-0.79,0.7,0.72,-0.61,-0.28,-0.74,-1,0.87,0.01,-0.07],\n", - "[-0.18,-0.32,-0.74,0.35,0.19,-0.62,-0.77,0.61,0.84,0.04,-1,0.61,-0.19,0.1,-0.08,-0.69,0.2,0.79,0.41,0.93],\n", - "[0.34,-0.05,-0.51,-0.86,-0.25,-0.43,-0.78,-0.15,0.58,-0.74,0.3,-0.16,0.52,-0.51,-0.38,-0.23,0.47,0.2,-0.15,0.93],\n", - "[-0.12,0.62,0.77,0.59,0.9,-0.44,0.29,-0.31,-0.65,0.03,0.11,0.36,-0.09,-0.37,0.16,-0.45,0.67,0.62,0.19,-0.91],\n", - "[-0.82,-0.41,-0.83,-0.57,0.36,-0.88,0.38,0.32,-0.63,-0.1,-0.22,0.89,-0.5,0.35,-0.73,-0.33,-0.88,-0.78,-0.48,0.42],\n", - "[-0.12,-0.88,0.09,-0.13,-0.74,-0.55,0.71,0.83,0.2,0,-0.06,0.59,0.1,-0.85,0.52,0.04,0.62,-0.83,0.19,-0.57],\n", - "[0.31,0.27,-0.83,0.24,0.2,-0.33,-0.13,-0.45,0.33,-0.85,0.07,0.72,0.56,-0.73,0.28,-0.77,-0.36,-0.13,0.69,0.3],\n", - "[-0.64,-0.74,-0.15,-0.25,0.04,-0,-0.92,-0.23,-0.12,0.32,0.25,0.07,-0.84,-0.52,-0.28,-0.87,-0.65,0.88,-0.04,-0.49],\n", - "[-0.6,0.16,0.69,-0.89,-0.84,-0.92,-0.81,0.06,-0.59,-0.63,0.76,0.77,0.78,0.66,-0.57,0.45,0.44,0.11,-0.49,-0.3],\n", - "[-0.47,-0.9,0.52,0.53,-0.29,0.43,-0.04,-0.55,0.81,-0.97,-0.65,-0.71,-0.18,-0.66,-0.12,-0.09,0.15,0.58,0.41,-0.09],\n", - "[-0.21,0.95,0.26,0.19,0.56,0.4,-1,0.18,-0.58,-0.85,0.12,0.64,-0.33,0.18,0.45,0.77,-0.3,-0.45,-0.96,0.72],\n", - "[-1,-0.65,-0.89,0.59,-0.09,0,-0.21,0.26,0.9,-0.23,-0.43,0.84,-0.82,-0.67,-0.75,-0.25,0.58,0.21,0.74,-0.19],\n", - "[0.13,0.69,-0.99,0.41,-0.39,-0.84,0.24,-0.8,-0.64,0.2,-0.93,0.49,-0.54,-0.88,0.95,0.98,-0.6,0.13,0.68,-0.41],\n", - "[0.41,-0.31,-0.12,0.46,-0.12,0.44,-0.95,0.55,-0.66,-0.46,0.31,-0.27,-0.34,0.89,0.79,-0.43,0.82,-0.14,0.57,0.04],\n", - "[0.39,0.45,-0.17,-0.83,-0.04,0.14,-0.95,0.34,0.58,-0.62,-0.23,-0.35,0.46,-0.13,0.04,-0.48,-0.6,-0.39,-0.71,0.7]\n", - "]\n", - "\n", - "\n", - "\n", - "24\n", - "\n", - "WeightedRuleNeuron:24:lstm1_out__n(T) :- {20, 1} x(T), {20, 20} lstm1_out(Z), @next(Z, T). [transformation=tanh]\n", - "val: [-0.02,-0.01,-0.01,-0.01,-0.01,-0,-0.01,0.01,-0,-0,-0.01,0,-0.01,0.01,0,0.01,0,0.01,-0.01,0.01]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Tanh\n", - "\n", - "\n", - "\n", - "24->3\n", - "\n", - "\n", - "6:w6:[20, 1]:[-0.36,0.99,0.89,-0.51,-0.12,0.21,0.43,0.69,0.25,0.9,-0.71,-0.87,0.46,-0.52,-0.22,-0.25,-0.69,-0.65,0.31,0.63]\n", - "\n", - "\n", - "\n", - "24->4\n", - "\n", - "\n", - "7:w7:[20, 20]:[\n", - "[-0.54,0.2,0.56,-0.14,0.56,0.07,0.23,0.01,-0.18,0.53,-0.57,0.18,0.58,0.47,-0.76,0.77,0.83,-0.63,-0.53,0.07],\n", - "[-0.1,0.37,-0.13,-0.81,0.3,0.77,0.83,-0.26,-0.99,-0.62,0.06,0.32,0.34,-0.62,0.38,0.02,-0.99,-0.82,-0.36,0.61],\n", - "[0.95,-0.42,-0.69,-0.37,-0.84,1,-0.57,0.04,-0.34,0.69,0.27,0.85,-0.13,-0.5,-0.98,0.67,0.23,-0.18,0.24,0.94],\n", - "[0.5,0.43,-0.75,-0.58,-0.11,-0.14,-0.51,-0.75,-0.46,0.13,0.48,-0.94,-0.91,-0.93,-0.14,0.3,-0.57,-0.67,-0.2,0.19],\n", - "[0.01,0.94,-0.07,-0.13,0.64,0.36,-0.91,-0.8,-0.93,-0.45,1,-0.68,0.38,-0.87,0.18,-0.07,0.68,0.2,0.65,0.09],\n", - "[-0.61,0.04,0,-0.81,0.17,-0.29,0.84,0.87,-0.81,0.23,0.99,0.65,0.53,0.89,0.05,0.06,0.4,0.16,0.6,0.37],\n", - "[-0.35,0.62,-0.26,-0.8,-0.58,0.17,0.26,-0.48,-0.27,-0.69,-0.51,-0.94,0.98,0.82,0.49,0.44,0.67,-0.12,0.01,0.83],\n", - "[0.97,-0.29,-0.33,0.67,-0.04,-0.92,-0.48,-0.73,0.13,-0.33,0.15,0.14,-0.09,0.43,-0.32,-0.7,-0.25,-0.14,-0,-0.98],\n", - "[-0.17,0.51,-0.5,-0.26,-0.79,-0.2,-0.59,0.52,0.94,0.86,0.37,0.74,-0.53,0.85,0.81,-0.81,0.59,0.25,-0.18,-0.28],\n", - "[-0.85,0.08,0.71,-0.22,0.59,0.57,0.02,-0.79,0.6,0.11,-0.52,0.51,-0.71,0.12,-0.41,0.4,-0.64,0.92,0,0.63],\n", - "[-0.88,0.53,-0.2,0.95,-0.52,0.08,0.37,-0.92,0.29,0.11,0.82,-0.2,-0.77,0.48,0.55,-0.69,-0.66,0.79,-0.12,0.37],\n", - "[0.48,-0.96,-0.65,-0.1,0.19,0.59,0.15,0.15,-0.67,0.36,0.39,-0.89,-0.46,-0.36,-0.73,0.9,0.39,-0.09,0.55,0.05],\n", - "[-0.93,0.73,-0.93,0.49,-0.84,-0.3,0.59,-0.53,-0.62,-0.5,0.7,-0.36,-0.11,-0.07,0.33,-0.02,-0.82,0.08,-0.39,-0.42],\n", - "[0.56,0.72,-0.78,-0.33,-0.41,0.48,-0.55,0.85,-0.86,0.3,-0.44,-0.13,0.02,0.38,0.61,0.23,-0.18,0.76,0.2,0.1],\n", - "[0.63,0.68,-0.79,-0.77,0.54,0.82,-0.85,-0.01,0.73,0.54,-0.56,-0.16,-0.71,-0.32,-0.27,0.06,-0.86,0.12,-0.04,0.09],\n", - "[-0.39,0.38,-0.2,0.25,0.92,-0.15,0.26,0.3,-0.23,-0.23,-0.35,-0.85,-0.63,-0.57,0.82,-0.76,-0.98,0.5,-0.92,-0.32],\n", - "[0.88,0.34,-0.24,0.28,0.42,-0.88,0.4,-0.75,0.41,0.15,0.98,-0.22,0.67,0.74,0.66,-0.88,0.16,-0.66,-0.57,0.69],\n", - "[-0.65,-0.67,0.9,0.05,0.36,-0.14,-0.46,0.66,-0.96,-0.93,0.91,-0.13,-0.46,-0.69,-0.66,0.62,-0.16,0.82,0.04,-0.86],\n", - "[0.36,0.86,-0.54,0.12,0.56,0.23,-0.31,-0.53,-0.65,-0.24,0.41,0.47,-0.79,0.44,-0.99,-0.3,-0.49,-0.36,0.43,0.87],\n", - "[0.14,0.96,0.41,-0.47,-0.76,0.42,0.2,-0.48,-0.73,0.62,-0.6,0.74,-0.38,-0,-0.21,0.15,-0.13,-0.76,0.38,0.64]\n", - "]\n", - "\n", - "\n", - "\n", - "27\n", - "\n", - "RuleNeuron:27:lstm1_out__right(T) :- lstm1_out__i(T), lstm1_out__n(T). [transformation=identity, combination=elproduct]\n", - "val: [-0.01,-0.01,-0,-0.01,-0,-0,-0.01,0.01,-0,-0,-0.01,0,-0,0,0,0,0,0.01,-0.01,0.01]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "27->21\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "27->24\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "34\n", - "\n", - "RuleNeuron:34:lstm1_out(T) :- lstm1_out__o(T), tanh(lstm1_out__c(T)). [transformation=identity, combination=elproduct]\n", - "val: [-0,-0,-0,-0,-0,-0,-0,0,-0,-0,-0,0,-0,0,0,0,0,0,-0,0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "34->8\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "34->27\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "37\n", - "\n", - "WeightedRuleNeuron:37:lstm1_out__o(T) :- {20, 1} x(T), {20, 20} lstm1_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.51,0.5,0.5,0.5,0.49,0.5,0.49,0.51,0.51,0.5,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "37->2\n", - "\n", - "\n", - "4:w4:[20, 1]:[0.61,0.97,-0.44,-0.47,-0.14,-0.94,0.63,-0.86,-0.09,-0.8,0.36,0.42,-0.65,0.19,0.91,0.76,-0.61,-0.6,-0.19,0.43]\n", - "\n", - "\n", - "\n", - "37->34\n", - "\n", - "\n", - "5:w5:[20, 20]:[\n", - "[0.72,0.53,0.15,-0.5,0.14,-0.67,-0.08,-0.83,-0.83,0.2,0.92,0.41,0.24,-0.61,0.99,-0.06,-0.81,-0.67,0.66,0.42],\n", - "[-0.7,-0.76,-0.46,-0.8,0.55,0.47,0.68,0.55,-0.86,-0.61,-0.04,0.39,0.06,0.38,0.09,0.78,0.02,-0.28,0.42,-0.06],\n", - "[0.74,-0.76,-0.53,-0.18,0.39,0.97,0.1,-0.1,0.78,-0.23,-0.44,0.32,0.68,0.21,-0.65,0.05,0.82,-0.84,0.88,0.7],\n", - "[-0.13,-0.75,0.84,0.86,0.82,-0.42,0.52,-0.75,0.55,0.04,0.18,-0.79,0.15,0.9,0.97,0.55,0.21,-0.91,-0.74,-0.77],\n", - "[-0.74,0.57,0.26,0.94,-0.7,-0.2,-0.3,-0.82,-0.65,-0.77,-0.71,0.16,0.94,-0.09,0.62,-0.33,-0.49,-0.09,0.61,0.65],\n", - "[-0.14,-0.15,-0.85,-0.76,0.9,0.57,-0.45,0.93,0.34,0.23,0.95,-0.49,0.1,0.73,0.91,-0.66,-0.28,-0.82,0.1,0.94],\n", - "[0.49,0.94,-0.85,-0.41,0.37,0.94,0.62,0.49,0.67,0.78,0.29,-0.45,-0.42,-0.71,-0.02,-0.9,-0.15,-0.63,0.42,-0.53],\n", - "[-0.67,0.19,-0.06,-0.64,0.06,0.84,0.12,0.33,0.35,0.36,-0.02,0.25,0.33,-0.21,0.59,-0.65,-0.53,-0.97,-0.27,0.32],\n", - "[-0.46,0.27,-0.21,0.33,-0.38,0.65,0.57,-0.73,0.07,0.3,-0.45,0.96,0.51,0.77,-0.05,0.2,0.16,0.35,0.72,-0.2],\n", - "[0.41,-0.02,-0.53,-0.54,-0.21,0.94,0.41,-0.11,-0.04,0.94,-0.7,-0.46,-0.74,0.67,-0.41,0.81,-0.59,0.52,-0.89,0.8],\n", - "[0.21,0.89,-0.28,-0.78,-0.59,-0.95,0.83,0.46,-0.89,-0.61,-0.06,0.49,-0.5,-0.3,0.68,-0.29,0.16,0.1,0.53,0.67],\n", - "[-0.37,0.98,0.6,0.36,0.62,0.49,0.9,0.3,-0.28,0.35,-0.39,0.92,0.26,0.39,-0.84,0.28,-0.38,-0.93,0.26,0.4],\n", - "[-0.71,0.99,0.24,0.4,-0.4,-0.97,-0.43,0.83,-0.76,0.95,0.77,-0.31,0.49,0.96,0.73,-0.15,0.2,-0.97,0.68,0.13],\n", - "[0.87,-0.44,-0,-0.86,-0.06,-0.62,0.65,0.02,-0.77,0.04,-0.14,-0.73,-0.8,-0.05,0.45,-0.51,-0.43,0.69,0.15,-0.28],\n", - "[-0.13,-0.97,0.27,-0.69,-0.61,0.7,-0.95,-0.56,-0.85,-0.64,-0.9,0.75,-0.48,-0.89,0.62,-0.07,-0.51,-0.04,-0.86,0.45],\n", - "[-0.35,-0.88,-0.06,-0.84,0.41,0.05,-0.44,-0.92,0.19,0.44,-0.63,-0.82,-0.97,0.16,-0.67,0.81,-0.8,0.97,0.08,0.9],\n", - "[0.34,-0.09,-0.48,-0.71,-0.05,-0.36,-0.68,-0.18,-0.98,0.4,-0.13,0.31,-0.92,0.58,0.61,-0.82,-0.99,-0.16,0.95,-0.38],\n", - "[-0.32,-0.58,0.36,-0.64,-0.58,-0.64,-0.23,-0.78,0.39,0.97,0.95,0.88,0.06,0.48,-0.99,-0.39,-0.18,0.02,0.96,-0.99],\n", - "[0.35,0.32,0.94,0.32,0.35,-0.47,-0.51,0.25,-0.41,0.99,-0.89,-0.97,-0.3,-0.98,0.23,-0.44,-0.47,0.23,-0.23,0.83],\n", - "[-0.75,0.32,-0.68,-0.13,0.28,0.75,-0.69,0.88,0.42,0.49,-0.49,-0.46,-0.7,0.3,0.49,0.82,-0.68,0.18,-0.11,0.33]\n", - "]\n", - "\n", - "\n", - "\n", - "40\n", - "\n", - "WeightedRuleNeuron:40:lstm1_out__f(T) :- {20, 1} x(T), {20, 20} lstm1_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.49,0.5,0.49,0.5,0.5,0.51,0.5,0.5,0.5,0.5,0.5,0.49,0.5,0.5,0.5,0.5,0.51,0.5,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "40->2\n", - "\n", - "\n", - "2:w2:[20, 1]:[-0.74,0.22,-0.98,0.54,0.43,-0.24,-0.29,-0.31,0.35,0.08,0.83,-0.93,-0.47,-0.5,0.19,-0.54,0.07,-0.44,0.44,-0.2]\n", - "\n", - "\n", - "\n", - "40->34\n", - "\n", - "\n", - "3:w3:[20, 20]:[\n", - "[-0.03,0.55,0.74,-0.99,-0.63,0.45,0.79,-0.05,0.19,0.11,-0.54,0.93,0.56,-0.38,0.35,0.26,0.1,0.88,-0.67,0.08],\n", - "[0.83,-0.1,-0.16,0.02,0.53,0.47,-0.19,0.19,-0.78,-0.65,-0.13,-0.81,0.47,-0.47,-0.39,0.04,0.19,0.64,-0.11,-0.57],\n", - "[-0.09,0.25,0.99,0.77,-0.45,0.28,-0.85,0.02,-0.06,-0.42,0.74,-0.95,0.9,0.44,-0.87,-0.69,0.67,-0.23,-0.52,0.72],\n", - "[-0.69,0.19,-0.84,0.97,-0.05,-0.21,-0.8,0.43,-0.62,-0.18,-0.45,0.49,0.25,-0.99,-0.27,-0.89,-0.53,0.11,-0.82,0.22],\n", - "[-0.28,-0.31,-0.81,-0.33,0.16,-0.8,0.65,-0.58,-0.19,0.95,-0.09,-0.22,-0.24,-0.13,-0.94,0.67,0.43,-0.65,-0.01,-0.8],\n", - "[0.06,-0.98,0.49,0.02,0.12,0.07,-0.91,-0.6,-0.15,-0.98,-0.75,-0.75,-0.87,0.79,-0.7,0.02,0.64,-0.24,0.98,0.69],\n", - "[-0.11,-0.98,-0.96,0.9,-0.86,0.21,-0.39,-0.17,-0.27,0.23,-0.37,-0.83,-0.33,0.97,0.38,-0.64,-0.53,0.49,-0.52,-0.8],\n", - "[0.28,0.07,-0.13,0.67,-0.35,-0.13,0.7,-0.06,0.76,0.15,0.13,-0.39,0.94,-0.68,-0.17,0.25,-0.81,0.44,-0.99,-0.55],\n", - "[-0.41,0.46,0.79,-0.01,0.19,0.67,0.85,0.86,-0.45,-0.25,0.47,0.38,0.8,-0.84,0.28,0.89,0.95,-0.43,0.2,0.26],\n", - "[-0.23,-0.43,0.71,-0.34,-0.4,0.87,0.56,-0.98,-0.58,-0.83,1,-0.09,0.32,0.51,-0.38,-0.35,0.88,0.35,0.94,0.98],\n", - "[-0.35,-0.66,-0.74,-0.31,0.69,-0.82,0.14,-0.62,-0.23,0.95,0.53,0.6,0.69,-0.32,-0.29,0.63,0.15,0.18,-0.39,0.31],\n", - "[0.22,0.72,0.65,-0.58,-0.03,0.29,0.66,0.44,-0.57,-0.94,-0.86,-0.23,0.42,0.89,-0.24,-0.09,-0.18,0.64,-0.43,-0.68],\n", - "[-0.68,0.64,-0.59,-0.43,-0.18,0.43,0.72,0.41,0.03,-0.36,0.82,0.51,0.4,0.08,0.82,0.97,-0.1,0.6,0.28,-0.51],\n", - "[-0.31,0.7,-0.16,-0.97,-0.42,0.75,0.14,-0.51,0.47,0.85,-0.17,0.86,-0.79,0.6,-0.67,-0.43,0.63,0.28,0.67,-0.69],\n", - "[-0.17,0.33,-0.8,-0.86,-0.7,0.31,0.93,-0.26,-1,-0.9,0.6,-0.24,-0.11,0.51,0.31,0.82,0.22,-0.91,0.87,0.65],\n", - "[-0.32,-0.4,0.4,0.72,-0.22,0.33,-0.36,0.01,0.27,-0.96,-0.92,0.91,-0.18,-0.61,-0.7,-0.72,-0.65,0.44,0.39,-0.93],\n", - "[0.32,0.82,-0.1,-0.45,-0.69,0.17,0.22,0.53,-0.43,-0.53,-0.75,0.03,0.05,-0.73,0.56,0.46,-0.98,-0.4,-0.74,-0.95],\n", - "[0.47,-0.88,-0.25,-0.7,0.99,-0.14,-0.03,0.42,-0.84,-0.12,0.43,0.88,0.15,-0.91,-0.47,0.48,-0.47,-0.74,-0.4,0.46],\n", - "[0.24,0.98,0.53,-0.2,-0.92,-0.49,-0.44,-0.9,-0.03,-0.34,-0.12,-0.1,-0.94,-0.7,-0.95,0.99,0.54,-0.99,-0.3,-0.18],\n", - "[-0.91,-0.67,0.45,0.64,-0.26,0.99,-0.61,0.01,0.09,-0.64,-0.66,-0.15,0.02,0.81,-0.64,-0.7,-0.99,0.43,-0.81,0.39]\n", - "]\n", - "\n", - "\n", - "\n", - "43\n", - "\n", - "RuleNeuron:43:lstm1_out__left(T) :- lstm1_out__f(T), lstm1_out__c(Z), @next(Z, T). [transformation=identity, combination=elproduct]\n", - "val: [-0,-0,-0,-0,-0,-0,-0,0,-0,-0,-0,0,-0,0,0,0,0,0,-0,0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "43->27\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "43->40\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "46\n", - "\n", - "WeightedRuleNeuron:46:lstm1_out__i(T) :- {20, 1} x(T), {20, 20} lstm1_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.5,0.49,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.49,0.5,0.49,0.5,0.5,0.51,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "46->2\n", - "\n", - "\n", - "0:w0:[20, 1]:[0.75,0.98,0.3,0.15,0.74,0.4,-0.45,0.34,0.4,-0.25,-0.22,-0.14,0.35,-0.51,-0.06,0.93,0.37,-0.03,-0.62,0.86]\n", - "\n", - "\n", - "\n", - "46->34\n", - "\n", - "\n", - "1:w1:[20, 20]:[\n", - "[-0.28,0.14,-0.67,-0.56,0.68,0.25,-0.87,0.41,-0.87,-0.13,-0.83,0.91,-0.52,0.74,-0.95,-0.12,0.16,0.99,-0.73,-0.75],\n", - "[0.46,-0.44,0.42,0.58,-0.59,0.89,0.97,-0.77,-0.81,0.21,-0.76,0.75,0.65,-0.61,-0.45,-0.17,-0.73,-0.73,0.21,-0.91],\n", - "[0.34,0.68,0.31,0.37,-0.27,0.02,-0.48,0.62,0.4,-0.19,0.94,-0.25,-0.4,0.49,-0.19,-0.57,0.26,0.31,-0.17,-0.63],\n", - "[0.26,0.33,0.6,-0.97,-0.92,-0.17,-0.23,0.78,0.24,0.17,0.46,-0.27,-0.84,-0.55,-0.34,-0.19,0.15,0.87,-0.61,-0.96],\n", - "[0.31,0.56,0.81,-0.72,-0.4,-0.89,0.47,-0.66,-0.61,-0.12,-0.57,0.82,0.5,0.12,0.48,-0.31,-0.28,-0.4,0.79,-0.86],\n", - "[-0.15,0.61,-0.65,-0.64,-0.56,0.09,-0.34,0.16,-0.81,-0.2,-0.79,0.7,0.72,-0.61,-0.28,-0.74,-1,0.87,0.01,-0.07],\n", - "[-0.18,-0.32,-0.74,0.35,0.19,-0.62,-0.77,0.61,0.84,0.04,-1,0.61,-0.19,0.1,-0.08,-0.69,0.2,0.79,0.41,0.93],\n", - "[0.34,-0.05,-0.51,-0.86,-0.25,-0.43,-0.78,-0.15,0.58,-0.74,0.3,-0.16,0.52,-0.51,-0.38,-0.23,0.47,0.2,-0.15,0.93],\n", - "[-0.12,0.62,0.77,0.59,0.9,-0.44,0.29,-0.31,-0.65,0.03,0.11,0.36,-0.09,-0.37,0.16,-0.45,0.67,0.62,0.19,-0.91],\n", - "[-0.82,-0.41,-0.83,-0.57,0.36,-0.88,0.38,0.32,-0.63,-0.1,-0.22,0.89,-0.5,0.35,-0.73,-0.33,-0.88,-0.78,-0.48,0.42],\n", - "[-0.12,-0.88,0.09,-0.13,-0.74,-0.55,0.71,0.83,0.2,0,-0.06,0.59,0.1,-0.85,0.52,0.04,0.62,-0.83,0.19,-0.57],\n", - "[0.31,0.27,-0.83,0.24,0.2,-0.33,-0.13,-0.45,0.33,-0.85,0.07,0.72,0.56,-0.73,0.28,-0.77,-0.36,-0.13,0.69,0.3],\n", - "[-0.64,-0.74,-0.15,-0.25,0.04,-0,-0.92,-0.23,-0.12,0.32,0.25,0.07,-0.84,-0.52,-0.28,-0.87,-0.65,0.88,-0.04,-0.49],\n", - "[-0.6,0.16,0.69,-0.89,-0.84,-0.92,-0.81,0.06,-0.59,-0.63,0.76,0.77,0.78,0.66,-0.57,0.45,0.44,0.11,-0.49,-0.3],\n", - "[-0.47,-0.9,0.52,0.53,-0.29,0.43,-0.04,-0.55,0.81,-0.97,-0.65,-0.71,-0.18,-0.66,-0.12,-0.09,0.15,0.58,0.41,-0.09],\n", - "[-0.21,0.95,0.26,0.19,0.56,0.4,-1,0.18,-0.58,-0.85,0.12,0.64,-0.33,0.18,0.45,0.77,-0.3,-0.45,-0.96,0.72],\n", - "[-1,-0.65,-0.89,0.59,-0.09,0,-0.21,0.26,0.9,-0.23,-0.43,0.84,-0.82,-0.67,-0.75,-0.25,0.58,0.21,0.74,-0.19],\n", - "[0.13,0.69,-0.99,0.41,-0.39,-0.84,0.24,-0.8,-0.64,0.2,-0.93,0.49,-0.54,-0.88,0.95,0.98,-0.6,0.13,0.68,-0.41],\n", - "[0.41,-0.31,-0.12,0.46,-0.12,0.44,-0.95,0.55,-0.66,-0.46,0.31,-0.27,-0.34,0.89,0.79,-0.43,0.82,-0.14,0.57,0.04],\n", - "[0.39,0.45,-0.17,-0.83,-0.04,0.14,-0.95,0.34,0.58,-0.62,-0.23,-0.35,0.46,-0.13,0.04,-0.48,-0.6,-0.39,-0.71,0.7]\n", - "]\n", - "\n", - "\n", - "\n", - "49\n", - "\n", - "WeightedRuleNeuron:49:lstm1_out__n(T) :- {20, 1} x(T), {20, 20} lstm1_out(Z), @next(Z, T). [transformation=tanh]\n", - "val: [-0.03,-0.03,-0.02,-0.02,-0,-0,-0.02,0.01,-0.01,-0.01,-0.02,0.01,-0.01,0.02,-0,0.02,-0,0.02,-0.03,0.03]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Tanh\n", - "\n", - "\n", - "\n", - "49->2\n", - "\n", - "\n", - "6:w6:[20, 1]:[-0.36,0.99,0.89,-0.51,-0.12,0.21,0.43,0.69,0.25,0.9,-0.71,-0.87,0.46,-0.52,-0.22,-0.25,-0.69,-0.65,0.31,0.63]\n", - "\n", - "\n", - "\n", - "49->34\n", - "\n", - "\n", - "7:w7:[20, 20]:[\n", - "[-0.54,0.2,0.56,-0.14,0.56,0.07,0.23,0.01,-0.18,0.53,-0.57,0.18,0.58,0.47,-0.76,0.77,0.83,-0.63,-0.53,0.07],\n", - "[-0.1,0.37,-0.13,-0.81,0.3,0.77,0.83,-0.26,-0.99,-0.62,0.06,0.32,0.34,-0.62,0.38,0.02,-0.99,-0.82,-0.36,0.61],\n", - "[0.95,-0.42,-0.69,-0.37,-0.84,1,-0.57,0.04,-0.34,0.69,0.27,0.85,-0.13,-0.5,-0.98,0.67,0.23,-0.18,0.24,0.94],\n", - "[0.5,0.43,-0.75,-0.58,-0.11,-0.14,-0.51,-0.75,-0.46,0.13,0.48,-0.94,-0.91,-0.93,-0.14,0.3,-0.57,-0.67,-0.2,0.19],\n", - "[0.01,0.94,-0.07,-0.13,0.64,0.36,-0.91,-0.8,-0.93,-0.45,1,-0.68,0.38,-0.87,0.18,-0.07,0.68,0.2,0.65,0.09],\n", - "[-0.61,0.04,0,-0.81,0.17,-0.29,0.84,0.87,-0.81,0.23,0.99,0.65,0.53,0.89,0.05,0.06,0.4,0.16,0.6,0.37],\n", - "[-0.35,0.62,-0.26,-0.8,-0.58,0.17,0.26,-0.48,-0.27,-0.69,-0.51,-0.94,0.98,0.82,0.49,0.44,0.67,-0.12,0.01,0.83],\n", - "[0.97,-0.29,-0.33,0.67,-0.04,-0.92,-0.48,-0.73,0.13,-0.33,0.15,0.14,-0.09,0.43,-0.32,-0.7,-0.25,-0.14,-0,-0.98],\n", - "[-0.17,0.51,-0.5,-0.26,-0.79,-0.2,-0.59,0.52,0.94,0.86,0.37,0.74,-0.53,0.85,0.81,-0.81,0.59,0.25,-0.18,-0.28],\n", - "[-0.85,0.08,0.71,-0.22,0.59,0.57,0.02,-0.79,0.6,0.11,-0.52,0.51,-0.71,0.12,-0.41,0.4,-0.64,0.92,0,0.63],\n", - "[-0.88,0.53,-0.2,0.95,-0.52,0.08,0.37,-0.92,0.29,0.11,0.82,-0.2,-0.77,0.48,0.55,-0.69,-0.66,0.79,-0.12,0.37],\n", - "[0.48,-0.96,-0.65,-0.1,0.19,0.59,0.15,0.15,-0.67,0.36,0.39,-0.89,-0.46,-0.36,-0.73,0.9,0.39,-0.09,0.55,0.05],\n", - "[-0.93,0.73,-0.93,0.49,-0.84,-0.3,0.59,-0.53,-0.62,-0.5,0.7,-0.36,-0.11,-0.07,0.33,-0.02,-0.82,0.08,-0.39,-0.42],\n", - "[0.56,0.72,-0.78,-0.33,-0.41,0.48,-0.55,0.85,-0.86,0.3,-0.44,-0.13,0.02,0.38,0.61,0.23,-0.18,0.76,0.2,0.1],\n", - "[0.63,0.68,-0.79,-0.77,0.54,0.82,-0.85,-0.01,0.73,0.54,-0.56,-0.16,-0.71,-0.32,-0.27,0.06,-0.86,0.12,-0.04,0.09],\n", - "[-0.39,0.38,-0.2,0.25,0.92,-0.15,0.26,0.3,-0.23,-0.23,-0.35,-0.85,-0.63,-0.57,0.82,-0.76,-0.98,0.5,-0.92,-0.32],\n", - "[0.88,0.34,-0.24,0.28,0.42,-0.88,0.4,-0.75,0.41,0.15,0.98,-0.22,0.67,0.74,0.66,-0.88,0.16,-0.66,-0.57,0.69],\n", - "[-0.65,-0.67,0.9,0.05,0.36,-0.14,-0.46,0.66,-0.96,-0.93,0.91,-0.13,-0.46,-0.69,-0.66,0.62,-0.16,0.82,0.04,-0.86],\n", - "[0.36,0.86,-0.54,0.12,0.56,0.23,-0.31,-0.53,-0.65,-0.24,0.41,0.47,-0.79,0.44,-0.99,-0.3,-0.49,-0.36,0.43,0.87],\n", - "[0.14,0.96,0.41,-0.47,-0.76,0.42,0.2,-0.48,-0.73,0.62,-0.6,0.74,-0.38,-0,-0.21,0.15,-0.13,-0.76,0.38,0.64]\n", - "]\n", - "\n", - "\n", - "\n", - "52\n", - "\n", - "RuleNeuron:52:lstm1_out__right(T) :- lstm1_out__i(T), lstm1_out__n(T). [transformation=identity, combination=elproduct]\n", - "val: [-0.02,-0.02,-0.01,-0.01,-0,-0,-0.01,0,-0.01,-0,-0.01,0.01,-0.01,0.01,-0,0.01,-0,0.01,-0.02,0.01]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "52->46\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "52->49\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "55\n", - "\n", - "RuleNeuron:55:lstm1_out__c(T) :- lstm1_out__left(T), lstm1_out__right(T). [transformation=identity]\n", - "val: [-0.02,-0.02,-0.01,-0.01,-0,-0,-0.01,0.01,-0.01,-0,-0.01,0.01,-0.01,0.01,-0,0.01,0,0.01,-0.02,0.02]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum\n", - "\n", - "\n", - "\n", - "55->43\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "55->52\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "59\n", - "\n", - "RuleNeuron:59:lstm1_out(T) :- lstm1_out__o(T), tanh(lstm1_out__c(T)). [transformation=identity, combination=elproduct]\n", - "val: [-0.01,-0.01,-0.01,-0.01,-0,-0,-0.01,0,-0,-0,-0.01,0,-0,0.01,-0,0,0,0.01,-0.01,0.01]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "59->37\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "59->55\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "62\n", - "\n", - "WeightedRuleNeuron:62:lstm1_out__o(T) :- {20, 1} x(T), {20, 20} lstm1_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.51,0.5,0.5,0.5,0.51,0.5,0.49,0.51,0.51,0.5,0.49,0.5,0.49,0.5,0.49,0.51,0.52,0.51,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "62->1\n", - "\n", - "\n", - "4:w4:[20, 1]:[0.61,0.97,-0.44,-0.47,-0.14,-0.94,0.63,-0.86,-0.09,-0.8,0.36,0.42,-0.65,0.19,0.91,0.76,-0.61,-0.6,-0.19,0.43]\n", - "\n", - "\n", - "\n", - "62->59\n", - "\n", - "\n", - "5:w5:[20, 20]:[\n", - "[0.72,0.53,0.15,-0.5,0.14,-0.67,-0.08,-0.83,-0.83,0.2,0.92,0.41,0.24,-0.61,0.99,-0.06,-0.81,-0.67,0.66,0.42],\n", - "[-0.7,-0.76,-0.46,-0.8,0.55,0.47,0.68,0.55,-0.86,-0.61,-0.04,0.39,0.06,0.38,0.09,0.78,0.02,-0.28,0.42,-0.06],\n", - "[0.74,-0.76,-0.53,-0.18,0.39,0.97,0.1,-0.1,0.78,-0.23,-0.44,0.32,0.68,0.21,-0.65,0.05,0.82,-0.84,0.88,0.7],\n", - "[-0.13,-0.75,0.84,0.86,0.82,-0.42,0.52,-0.75,0.55,0.04,0.18,-0.79,0.15,0.9,0.97,0.55,0.21,-0.91,-0.74,-0.77],\n", - "[-0.74,0.57,0.26,0.94,-0.7,-0.2,-0.3,-0.82,-0.65,-0.77,-0.71,0.16,0.94,-0.09,0.62,-0.33,-0.49,-0.09,0.61,0.65],\n", - "[-0.14,-0.15,-0.85,-0.76,0.9,0.57,-0.45,0.93,0.34,0.23,0.95,-0.49,0.1,0.73,0.91,-0.66,-0.28,-0.82,0.1,0.94],\n", - "[0.49,0.94,-0.85,-0.41,0.37,0.94,0.62,0.49,0.67,0.78,0.29,-0.45,-0.42,-0.71,-0.02,-0.9,-0.15,-0.63,0.42,-0.53],\n", - "[-0.67,0.19,-0.06,-0.64,0.06,0.84,0.12,0.33,0.35,0.36,-0.02,0.25,0.33,-0.21,0.59,-0.65,-0.53,-0.97,-0.27,0.32],\n", - "[-0.46,0.27,-0.21,0.33,-0.38,0.65,0.57,-0.73,0.07,0.3,-0.45,0.96,0.51,0.77,-0.05,0.2,0.16,0.35,0.72,-0.2],\n", - "[0.41,-0.02,-0.53,-0.54,-0.21,0.94,0.41,-0.11,-0.04,0.94,-0.7,-0.46,-0.74,0.67,-0.41,0.81,-0.59,0.52,-0.89,0.8],\n", - "[0.21,0.89,-0.28,-0.78,-0.59,-0.95,0.83,0.46,-0.89,-0.61,-0.06,0.49,-0.5,-0.3,0.68,-0.29,0.16,0.1,0.53,0.67],\n", - "[-0.37,0.98,0.6,0.36,0.62,0.49,0.9,0.3,-0.28,0.35,-0.39,0.92,0.26,0.39,-0.84,0.28,-0.38,-0.93,0.26,0.4],\n", - "[-0.71,0.99,0.24,0.4,-0.4,-0.97,-0.43,0.83,-0.76,0.95,0.77,-0.31,0.49,0.96,0.73,-0.15,0.2,-0.97,0.68,0.13],\n", - "[0.87,-0.44,-0,-0.86,-0.06,-0.62,0.65,0.02,-0.77,0.04,-0.14,-0.73,-0.8,-0.05,0.45,-0.51,-0.43,0.69,0.15,-0.28],\n", - "[-0.13,-0.97,0.27,-0.69,-0.61,0.7,-0.95,-0.56,-0.85,-0.64,-0.9,0.75,-0.48,-0.89,0.62,-0.07,-0.51,-0.04,-0.86,0.45],\n", - "[-0.35,-0.88,-0.06,-0.84,0.41,0.05,-0.44,-0.92,0.19,0.44,-0.63,-0.82,-0.97,0.16,-0.67,0.81,-0.8,0.97,0.08,0.9],\n", - "[0.34,-0.09,-0.48,-0.71,-0.05,-0.36,-0.68,-0.18,-0.98,0.4,-0.13,0.31,-0.92,0.58,0.61,-0.82,-0.99,-0.16,0.95,-0.38],\n", - "[-0.32,-0.58,0.36,-0.64,-0.58,-0.64,-0.23,-0.78,0.39,0.97,0.95,0.88,0.06,0.48,-0.99,-0.39,-0.18,0.02,0.96,-0.99],\n", - "[0.35,0.32,0.94,0.32,0.35,-0.47,-0.51,0.25,-0.41,0.99,-0.89,-0.97,-0.3,-0.98,0.23,-0.44,-0.47,0.23,-0.23,0.83],\n", - "[-0.75,0.32,-0.68,-0.13,0.28,0.75,-0.69,0.88,0.42,0.49,-0.49,-0.46,-0.7,0.3,0.49,0.82,-0.68,0.18,-0.11,0.33]\n", - "]\n", - "\n", - "\n", - "\n", - "65\n", - "\n", - "WeightedRuleNeuron:65:lstm1_out__f(T) :- {20, 1} x(T), {20, 20} lstm1_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.49,0.5,0.49,0.5,0.5,0.51,0.51,0.49,0.5,0.49,0.5,0.49,0.49,0.5,0.5,0.5,0.52,0.49,0.5,0.51]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "65->1\n", - "\n", - "\n", - "2:w2:[20, 1]:[-0.74,0.22,-0.98,0.54,0.43,-0.24,-0.29,-0.31,0.35,0.08,0.83,-0.93,-0.47,-0.5,0.19,-0.54,0.07,-0.44,0.44,-0.2]\n", - "\n", - "\n", - "\n", - "65->59\n", - "\n", - "\n", - "3:w3:[20, 20]:[\n", - "[-0.03,0.55,0.74,-0.99,-0.63,0.45,0.79,-0.05,0.19,0.11,-0.54,0.93,0.56,-0.38,0.35,0.26,0.1,0.88,-0.67,0.08],\n", - "[0.83,-0.1,-0.16,0.02,0.53,0.47,-0.19,0.19,-0.78,-0.65,-0.13,-0.81,0.47,-0.47,-0.39,0.04,0.19,0.64,-0.11,-0.57],\n", - "[-0.09,0.25,0.99,0.77,-0.45,0.28,-0.85,0.02,-0.06,-0.42,0.74,-0.95,0.9,0.44,-0.87,-0.69,0.67,-0.23,-0.52,0.72],\n", - "[-0.69,0.19,-0.84,0.97,-0.05,-0.21,-0.8,0.43,-0.62,-0.18,-0.45,0.49,0.25,-0.99,-0.27,-0.89,-0.53,0.11,-0.82,0.22],\n", - "[-0.28,-0.31,-0.81,-0.33,0.16,-0.8,0.65,-0.58,-0.19,0.95,-0.09,-0.22,-0.24,-0.13,-0.94,0.67,0.43,-0.65,-0.01,-0.8],\n", - "[0.06,-0.98,0.49,0.02,0.12,0.07,-0.91,-0.6,-0.15,-0.98,-0.75,-0.75,-0.87,0.79,-0.7,0.02,0.64,-0.24,0.98,0.69],\n", - "[-0.11,-0.98,-0.96,0.9,-0.86,0.21,-0.39,-0.17,-0.27,0.23,-0.37,-0.83,-0.33,0.97,0.38,-0.64,-0.53,0.49,-0.52,-0.8],\n", - "[0.28,0.07,-0.13,0.67,-0.35,-0.13,0.7,-0.06,0.76,0.15,0.13,-0.39,0.94,-0.68,-0.17,0.25,-0.81,0.44,-0.99,-0.55],\n", - "[-0.41,0.46,0.79,-0.01,0.19,0.67,0.85,0.86,-0.45,-0.25,0.47,0.38,0.8,-0.84,0.28,0.89,0.95,-0.43,0.2,0.26],\n", - "[-0.23,-0.43,0.71,-0.34,-0.4,0.87,0.56,-0.98,-0.58,-0.83,1,-0.09,0.32,0.51,-0.38,-0.35,0.88,0.35,0.94,0.98],\n", - "[-0.35,-0.66,-0.74,-0.31,0.69,-0.82,0.14,-0.62,-0.23,0.95,0.53,0.6,0.69,-0.32,-0.29,0.63,0.15,0.18,-0.39,0.31],\n", - "[0.22,0.72,0.65,-0.58,-0.03,0.29,0.66,0.44,-0.57,-0.94,-0.86,-0.23,0.42,0.89,-0.24,-0.09,-0.18,0.64,-0.43,-0.68],\n", - "[-0.68,0.64,-0.59,-0.43,-0.18,0.43,0.72,0.41,0.03,-0.36,0.82,0.51,0.4,0.08,0.82,0.97,-0.1,0.6,0.28,-0.51],\n", - "[-0.31,0.7,-0.16,-0.97,-0.42,0.75,0.14,-0.51,0.47,0.85,-0.17,0.86,-0.79,0.6,-0.67,-0.43,0.63,0.28,0.67,-0.69],\n", - "[-0.17,0.33,-0.8,-0.86,-0.7,0.31,0.93,-0.26,-1,-0.9,0.6,-0.24,-0.11,0.51,0.31,0.82,0.22,-0.91,0.87,0.65],\n", - "[-0.32,-0.4,0.4,0.72,-0.22,0.33,-0.36,0.01,0.27,-0.96,-0.92,0.91,-0.18,-0.61,-0.7,-0.72,-0.65,0.44,0.39,-0.93],\n", - "[0.32,0.82,-0.1,-0.45,-0.69,0.17,0.22,0.53,-0.43,-0.53,-0.75,0.03,0.05,-0.73,0.56,0.46,-0.98,-0.4,-0.74,-0.95],\n", - "[0.47,-0.88,-0.25,-0.7,0.99,-0.14,-0.03,0.42,-0.84,-0.12,0.43,0.88,0.15,-0.91,-0.47,0.48,-0.47,-0.74,-0.4,0.46],\n", - "[0.24,0.98,0.53,-0.2,-0.92,-0.49,-0.44,-0.9,-0.03,-0.34,-0.12,-0.1,-0.94,-0.7,-0.95,0.99,0.54,-0.99,-0.3,-0.18],\n", - "[-0.91,-0.67,0.45,0.64,-0.26,0.99,-0.61,0.01,0.09,-0.64,-0.66,-0.15,0.02,0.81,-0.64,-0.7,-0.99,0.43,-0.81,0.39]\n", - "]\n", - "\n", - "\n", - "\n", - "68\n", - "\n", - "RuleNeuron:68:lstm1_out__left(T) :- lstm1_out__f(T), lstm1_out__c(Z), @next(Z, T). [transformation=identity, combination=elproduct]\n", - "val: [-0.01,-0.01,-0,-0.01,-0,-0,-0.01,0,-0,-0,-0.01,0,-0,0.01,-0,0,0,0.01,-0.01,0.01]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "68->55\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "68->65\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "71\n", - "\n", - "WeightedRuleNeuron:71:lstm1_out__i(T) :- {20, 1} x(T), {20, 20} lstm1_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.49,0.48,0.51,0.51,0.5,0.51,0.5,0.5,0.51,0.49,0.49,0.5,0.49,0.49,0.51,0.51,0.49]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "71->1\n", - "\n", - "\n", - "0:w0:[20, 1]:[0.75,0.98,0.3,0.15,0.74,0.4,-0.45,0.34,0.4,-0.25,-0.22,-0.14,0.35,-0.51,-0.06,0.93,0.37,-0.03,-0.62,0.86]\n", - "\n", - "\n", - "\n", - "71->59\n", - "\n", - "\n", - "1:w1:[20, 20]:[\n", - "[-0.28,0.14,-0.67,-0.56,0.68,0.25,-0.87,0.41,-0.87,-0.13,-0.83,0.91,-0.52,0.74,-0.95,-0.12,0.16,0.99,-0.73,-0.75],\n", - "[0.46,-0.44,0.42,0.58,-0.59,0.89,0.97,-0.77,-0.81,0.21,-0.76,0.75,0.65,-0.61,-0.45,-0.17,-0.73,-0.73,0.21,-0.91],\n", - "[0.34,0.68,0.31,0.37,-0.27,0.02,-0.48,0.62,0.4,-0.19,0.94,-0.25,-0.4,0.49,-0.19,-0.57,0.26,0.31,-0.17,-0.63],\n", - "[0.26,0.33,0.6,-0.97,-0.92,-0.17,-0.23,0.78,0.24,0.17,0.46,-0.27,-0.84,-0.55,-0.34,-0.19,0.15,0.87,-0.61,-0.96],\n", - "[0.31,0.56,0.81,-0.72,-0.4,-0.89,0.47,-0.66,-0.61,-0.12,-0.57,0.82,0.5,0.12,0.48,-0.31,-0.28,-0.4,0.79,-0.86],\n", - "[-0.15,0.61,-0.65,-0.64,-0.56,0.09,-0.34,0.16,-0.81,-0.2,-0.79,0.7,0.72,-0.61,-0.28,-0.74,-1,0.87,0.01,-0.07],\n", - "[-0.18,-0.32,-0.74,0.35,0.19,-0.62,-0.77,0.61,0.84,0.04,-1,0.61,-0.19,0.1,-0.08,-0.69,0.2,0.79,0.41,0.93],\n", - "[0.34,-0.05,-0.51,-0.86,-0.25,-0.43,-0.78,-0.15,0.58,-0.74,0.3,-0.16,0.52,-0.51,-0.38,-0.23,0.47,0.2,-0.15,0.93],\n", - "[-0.12,0.62,0.77,0.59,0.9,-0.44,0.29,-0.31,-0.65,0.03,0.11,0.36,-0.09,-0.37,0.16,-0.45,0.67,0.62,0.19,-0.91],\n", - "[-0.82,-0.41,-0.83,-0.57,0.36,-0.88,0.38,0.32,-0.63,-0.1,-0.22,0.89,-0.5,0.35,-0.73,-0.33,-0.88,-0.78,-0.48,0.42],\n", - "[-0.12,-0.88,0.09,-0.13,-0.74,-0.55,0.71,0.83,0.2,0,-0.06,0.59,0.1,-0.85,0.52,0.04,0.62,-0.83,0.19,-0.57],\n", - "[0.31,0.27,-0.83,0.24,0.2,-0.33,-0.13,-0.45,0.33,-0.85,0.07,0.72,0.56,-0.73,0.28,-0.77,-0.36,-0.13,0.69,0.3],\n", - "[-0.64,-0.74,-0.15,-0.25,0.04,-0,-0.92,-0.23,-0.12,0.32,0.25,0.07,-0.84,-0.52,-0.28,-0.87,-0.65,0.88,-0.04,-0.49],\n", - "[-0.6,0.16,0.69,-0.89,-0.84,-0.92,-0.81,0.06,-0.59,-0.63,0.76,0.77,0.78,0.66,-0.57,0.45,0.44,0.11,-0.49,-0.3],\n", - "[-0.47,-0.9,0.52,0.53,-0.29,0.43,-0.04,-0.55,0.81,-0.97,-0.65,-0.71,-0.18,-0.66,-0.12,-0.09,0.15,0.58,0.41,-0.09],\n", - "[-0.21,0.95,0.26,0.19,0.56,0.4,-1,0.18,-0.58,-0.85,0.12,0.64,-0.33,0.18,0.45,0.77,-0.3,-0.45,-0.96,0.72],\n", - "[-1,-0.65,-0.89,0.59,-0.09,0,-0.21,0.26,0.9,-0.23,-0.43,0.84,-0.82,-0.67,-0.75,-0.25,0.58,0.21,0.74,-0.19],\n", - "[0.13,0.69,-0.99,0.41,-0.39,-0.84,0.24,-0.8,-0.64,0.2,-0.93,0.49,-0.54,-0.88,0.95,0.98,-0.6,0.13,0.68,-0.41],\n", - "[0.41,-0.31,-0.12,0.46,-0.12,0.44,-0.95,0.55,-0.66,-0.46,0.31,-0.27,-0.34,0.89,0.79,-0.43,0.82,-0.14,0.57,0.04],\n", - "[0.39,0.45,-0.17,-0.83,-0.04,0.14,-0.95,0.34,0.58,-0.62,-0.23,-0.35,0.46,-0.13,0.04,-0.48,-0.6,-0.39,-0.71,0.7]\n", - "]\n", - "\n", - "\n", - "\n", - "74\n", - "\n", - "WeightedRuleNeuron:74:lstm1_out__n(T) :- {20, 1} x(T), {20, 20} lstm1_out(Z), @next(Z, T). [transformation=tanh]\n", - "val: [-0.06,-0.06,-0.02,-0.04,0,-0.01,-0.03,0.01,-0.02,-0.01,-0.03,0.02,-0.02,0.04,-0.01,0.03,-0.01,0.05,-0.06,0.05]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Tanh\n", - "\n", - "\n", - "\n", - "74->1\n", - "\n", - "\n", - "6:w6:[20, 1]:[-0.36,0.99,0.89,-0.51,-0.12,0.21,0.43,0.69,0.25,0.9,-0.71,-0.87,0.46,-0.52,-0.22,-0.25,-0.69,-0.65,0.31,0.63]\n", - "\n", - "\n", - "\n", - "74->59\n", - "\n", - "\n", - "7:w7:[20, 20]:[\n", - "[-0.54,0.2,0.56,-0.14,0.56,0.07,0.23,0.01,-0.18,0.53,-0.57,0.18,0.58,0.47,-0.76,0.77,0.83,-0.63,-0.53,0.07],\n", - "[-0.1,0.37,-0.13,-0.81,0.3,0.77,0.83,-0.26,-0.99,-0.62,0.06,0.32,0.34,-0.62,0.38,0.02,-0.99,-0.82,-0.36,0.61],\n", - "[0.95,-0.42,-0.69,-0.37,-0.84,1,-0.57,0.04,-0.34,0.69,0.27,0.85,-0.13,-0.5,-0.98,0.67,0.23,-0.18,0.24,0.94],\n", - "[0.5,0.43,-0.75,-0.58,-0.11,-0.14,-0.51,-0.75,-0.46,0.13,0.48,-0.94,-0.91,-0.93,-0.14,0.3,-0.57,-0.67,-0.2,0.19],\n", - "[0.01,0.94,-0.07,-0.13,0.64,0.36,-0.91,-0.8,-0.93,-0.45,1,-0.68,0.38,-0.87,0.18,-0.07,0.68,0.2,0.65,0.09],\n", - "[-0.61,0.04,0,-0.81,0.17,-0.29,0.84,0.87,-0.81,0.23,0.99,0.65,0.53,0.89,0.05,0.06,0.4,0.16,0.6,0.37],\n", - "[-0.35,0.62,-0.26,-0.8,-0.58,0.17,0.26,-0.48,-0.27,-0.69,-0.51,-0.94,0.98,0.82,0.49,0.44,0.67,-0.12,0.01,0.83],\n", - "[0.97,-0.29,-0.33,0.67,-0.04,-0.92,-0.48,-0.73,0.13,-0.33,0.15,0.14,-0.09,0.43,-0.32,-0.7,-0.25,-0.14,-0,-0.98],\n", - "[-0.17,0.51,-0.5,-0.26,-0.79,-0.2,-0.59,0.52,0.94,0.86,0.37,0.74,-0.53,0.85,0.81,-0.81,0.59,0.25,-0.18,-0.28],\n", - "[-0.85,0.08,0.71,-0.22,0.59,0.57,0.02,-0.79,0.6,0.11,-0.52,0.51,-0.71,0.12,-0.41,0.4,-0.64,0.92,0,0.63],\n", - "[-0.88,0.53,-0.2,0.95,-0.52,0.08,0.37,-0.92,0.29,0.11,0.82,-0.2,-0.77,0.48,0.55,-0.69,-0.66,0.79,-0.12,0.37],\n", - "[0.48,-0.96,-0.65,-0.1,0.19,0.59,0.15,0.15,-0.67,0.36,0.39,-0.89,-0.46,-0.36,-0.73,0.9,0.39,-0.09,0.55,0.05],\n", - "[-0.93,0.73,-0.93,0.49,-0.84,-0.3,0.59,-0.53,-0.62,-0.5,0.7,-0.36,-0.11,-0.07,0.33,-0.02,-0.82,0.08,-0.39,-0.42],\n", - "[0.56,0.72,-0.78,-0.33,-0.41,0.48,-0.55,0.85,-0.86,0.3,-0.44,-0.13,0.02,0.38,0.61,0.23,-0.18,0.76,0.2,0.1],\n", - "[0.63,0.68,-0.79,-0.77,0.54,0.82,-0.85,-0.01,0.73,0.54,-0.56,-0.16,-0.71,-0.32,-0.27,0.06,-0.86,0.12,-0.04,0.09],\n", - "[-0.39,0.38,-0.2,0.25,0.92,-0.15,0.26,0.3,-0.23,-0.23,-0.35,-0.85,-0.63,-0.57,0.82,-0.76,-0.98,0.5,-0.92,-0.32],\n", - "[0.88,0.34,-0.24,0.28,0.42,-0.88,0.4,-0.75,0.41,0.15,0.98,-0.22,0.67,0.74,0.66,-0.88,0.16,-0.66,-0.57,0.69],\n", - "[-0.65,-0.67,0.9,0.05,0.36,-0.14,-0.46,0.66,-0.96,-0.93,0.91,-0.13,-0.46,-0.69,-0.66,0.62,-0.16,0.82,0.04,-0.86],\n", - "[0.36,0.86,-0.54,0.12,0.56,0.23,-0.31,-0.53,-0.65,-0.24,0.41,0.47,-0.79,0.44,-0.99,-0.3,-0.49,-0.36,0.43,0.87],\n", - "[0.14,0.96,0.41,-0.47,-0.76,0.42,0.2,-0.48,-0.73,0.62,-0.6,0.74,-0.38,-0,-0.21,0.15,-0.13,-0.76,0.38,0.64]\n", - "]\n", - "\n", - "\n", - "\n", - "77\n", - "\n", - "RuleNeuron:77:lstm1_out__right(T) :- lstm1_out__i(T), lstm1_out__n(T). [transformation=identity, combination=elproduct]\n", - "val: [-0.03,-0.03,-0.01,-0.02,0,-0,-0.02,0.01,-0.01,-0.01,-0.01,0.01,-0.01,0.02,-0,0.02,-0.01,0.02,-0.03,0.02]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "77->71\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "77->74\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "80\n", - "\n", - "RuleNeuron:80:lstm1_out__c(T) :- lstm1_out__left(T), lstm1_out__right(T). [transformation=identity]\n", - "val: [-0.04,-0.04,-0.02,-0.03,-0,-0,-0.02,0.01,-0.02,-0.01,-0.02,0.01,-0.02,0.02,-0,0.02,-0.01,0.03,-0.04,0.03]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum\n", - "\n", - "\n", - "\n", - "80->68\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "80->77\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "84\n", - "\n", - "RuleNeuron:84:lstm1_out(T) :- lstm1_out__o(T), tanh(lstm1_out__c(T)). [transformation=identity, combination=elproduct]\n", - "val: [-0.02,-0.02,-0.01,-0.01,-0,-0,-0.01,0,-0.01,-0,-0.01,0.01,-0.01,0.01,-0,0.01,-0,0.02,-0.02,0.02]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "84->62\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "84->80\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "87\n", - "\n", - "WeightedRuleNeuron:87:lstm1_out__o(T) :- {20, 1} x(T), {20, 20} lstm1_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.52,0.5,0.5,0.5,0.51,0.5,0.49,0.51,0.52,0.5,0.49,0.51,0.49,0.5,0.49,0.52,0.52,0.51,0.5,0.49]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "87->0\n", - "\n", - "\n", - "4:w4:[20, 1]:[0.61,0.97,-0.44,-0.47,-0.14,-0.94,0.63,-0.86,-0.09,-0.8,0.36,0.42,-0.65,0.19,0.91,0.76,-0.61,-0.6,-0.19,0.43]\n", - "\n", - "\n", - "\n", - "87->84\n", - "\n", - "\n", - "5:w5:[20, 20]:[\n", - "[0.72,0.53,0.15,-0.5,0.14,-0.67,-0.08,-0.83,-0.83,0.2,0.92,0.41,0.24,-0.61,0.99,-0.06,-0.81,-0.67,0.66,0.42],\n", - "[-0.7,-0.76,-0.46,-0.8,0.55,0.47,0.68,0.55,-0.86,-0.61,-0.04,0.39,0.06,0.38,0.09,0.78,0.02,-0.28,0.42,-0.06],\n", - "[0.74,-0.76,-0.53,-0.18,0.39,0.97,0.1,-0.1,0.78,-0.23,-0.44,0.32,0.68,0.21,-0.65,0.05,0.82,-0.84,0.88,0.7],\n", - "[-0.13,-0.75,0.84,0.86,0.82,-0.42,0.52,-0.75,0.55,0.04,0.18,-0.79,0.15,0.9,0.97,0.55,0.21,-0.91,-0.74,-0.77],\n", - "[-0.74,0.57,0.26,0.94,-0.7,-0.2,-0.3,-0.82,-0.65,-0.77,-0.71,0.16,0.94,-0.09,0.62,-0.33,-0.49,-0.09,0.61,0.65],\n", - "[-0.14,-0.15,-0.85,-0.76,0.9,0.57,-0.45,0.93,0.34,0.23,0.95,-0.49,0.1,0.73,0.91,-0.66,-0.28,-0.82,0.1,0.94],\n", - "[0.49,0.94,-0.85,-0.41,0.37,0.94,0.62,0.49,0.67,0.78,0.29,-0.45,-0.42,-0.71,-0.02,-0.9,-0.15,-0.63,0.42,-0.53],\n", - "[-0.67,0.19,-0.06,-0.64,0.06,0.84,0.12,0.33,0.35,0.36,-0.02,0.25,0.33,-0.21,0.59,-0.65,-0.53,-0.97,-0.27,0.32],\n", - "[-0.46,0.27,-0.21,0.33,-0.38,0.65,0.57,-0.73,0.07,0.3,-0.45,0.96,0.51,0.77,-0.05,0.2,0.16,0.35,0.72,-0.2],\n", - "[0.41,-0.02,-0.53,-0.54,-0.21,0.94,0.41,-0.11,-0.04,0.94,-0.7,-0.46,-0.74,0.67,-0.41,0.81,-0.59,0.52,-0.89,0.8],\n", - "[0.21,0.89,-0.28,-0.78,-0.59,-0.95,0.83,0.46,-0.89,-0.61,-0.06,0.49,-0.5,-0.3,0.68,-0.29,0.16,0.1,0.53,0.67],\n", - "[-0.37,0.98,0.6,0.36,0.62,0.49,0.9,0.3,-0.28,0.35,-0.39,0.92,0.26,0.39,-0.84,0.28,-0.38,-0.93,0.26,0.4],\n", - "[-0.71,0.99,0.24,0.4,-0.4,-0.97,-0.43,0.83,-0.76,0.95,0.77,-0.31,0.49,0.96,0.73,-0.15,0.2,-0.97,0.68,0.13],\n", - "[0.87,-0.44,-0,-0.86,-0.06,-0.62,0.65,0.02,-0.77,0.04,-0.14,-0.73,-0.8,-0.05,0.45,-0.51,-0.43,0.69,0.15,-0.28],\n", - "[-0.13,-0.97,0.27,-0.69,-0.61,0.7,-0.95,-0.56,-0.85,-0.64,-0.9,0.75,-0.48,-0.89,0.62,-0.07,-0.51,-0.04,-0.86,0.45],\n", - "[-0.35,-0.88,-0.06,-0.84,0.41,0.05,-0.44,-0.92,0.19,0.44,-0.63,-0.82,-0.97,0.16,-0.67,0.81,-0.8,0.97,0.08,0.9],\n", - "[0.34,-0.09,-0.48,-0.71,-0.05,-0.36,-0.68,-0.18,-0.98,0.4,-0.13,0.31,-0.92,0.58,0.61,-0.82,-0.99,-0.16,0.95,-0.38],\n", - "[-0.32,-0.58,0.36,-0.64,-0.58,-0.64,-0.23,-0.78,0.39,0.97,0.95,0.88,0.06,0.48,-0.99,-0.39,-0.18,0.02,0.96,-0.99],\n", - "[0.35,0.32,0.94,0.32,0.35,-0.47,-0.51,0.25,-0.41,0.99,-0.89,-0.97,-0.3,-0.98,0.23,-0.44,-0.47,0.23,-0.23,0.83],\n", - "[-0.75,0.32,-0.68,-0.13,0.28,0.75,-0.69,0.88,0.42,0.49,-0.49,-0.46,-0.7,0.3,0.49,0.82,-0.68,0.18,-0.11,0.33]\n", - "]\n", - "\n", - "\n", - "\n", - "90\n", - "\n", - "WeightedRuleNeuron:90:lstm1_out__f(T) :- {20, 1} x(T), {20, 20} lstm1_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.49,0.49,0.5,0.49,0.5,0.52,0.51,0.5,0.5,0.5,0.5,0.49,0.5,0.5,0.5,0.5,0.52,0.49,0.51,0.51]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "90->0\n", - "\n", - "\n", - "2:w2:[20, 1]:[-0.74,0.22,-0.98,0.54,0.43,-0.24,-0.29,-0.31,0.35,0.08,0.83,-0.93,-0.47,-0.5,0.19,-0.54,0.07,-0.44,0.44,-0.2]\n", - "\n", - "\n", - "\n", - "90->84\n", - "\n", - "\n", - "3:w3:[20, 20]:[\n", - "[-0.03,0.55,0.74,-0.99,-0.63,0.45,0.79,-0.05,0.19,0.11,-0.54,0.93,0.56,-0.38,0.35,0.26,0.1,0.88,-0.67,0.08],\n", - "[0.83,-0.1,-0.16,0.02,0.53,0.47,-0.19,0.19,-0.78,-0.65,-0.13,-0.81,0.47,-0.47,-0.39,0.04,0.19,0.64,-0.11,-0.57],\n", - "[-0.09,0.25,0.99,0.77,-0.45,0.28,-0.85,0.02,-0.06,-0.42,0.74,-0.95,0.9,0.44,-0.87,-0.69,0.67,-0.23,-0.52,0.72],\n", - "[-0.69,0.19,-0.84,0.97,-0.05,-0.21,-0.8,0.43,-0.62,-0.18,-0.45,0.49,0.25,-0.99,-0.27,-0.89,-0.53,0.11,-0.82,0.22],\n", - "[-0.28,-0.31,-0.81,-0.33,0.16,-0.8,0.65,-0.58,-0.19,0.95,-0.09,-0.22,-0.24,-0.13,-0.94,0.67,0.43,-0.65,-0.01,-0.8],\n", - "[0.06,-0.98,0.49,0.02,0.12,0.07,-0.91,-0.6,-0.15,-0.98,-0.75,-0.75,-0.87,0.79,-0.7,0.02,0.64,-0.24,0.98,0.69],\n", - "[-0.11,-0.98,-0.96,0.9,-0.86,0.21,-0.39,-0.17,-0.27,0.23,-0.37,-0.83,-0.33,0.97,0.38,-0.64,-0.53,0.49,-0.52,-0.8],\n", - "[0.28,0.07,-0.13,0.67,-0.35,-0.13,0.7,-0.06,0.76,0.15,0.13,-0.39,0.94,-0.68,-0.17,0.25,-0.81,0.44,-0.99,-0.55],\n", - "[-0.41,0.46,0.79,-0.01,0.19,0.67,0.85,0.86,-0.45,-0.25,0.47,0.38,0.8,-0.84,0.28,0.89,0.95,-0.43,0.2,0.26],\n", - "[-0.23,-0.43,0.71,-0.34,-0.4,0.87,0.56,-0.98,-0.58,-0.83,1,-0.09,0.32,0.51,-0.38,-0.35,0.88,0.35,0.94,0.98],\n", - "[-0.35,-0.66,-0.74,-0.31,0.69,-0.82,0.14,-0.62,-0.23,0.95,0.53,0.6,0.69,-0.32,-0.29,0.63,0.15,0.18,-0.39,0.31],\n", - "[0.22,0.72,0.65,-0.58,-0.03,0.29,0.66,0.44,-0.57,-0.94,-0.86,-0.23,0.42,0.89,-0.24,-0.09,-0.18,0.64,-0.43,-0.68],\n", - "[-0.68,0.64,-0.59,-0.43,-0.18,0.43,0.72,0.41,0.03,-0.36,0.82,0.51,0.4,0.08,0.82,0.97,-0.1,0.6,0.28,-0.51],\n", - "[-0.31,0.7,-0.16,-0.97,-0.42,0.75,0.14,-0.51,0.47,0.85,-0.17,0.86,-0.79,0.6,-0.67,-0.43,0.63,0.28,0.67,-0.69],\n", - "[-0.17,0.33,-0.8,-0.86,-0.7,0.31,0.93,-0.26,-1,-0.9,0.6,-0.24,-0.11,0.51,0.31,0.82,0.22,-0.91,0.87,0.65],\n", - "[-0.32,-0.4,0.4,0.72,-0.22,0.33,-0.36,0.01,0.27,-0.96,-0.92,0.91,-0.18,-0.61,-0.7,-0.72,-0.65,0.44,0.39,-0.93],\n", - "[0.32,0.82,-0.1,-0.45,-0.69,0.17,0.22,0.53,-0.43,-0.53,-0.75,0.03,0.05,-0.73,0.56,0.46,-0.98,-0.4,-0.74,-0.95],\n", - "[0.47,-0.88,-0.25,-0.7,0.99,-0.14,-0.03,0.42,-0.84,-0.12,0.43,0.88,0.15,-0.91,-0.47,0.48,-0.47,-0.74,-0.4,0.46],\n", - "[0.24,0.98,0.53,-0.2,-0.92,-0.49,-0.44,-0.9,-0.03,-0.34,-0.12,-0.1,-0.94,-0.7,-0.95,0.99,0.54,-0.99,-0.3,-0.18],\n", - "[-0.91,-0.67,0.45,0.64,-0.26,0.99,-0.61,0.01,0.09,-0.64,-0.66,-0.15,0.02,0.81,-0.64,-0.7,-0.99,0.43,-0.81,0.39]\n", - "]\n", - "\n", - "\n", - "\n", - "93\n", - "\n", - "RuleNeuron:93:lstm1_out__left(T) :- lstm1_out__f(T), lstm1_out__c(Z), @next(Z, T). [transformation=identity, combination=elproduct]\n", - "val: [-0.02,-0.02,-0.01,-0.01,-0,-0,-0.01,0,-0.01,-0,-0.01,0.01,-0.01,0.01,-0,0.01,-0,0.02,-0.02,0.02]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "93->80\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "93->90\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "96\n", - "\n", - "WeightedRuleNeuron:96:lstm1_out__i(T) :- {20, 1} x(T), {20, 20} lstm1_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.49,0.48,0.5,0.51,0.5,0.51,0.5,0.5,0.51,0.49,0.49,0.5,0.49,0.49,0.5,0.52,0.49]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "96->0\n", - "\n", - "\n", - "0:w0:[20, 1]:[0.75,0.98,0.3,0.15,0.74,0.4,-0.45,0.34,0.4,-0.25,-0.22,-0.14,0.35,-0.51,-0.06,0.93,0.37,-0.03,-0.62,0.86]\n", - "\n", - "\n", - "\n", - "96->84\n", - "\n", - "\n", - "1:w1:[20, 20]:[\n", - "[-0.28,0.14,-0.67,-0.56,0.68,0.25,-0.87,0.41,-0.87,-0.13,-0.83,0.91,-0.52,0.74,-0.95,-0.12,0.16,0.99,-0.73,-0.75],\n", - "[0.46,-0.44,0.42,0.58,-0.59,0.89,0.97,-0.77,-0.81,0.21,-0.76,0.75,0.65,-0.61,-0.45,-0.17,-0.73,-0.73,0.21,-0.91],\n", - "[0.34,0.68,0.31,0.37,-0.27,0.02,-0.48,0.62,0.4,-0.19,0.94,-0.25,-0.4,0.49,-0.19,-0.57,0.26,0.31,-0.17,-0.63],\n", - "[0.26,0.33,0.6,-0.97,-0.92,-0.17,-0.23,0.78,0.24,0.17,0.46,-0.27,-0.84,-0.55,-0.34,-0.19,0.15,0.87,-0.61,-0.96],\n", - "[0.31,0.56,0.81,-0.72,-0.4,-0.89,0.47,-0.66,-0.61,-0.12,-0.57,0.82,0.5,0.12,0.48,-0.31,-0.28,-0.4,0.79,-0.86],\n", - "[-0.15,0.61,-0.65,-0.64,-0.56,0.09,-0.34,0.16,-0.81,-0.2,-0.79,0.7,0.72,-0.61,-0.28,-0.74,-1,0.87,0.01,-0.07],\n", - "[-0.18,-0.32,-0.74,0.35,0.19,-0.62,-0.77,0.61,0.84,0.04,-1,0.61,-0.19,0.1,-0.08,-0.69,0.2,0.79,0.41,0.93],\n", - "[0.34,-0.05,-0.51,-0.86,-0.25,-0.43,-0.78,-0.15,0.58,-0.74,0.3,-0.16,0.52,-0.51,-0.38,-0.23,0.47,0.2,-0.15,0.93],\n", - "[-0.12,0.62,0.77,0.59,0.9,-0.44,0.29,-0.31,-0.65,0.03,0.11,0.36,-0.09,-0.37,0.16,-0.45,0.67,0.62,0.19,-0.91],\n", - "[-0.82,-0.41,-0.83,-0.57,0.36,-0.88,0.38,0.32,-0.63,-0.1,-0.22,0.89,-0.5,0.35,-0.73,-0.33,-0.88,-0.78,-0.48,0.42],\n", - "[-0.12,-0.88,0.09,-0.13,-0.74,-0.55,0.71,0.83,0.2,0,-0.06,0.59,0.1,-0.85,0.52,0.04,0.62,-0.83,0.19,-0.57],\n", - "[0.31,0.27,-0.83,0.24,0.2,-0.33,-0.13,-0.45,0.33,-0.85,0.07,0.72,0.56,-0.73,0.28,-0.77,-0.36,-0.13,0.69,0.3],\n", - "[-0.64,-0.74,-0.15,-0.25,0.04,-0,-0.92,-0.23,-0.12,0.32,0.25,0.07,-0.84,-0.52,-0.28,-0.87,-0.65,0.88,-0.04,-0.49],\n", - "[-0.6,0.16,0.69,-0.89,-0.84,-0.92,-0.81,0.06,-0.59,-0.63,0.76,0.77,0.78,0.66,-0.57,0.45,0.44,0.11,-0.49,-0.3],\n", - "[-0.47,-0.9,0.52,0.53,-0.29,0.43,-0.04,-0.55,0.81,-0.97,-0.65,-0.71,-0.18,-0.66,-0.12,-0.09,0.15,0.58,0.41,-0.09],\n", - "[-0.21,0.95,0.26,0.19,0.56,0.4,-1,0.18,-0.58,-0.85,0.12,0.64,-0.33,0.18,0.45,0.77,-0.3,-0.45,-0.96,0.72],\n", - "[-1,-0.65,-0.89,0.59,-0.09,0,-0.21,0.26,0.9,-0.23,-0.43,0.84,-0.82,-0.67,-0.75,-0.25,0.58,0.21,0.74,-0.19],\n", - "[0.13,0.69,-0.99,0.41,-0.39,-0.84,0.24,-0.8,-0.64,0.2,-0.93,0.49,-0.54,-0.88,0.95,0.98,-0.6,0.13,0.68,-0.41],\n", - "[0.41,-0.31,-0.12,0.46,-0.12,0.44,-0.95,0.55,-0.66,-0.46,0.31,-0.27,-0.34,0.89,0.79,-0.43,0.82,-0.14,0.57,0.04],\n", - "[0.39,0.45,-0.17,-0.83,-0.04,0.14,-0.95,0.34,0.58,-0.62,-0.23,-0.35,0.46,-0.13,0.04,-0.48,-0.6,-0.39,-0.71,0.7]\n", - "]\n", - "\n", - "\n", - "\n", - "99\n", - "\n", - "WeightedRuleNeuron:99:lstm1_out__n(T) :- {20, 1} x(T), {20, 20} lstm1_out(Z), @next(Z, T). [transformation=tanh]\n", - "val: [-0.06,-0.07,0,-0.01,0.03,-0.01,-0.01,-0.02,-0.03,-0.02,-0.01,0.03,-0.02,0.04,-0.02,0.04,-0.05,0.06,-0.08,0.04]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Tanh\n", - "\n", - "\n", - "\n", - "99->0\n", - "\n", - "\n", - "6:w6:[20, 1]:[-0.36,0.99,0.89,-0.51,-0.12,0.21,0.43,0.69,0.25,0.9,-0.71,-0.87,0.46,-0.52,-0.22,-0.25,-0.69,-0.65,0.31,0.63]\n", - "\n", - "\n", - "\n", - "99->84\n", - "\n", - "\n", - "7:w7:[20, 20]:[\n", - "[-0.54,0.2,0.56,-0.14,0.56,0.07,0.23,0.01,-0.18,0.53,-0.57,0.18,0.58,0.47,-0.76,0.77,0.83,-0.63,-0.53,0.07],\n", - "[-0.1,0.37,-0.13,-0.81,0.3,0.77,0.83,-0.26,-0.99,-0.62,0.06,0.32,0.34,-0.62,0.38,0.02,-0.99,-0.82,-0.36,0.61],\n", - "[0.95,-0.42,-0.69,-0.37,-0.84,1,-0.57,0.04,-0.34,0.69,0.27,0.85,-0.13,-0.5,-0.98,0.67,0.23,-0.18,0.24,0.94],\n", - "[0.5,0.43,-0.75,-0.58,-0.11,-0.14,-0.51,-0.75,-0.46,0.13,0.48,-0.94,-0.91,-0.93,-0.14,0.3,-0.57,-0.67,-0.2,0.19],\n", - "[0.01,0.94,-0.07,-0.13,0.64,0.36,-0.91,-0.8,-0.93,-0.45,1,-0.68,0.38,-0.87,0.18,-0.07,0.68,0.2,0.65,0.09],\n", - "[-0.61,0.04,0,-0.81,0.17,-0.29,0.84,0.87,-0.81,0.23,0.99,0.65,0.53,0.89,0.05,0.06,0.4,0.16,0.6,0.37],\n", - "[-0.35,0.62,-0.26,-0.8,-0.58,0.17,0.26,-0.48,-0.27,-0.69,-0.51,-0.94,0.98,0.82,0.49,0.44,0.67,-0.12,0.01,0.83],\n", - "[0.97,-0.29,-0.33,0.67,-0.04,-0.92,-0.48,-0.73,0.13,-0.33,0.15,0.14,-0.09,0.43,-0.32,-0.7,-0.25,-0.14,-0,-0.98],\n", - "[-0.17,0.51,-0.5,-0.26,-0.79,-0.2,-0.59,0.52,0.94,0.86,0.37,0.74,-0.53,0.85,0.81,-0.81,0.59,0.25,-0.18,-0.28],\n", - "[-0.85,0.08,0.71,-0.22,0.59,0.57,0.02,-0.79,0.6,0.11,-0.52,0.51,-0.71,0.12,-0.41,0.4,-0.64,0.92,0,0.63],\n", - "[-0.88,0.53,-0.2,0.95,-0.52,0.08,0.37,-0.92,0.29,0.11,0.82,-0.2,-0.77,0.48,0.55,-0.69,-0.66,0.79,-0.12,0.37],\n", - "[0.48,-0.96,-0.65,-0.1,0.19,0.59,0.15,0.15,-0.67,0.36,0.39,-0.89,-0.46,-0.36,-0.73,0.9,0.39,-0.09,0.55,0.05],\n", - "[-0.93,0.73,-0.93,0.49,-0.84,-0.3,0.59,-0.53,-0.62,-0.5,0.7,-0.36,-0.11,-0.07,0.33,-0.02,-0.82,0.08,-0.39,-0.42],\n", - "[0.56,0.72,-0.78,-0.33,-0.41,0.48,-0.55,0.85,-0.86,0.3,-0.44,-0.13,0.02,0.38,0.61,0.23,-0.18,0.76,0.2,0.1],\n", - "[0.63,0.68,-0.79,-0.77,0.54,0.82,-0.85,-0.01,0.73,0.54,-0.56,-0.16,-0.71,-0.32,-0.27,0.06,-0.86,0.12,-0.04,0.09],\n", - "[-0.39,0.38,-0.2,0.25,0.92,-0.15,0.26,0.3,-0.23,-0.23,-0.35,-0.85,-0.63,-0.57,0.82,-0.76,-0.98,0.5,-0.92,-0.32],\n", - "[0.88,0.34,-0.24,0.28,0.42,-0.88,0.4,-0.75,0.41,0.15,0.98,-0.22,0.67,0.74,0.66,-0.88,0.16,-0.66,-0.57,0.69],\n", - "[-0.65,-0.67,0.9,0.05,0.36,-0.14,-0.46,0.66,-0.96,-0.93,0.91,-0.13,-0.46,-0.69,-0.66,0.62,-0.16,0.82,0.04,-0.86],\n", - "[0.36,0.86,-0.54,0.12,0.56,0.23,-0.31,-0.53,-0.65,-0.24,0.41,0.47,-0.79,0.44,-0.99,-0.3,-0.49,-0.36,0.43,0.87],\n", - "[0.14,0.96,0.41,-0.47,-0.76,0.42,0.2,-0.48,-0.73,0.62,-0.6,0.74,-0.38,-0,-0.21,0.15,-0.13,-0.76,0.38,0.64]\n", - "]\n", - "\n", - "\n", - "\n", - "102\n", - "\n", - "RuleNeuron:102:lstm1_out__right(T) :- lstm1_out__i(T), lstm1_out__n(T). [transformation=identity, combination=elproduct]\n", - "val: [-0.03,-0.04,0,-0.01,0.01,-0,-0.01,-0.01,-0.02,-0.01,-0.01,0.01,-0.01,0.02,-0.01,0.02,-0.03,0.03,-0.04,0.02]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "102->96\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "102->99\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "105\n", - "\n", - "RuleNeuron:105:lstm1_out__c(T) :- lstm1_out__left(T), lstm1_out__right(T). [transformation=identity]\n", - "val: [-0.05,-0.06,-0.01,-0.02,0.01,-0,-0.02,-0.01,-0.02,-0.01,-0.01,0.02,-0.02,0.03,-0.01,0.03,-0.03,0.04,-0.06,0.04]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum\n", - "\n", - "\n", - "\n", - "105->93\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "105->102\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "109\n", - "\n", - "RuleNeuron:109:lstm1_out(T) :- lstm1_out__o(T), tanh(lstm1_out__c(T)). [transformation=identity, combination=elproduct]\n", - "val: [-0.03,-0.03,-0,-0.01,0.01,-0,-0.01,-0,-0.01,-0.01,-0.01,0.01,-0.01,0.02,-0.01,0.02,-0.01,0.02,-0.03,0.02]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "109->87\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "109->105\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "116\n", - "\n", - "WeightedRuleNeuron:116:lstm2_out__o(T) :- {20, 20} lstm1_out(T), {20, 20} lstm2_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "116->4\n", - "\n", - "\n", - "13:w13:[20, 20]:[\n", - "[-0.39,-0.29,-0.09,-0.68,0.85,0.76,-0.02,0.19,-0.47,-0.1,0.87,0.37,0.15,-0.5,-0.75,-0.45,0.85,-0.88,0.4,-0.55],\n", - "[-0.52,0.24,0.04,-0.3,0.75,0.11,0.31,-0.73,0.4,-0.71,0.29,-0.83,-0.73,0.53,0.31,0.49,0.5,-0.37,-0.28,0.34],\n", - "[-0.92,0.48,-0.14,0.44,0.73,-0.74,-0.39,0.21,-0.09,0.57,-0.19,0.13,-0.26,-0.69,0.18,0.51,0.68,-0.14,-0.37,0.89],\n", - "[0.65,-0.71,-0.06,0.74,0.78,-0.6,-0.98,0.39,-0.26,-0.55,-0.18,0.44,0.28,-0.36,0.91,0.72,-0.57,0.91,-0.7,-0.91],\n", - "[-0.22,0.45,-0.72,-0.93,0.11,0.42,0.62,-0.93,0.6,-0.99,0.14,0.64,0.12,-0.68,-0.97,0.95,0.91,-0.64,-0.02,-0.43],\n", - "[-0.48,0.74,0.24,-0.38,-0.3,-0.16,0.89,0.26,-0.19,-0.57,0.16,-0.43,0.96,-0.86,-0.08,-0,0.39,-0.59,0.87,0.68],\n", - "[-0.8,0.32,0.74,0.25,0.92,0.76,-0.94,0.01,-0.49,0.47,0.1,0.56,-0.83,-0.31,-0.14,-0.17,0.35,0.26,0.29,-0.69],\n", - "[0.74,0.34,-0.74,0.45,-0.67,0.25,-0.93,0.37,-0.79,-0.66,-0.92,0.96,-0.98,-0.92,-0.85,0.95,-0.05,0.56,0.62,-0.37],\n", - "[0.22,-0.64,-0.18,0.48,-0.67,0.28,0.06,-0,0.57,-0.47,-0.15,-0.92,-0.57,0.93,0.69,-0.72,-0.87,0.49,-0.93,0.27],\n", - "[-0.98,0.78,-0.1,0.33,-0.42,0.07,0.89,-0.52,0.06,0.81,-0.65,0.65,-0.22,-0.63,0.18,0.18,-0.79,0.95,0.88,0.73],\n", - "[-0.54,0.96,-0.3,-0,0.8,0.07,-0.26,0.16,0.54,-0.02,0.53,-0.27,-0.26,0.81,0.87,0.9,0.99,-0.95,0.67,0.63],\n", - "[-0.25,0.73,0.42,-0.96,-0.21,-0.88,0.36,-0.14,0.44,-0.25,-0.45,-0.48,0.7,0.24,-0.87,0.43,-0.5,0.49,-0.27,0.08],\n", - "[-0.37,-0.89,0.2,0.63,0.81,0.4,-0.74,-0.06,0.14,0.7,0.36,0.19,-0.57,-0.74,-0.47,0.48,0.96,-0.36,-0.27,-0.81],\n", - "[-0.6,0.03,-0.3,0.4,-0.97,-0.32,0.65,-0.02,0.34,-0.42,-0.87,-0.3,0.94,-0.08,-0.8,-0.9,-0.24,0.3,-0.59,-0.18],\n", - "[-0.68,-0.87,-0.67,0.5,0.06,0.05,0.42,-0.55,-0.11,-0.08,0.36,-0.03,0.07,0.55,0.91,0.18,-0.93,0.93,0.77,-0.78],\n", - "[-0.74,0.64,-0.62,0.68,0.47,-0.1,-0.14,0.38,-0.66,-0.55,0.67,0.44,-0.67,0.84,-0.54,0.4,0.39,-1,0.78,0.75],\n", - "[0.35,-0.97,0.16,-0.8,-0.67,0.17,0.82,-0.91,0.75,0.17,-0.92,-0.09,0.75,0.68,-0.88,-0.49,-0.82,0.34,1,0.16],\n", - "[-0.98,-0.07,-0.36,0.17,-0.8,0.14,0.53,-0.21,0.66,0.31,0.84,-0.38,-0.98,0.72,0.32,0.77,0.79,-0.98,-0.23,-0.01],\n", - "[-0.8,-0.12,0.01,-0.2,-0.16,0.72,0.87,-0.2,-0.96,-0.79,-0.27,0.53,-0.84,0.31,-0.83,-0.7,0.27,-0.79,0.11,0.59],\n", - "[0.81,-0.99,0.27,0.25,-0.23,0.49,-0.56,-0.47,-0.83,0.11,-0.11,0.08,-0.87,-0.97,0.92,-0.13,0.55,0.55,0.32,-0.91]\n", - "]\n", - "\n", - "\n", - "\n", - "116->34\n", - "\n", - "\n", - "12:w12:[20, 20]:[\n", - "[-0.47,0.05,0.77,-0.82,0.69,-0.87,0.27,0.97,0.08,0.55,-0.66,-0.91,0.88,-0.82,0.07,-0.47,-0.05,0.09,0.5,-0.11],\n", - "[-0.67,0.38,-0.22,0.1,0.69,-0.92,0.38,0.12,-0.19,-0.69,-0.27,0.07,0.52,-0.33,-0.1,-0.14,-0.19,0.72,0.51,-0.04],\n", - "[-0.47,-0.54,-0.55,0.97,-0.46,-0.39,-0.27,-0.05,-0.2,0.54,-0.67,-0.03,-0.31,-0.83,-0.22,-0.36,0.4,-0.17,0.77,-0.14],\n", - "[0.3,0.87,0.14,-0.3,-0.21,-0.26,-0.62,-0.19,-0.64,-0.08,-0.15,0.46,0.42,0.3,0.8,0.09,0.39,-0.53,0.61,-0.9],\n", - "[-0.83,0.44,-0.56,0.14,0.6,-0.38,-0.71,-0.3,-0.6,0.85,-0.96,0.08,-0.04,-0.24,-0.94,0.5,-0.92,-0.99,0.89,0.68],\n", - "[-0.25,-0.83,-0.81,-0.13,0.12,-0.2,-0.28,-0.9,-0.13,-0.54,-0.42,0.26,0.11,-0.73,-0.33,0.21,0.55,0.89,-0.95,0.54],\n", - "[0.5,0.06,-0.35,0.01,0.72,0.93,0.6,-0.77,-0.2,0.95,-0.55,-0.03,-0.38,0.24,-0.56,-0.25,0.53,0.08,0.88,-0.37],\n", - "[-0.99,-0.54,0.28,-0.63,-0.52,-0.65,0.27,-0.57,-0.49,0.87,0.42,-0.02,-0.73,0.44,-0.42,0.07,-0.61,-0.56,0.26,-0.69],\n", - "[-0.5,-0.86,-0.32,-0.65,0.45,0.19,0.78,-0.29,0.61,-0.82,-0.26,-0.19,0.13,-0.47,0.55,0.14,-0.2,-0.55,-0.57,-0.87],\n", - "[-0.28,0.36,0.91,-0.47,0.92,0.79,0.99,-0.39,0.62,-0.27,0.74,0.12,0.93,-0.93,0.61,-0.42,-0.39,-0.27,0.58,-0.15],\n", - "[-0.36,-0.05,-0.59,0.26,0.08,-0.56,0.47,-0.41,0.42,0.89,0.36,-0.15,-0.92,-0.28,0.33,-0.87,-0.34,0.9,0.35,-0.85],\n", - "[0.35,-0.57,0.13,0.83,-0.49,0.53,-1,-0.14,-0.03,0.15,-0.25,-0.26,0.74,-0.42,0.18,0.62,-0.13,0.75,0.83,0.11],\n", - "[-0.66,0.15,0.8,0.93,0.93,0.01,-0.6,0.31,0.62,-0.65,0.17,-0.83,-0.25,-0.37,0.65,0.71,-0.6,0.45,-0.88,-0.26],\n", - "[-0.06,0.72,-0.4,-0.21,-0.86,-0.44,-0.69,0.51,0.97,-0.71,0.1,0.06,-0.51,-0.09,-0.88,0.68,0.43,0.71,0.86,-0.11],\n", - "[0.23,-0.34,-0.85,-0.67,-0.27,-0.32,-0.11,-0.06,-0.02,0.19,-0.97,-0.54,0.1,0.63,-0.82,-0.74,0.72,0.12,-0.41,-0.94],\n", - "[-0.06,0.22,-0.08,0.77,-0.53,0.71,0.82,-0.98,-0.67,-0.19,-0.94,0.16,-0.88,-0.99,0.83,0.24,-0.18,0.43,0.56,0.93],\n", - "[0.87,-0.53,-0.57,0.84,0.24,-0.66,-0.8,0.67,-0.8,-0.75,0.16,0.91,-0.34,-0.44,-0.88,0.17,0.82,-0.98,-0.42,0.53],\n", - "[-0.65,0.48,0.1,-0.42,-0.48,-0.46,0.83,-0.32,-0.32,0.08,0.7,0.13,-0.52,0.66,0.93,0.68,0.95,-0.44,0.3,0.72],\n", - "[-0.22,0.11,-0.02,0.71,-0.94,0.22,0.16,0.82,-0.22,-0.18,0.87,0.16,-0.01,0.28,0.44,-0.03,-0.87,-0.02,0.45,-0.59],\n", - "[-0.96,0.38,0.45,0.34,0.98,-0.37,-0.19,0.61,-0.29,0.58,0.01,-0.24,-0.73,-0.94,-0.26,0.1,-0.73,0.53,0.14,0.48]\n", - "]\n", - "\n", - "\n", - "\n", - "128\n", - "\n", - "WeightedRuleNeuron:128:lstm2_out__i(T) :- {20, 20} lstm1_out(T), {20, 20} lstm2_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "128->4\n", - "\n", - "\n", - "9:w9:[20, 20]:[\n", - "[-0.08,-0.9,0.89,0.73,-0.35,0.92,0.86,-0.84,0.7,0.28,0.12,-0.41,-0.54,-0.13,-0.99,0.04,-0.51,-0.42,0.34,0.32],\n", - "[0.54,0.4,0.8,0.55,-0.18,-0,0.51,0.89,-0.81,0.76,0.63,0.49,-0.18,-0.31,0.26,0.31,-0.66,-0.92,-0.29,0.77],\n", - "[-0.6,-0.32,-0.17,-0.44,0.25,-0.21,0.35,-0.35,-0.98,0.63,-0.47,0.24,0.1,0.06,-0.22,-0.38,-0.89,-0.12,0.95,0.29],\n", - "[-0.94,-0.15,-0.23,0.86,-0.52,0.54,0.38,0.62,-0.87,-0.38,-0.91,-0.8,0.78,0.87,-0.5,-0.94,-0.79,0.09,-0.08,-0.74],\n", - "[0.39,0.72,0.98,0.48,0.07,-0.84,0.12,-0.07,-0.4,0.32,-0.99,0.58,0.22,-0.53,-0.34,-0.14,-0.22,-0.94,-0.61,-0.97],\n", - "[0.66,0.57,0.08,0.97,-0.86,0.7,0.12,0.59,0.49,0.25,0.2,0.74,0.87,-0.4,0.2,0.81,0.11,0.02,0,0.33],\n", - "[-0.7,-0.66,0.69,-0,-0.64,0.31,0.95,0.21,-0.21,-0.61,0.49,-0.95,0.05,-0.76,-0.42,-0.55,0.1,0.22,0.05,-0.23],\n", - "[-0.71,-0.38,-0.37,-0.83,-0.24,-0.56,0.47,0.38,0.2,-0.8,0.71,-0.51,0.41,0.43,0.14,-0.29,-0.8,-0.54,-0.43,-0.57],\n", - "[0.53,0.97,-0.53,-0.09,-0.36,-0.71,0.94,0.02,0.48,0.37,0.73,0.05,-0.11,-0.3,-0.66,-0.85,0.34,-0.27,0.6,-0.44],\n", - "[0.67,0.89,-0.53,-0.59,-0.54,0.56,-0.39,0.81,0.67,0.98,-0.36,0.39,-0.81,-0.35,0.77,0.75,0.37,0.44,-0.91,-0.25],\n", - "[0.71,0.25,0.81,0.02,0.05,0.87,0.37,-0.39,0.45,0.58,-0.7,0.38,0.78,0.89,0.78,-0.67,0.01,0.64,-0.15,-0.2],\n", - "[-0.39,0.92,-0.95,0.4,-0.03,-0.71,0.02,-0.81,0.62,0.64,0.12,-0.35,-0.29,-0.18,0.89,-0.87,-0.02,-0.26,-0.58,-0.86],\n", - "[-0.54,0.62,-0.1,-0.64,-0.18,0.94,0.36,0.15,-0.3,0.45,-0.94,0.62,-0.03,0.89,-0.63,0.79,-0.69,-0.15,-0.89,0.5],\n", - "[-0.51,0.69,-0.59,-0.7,0.29,-0.92,0.62,-0.07,-0.18,0.88,-0.92,-0.23,-0.21,-0.71,-0.3,0.91,0.67,-0.31,-0.58,-0.58],\n", - "[-0.02,0.19,0.29,-0.86,-0.45,0.02,0.4,0.9,-0.37,-0.01,-0.3,-0.8,0.66,-0.6,0.23,0.03,-0.95,-0.12,0.43,0.85],\n", - "[-0.46,-0.83,0.56,-0.66,0.96,-0.27,-0.23,-0.81,-0.08,0.05,-0.44,0.73,0.17,0.37,-0.62,0.15,0.14,-0.43,-0.91,0.73],\n", - "[-0.34,0.16,-0.68,-0.82,0.74,-0.72,0.88,0.68,0.74,0.06,-0.33,-0.57,-0.54,0.49,0.96,0.27,0.08,0.11,-0.79,0.49],\n", - "[0.39,-0.83,0.16,-0.33,0.2,0.35,-0.64,0.9,0.3,-0.09,-0.92,-0.05,-0.84,0.18,-0.13,0.37,0.05,0.7,0.26,-0.28],\n", - "[-0.8,0.19,-0.7,0.1,-0.83,-0.92,0.95,0.86,0.89,-0.56,-0.02,0.14,0.26,0.76,0.18,0.29,-0.34,-0.75,-0.58,-0.95],\n", - "[0.51,0.09,0.65,-0.51,0.21,-0.62,0.69,-0.19,0.02,0.49,0,-0.01,0.71,0.83,-0.94,0.39,0.98,-0.7,-0.95,-0.82]\n", - "]\n", - "\n", - "\n", - "\n", - "128->34\n", - "\n", - "\n", - "8:w8:[20, 20]:[\n", - "[0.27,0.6,-0.82,0.69,-0.56,0.92,-0.02,0.9,0.98,0.9,0.49,-0.39,0.84,0.46,0.35,-0.67,0.41,-0.22,0.58,-0.26],\n", - "[-0.46,-0.96,-0.7,0.85,-0.16,-0.03,0.96,-0.02,-0.97,-0.51,0.4,0.14,0.43,0.28,0.67,-0.27,-0.88,0.3,-0.92,-0.75],\n", - "[0.89,-0.26,0.12,-0.45,0.42,-0.23,0.97,-0.29,-0.87,-0.7,0.58,-0.23,0.5,0.48,-0.76,-0.73,-0.41,-0.44,0.26,-0.46],\n", - "[0.19,0.29,0.3,-0.02,-0.97,0.99,0.62,-0.81,0.41,-0.77,-0.21,0.66,-0.61,0.2,-0.97,-0.77,-0.95,-0.48,0.96,-0.56],\n", - "[-0.61,-0.21,0.15,-0.07,-0.72,0.83,0.47,0.47,0.78,0.17,-0.6,-0.12,-0.26,-0.67,0.78,-0.54,-0.42,0.01,-0.15,0.29],\n", - "[-0.86,-0.58,-0.92,-0.78,-0.66,0.71,0.64,0.58,0.57,-0.21,0.95,0.7,-0.1,0.5,-0.36,-0.99,-0.84,-0.31,-0.34,-0.21],\n", - "[-0.48,0.52,-0.74,0.82,-0.76,-0.02,0.41,0.52,0.79,-0.39,-0.98,0.69,-0.14,-0,-0.67,-0.3,-0.55,-0.91,-0.97,0.94],\n", - "[-0.36,-0.32,-0.49,0.61,-0.67,-0.73,0.89,-0.58,0.88,0.97,0.98,0.13,-0.47,-0.13,0.99,-0.67,0.89,-0.93,-0.12,-0.48],\n", - "[0.2,-0.66,-0.01,-0.08,-0.54,0.47,0.65,0.41,-0.07,-0.01,0.39,-0.46,-0.54,-0.69,-0.38,-0.65,0.05,-0.91,0.74,0.79],\n", - "[0.48,0.09,0.71,-0.87,0.02,-0.23,-0.26,0.17,-0.54,-0.85,0.7,-0.45,0.14,0.91,-0.39,-0.76,-0.79,0.07,-0.23,0.75],\n", - "[0.61,0.89,0.85,0.52,-0.01,0.38,-0.34,0.67,0.46,-0.78,-0.92,0.08,-0.37,-0.38,0.36,0.88,0.12,0.24,0.81,-0.8],\n", - "[0.74,-0.7,0.5,-0.7,0.23,-0.46,-0.03,0.01,-0.61,-0.27,-0.62,-0.95,-0.4,0.53,-0.97,0.41,0.62,-0.84,0.91,0.98],\n", - "[-0.29,0.78,-0.58,0.26,0.45,-0.58,0.28,0.08,-0.29,0.5,0.31,-0.86,0.55,0.41,0.7,0.29,0.46,-0.24,-0.74,0.56],\n", - "[0.31,0.44,0.02,-0.96,0.43,-0.08,0.86,0.14,-0.57,0.07,-0.48,0.91,0.97,0.84,-0.06,-0.6,0.86,-0.65,-0.23,0.75],\n", - "[0.86,0.22,0.36,-0.23,0.14,0.93,0.8,0.78,0.65,-0.18,-0.99,-0.39,-0.07,0.18,0.4,0.83,-0.31,-0.18,-0.95,-0.43],\n", - "[0.83,-0.29,0.99,-0.63,-0.77,-0.98,-0.99,0.3,0.38,0.38,-0.31,-0.59,-0.28,-0.79,0.75,0.75,0.16,-0.79,-0.94,-0.01],\n", - "[0.51,0.49,-0.06,0.07,-0.07,-0.67,0.31,0.29,-0.84,0.66,-0.99,0.44,-0.17,-0.65,-0.81,0.13,-0.46,0.56,-0.75,0.61],\n", - "[0.5,0.25,-0.43,0.79,0.88,-0.35,-0.88,0.8,-0.87,0.22,-0.2,0.98,0.12,0.41,0.25,-0.49,0.3,0.55,-0.08,-0.96],\n", - "[0.49,-0.97,-0.72,-0.09,0.69,0,-0.2,0.88,-0.23,-0.02,0.12,-0.5,-0.53,-0.34,-0.78,-0.02,0.31,0.81,-0.37,-0.48],\n", - "[0.36,0.89,-0.91,-0.05,0.98,-0.04,0.38,0.96,0.39,-0.75,-0.28,0.83,0.29,0.69,0.43,0.57,0.17,-0.05,0.34,0.76]\n", - "]\n", - "\n", - "\n", - "\n", - "131\n", - "\n", - "WeightedRuleNeuron:131:lstm2_out__n(T) :- {20, 20} lstm1_out(T), {20, 20} lstm2_out(Z), @next(Z, T). [transformation=tanh]\n", - "val: [0,-0.01,-0,-0,-0.01,0,-0.01,-0,0,-0,-0.01,-0,-0.01,-0,0,0.01,0,-0.01,-0.01,0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Tanh\n", - "\n", - "\n", - "\n", - "131->4\n", - "\n", - "\n", - "15:w15:[20, 20]:[\n", - "[-0.74,-0.47,-0.56,0.04,-0.88,-1,-0.26,-0.18,-0.48,0.63,-0.6,-0.4,0.54,-0.6,-0.59,-0.04,-0.18,-0.51,0.79,-0.63],\n", - "[-0.53,0.89,0.75,-0.08,-0.36,0.6,0.12,0.29,-0.22,0.79,-0.06,-0.07,0.31,-0.3,0.02,0.76,-0.21,0.93,-0.06,0.17],\n", - "[-0.26,0.3,0.81,-0.51,-0.96,-0.77,0.67,-0.11,-0.63,-0.37,-0.17,-0.17,0.75,0.81,-0.14,0.37,0.4,-0.74,0.43,-0.95],\n", - "[-0.91,-0.03,0.96,0.94,0.53,-0.02,-0.59,0.34,-0.99,-0.81,0.74,0.54,-0.54,-0.54,-0.3,-0.01,0.81,-0.49,0.11,-0.69],\n", - "[-0.76,-0.79,0.17,0.67,0.18,-0.54,-0.21,-0.7,0.69,0.25,0.38,0.02,0.85,-0.11,-0.67,0.04,-0.74,-0.79,0.94,0.67],\n", - "[-0.47,0.87,0.16,0.99,0.18,-0.96,0.63,-0.13,-0.51,0.98,0.02,-0.37,-0.52,-0.75,-0.33,-0.25,-0.77,0.17,0.69,0.22],\n", - "[-0.25,0.73,-0.63,-0.52,-0.35,-0.35,-0.8,0.86,0.07,-0.91,0.67,-0.67,-0.7,0.66,-0.58,-0.9,0.14,0.7,0.6,0.92],\n", - "[-0.05,0.96,-0.65,0.58,1,0.2,0.33,0.6,-0.85,-0.34,-0.12,-0.04,-0.81,0.66,0.78,-0.73,0.99,0.25,-0.6,-0.33],\n", - "[0.03,0.34,0.39,0.63,0.76,-0.49,-0.94,0.7,-0.25,0.07,0.19,0.17,-0.29,-0.84,-0.01,-0.01,1,0.2,-0.36,-0.24],\n", - "[0.54,0.91,0.25,-0.37,0.67,0.28,-0.67,-0.69,0.93,0.66,0.01,-0.57,-0.49,-0.73,-0.78,-0.69,0.71,-0.22,-0.48,-0.99],\n", - "[-0.54,0.46,-0.72,0.67,0.94,-0.05,-0.58,0.08,-0.95,-0.03,0.35,0.67,-0.54,0.56,-0.93,0.02,0.03,0.05,-0.92,0.59],\n", - "[0.33,0.34,-0.72,0.37,0.41,0.86,-0.37,0.63,0.62,0.91,0.25,0.39,0.15,-0.9,-0.82,0.72,0.88,-0.88,0.04,0.35],\n", - "[-0.42,0.16,0.79,0.06,0.01,0.89,-0.61,-0.42,-0.08,-0.86,-0.22,0.68,-0.42,0.27,-0.84,-0.96,0.94,0.37,0,0.98],\n", - "[-0.55,-0.43,-0.65,0.48,0.04,-0.15,-0.68,-0.35,0.82,-0.79,0.86,0.81,0.9,-0.83,-0.35,0.8,-0.63,0.13,-0.12,0.81],\n", - "[0.97,0.4,-0.17,-0.29,-0.27,0.46,-0.73,-0.12,0.89,-0.56,-0.33,-0.95,-0.98,0.32,-0.32,0.91,0.06,0.43,0.44,0.03],\n", - "[-0.45,-0.75,-0.26,-0.15,0.73,-0.47,-0.7,-0.71,-0.58,-0.56,-0.78,-0.74,-0.58,-0.18,-0.28,0.9,0.27,0.05,-0.6,-0.69],\n", - "[-0.61,0.68,0.47,-0.56,-0.2,0.76,-0.58,0.48,-0.21,0.1,0.24,0.48,-0.7,-0.92,0.11,0.53,0.17,0.69,-0.44,-0.66],\n", - "[0.81,0.3,0.93,0.92,-0.76,0.54,-0.08,-0.46,-0.7,-0.25,0.47,0.24,-0.81,0.75,0.79,0.76,-0.74,-0.8,-0.74,0.86],\n", - "[-0.22,-0.04,-0.43,0.85,0.2,0.08,0.25,-0.59,0.31,0.44,0.02,0.56,-0.45,0.4,0.72,0.2,0.8,-0.75,-0.69,0.51],\n", - "[0.24,0.97,0.02,-0.81,-0.57,-0.01,0.13,0.31,0.69,0.83,-0.05,0.78,0.49,-0.42,-0.91,0.04,-0.74,0.53,-1,-0.15]\n", - "]\n", - "\n", - "\n", - "\n", - "131->34\n", - "\n", - "\n", - "14:w14:[20, 20]:[\n", - "[-0.47,0.48,-0.24,0.29,-0.5,-0.41,-0.41,-0.06,0.83,0.5,0.06,0.26,-0.83,-0.03,-0.29,0.88,0.44,-0.77,0.05,-0.03],\n", - "[-0.94,-0.46,0.73,-0.97,-0.27,-0.19,0.84,0.06,0.82,0.02,0.06,0.21,0.65,-0.05,0.2,-0.13,0.33,-0.37,0.67,0.54],\n", - "[-0.68,-0.28,-0.63,-0.22,-0.34,-0.55,0.73,-0.64,0.09,-0.73,-0.36,-0.7,0.3,-0.87,-0.64,-0.25,0.06,0.82,-0.37,-0.47],\n", - "[0,0.81,-0.2,-0.95,0.47,-0.35,0.83,0.68,-0,0.01,-0.07,0.73,-0.09,-0.66,-0.36,0.18,-0.45,-0.85,0.36,-0.8],\n", - "[0.78,-0.98,0.25,-0.9,-0.23,-0.43,0.95,-0.49,0.94,0.85,-0.4,-0.57,0.8,0.74,-0.91,0.87,-0.9,0.61,-0.06,-0.98],\n", - "[-0.98,-0.54,-0.49,0.86,0.68,0.5,-1,-0.45,-0.42,0.7,0.02,0.13,-0.18,0,0.49,-0.55,-0.23,0.12,-0.04,-0.34],\n", - "[-0.62,-0.94,-0.22,0.13,-0.31,0.55,0.62,-0.85,-0.75,0.29,0.59,0.39,0.67,-0.23,-0.45,-0.16,0.23,-0.45,0.5,0.49],\n", - "[0.72,0.75,-0.94,-0.8,0.57,0.53,0.07,-0.44,0.81,0.71,-0.32,-0.18,-0.62,-0.3,0.5,0.99,-0.05,-0.84,-0.13,0.89],\n", - "[-0.77,0.35,-0.56,0.47,0.03,0.2,0.02,-0.35,0.87,0.96,-0.93,0.16,-0.61,0.45,0.7,0.31,-0.33,-0.27,0.92,0.87],\n", - "[0.95,0.66,0.35,0.57,-0.56,-0.65,0.63,0.6,-0.97,-0.54,-0.56,0.41,0.46,-0.97,0.04,0.4,-0.4,0.39,-0.88,0.27],\n", - "[0.24,0.84,0.92,0.77,-0.89,-0.49,0.55,0.29,0.82,0.17,0.66,0.7,-0.35,-0.23,-0.18,0.63,0.37,0.62,-0.79,-0.47],\n", - "[0.59,0.62,-0.31,-0.32,-0.5,0.02,0.61,0.71,-0.64,-0.23,0.09,-0.91,-0.59,0.88,-0.53,0.57,0.74,0.73,-0.43,-0.55],\n", - "[-0.93,0.97,0.7,0.38,0.07,0.05,0.6,0.56,-0.31,0.4,-0.35,-0.4,0.87,0.83,0.06,-0.86,0.57,-0.25,0.36,-0.33],\n", - "[-0.63,-0.56,0.64,0.88,-0.41,0.81,-0.2,-0.62,-0.82,-0.97,-0.42,0.66,-0.43,-0.64,0.74,-0.14,-0.09,-0.17,-0.47,0.92],\n", - "[0.58,0.53,0.39,-0.41,-0.26,0.62,0.75,0.97,0.56,-0.72,0.29,0.2,0.82,0.61,0.74,0.19,-0.75,-0.58,-0.48,0.53],\n", - "[0.59,1,-0.75,0.91,0.81,0.81,-0.11,0.86,0.75,0.42,-0.18,-0.98,-0.81,-0.9,0.11,-0.44,0.05,0.63,0.99,0.99],\n", - "[0.53,0.64,0.47,-0.78,-0.38,-0.5,-0.98,0.35,-0.27,-0.01,0.23,-0.55,0.05,0.89,0.53,-0.62,0.45,-0.4,0.15,0.46],\n", - "[-0.12,-0.7,-0.26,0.05,0.6,0.07,-0.67,-0.47,-0.32,-0.12,-0.26,0.88,-0.23,-0.36,-0.14,-0.48,0.51,0.74,-0.36,0.84],\n", - "[-0.56,0.26,-0.53,0.61,-0.18,-0.81,-0.28,-0.01,-0.47,-0.06,0.51,-0.72,0.61,-0.25,-0.34,0.79,-0.81,-0.42,0.65,0.79],\n", - "[-0.06,0.58,0.99,-0.95,0.28,-0.74,0.6,0.17,-0.4,-0.69,-0.37,-0.69,0.44,0.93,-0.94,-0.97,0.01,-0.35,0.91,0.47]\n", - "]\n", - "\n", - "\n", - "\n", - "134\n", - "\n", - "RuleNeuron:134:lstm2_out__right(T) :- lstm2_out__i(T), lstm2_out__n(T). [transformation=identity, combination=elproduct]\n", - "val: [0,-0,-0,-0,-0,0,-0,-0,0,-0,-0,-0,-0,-0,0,0,0,-0,-0,0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "134->128\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "134->131\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "141\n", - "\n", - "RuleNeuron:141:lstm2_out(T) :- lstm2_out__o(T), tanh(lstm2_out__c(T)). [transformation=identity, combination=elproduct]\n", - "val: [0,-0,-0,-0,-0,0,-0,-0,0,-0,-0,-0,-0,-0,0,0,0,-0,-0,0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "141->116\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "141->134\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "144\n", - "\n", - "WeightedRuleNeuron:144:lstm2_out__o(T) :- {20, 20} lstm1_out(T), {20, 20} lstm2_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.5,0.5,0.49,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.51,0.49,0.5,0.5,0.5,0.5,0.51]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "144->59\n", - "\n", - "\n", - "12:w12:[20, 20]:[\n", - "[-0.47,0.05,0.77,-0.82,0.69,-0.87,0.27,0.97,0.08,0.55,-0.66,-0.91,0.88,-0.82,0.07,-0.47,-0.05,0.09,0.5,-0.11],\n", - "[-0.67,0.38,-0.22,0.1,0.69,-0.92,0.38,0.12,-0.19,-0.69,-0.27,0.07,0.52,-0.33,-0.1,-0.14,-0.19,0.72,0.51,-0.04],\n", - "[-0.47,-0.54,-0.55,0.97,-0.46,-0.39,-0.27,-0.05,-0.2,0.54,-0.67,-0.03,-0.31,-0.83,-0.22,-0.36,0.4,-0.17,0.77,-0.14],\n", - "[0.3,0.87,0.14,-0.3,-0.21,-0.26,-0.62,-0.19,-0.64,-0.08,-0.15,0.46,0.42,0.3,0.8,0.09,0.39,-0.53,0.61,-0.9],\n", - "[-0.83,0.44,-0.56,0.14,0.6,-0.38,-0.71,-0.3,-0.6,0.85,-0.96,0.08,-0.04,-0.24,-0.94,0.5,-0.92,-0.99,0.89,0.68],\n", - "[-0.25,-0.83,-0.81,-0.13,0.12,-0.2,-0.28,-0.9,-0.13,-0.54,-0.42,0.26,0.11,-0.73,-0.33,0.21,0.55,0.89,-0.95,0.54],\n", - "[0.5,0.06,-0.35,0.01,0.72,0.93,0.6,-0.77,-0.2,0.95,-0.55,-0.03,-0.38,0.24,-0.56,-0.25,0.53,0.08,0.88,-0.37],\n", - "[-0.99,-0.54,0.28,-0.63,-0.52,-0.65,0.27,-0.57,-0.49,0.87,0.42,-0.02,-0.73,0.44,-0.42,0.07,-0.61,-0.56,0.26,-0.69],\n", - "[-0.5,-0.86,-0.32,-0.65,0.45,0.19,0.78,-0.29,0.61,-0.82,-0.26,-0.19,0.13,-0.47,0.55,0.14,-0.2,-0.55,-0.57,-0.87],\n", - "[-0.28,0.36,0.91,-0.47,0.92,0.79,0.99,-0.39,0.62,-0.27,0.74,0.12,0.93,-0.93,0.61,-0.42,-0.39,-0.27,0.58,-0.15],\n", - "[-0.36,-0.05,-0.59,0.26,0.08,-0.56,0.47,-0.41,0.42,0.89,0.36,-0.15,-0.92,-0.28,0.33,-0.87,-0.34,0.9,0.35,-0.85],\n", - "[0.35,-0.57,0.13,0.83,-0.49,0.53,-1,-0.14,-0.03,0.15,-0.25,-0.26,0.74,-0.42,0.18,0.62,-0.13,0.75,0.83,0.11],\n", - "[-0.66,0.15,0.8,0.93,0.93,0.01,-0.6,0.31,0.62,-0.65,0.17,-0.83,-0.25,-0.37,0.65,0.71,-0.6,0.45,-0.88,-0.26],\n", - "[-0.06,0.72,-0.4,-0.21,-0.86,-0.44,-0.69,0.51,0.97,-0.71,0.1,0.06,-0.51,-0.09,-0.88,0.68,0.43,0.71,0.86,-0.11],\n", - "[0.23,-0.34,-0.85,-0.67,-0.27,-0.32,-0.11,-0.06,-0.02,0.19,-0.97,-0.54,0.1,0.63,-0.82,-0.74,0.72,0.12,-0.41,-0.94],\n", - "[-0.06,0.22,-0.08,0.77,-0.53,0.71,0.82,-0.98,-0.67,-0.19,-0.94,0.16,-0.88,-0.99,0.83,0.24,-0.18,0.43,0.56,0.93],\n", - "[0.87,-0.53,-0.57,0.84,0.24,-0.66,-0.8,0.67,-0.8,-0.75,0.16,0.91,-0.34,-0.44,-0.88,0.17,0.82,-0.98,-0.42,0.53],\n", - "[-0.65,0.48,0.1,-0.42,-0.48,-0.46,0.83,-0.32,-0.32,0.08,0.7,0.13,-0.52,0.66,0.93,0.68,0.95,-0.44,0.3,0.72],\n", - "[-0.22,0.11,-0.02,0.71,-0.94,0.22,0.16,0.82,-0.22,-0.18,0.87,0.16,-0.01,0.28,0.44,-0.03,-0.87,-0.02,0.45,-0.59],\n", - "[-0.96,0.38,0.45,0.34,0.98,-0.37,-0.19,0.61,-0.29,0.58,0.01,-0.24,-0.73,-0.94,-0.26,0.1,-0.73,0.53,0.14,0.48]\n", - "]\n", - "\n", - "\n", - "\n", - "144->141\n", - "\n", - "\n", - "13:w13:[20, 20]:[\n", - "[-0.39,-0.29,-0.09,-0.68,0.85,0.76,-0.02,0.19,-0.47,-0.1,0.87,0.37,0.15,-0.5,-0.75,-0.45,0.85,-0.88,0.4,-0.55],\n", - "[-0.52,0.24,0.04,-0.3,0.75,0.11,0.31,-0.73,0.4,-0.71,0.29,-0.83,-0.73,0.53,0.31,0.49,0.5,-0.37,-0.28,0.34],\n", - "[-0.92,0.48,-0.14,0.44,0.73,-0.74,-0.39,0.21,-0.09,0.57,-0.19,0.13,-0.26,-0.69,0.18,0.51,0.68,-0.14,-0.37,0.89],\n", - "[0.65,-0.71,-0.06,0.74,0.78,-0.6,-0.98,0.39,-0.26,-0.55,-0.18,0.44,0.28,-0.36,0.91,0.72,-0.57,0.91,-0.7,-0.91],\n", - "[-0.22,0.45,-0.72,-0.93,0.11,0.42,0.62,-0.93,0.6,-0.99,0.14,0.64,0.12,-0.68,-0.97,0.95,0.91,-0.64,-0.02,-0.43],\n", - "[-0.48,0.74,0.24,-0.38,-0.3,-0.16,0.89,0.26,-0.19,-0.57,0.16,-0.43,0.96,-0.86,-0.08,-0,0.39,-0.59,0.87,0.68],\n", - "[-0.8,0.32,0.74,0.25,0.92,0.76,-0.94,0.01,-0.49,0.47,0.1,0.56,-0.83,-0.31,-0.14,-0.17,0.35,0.26,0.29,-0.69],\n", - "[0.74,0.34,-0.74,0.45,-0.67,0.25,-0.93,0.37,-0.79,-0.66,-0.92,0.96,-0.98,-0.92,-0.85,0.95,-0.05,0.56,0.62,-0.37],\n", - "[0.22,-0.64,-0.18,0.48,-0.67,0.28,0.06,-0,0.57,-0.47,-0.15,-0.92,-0.57,0.93,0.69,-0.72,-0.87,0.49,-0.93,0.27],\n", - "[-0.98,0.78,-0.1,0.33,-0.42,0.07,0.89,-0.52,0.06,0.81,-0.65,0.65,-0.22,-0.63,0.18,0.18,-0.79,0.95,0.88,0.73],\n", - "[-0.54,0.96,-0.3,-0,0.8,0.07,-0.26,0.16,0.54,-0.02,0.53,-0.27,-0.26,0.81,0.87,0.9,0.99,-0.95,0.67,0.63],\n", - "[-0.25,0.73,0.42,-0.96,-0.21,-0.88,0.36,-0.14,0.44,-0.25,-0.45,-0.48,0.7,0.24,-0.87,0.43,-0.5,0.49,-0.27,0.08],\n", - "[-0.37,-0.89,0.2,0.63,0.81,0.4,-0.74,-0.06,0.14,0.7,0.36,0.19,-0.57,-0.74,-0.47,0.48,0.96,-0.36,-0.27,-0.81],\n", - "[-0.6,0.03,-0.3,0.4,-0.97,-0.32,0.65,-0.02,0.34,-0.42,-0.87,-0.3,0.94,-0.08,-0.8,-0.9,-0.24,0.3,-0.59,-0.18],\n", - "[-0.68,-0.87,-0.67,0.5,0.06,0.05,0.42,-0.55,-0.11,-0.08,0.36,-0.03,0.07,0.55,0.91,0.18,-0.93,0.93,0.77,-0.78],\n", - "[-0.74,0.64,-0.62,0.68,0.47,-0.1,-0.14,0.38,-0.66,-0.55,0.67,0.44,-0.67,0.84,-0.54,0.4,0.39,-1,0.78,0.75],\n", - "[0.35,-0.97,0.16,-0.8,-0.67,0.17,0.82,-0.91,0.75,0.17,-0.92,-0.09,0.75,0.68,-0.88,-0.49,-0.82,0.34,1,0.16],\n", - "[-0.98,-0.07,-0.36,0.17,-0.8,0.14,0.53,-0.21,0.66,0.31,0.84,-0.38,-0.98,0.72,0.32,0.77,0.79,-0.98,-0.23,-0.01],\n", - "[-0.8,-0.12,0.01,-0.2,-0.16,0.72,0.87,-0.2,-0.96,-0.79,-0.27,0.53,-0.84,0.31,-0.83,-0.7,0.27,-0.79,0.11,0.59],\n", - "[0.81,-0.99,0.27,0.25,-0.23,0.49,-0.56,-0.47,-0.83,0.11,-0.11,0.08,-0.87,-0.97,0.92,-0.13,0.55,0.55,0.32,-0.91]\n", - "]\n", - "\n", - "\n", - "\n", - "147\n", - "\n", - "WeightedRuleNeuron:147:lstm2_out__f(T) :- {20, 20} lstm1_out(T), {20, 20} lstm2_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "147->59\n", - "\n", - "\n", - "10:w10:[20, 20]:[\n", - "[-0.75,-0,0.21,-0.16,0.61,0.09,0.59,-0.02,-0.51,-0.54,0.27,-0.79,0.91,-0.16,-0.75,-0.4,0.96,-0.84,0.12,-0.66],\n", - "[0.93,-0.85,-0.91,-0.17,-0.1,0.2,-0.86,-0.26,-0.12,-0.1,0.35,-0.69,-0.83,0.09,-0.04,-0.45,-0.45,-0.67,-0.84,0.99],\n", - "[0.13,0.6,0.13,-0.09,-0.54,0.13,0.43,0.39,-0.07,0.09,-0.47,-0.35,0.33,0.52,0.95,0.83,0.16,0.43,0.38,0.67],\n", - "[0.02,0.82,-0.23,0.17,-0.66,-0.56,-0.46,-0.11,0.75,-0.51,-0.13,0.29,0.12,-0.01,-0.9,-0.85,0.71,0.19,-0.62,0.94],\n", - "[0.72,-0.49,-0.12,0.99,-0.2,-0.11,-0.91,-0,0.5,0.26,0.89,0.62,-0.38,0.47,-0.26,0.86,0.13,-0.96,0.32,0.99],\n", - "[0.55,0.81,0.77,0.08,-0.71,-0.38,0.12,0.35,0.96,-0.96,0.86,0.83,-0.66,-0.36,0.77,0.24,0.33,0.69,-0.83,0.43],\n", - "[0.96,0.3,-0.72,0.15,0.47,-0.52,0.51,0.47,-0.89,-0.53,0.01,-0.12,-0.47,-0.3,0.33,0.67,0.37,-0.4,0.37,0.89],\n", - "[-0.05,-0.46,-0.24,0.2,-0.06,0.14,-0.03,0.54,0.11,-0.37,-0.34,0.53,0.5,0.7,0.29,-0.31,-0.14,-0.94,0.81,0.36],\n", - "[-0.91,-0.23,0.46,0.38,0.6,0.1,0.26,-0.04,-0.85,0.2,-0.79,-0.92,-0.88,0.9,0.01,0.36,-0.9,-0.75,-0.51,-0.54],\n", - "[0.9,0.37,0.93,-0.33,-0.59,-0.6,0.99,0.83,-0.97,-0.29,-0.83,-0.99,0.61,0.75,-0.02,-0.57,-0.44,-0.56,0.98,-0.64],\n", - "[-0.87,0.22,0.59,-0.3,0.28,-0.27,0.12,0.8,0.58,-0.66,0.12,-0.44,0.82,-0.29,0.83,0.41,0.17,0.85,-0.54,0.02],\n", - "[0.15,0.52,-0.44,-0.67,0.94,0.89,0.12,0.03,0.19,-0.99,-0.65,0.23,0.62,0.05,-0.01,-0.03,0.8,-0.42,-0.87,0.77],\n", - "[0.58,-0.94,0.5,-0.4,-0.28,0.17,0.87,-0.85,0.85,-0.77,-0.06,0.86,-0.26,0.99,-0.27,0.44,-0.32,0.63,0.95,0.57],\n", - "[-0.15,0.89,-0.87,0.84,0.44,-0.9,-0.16,0.06,-0.64,0.94,-0.87,-0.62,-0.47,-0.13,0.61,0.08,0.96,-0.79,0.77,-0.63],\n", - "[-0.97,0.15,0.5,0.86,0.67,-0.75,-0.16,0.06,0.67,-0.75,0.91,0.71,-0.48,-0.67,-0.2,-0.22,0.02,-0.26,-0.58,0.1],\n", - "[0.12,-0.19,0.22,0.12,0.36,-0.8,-0.11,-0.57,-0.19,0.88,-0.39,-0.73,0.16,0.9,-0.74,0.65,0.58,0.93,-0.19,-0.78],\n", - "[0.58,0.2,-0.94,-0.2,0.42,0.7,-0.08,-0.09,0.05,-0.15,-0.68,-0.33,-0.57,0.71,0.01,-0.07,-0.75,-0.01,-0.26,0.61],\n", - "[0.21,-0.04,-0.6,-0.02,0.48,-0.83,-0.99,-0.24,0.83,0.33,0.94,-0.32,-0.08,0.97,-0.08,-0.3,-0.95,-0.14,-0.98,0.82],\n", - "[-0.34,-0.28,0.34,-0.18,-0.96,0.24,-0.4,0.8,0.88,0.84,0.61,0.24,0.13,0.63,-0.76,0.27,0.38,0.23,0.41,0.87],\n", - "[-0.01,0.32,0.24,-0.35,0.54,-0.6,0.73,0.2,0.83,-0.89,-0.1,-0.28,-0,-0.44,0.91,-0.45,-0.18,-0.94,0.07,-0.21]\n", - "]\n", - "\n", - "\n", - "\n", - "147->141\n", - "\n", - "\n", - "11:w11:[20, 20]:[\n", - "[0.27,0.56,-0.52,0.25,-0.59,-0.36,-0.1,-0.31,0.33,0.71,-0.34,0.55,-0.23,0.54,-0.88,0.19,-0.37,0.54,-0.74,0.88],\n", - "[0.11,-0.46,-0.42,0.95,0.43,-0.54,-0.94,0.82,-0.05,0.5,-0.54,-0.38,0.94,-0.73,0.01,-0.05,0.07,-0.09,0.94,0.31],\n", - "[-0.76,0.76,0.67,0.55,-0.53,0.22,0.62,-0.03,0.28,0,0.15,-0.54,0.79,0.98,0.57,-0.14,-0.86,-0.18,0.75,0.07],\n", - "[-0.41,-0.49,0.28,-0.26,-0.21,0.05,0.98,0.71,-0.82,0.8,0.33,0.88,-0.63,0.83,-0.24,-0.5,0.98,-0.55,-0.81,0.58],\n", - "[0.48,-0.18,-0.37,0.74,-0.61,0.43,0.81,-0.74,0.32,0.42,0.92,0.2,0.17,0.26,-0.55,-0.14,0.28,0.95,-0.57,-0.91],\n", - "[0.12,0.63,0.54,-0.48,-0.27,0.84,0.78,-0.83,0.15,0.97,0.88,0.38,0.62,0.74,0.89,-0.08,0.64,0.89,-0.03,0.39],\n", - "[-0.24,-0.22,0.78,-0.93,0.02,0.14,0.83,0.06,0.25,-0.33,0.14,0.41,0.46,-0.04,-0.83,0.94,-0.06,-0.06,0.83,-0.73],\n", - "[0.84,0.45,0.1,0.8,-0.46,-0.84,-0.14,0.51,-0.44,0.84,0.34,0.33,-0.97,-0.79,-0.14,0.52,0.6,0.3,0.19,-0.68],\n", - "[0.15,0.57,-0.26,0.95,0.31,-0.82,0.04,0.09,0.67,0.69,0.43,0.64,-0.7,-0.71,0.83,-0.86,-0.98,0.36,-0.54,0.68],\n", - "[0.01,0.14,0.2,-0.04,0.09,-0.76,0.32,-0.14,-0.79,0.69,0.7,0.6,-0.16,0.37,-0.81,0.51,-0.92,0.68,0.85,0.35],\n", - "[0.49,-0.17,0.16,-0.94,-0.82,-0.14,0.43,-0.08,0.81,0.44,0.49,0.17,-0.18,0.37,-0.33,0.35,0.47,0.59,-0.03,-0.14],\n", - "[0.19,-0.42,0.9,0.4,-0.15,0.13,0.9,-0.41,0.88,-0.48,-0.77,0.27,-0.72,-0.54,-0.13,-0.55,0.94,-0.66,-0.24,0.63],\n", - "[0.15,0.63,0.84,-0.37,0.31,-0.15,-0.95,-0.21,0.89,-0.33,-0.38,-0.7,0.63,-0.4,-0.8,0.82,-0.87,-0.06,0.22,0.98],\n", - "[0.3,0.58,-0.63,0.7,0.11,0.78,0.33,0.66,-0.15,-0.04,0.02,0.13,-0.22,0.14,-0.65,0.53,0.44,0.95,-0.28,-0.85],\n", - "[-0.6,0.53,-0.94,-0.1,0.58,0.28,0.89,0.15,-0.66,-0.03,-0.46,-0.08,0.77,0.22,-0.82,0.51,-0.31,0.3,-0.55,0.68],\n", - "[0.39,-0.21,-0.06,-0.61,0.45,-0.5,0.08,-0.48,-0.68,0.05,-0.51,-0.17,0.03,-0.68,-0.57,-0.11,-0.29,-0.94,-0.43,-0.12],\n", - "[-0.41,0.07,0.79,0.69,-0.62,0.43,-0.17,0.78,-0.84,0.57,0.86,0.46,-0.74,0.04,-0.48,0.84,-0.03,0.39,-0.9,-0.69],\n", - "[-0.65,-0.07,0.43,0.25,-0.08,-0.69,0.21,-0.04,0.6,0.87,-0.81,0.34,-0.26,-0.13,0.53,-0.24,0.22,-0.49,-0.11,-0.53],\n", - "[-0.42,-0.6,-0.8,0.08,0.81,0.83,0.26,0.18,0.68,0.05,-0.18,0.04,-0.25,0.69,-0.72,0.73,-1,0.27,0.56,-0.19],\n", - "[0.35,0.1,0.5,0.69,0.16,-0.62,-0.55,-0.59,0.19,-0.79,-0.03,0.64,-0.24,-0.42,-0.31,-0.43,0.99,-0.51,-0.08,0.7]\n", - "]\n", - "\n", - "\n", - "\n", - "150\n", - "\n", - "RuleNeuron:150:lstm2_out__left(T) :- lstm2_out__f(T), lstm2_out__c(Z), @next(Z, T). [transformation=identity, combination=elproduct]\n", - "val: [0,-0,-0,-0,-0,0,-0,-0,0,-0,-0,-0,-0,-0,0,0,0,-0,-0,0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "150->134\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "150->147\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "153\n", - "\n", - "WeightedRuleNeuron:153:lstm2_out__i(T) :- {20, 20} lstm1_out(T), {20, 20} lstm2_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.5,0.51,0.51,0.5,0.5,0.5,0.5,0.5,0.5,0.49,0.5,0.5,0.5,0.5,0.5,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "153->59\n", - "\n", - "\n", - "8:w8:[20, 20]:[\n", - "[0.27,0.6,-0.82,0.69,-0.56,0.92,-0.02,0.9,0.98,0.9,0.49,-0.39,0.84,0.46,0.35,-0.67,0.41,-0.22,0.58,-0.26],\n", - "[-0.46,-0.96,-0.7,0.85,-0.16,-0.03,0.96,-0.02,-0.97,-0.51,0.4,0.14,0.43,0.28,0.67,-0.27,-0.88,0.3,-0.92,-0.75],\n", - "[0.89,-0.26,0.12,-0.45,0.42,-0.23,0.97,-0.29,-0.87,-0.7,0.58,-0.23,0.5,0.48,-0.76,-0.73,-0.41,-0.44,0.26,-0.46],\n", - "[0.19,0.29,0.3,-0.02,-0.97,0.99,0.62,-0.81,0.41,-0.77,-0.21,0.66,-0.61,0.2,-0.97,-0.77,-0.95,-0.48,0.96,-0.56],\n", - "[-0.61,-0.21,0.15,-0.07,-0.72,0.83,0.47,0.47,0.78,0.17,-0.6,-0.12,-0.26,-0.67,0.78,-0.54,-0.42,0.01,-0.15,0.29],\n", - "[-0.86,-0.58,-0.92,-0.78,-0.66,0.71,0.64,0.58,0.57,-0.21,0.95,0.7,-0.1,0.5,-0.36,-0.99,-0.84,-0.31,-0.34,-0.21],\n", - "[-0.48,0.52,-0.74,0.82,-0.76,-0.02,0.41,0.52,0.79,-0.39,-0.98,0.69,-0.14,-0,-0.67,-0.3,-0.55,-0.91,-0.97,0.94],\n", - "[-0.36,-0.32,-0.49,0.61,-0.67,-0.73,0.89,-0.58,0.88,0.97,0.98,0.13,-0.47,-0.13,0.99,-0.67,0.89,-0.93,-0.12,-0.48],\n", - "[0.2,-0.66,-0.01,-0.08,-0.54,0.47,0.65,0.41,-0.07,-0.01,0.39,-0.46,-0.54,-0.69,-0.38,-0.65,0.05,-0.91,0.74,0.79],\n", - "[0.48,0.09,0.71,-0.87,0.02,-0.23,-0.26,0.17,-0.54,-0.85,0.7,-0.45,0.14,0.91,-0.39,-0.76,-0.79,0.07,-0.23,0.75],\n", - "[0.61,0.89,0.85,0.52,-0.01,0.38,-0.34,0.67,0.46,-0.78,-0.92,0.08,-0.37,-0.38,0.36,0.88,0.12,0.24,0.81,-0.8],\n", - "[0.74,-0.7,0.5,-0.7,0.23,-0.46,-0.03,0.01,-0.61,-0.27,-0.62,-0.95,-0.4,0.53,-0.97,0.41,0.62,-0.84,0.91,0.98],\n", - "[-0.29,0.78,-0.58,0.26,0.45,-0.58,0.28,0.08,-0.29,0.5,0.31,-0.86,0.55,0.41,0.7,0.29,0.46,-0.24,-0.74,0.56],\n", - "[0.31,0.44,0.02,-0.96,0.43,-0.08,0.86,0.14,-0.57,0.07,-0.48,0.91,0.97,0.84,-0.06,-0.6,0.86,-0.65,-0.23,0.75],\n", - "[0.86,0.22,0.36,-0.23,0.14,0.93,0.8,0.78,0.65,-0.18,-0.99,-0.39,-0.07,0.18,0.4,0.83,-0.31,-0.18,-0.95,-0.43],\n", - "[0.83,-0.29,0.99,-0.63,-0.77,-0.98,-0.99,0.3,0.38,0.38,-0.31,-0.59,-0.28,-0.79,0.75,0.75,0.16,-0.79,-0.94,-0.01],\n", - "[0.51,0.49,-0.06,0.07,-0.07,-0.67,0.31,0.29,-0.84,0.66,-0.99,0.44,-0.17,-0.65,-0.81,0.13,-0.46,0.56,-0.75,0.61],\n", - "[0.5,0.25,-0.43,0.79,0.88,-0.35,-0.88,0.8,-0.87,0.22,-0.2,0.98,0.12,0.41,0.25,-0.49,0.3,0.55,-0.08,-0.96],\n", - "[0.49,-0.97,-0.72,-0.09,0.69,0,-0.2,0.88,-0.23,-0.02,0.12,-0.5,-0.53,-0.34,-0.78,-0.02,0.31,0.81,-0.37,-0.48],\n", - "[0.36,0.89,-0.91,-0.05,0.98,-0.04,0.38,0.96,0.39,-0.75,-0.28,0.83,0.29,0.69,0.43,0.57,0.17,-0.05,0.34,0.76]\n", - "]\n", - "\n", - "\n", - "\n", - "153->141\n", - "\n", - "\n", - "9:w9:[20, 20]:[\n", - "[-0.08,-0.9,0.89,0.73,-0.35,0.92,0.86,-0.84,0.7,0.28,0.12,-0.41,-0.54,-0.13,-0.99,0.04,-0.51,-0.42,0.34,0.32],\n", - "[0.54,0.4,0.8,0.55,-0.18,-0,0.51,0.89,-0.81,0.76,0.63,0.49,-0.18,-0.31,0.26,0.31,-0.66,-0.92,-0.29,0.77],\n", - "[-0.6,-0.32,-0.17,-0.44,0.25,-0.21,0.35,-0.35,-0.98,0.63,-0.47,0.24,0.1,0.06,-0.22,-0.38,-0.89,-0.12,0.95,0.29],\n", - "[-0.94,-0.15,-0.23,0.86,-0.52,0.54,0.38,0.62,-0.87,-0.38,-0.91,-0.8,0.78,0.87,-0.5,-0.94,-0.79,0.09,-0.08,-0.74],\n", - "[0.39,0.72,0.98,0.48,0.07,-0.84,0.12,-0.07,-0.4,0.32,-0.99,0.58,0.22,-0.53,-0.34,-0.14,-0.22,-0.94,-0.61,-0.97],\n", - "[0.66,0.57,0.08,0.97,-0.86,0.7,0.12,0.59,0.49,0.25,0.2,0.74,0.87,-0.4,0.2,0.81,0.11,0.02,0,0.33],\n", - "[-0.7,-0.66,0.69,-0,-0.64,0.31,0.95,0.21,-0.21,-0.61,0.49,-0.95,0.05,-0.76,-0.42,-0.55,0.1,0.22,0.05,-0.23],\n", - "[-0.71,-0.38,-0.37,-0.83,-0.24,-0.56,0.47,0.38,0.2,-0.8,0.71,-0.51,0.41,0.43,0.14,-0.29,-0.8,-0.54,-0.43,-0.57],\n", - "[0.53,0.97,-0.53,-0.09,-0.36,-0.71,0.94,0.02,0.48,0.37,0.73,0.05,-0.11,-0.3,-0.66,-0.85,0.34,-0.27,0.6,-0.44],\n", - "[0.67,0.89,-0.53,-0.59,-0.54,0.56,-0.39,0.81,0.67,0.98,-0.36,0.39,-0.81,-0.35,0.77,0.75,0.37,0.44,-0.91,-0.25],\n", - "[0.71,0.25,0.81,0.02,0.05,0.87,0.37,-0.39,0.45,0.58,-0.7,0.38,0.78,0.89,0.78,-0.67,0.01,0.64,-0.15,-0.2],\n", - "[-0.39,0.92,-0.95,0.4,-0.03,-0.71,0.02,-0.81,0.62,0.64,0.12,-0.35,-0.29,-0.18,0.89,-0.87,-0.02,-0.26,-0.58,-0.86],\n", - "[-0.54,0.62,-0.1,-0.64,-0.18,0.94,0.36,0.15,-0.3,0.45,-0.94,0.62,-0.03,0.89,-0.63,0.79,-0.69,-0.15,-0.89,0.5],\n", - "[-0.51,0.69,-0.59,-0.7,0.29,-0.92,0.62,-0.07,-0.18,0.88,-0.92,-0.23,-0.21,-0.71,-0.3,0.91,0.67,-0.31,-0.58,-0.58],\n", - "[-0.02,0.19,0.29,-0.86,-0.45,0.02,0.4,0.9,-0.37,-0.01,-0.3,-0.8,0.66,-0.6,0.23,0.03,-0.95,-0.12,0.43,0.85],\n", - "[-0.46,-0.83,0.56,-0.66,0.96,-0.27,-0.23,-0.81,-0.08,0.05,-0.44,0.73,0.17,0.37,-0.62,0.15,0.14,-0.43,-0.91,0.73],\n", - "[-0.34,0.16,-0.68,-0.82,0.74,-0.72,0.88,0.68,0.74,0.06,-0.33,-0.57,-0.54,0.49,0.96,0.27,0.08,0.11,-0.79,0.49],\n", - "[0.39,-0.83,0.16,-0.33,0.2,0.35,-0.64,0.9,0.3,-0.09,-0.92,-0.05,-0.84,0.18,-0.13,0.37,0.05,0.7,0.26,-0.28],\n", - "[-0.8,0.19,-0.7,0.1,-0.83,-0.92,0.95,0.86,0.89,-0.56,-0.02,0.14,0.26,0.76,0.18,0.29,-0.34,-0.75,-0.58,-0.95],\n", - "[0.51,0.09,0.65,-0.51,0.21,-0.62,0.69,-0.19,0.02,0.49,0,-0.01,0.71,0.83,-0.94,0.39,0.98,-0.7,-0.95,-0.82]\n", - "]\n", - "\n", - "\n", - "\n", - "156\n", - "\n", - "WeightedRuleNeuron:156:lstm2_out__n(T) :- {20, 20} lstm1_out(T), {20, 20} lstm2_out(Z), @next(Z, T). [transformation=tanh]\n", - "val: [0.01,-0.02,0,0,-0.02,-0.01,-0.03,-0,0.01,-0.01,-0.02,0,-0.02,-0.01,0.01,0.01,0.01,-0.02,-0.02,0.01]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Tanh\n", - "\n", - "\n", - "\n", - "156->59\n", - "\n", - "\n", - "14:w14:[20, 20]:[\n", - "[-0.47,0.48,-0.24,0.29,-0.5,-0.41,-0.41,-0.06,0.83,0.5,0.06,0.26,-0.83,-0.03,-0.29,0.88,0.44,-0.77,0.05,-0.03],\n", - "[-0.94,-0.46,0.73,-0.97,-0.27,-0.19,0.84,0.06,0.82,0.02,0.06,0.21,0.65,-0.05,0.2,-0.13,0.33,-0.37,0.67,0.54],\n", - "[-0.68,-0.28,-0.63,-0.22,-0.34,-0.55,0.73,-0.64,0.09,-0.73,-0.36,-0.7,0.3,-0.87,-0.64,-0.25,0.06,0.82,-0.37,-0.47],\n", - "[0,0.81,-0.2,-0.95,0.47,-0.35,0.83,0.68,-0,0.01,-0.07,0.73,-0.09,-0.66,-0.36,0.18,-0.45,-0.85,0.36,-0.8],\n", - "[0.78,-0.98,0.25,-0.9,-0.23,-0.43,0.95,-0.49,0.94,0.85,-0.4,-0.57,0.8,0.74,-0.91,0.87,-0.9,0.61,-0.06,-0.98],\n", - "[-0.98,-0.54,-0.49,0.86,0.68,0.5,-1,-0.45,-0.42,0.7,0.02,0.13,-0.18,0,0.49,-0.55,-0.23,0.12,-0.04,-0.34],\n", - "[-0.62,-0.94,-0.22,0.13,-0.31,0.55,0.62,-0.85,-0.75,0.29,0.59,0.39,0.67,-0.23,-0.45,-0.16,0.23,-0.45,0.5,0.49],\n", - "[0.72,0.75,-0.94,-0.8,0.57,0.53,0.07,-0.44,0.81,0.71,-0.32,-0.18,-0.62,-0.3,0.5,0.99,-0.05,-0.84,-0.13,0.89],\n", - "[-0.77,0.35,-0.56,0.47,0.03,0.2,0.02,-0.35,0.87,0.96,-0.93,0.16,-0.61,0.45,0.7,0.31,-0.33,-0.27,0.92,0.87],\n", - "[0.95,0.66,0.35,0.57,-0.56,-0.65,0.63,0.6,-0.97,-0.54,-0.56,0.41,0.46,-0.97,0.04,0.4,-0.4,0.39,-0.88,0.27],\n", - "[0.24,0.84,0.92,0.77,-0.89,-0.49,0.55,0.29,0.82,0.17,0.66,0.7,-0.35,-0.23,-0.18,0.63,0.37,0.62,-0.79,-0.47],\n", - "[0.59,0.62,-0.31,-0.32,-0.5,0.02,0.61,0.71,-0.64,-0.23,0.09,-0.91,-0.59,0.88,-0.53,0.57,0.74,0.73,-0.43,-0.55],\n", - "[-0.93,0.97,0.7,0.38,0.07,0.05,0.6,0.56,-0.31,0.4,-0.35,-0.4,0.87,0.83,0.06,-0.86,0.57,-0.25,0.36,-0.33],\n", - "[-0.63,-0.56,0.64,0.88,-0.41,0.81,-0.2,-0.62,-0.82,-0.97,-0.42,0.66,-0.43,-0.64,0.74,-0.14,-0.09,-0.17,-0.47,0.92],\n", - "[0.58,0.53,0.39,-0.41,-0.26,0.62,0.75,0.97,0.56,-0.72,0.29,0.2,0.82,0.61,0.74,0.19,-0.75,-0.58,-0.48,0.53],\n", - "[0.59,1,-0.75,0.91,0.81,0.81,-0.11,0.86,0.75,0.42,-0.18,-0.98,-0.81,-0.9,0.11,-0.44,0.05,0.63,0.99,0.99],\n", - "[0.53,0.64,0.47,-0.78,-0.38,-0.5,-0.98,0.35,-0.27,-0.01,0.23,-0.55,0.05,0.89,0.53,-0.62,0.45,-0.4,0.15,0.46],\n", - "[-0.12,-0.7,-0.26,0.05,0.6,0.07,-0.67,-0.47,-0.32,-0.12,-0.26,0.88,-0.23,-0.36,-0.14,-0.48,0.51,0.74,-0.36,0.84],\n", - "[-0.56,0.26,-0.53,0.61,-0.18,-0.81,-0.28,-0.01,-0.47,-0.06,0.51,-0.72,0.61,-0.25,-0.34,0.79,-0.81,-0.42,0.65,0.79],\n", - "[-0.06,0.58,0.99,-0.95,0.28,-0.74,0.6,0.17,-0.4,-0.69,-0.37,-0.69,0.44,0.93,-0.94,-0.97,0.01,-0.35,0.91,0.47]\n", - "]\n", - "\n", - "\n", - "\n", - "156->141\n", - "\n", - "\n", - "15:w15:[20, 20]:[\n", - "[-0.74,-0.47,-0.56,0.04,-0.88,-1,-0.26,-0.18,-0.48,0.63,-0.6,-0.4,0.54,-0.6,-0.59,-0.04,-0.18,-0.51,0.79,-0.63],\n", - "[-0.53,0.89,0.75,-0.08,-0.36,0.6,0.12,0.29,-0.22,0.79,-0.06,-0.07,0.31,-0.3,0.02,0.76,-0.21,0.93,-0.06,0.17],\n", - "[-0.26,0.3,0.81,-0.51,-0.96,-0.77,0.67,-0.11,-0.63,-0.37,-0.17,-0.17,0.75,0.81,-0.14,0.37,0.4,-0.74,0.43,-0.95],\n", - "[-0.91,-0.03,0.96,0.94,0.53,-0.02,-0.59,0.34,-0.99,-0.81,0.74,0.54,-0.54,-0.54,-0.3,-0.01,0.81,-0.49,0.11,-0.69],\n", - "[-0.76,-0.79,0.17,0.67,0.18,-0.54,-0.21,-0.7,0.69,0.25,0.38,0.02,0.85,-0.11,-0.67,0.04,-0.74,-0.79,0.94,0.67],\n", - "[-0.47,0.87,0.16,0.99,0.18,-0.96,0.63,-0.13,-0.51,0.98,0.02,-0.37,-0.52,-0.75,-0.33,-0.25,-0.77,0.17,0.69,0.22],\n", - "[-0.25,0.73,-0.63,-0.52,-0.35,-0.35,-0.8,0.86,0.07,-0.91,0.67,-0.67,-0.7,0.66,-0.58,-0.9,0.14,0.7,0.6,0.92],\n", - "[-0.05,0.96,-0.65,0.58,1,0.2,0.33,0.6,-0.85,-0.34,-0.12,-0.04,-0.81,0.66,0.78,-0.73,0.99,0.25,-0.6,-0.33],\n", - "[0.03,0.34,0.39,0.63,0.76,-0.49,-0.94,0.7,-0.25,0.07,0.19,0.17,-0.29,-0.84,-0.01,-0.01,1,0.2,-0.36,-0.24],\n", - "[0.54,0.91,0.25,-0.37,0.67,0.28,-0.67,-0.69,0.93,0.66,0.01,-0.57,-0.49,-0.73,-0.78,-0.69,0.71,-0.22,-0.48,-0.99],\n", - "[-0.54,0.46,-0.72,0.67,0.94,-0.05,-0.58,0.08,-0.95,-0.03,0.35,0.67,-0.54,0.56,-0.93,0.02,0.03,0.05,-0.92,0.59],\n", - "[0.33,0.34,-0.72,0.37,0.41,0.86,-0.37,0.63,0.62,0.91,0.25,0.39,0.15,-0.9,-0.82,0.72,0.88,-0.88,0.04,0.35],\n", - "[-0.42,0.16,0.79,0.06,0.01,0.89,-0.61,-0.42,-0.08,-0.86,-0.22,0.68,-0.42,0.27,-0.84,-0.96,0.94,0.37,0,0.98],\n", - "[-0.55,-0.43,-0.65,0.48,0.04,-0.15,-0.68,-0.35,0.82,-0.79,0.86,0.81,0.9,-0.83,-0.35,0.8,-0.63,0.13,-0.12,0.81],\n", - "[0.97,0.4,-0.17,-0.29,-0.27,0.46,-0.73,-0.12,0.89,-0.56,-0.33,-0.95,-0.98,0.32,-0.32,0.91,0.06,0.43,0.44,0.03],\n", - "[-0.45,-0.75,-0.26,-0.15,0.73,-0.47,-0.7,-0.71,-0.58,-0.56,-0.78,-0.74,-0.58,-0.18,-0.28,0.9,0.27,0.05,-0.6,-0.69],\n", - "[-0.61,0.68,0.47,-0.56,-0.2,0.76,-0.58,0.48,-0.21,0.1,0.24,0.48,-0.7,-0.92,0.11,0.53,0.17,0.69,-0.44,-0.66],\n", - "[0.81,0.3,0.93,0.92,-0.76,0.54,-0.08,-0.46,-0.7,-0.25,0.47,0.24,-0.81,0.75,0.79,0.76,-0.74,-0.8,-0.74,0.86],\n", - "[-0.22,-0.04,-0.43,0.85,0.2,0.08,0.25,-0.59,0.31,0.44,0.02,0.56,-0.45,0.4,0.72,0.2,0.8,-0.75,-0.69,0.51],\n", - "[0.24,0.97,0.02,-0.81,-0.57,-0.01,0.13,0.31,0.69,0.83,-0.05,0.78,0.49,-0.42,-0.91,0.04,-0.74,0.53,-1,-0.15]\n", - "]\n", - "\n", - "\n", - "\n", - "159\n", - "\n", - "RuleNeuron:159:lstm2_out__right(T) :- lstm2_out__i(T), lstm2_out__n(T). [transformation=identity, combination=elproduct]\n", - "val: [0.01,-0.01,0,0,-0.01,-0,-0.01,-0,0,-0,-0.01,0,-0.01,-0,0.01,0,0,-0.01,-0.01,0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "159->153\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "159->156\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "162\n", - "\n", - "RuleNeuron:162:lstm2_out__c(T) :- lstm2_out__left(T), lstm2_out__right(T). [transformation=identity]\n", - "val: [0.01,-0.01,0,-0,-0.01,-0,-0.02,-0,0,-0,-0.01,0,-0.01,-0,0.01,0,0,-0.01,-0.01,0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum\n", - "\n", - "\n", - "\n", - "162->150\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "162->159\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "166\n", - "\n", - "RuleNeuron:166:lstm2_out(T) :- lstm2_out__o(T), tanh(lstm2_out__c(T)). [transformation=identity, combination=elproduct]\n", - "val: [0,-0,0,-0,-0,-0,-0.01,-0,0,-0,-0.01,0,-0.01,-0,0,0,0,-0.01,-0,0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "166->144\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "166->162\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "169\n", - "\n", - "WeightedRuleNeuron:169:lstm2_out__o(T) :- {20, 20} lstm1_out(T), {20, 20} lstm2_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.49,0.5,0.5,0.5,0.51,0.48,0.51,0.5,0.5,0.51,0.5,0.5,0.51,0.51,0.48,0.5,0.51,0.5,0.49,0.51]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "169->84\n", - "\n", - "\n", - "12:w12:[20, 20]:[\n", - "[-0.47,0.05,0.77,-0.82,0.69,-0.87,0.27,0.97,0.08,0.55,-0.66,-0.91,0.88,-0.82,0.07,-0.47,-0.05,0.09,0.5,-0.11],\n", - "[-0.67,0.38,-0.22,0.1,0.69,-0.92,0.38,0.12,-0.19,-0.69,-0.27,0.07,0.52,-0.33,-0.1,-0.14,-0.19,0.72,0.51,-0.04],\n", - "[-0.47,-0.54,-0.55,0.97,-0.46,-0.39,-0.27,-0.05,-0.2,0.54,-0.67,-0.03,-0.31,-0.83,-0.22,-0.36,0.4,-0.17,0.77,-0.14],\n", - "[0.3,0.87,0.14,-0.3,-0.21,-0.26,-0.62,-0.19,-0.64,-0.08,-0.15,0.46,0.42,0.3,0.8,0.09,0.39,-0.53,0.61,-0.9],\n", - "[-0.83,0.44,-0.56,0.14,0.6,-0.38,-0.71,-0.3,-0.6,0.85,-0.96,0.08,-0.04,-0.24,-0.94,0.5,-0.92,-0.99,0.89,0.68],\n", - "[-0.25,-0.83,-0.81,-0.13,0.12,-0.2,-0.28,-0.9,-0.13,-0.54,-0.42,0.26,0.11,-0.73,-0.33,0.21,0.55,0.89,-0.95,0.54],\n", - "[0.5,0.06,-0.35,0.01,0.72,0.93,0.6,-0.77,-0.2,0.95,-0.55,-0.03,-0.38,0.24,-0.56,-0.25,0.53,0.08,0.88,-0.37],\n", - "[-0.99,-0.54,0.28,-0.63,-0.52,-0.65,0.27,-0.57,-0.49,0.87,0.42,-0.02,-0.73,0.44,-0.42,0.07,-0.61,-0.56,0.26,-0.69],\n", - "[-0.5,-0.86,-0.32,-0.65,0.45,0.19,0.78,-0.29,0.61,-0.82,-0.26,-0.19,0.13,-0.47,0.55,0.14,-0.2,-0.55,-0.57,-0.87],\n", - "[-0.28,0.36,0.91,-0.47,0.92,0.79,0.99,-0.39,0.62,-0.27,0.74,0.12,0.93,-0.93,0.61,-0.42,-0.39,-0.27,0.58,-0.15],\n", - "[-0.36,-0.05,-0.59,0.26,0.08,-0.56,0.47,-0.41,0.42,0.89,0.36,-0.15,-0.92,-0.28,0.33,-0.87,-0.34,0.9,0.35,-0.85],\n", - "[0.35,-0.57,0.13,0.83,-0.49,0.53,-1,-0.14,-0.03,0.15,-0.25,-0.26,0.74,-0.42,0.18,0.62,-0.13,0.75,0.83,0.11],\n", - "[-0.66,0.15,0.8,0.93,0.93,0.01,-0.6,0.31,0.62,-0.65,0.17,-0.83,-0.25,-0.37,0.65,0.71,-0.6,0.45,-0.88,-0.26],\n", - "[-0.06,0.72,-0.4,-0.21,-0.86,-0.44,-0.69,0.51,0.97,-0.71,0.1,0.06,-0.51,-0.09,-0.88,0.68,0.43,0.71,0.86,-0.11],\n", - "[0.23,-0.34,-0.85,-0.67,-0.27,-0.32,-0.11,-0.06,-0.02,0.19,-0.97,-0.54,0.1,0.63,-0.82,-0.74,0.72,0.12,-0.41,-0.94],\n", - "[-0.06,0.22,-0.08,0.77,-0.53,0.71,0.82,-0.98,-0.67,-0.19,-0.94,0.16,-0.88,-0.99,0.83,0.24,-0.18,0.43,0.56,0.93],\n", - "[0.87,-0.53,-0.57,0.84,0.24,-0.66,-0.8,0.67,-0.8,-0.75,0.16,0.91,-0.34,-0.44,-0.88,0.17,0.82,-0.98,-0.42,0.53],\n", - "[-0.65,0.48,0.1,-0.42,-0.48,-0.46,0.83,-0.32,-0.32,0.08,0.7,0.13,-0.52,0.66,0.93,0.68,0.95,-0.44,0.3,0.72],\n", - "[-0.22,0.11,-0.02,0.71,-0.94,0.22,0.16,0.82,-0.22,-0.18,0.87,0.16,-0.01,0.28,0.44,-0.03,-0.87,-0.02,0.45,-0.59],\n", - "[-0.96,0.38,0.45,0.34,0.98,-0.37,-0.19,0.61,-0.29,0.58,0.01,-0.24,-0.73,-0.94,-0.26,0.1,-0.73,0.53,0.14,0.48]\n", - "]\n", - "\n", - "\n", - "\n", - "169->166\n", - "\n", - "\n", - "13:w13:[20, 20]:[\n", - "[-0.39,-0.29,-0.09,-0.68,0.85,0.76,-0.02,0.19,-0.47,-0.1,0.87,0.37,0.15,-0.5,-0.75,-0.45,0.85,-0.88,0.4,-0.55],\n", - "[-0.52,0.24,0.04,-0.3,0.75,0.11,0.31,-0.73,0.4,-0.71,0.29,-0.83,-0.73,0.53,0.31,0.49,0.5,-0.37,-0.28,0.34],\n", - "[-0.92,0.48,-0.14,0.44,0.73,-0.74,-0.39,0.21,-0.09,0.57,-0.19,0.13,-0.26,-0.69,0.18,0.51,0.68,-0.14,-0.37,0.89],\n", - "[0.65,-0.71,-0.06,0.74,0.78,-0.6,-0.98,0.39,-0.26,-0.55,-0.18,0.44,0.28,-0.36,0.91,0.72,-0.57,0.91,-0.7,-0.91],\n", - "[-0.22,0.45,-0.72,-0.93,0.11,0.42,0.62,-0.93,0.6,-0.99,0.14,0.64,0.12,-0.68,-0.97,0.95,0.91,-0.64,-0.02,-0.43],\n", - "[-0.48,0.74,0.24,-0.38,-0.3,-0.16,0.89,0.26,-0.19,-0.57,0.16,-0.43,0.96,-0.86,-0.08,-0,0.39,-0.59,0.87,0.68],\n", - "[-0.8,0.32,0.74,0.25,0.92,0.76,-0.94,0.01,-0.49,0.47,0.1,0.56,-0.83,-0.31,-0.14,-0.17,0.35,0.26,0.29,-0.69],\n", - "[0.74,0.34,-0.74,0.45,-0.67,0.25,-0.93,0.37,-0.79,-0.66,-0.92,0.96,-0.98,-0.92,-0.85,0.95,-0.05,0.56,0.62,-0.37],\n", - "[0.22,-0.64,-0.18,0.48,-0.67,0.28,0.06,-0,0.57,-0.47,-0.15,-0.92,-0.57,0.93,0.69,-0.72,-0.87,0.49,-0.93,0.27],\n", - "[-0.98,0.78,-0.1,0.33,-0.42,0.07,0.89,-0.52,0.06,0.81,-0.65,0.65,-0.22,-0.63,0.18,0.18,-0.79,0.95,0.88,0.73],\n", - "[-0.54,0.96,-0.3,-0,0.8,0.07,-0.26,0.16,0.54,-0.02,0.53,-0.27,-0.26,0.81,0.87,0.9,0.99,-0.95,0.67,0.63],\n", - "[-0.25,0.73,0.42,-0.96,-0.21,-0.88,0.36,-0.14,0.44,-0.25,-0.45,-0.48,0.7,0.24,-0.87,0.43,-0.5,0.49,-0.27,0.08],\n", - "[-0.37,-0.89,0.2,0.63,0.81,0.4,-0.74,-0.06,0.14,0.7,0.36,0.19,-0.57,-0.74,-0.47,0.48,0.96,-0.36,-0.27,-0.81],\n", - "[-0.6,0.03,-0.3,0.4,-0.97,-0.32,0.65,-0.02,0.34,-0.42,-0.87,-0.3,0.94,-0.08,-0.8,-0.9,-0.24,0.3,-0.59,-0.18],\n", - "[-0.68,-0.87,-0.67,0.5,0.06,0.05,0.42,-0.55,-0.11,-0.08,0.36,-0.03,0.07,0.55,0.91,0.18,-0.93,0.93,0.77,-0.78],\n", - "[-0.74,0.64,-0.62,0.68,0.47,-0.1,-0.14,0.38,-0.66,-0.55,0.67,0.44,-0.67,0.84,-0.54,0.4,0.39,-1,0.78,0.75],\n", - "[0.35,-0.97,0.16,-0.8,-0.67,0.17,0.82,-0.91,0.75,0.17,-0.92,-0.09,0.75,0.68,-0.88,-0.49,-0.82,0.34,1,0.16],\n", - "[-0.98,-0.07,-0.36,0.17,-0.8,0.14,0.53,-0.21,0.66,0.31,0.84,-0.38,-0.98,0.72,0.32,0.77,0.79,-0.98,-0.23,-0.01],\n", - "[-0.8,-0.12,0.01,-0.2,-0.16,0.72,0.87,-0.2,-0.96,-0.79,-0.27,0.53,-0.84,0.31,-0.83,-0.7,0.27,-0.79,0.11,0.59],\n", - "[0.81,-0.99,0.27,0.25,-0.23,0.49,-0.56,-0.47,-0.83,0.11,-0.11,0.08,-0.87,-0.97,0.92,-0.13,0.55,0.55,0.32,-0.91]\n", - "]\n", - "\n", - "\n", - "\n", - "172\n", - "\n", - "WeightedRuleNeuron:172:lstm2_out__f(T) :- {20, 20} lstm1_out(T), {20, 20} lstm2_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.49,0.5,0.5,0.49,0.51,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.49,0.51,0.5,0.49,0.51,0.49,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "172->84\n", - "\n", - "\n", - "10:w10:[20, 20]:[\n", - "[-0.75,-0,0.21,-0.16,0.61,0.09,0.59,-0.02,-0.51,-0.54,0.27,-0.79,0.91,-0.16,-0.75,-0.4,0.96,-0.84,0.12,-0.66],\n", - "[0.93,-0.85,-0.91,-0.17,-0.1,0.2,-0.86,-0.26,-0.12,-0.1,0.35,-0.69,-0.83,0.09,-0.04,-0.45,-0.45,-0.67,-0.84,0.99],\n", - "[0.13,0.6,0.13,-0.09,-0.54,0.13,0.43,0.39,-0.07,0.09,-0.47,-0.35,0.33,0.52,0.95,0.83,0.16,0.43,0.38,0.67],\n", - "[0.02,0.82,-0.23,0.17,-0.66,-0.56,-0.46,-0.11,0.75,-0.51,-0.13,0.29,0.12,-0.01,-0.9,-0.85,0.71,0.19,-0.62,0.94],\n", - "[0.72,-0.49,-0.12,0.99,-0.2,-0.11,-0.91,-0,0.5,0.26,0.89,0.62,-0.38,0.47,-0.26,0.86,0.13,-0.96,0.32,0.99],\n", - "[0.55,0.81,0.77,0.08,-0.71,-0.38,0.12,0.35,0.96,-0.96,0.86,0.83,-0.66,-0.36,0.77,0.24,0.33,0.69,-0.83,0.43],\n", - "[0.96,0.3,-0.72,0.15,0.47,-0.52,0.51,0.47,-0.89,-0.53,0.01,-0.12,-0.47,-0.3,0.33,0.67,0.37,-0.4,0.37,0.89],\n", - "[-0.05,-0.46,-0.24,0.2,-0.06,0.14,-0.03,0.54,0.11,-0.37,-0.34,0.53,0.5,0.7,0.29,-0.31,-0.14,-0.94,0.81,0.36],\n", - "[-0.91,-0.23,0.46,0.38,0.6,0.1,0.26,-0.04,-0.85,0.2,-0.79,-0.92,-0.88,0.9,0.01,0.36,-0.9,-0.75,-0.51,-0.54],\n", - "[0.9,0.37,0.93,-0.33,-0.59,-0.6,0.99,0.83,-0.97,-0.29,-0.83,-0.99,0.61,0.75,-0.02,-0.57,-0.44,-0.56,0.98,-0.64],\n", - "[-0.87,0.22,0.59,-0.3,0.28,-0.27,0.12,0.8,0.58,-0.66,0.12,-0.44,0.82,-0.29,0.83,0.41,0.17,0.85,-0.54,0.02],\n", - "[0.15,0.52,-0.44,-0.67,0.94,0.89,0.12,0.03,0.19,-0.99,-0.65,0.23,0.62,0.05,-0.01,-0.03,0.8,-0.42,-0.87,0.77],\n", - "[0.58,-0.94,0.5,-0.4,-0.28,0.17,0.87,-0.85,0.85,-0.77,-0.06,0.86,-0.26,0.99,-0.27,0.44,-0.32,0.63,0.95,0.57],\n", - "[-0.15,0.89,-0.87,0.84,0.44,-0.9,-0.16,0.06,-0.64,0.94,-0.87,-0.62,-0.47,-0.13,0.61,0.08,0.96,-0.79,0.77,-0.63],\n", - "[-0.97,0.15,0.5,0.86,0.67,-0.75,-0.16,0.06,0.67,-0.75,0.91,0.71,-0.48,-0.67,-0.2,-0.22,0.02,-0.26,-0.58,0.1],\n", - "[0.12,-0.19,0.22,0.12,0.36,-0.8,-0.11,-0.57,-0.19,0.88,-0.39,-0.73,0.16,0.9,-0.74,0.65,0.58,0.93,-0.19,-0.78],\n", - "[0.58,0.2,-0.94,-0.2,0.42,0.7,-0.08,-0.09,0.05,-0.15,-0.68,-0.33,-0.57,0.71,0.01,-0.07,-0.75,-0.01,-0.26,0.61],\n", - "[0.21,-0.04,-0.6,-0.02,0.48,-0.83,-0.99,-0.24,0.83,0.33,0.94,-0.32,-0.08,0.97,-0.08,-0.3,-0.95,-0.14,-0.98,0.82],\n", - "[-0.34,-0.28,0.34,-0.18,-0.96,0.24,-0.4,0.8,0.88,0.84,0.61,0.24,0.13,0.63,-0.76,0.27,0.38,0.23,0.41,0.87],\n", - "[-0.01,0.32,0.24,-0.35,0.54,-0.6,0.73,0.2,0.83,-0.89,-0.1,-0.28,-0,-0.44,0.91,-0.45,-0.18,-0.94,0.07,-0.21]\n", - "]\n", - "\n", - "\n", - "\n", - "172->166\n", - "\n", - "\n", - "11:w11:[20, 20]:[\n", - "[0.27,0.56,-0.52,0.25,-0.59,-0.36,-0.1,-0.31,0.33,0.71,-0.34,0.55,-0.23,0.54,-0.88,0.19,-0.37,0.54,-0.74,0.88],\n", - "[0.11,-0.46,-0.42,0.95,0.43,-0.54,-0.94,0.82,-0.05,0.5,-0.54,-0.38,0.94,-0.73,0.01,-0.05,0.07,-0.09,0.94,0.31],\n", - "[-0.76,0.76,0.67,0.55,-0.53,0.22,0.62,-0.03,0.28,0,0.15,-0.54,0.79,0.98,0.57,-0.14,-0.86,-0.18,0.75,0.07],\n", - "[-0.41,-0.49,0.28,-0.26,-0.21,0.05,0.98,0.71,-0.82,0.8,0.33,0.88,-0.63,0.83,-0.24,-0.5,0.98,-0.55,-0.81,0.58],\n", - "[0.48,-0.18,-0.37,0.74,-0.61,0.43,0.81,-0.74,0.32,0.42,0.92,0.2,0.17,0.26,-0.55,-0.14,0.28,0.95,-0.57,-0.91],\n", - "[0.12,0.63,0.54,-0.48,-0.27,0.84,0.78,-0.83,0.15,0.97,0.88,0.38,0.62,0.74,0.89,-0.08,0.64,0.89,-0.03,0.39],\n", - "[-0.24,-0.22,0.78,-0.93,0.02,0.14,0.83,0.06,0.25,-0.33,0.14,0.41,0.46,-0.04,-0.83,0.94,-0.06,-0.06,0.83,-0.73],\n", - "[0.84,0.45,0.1,0.8,-0.46,-0.84,-0.14,0.51,-0.44,0.84,0.34,0.33,-0.97,-0.79,-0.14,0.52,0.6,0.3,0.19,-0.68],\n", - "[0.15,0.57,-0.26,0.95,0.31,-0.82,0.04,0.09,0.67,0.69,0.43,0.64,-0.7,-0.71,0.83,-0.86,-0.98,0.36,-0.54,0.68],\n", - "[0.01,0.14,0.2,-0.04,0.09,-0.76,0.32,-0.14,-0.79,0.69,0.7,0.6,-0.16,0.37,-0.81,0.51,-0.92,0.68,0.85,0.35],\n", - "[0.49,-0.17,0.16,-0.94,-0.82,-0.14,0.43,-0.08,0.81,0.44,0.49,0.17,-0.18,0.37,-0.33,0.35,0.47,0.59,-0.03,-0.14],\n", - "[0.19,-0.42,0.9,0.4,-0.15,0.13,0.9,-0.41,0.88,-0.48,-0.77,0.27,-0.72,-0.54,-0.13,-0.55,0.94,-0.66,-0.24,0.63],\n", - "[0.15,0.63,0.84,-0.37,0.31,-0.15,-0.95,-0.21,0.89,-0.33,-0.38,-0.7,0.63,-0.4,-0.8,0.82,-0.87,-0.06,0.22,0.98],\n", - "[0.3,0.58,-0.63,0.7,0.11,0.78,0.33,0.66,-0.15,-0.04,0.02,0.13,-0.22,0.14,-0.65,0.53,0.44,0.95,-0.28,-0.85],\n", - "[-0.6,0.53,-0.94,-0.1,0.58,0.28,0.89,0.15,-0.66,-0.03,-0.46,-0.08,0.77,0.22,-0.82,0.51,-0.31,0.3,-0.55,0.68],\n", - "[0.39,-0.21,-0.06,-0.61,0.45,-0.5,0.08,-0.48,-0.68,0.05,-0.51,-0.17,0.03,-0.68,-0.57,-0.11,-0.29,-0.94,-0.43,-0.12],\n", - "[-0.41,0.07,0.79,0.69,-0.62,0.43,-0.17,0.78,-0.84,0.57,0.86,0.46,-0.74,0.04,-0.48,0.84,-0.03,0.39,-0.9,-0.69],\n", - "[-0.65,-0.07,0.43,0.25,-0.08,-0.69,0.21,-0.04,0.6,0.87,-0.81,0.34,-0.26,-0.13,0.53,-0.24,0.22,-0.49,-0.11,-0.53],\n", - "[-0.42,-0.6,-0.8,0.08,0.81,0.83,0.26,0.18,0.68,0.05,-0.18,0.04,-0.25,0.69,-0.72,0.73,-1,0.27,0.56,-0.19],\n", - "[0.35,0.1,0.5,0.69,0.16,-0.62,-0.55,-0.59,0.19,-0.79,-0.03,0.64,-0.24,-0.42,-0.31,-0.43,0.99,-0.51,-0.08,0.7]\n", - "]\n", - "\n", - "\n", - "\n", - "175\n", - "\n", - "RuleNeuron:175:lstm2_out__left(T) :- lstm2_out__f(T), lstm2_out__c(Z), @next(Z, T). [transformation=identity, combination=elproduct]\n", - "val: [0,-0,0,-0,-0,-0,-0.01,-0,0,-0,-0.01,0,-0.01,-0,0,0,0,-0.01,-0,0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "175->162\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "175->172\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "178\n", - "\n", - "WeightedRuleNeuron:178:lstm2_out__i(T) :- {20, 20} lstm1_out(T), {20, 20} lstm2_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.5,0.52,0.51,0.51,0.5,0.49,0.51,0.5,0.51,0.48,0.5,0.5,0.5,0.49,0.51,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "178->84\n", - "\n", - "\n", - "8:w8:[20, 20]:[\n", - "[0.27,0.6,-0.82,0.69,-0.56,0.92,-0.02,0.9,0.98,0.9,0.49,-0.39,0.84,0.46,0.35,-0.67,0.41,-0.22,0.58,-0.26],\n", - "[-0.46,-0.96,-0.7,0.85,-0.16,-0.03,0.96,-0.02,-0.97,-0.51,0.4,0.14,0.43,0.28,0.67,-0.27,-0.88,0.3,-0.92,-0.75],\n", - "[0.89,-0.26,0.12,-0.45,0.42,-0.23,0.97,-0.29,-0.87,-0.7,0.58,-0.23,0.5,0.48,-0.76,-0.73,-0.41,-0.44,0.26,-0.46],\n", - "[0.19,0.29,0.3,-0.02,-0.97,0.99,0.62,-0.81,0.41,-0.77,-0.21,0.66,-0.61,0.2,-0.97,-0.77,-0.95,-0.48,0.96,-0.56],\n", - "[-0.61,-0.21,0.15,-0.07,-0.72,0.83,0.47,0.47,0.78,0.17,-0.6,-0.12,-0.26,-0.67,0.78,-0.54,-0.42,0.01,-0.15,0.29],\n", - "[-0.86,-0.58,-0.92,-0.78,-0.66,0.71,0.64,0.58,0.57,-0.21,0.95,0.7,-0.1,0.5,-0.36,-0.99,-0.84,-0.31,-0.34,-0.21],\n", - "[-0.48,0.52,-0.74,0.82,-0.76,-0.02,0.41,0.52,0.79,-0.39,-0.98,0.69,-0.14,-0,-0.67,-0.3,-0.55,-0.91,-0.97,0.94],\n", - "[-0.36,-0.32,-0.49,0.61,-0.67,-0.73,0.89,-0.58,0.88,0.97,0.98,0.13,-0.47,-0.13,0.99,-0.67,0.89,-0.93,-0.12,-0.48],\n", - "[0.2,-0.66,-0.01,-0.08,-0.54,0.47,0.65,0.41,-0.07,-0.01,0.39,-0.46,-0.54,-0.69,-0.38,-0.65,0.05,-0.91,0.74,0.79],\n", - "[0.48,0.09,0.71,-0.87,0.02,-0.23,-0.26,0.17,-0.54,-0.85,0.7,-0.45,0.14,0.91,-0.39,-0.76,-0.79,0.07,-0.23,0.75],\n", - "[0.61,0.89,0.85,0.52,-0.01,0.38,-0.34,0.67,0.46,-0.78,-0.92,0.08,-0.37,-0.38,0.36,0.88,0.12,0.24,0.81,-0.8],\n", - "[0.74,-0.7,0.5,-0.7,0.23,-0.46,-0.03,0.01,-0.61,-0.27,-0.62,-0.95,-0.4,0.53,-0.97,0.41,0.62,-0.84,0.91,0.98],\n", - "[-0.29,0.78,-0.58,0.26,0.45,-0.58,0.28,0.08,-0.29,0.5,0.31,-0.86,0.55,0.41,0.7,0.29,0.46,-0.24,-0.74,0.56],\n", - "[0.31,0.44,0.02,-0.96,0.43,-0.08,0.86,0.14,-0.57,0.07,-0.48,0.91,0.97,0.84,-0.06,-0.6,0.86,-0.65,-0.23,0.75],\n", - "[0.86,0.22,0.36,-0.23,0.14,0.93,0.8,0.78,0.65,-0.18,-0.99,-0.39,-0.07,0.18,0.4,0.83,-0.31,-0.18,-0.95,-0.43],\n", - "[0.83,-0.29,0.99,-0.63,-0.77,-0.98,-0.99,0.3,0.38,0.38,-0.31,-0.59,-0.28,-0.79,0.75,0.75,0.16,-0.79,-0.94,-0.01],\n", - "[0.51,0.49,-0.06,0.07,-0.07,-0.67,0.31,0.29,-0.84,0.66,-0.99,0.44,-0.17,-0.65,-0.81,0.13,-0.46,0.56,-0.75,0.61],\n", - "[0.5,0.25,-0.43,0.79,0.88,-0.35,-0.88,0.8,-0.87,0.22,-0.2,0.98,0.12,0.41,0.25,-0.49,0.3,0.55,-0.08,-0.96],\n", - "[0.49,-0.97,-0.72,-0.09,0.69,0,-0.2,0.88,-0.23,-0.02,0.12,-0.5,-0.53,-0.34,-0.78,-0.02,0.31,0.81,-0.37,-0.48],\n", - "[0.36,0.89,-0.91,-0.05,0.98,-0.04,0.38,0.96,0.39,-0.75,-0.28,0.83,0.29,0.69,0.43,0.57,0.17,-0.05,0.34,0.76]\n", - "]\n", - "\n", - "\n", - "\n", - "178->166\n", - "\n", - "\n", - "9:w9:[20, 20]:[\n", - "[-0.08,-0.9,0.89,0.73,-0.35,0.92,0.86,-0.84,0.7,0.28,0.12,-0.41,-0.54,-0.13,-0.99,0.04,-0.51,-0.42,0.34,0.32],\n", - "[0.54,0.4,0.8,0.55,-0.18,-0,0.51,0.89,-0.81,0.76,0.63,0.49,-0.18,-0.31,0.26,0.31,-0.66,-0.92,-0.29,0.77],\n", - "[-0.6,-0.32,-0.17,-0.44,0.25,-0.21,0.35,-0.35,-0.98,0.63,-0.47,0.24,0.1,0.06,-0.22,-0.38,-0.89,-0.12,0.95,0.29],\n", - "[-0.94,-0.15,-0.23,0.86,-0.52,0.54,0.38,0.62,-0.87,-0.38,-0.91,-0.8,0.78,0.87,-0.5,-0.94,-0.79,0.09,-0.08,-0.74],\n", - "[0.39,0.72,0.98,0.48,0.07,-0.84,0.12,-0.07,-0.4,0.32,-0.99,0.58,0.22,-0.53,-0.34,-0.14,-0.22,-0.94,-0.61,-0.97],\n", - "[0.66,0.57,0.08,0.97,-0.86,0.7,0.12,0.59,0.49,0.25,0.2,0.74,0.87,-0.4,0.2,0.81,0.11,0.02,0,0.33],\n", - "[-0.7,-0.66,0.69,-0,-0.64,0.31,0.95,0.21,-0.21,-0.61,0.49,-0.95,0.05,-0.76,-0.42,-0.55,0.1,0.22,0.05,-0.23],\n", - "[-0.71,-0.38,-0.37,-0.83,-0.24,-0.56,0.47,0.38,0.2,-0.8,0.71,-0.51,0.41,0.43,0.14,-0.29,-0.8,-0.54,-0.43,-0.57],\n", - "[0.53,0.97,-0.53,-0.09,-0.36,-0.71,0.94,0.02,0.48,0.37,0.73,0.05,-0.11,-0.3,-0.66,-0.85,0.34,-0.27,0.6,-0.44],\n", - "[0.67,0.89,-0.53,-0.59,-0.54,0.56,-0.39,0.81,0.67,0.98,-0.36,0.39,-0.81,-0.35,0.77,0.75,0.37,0.44,-0.91,-0.25],\n", - "[0.71,0.25,0.81,0.02,0.05,0.87,0.37,-0.39,0.45,0.58,-0.7,0.38,0.78,0.89,0.78,-0.67,0.01,0.64,-0.15,-0.2],\n", - "[-0.39,0.92,-0.95,0.4,-0.03,-0.71,0.02,-0.81,0.62,0.64,0.12,-0.35,-0.29,-0.18,0.89,-0.87,-0.02,-0.26,-0.58,-0.86],\n", - "[-0.54,0.62,-0.1,-0.64,-0.18,0.94,0.36,0.15,-0.3,0.45,-0.94,0.62,-0.03,0.89,-0.63,0.79,-0.69,-0.15,-0.89,0.5],\n", - "[-0.51,0.69,-0.59,-0.7,0.29,-0.92,0.62,-0.07,-0.18,0.88,-0.92,-0.23,-0.21,-0.71,-0.3,0.91,0.67,-0.31,-0.58,-0.58],\n", - "[-0.02,0.19,0.29,-0.86,-0.45,0.02,0.4,0.9,-0.37,-0.01,-0.3,-0.8,0.66,-0.6,0.23,0.03,-0.95,-0.12,0.43,0.85],\n", - "[-0.46,-0.83,0.56,-0.66,0.96,-0.27,-0.23,-0.81,-0.08,0.05,-0.44,0.73,0.17,0.37,-0.62,0.15,0.14,-0.43,-0.91,0.73],\n", - "[-0.34,0.16,-0.68,-0.82,0.74,-0.72,0.88,0.68,0.74,0.06,-0.33,-0.57,-0.54,0.49,0.96,0.27,0.08,0.11,-0.79,0.49],\n", - "[0.39,-0.83,0.16,-0.33,0.2,0.35,-0.64,0.9,0.3,-0.09,-0.92,-0.05,-0.84,0.18,-0.13,0.37,0.05,0.7,0.26,-0.28],\n", - "[-0.8,0.19,-0.7,0.1,-0.83,-0.92,0.95,0.86,0.89,-0.56,-0.02,0.14,0.26,0.76,0.18,0.29,-0.34,-0.75,-0.58,-0.95],\n", - "[0.51,0.09,0.65,-0.51,0.21,-0.62,0.69,-0.19,0.02,0.49,0,-0.01,0.71,0.83,-0.94,0.39,0.98,-0.7,-0.95,-0.82]\n", - "]\n", - "\n", - "\n", - "\n", - "181\n", - "\n", - "WeightedRuleNeuron:181:lstm2_out__n(T) :- {20, 20} lstm1_out(T), {20, 20} lstm2_out(Z), @next(Z, T). [transformation=tanh]\n", - "val: [0.02,-0.04,0,0,-0.03,-0.02,-0.05,-0,0.01,-0.02,-0.05,0.01,-0.04,-0.01,0.03,0,0.01,-0.04,-0.04,0.01]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Tanh\n", - "\n", - "\n", - "\n", - "181->84\n", - "\n", - "\n", - "14:w14:[20, 20]:[\n", - "[-0.47,0.48,-0.24,0.29,-0.5,-0.41,-0.41,-0.06,0.83,0.5,0.06,0.26,-0.83,-0.03,-0.29,0.88,0.44,-0.77,0.05,-0.03],\n", - "[-0.94,-0.46,0.73,-0.97,-0.27,-0.19,0.84,0.06,0.82,0.02,0.06,0.21,0.65,-0.05,0.2,-0.13,0.33,-0.37,0.67,0.54],\n", - "[-0.68,-0.28,-0.63,-0.22,-0.34,-0.55,0.73,-0.64,0.09,-0.73,-0.36,-0.7,0.3,-0.87,-0.64,-0.25,0.06,0.82,-0.37,-0.47],\n", - "[0,0.81,-0.2,-0.95,0.47,-0.35,0.83,0.68,-0,0.01,-0.07,0.73,-0.09,-0.66,-0.36,0.18,-0.45,-0.85,0.36,-0.8],\n", - "[0.78,-0.98,0.25,-0.9,-0.23,-0.43,0.95,-0.49,0.94,0.85,-0.4,-0.57,0.8,0.74,-0.91,0.87,-0.9,0.61,-0.06,-0.98],\n", - "[-0.98,-0.54,-0.49,0.86,0.68,0.5,-1,-0.45,-0.42,0.7,0.02,0.13,-0.18,0,0.49,-0.55,-0.23,0.12,-0.04,-0.34],\n", - "[-0.62,-0.94,-0.22,0.13,-0.31,0.55,0.62,-0.85,-0.75,0.29,0.59,0.39,0.67,-0.23,-0.45,-0.16,0.23,-0.45,0.5,0.49],\n", - "[0.72,0.75,-0.94,-0.8,0.57,0.53,0.07,-0.44,0.81,0.71,-0.32,-0.18,-0.62,-0.3,0.5,0.99,-0.05,-0.84,-0.13,0.89],\n", - "[-0.77,0.35,-0.56,0.47,0.03,0.2,0.02,-0.35,0.87,0.96,-0.93,0.16,-0.61,0.45,0.7,0.31,-0.33,-0.27,0.92,0.87],\n", - "[0.95,0.66,0.35,0.57,-0.56,-0.65,0.63,0.6,-0.97,-0.54,-0.56,0.41,0.46,-0.97,0.04,0.4,-0.4,0.39,-0.88,0.27],\n", - "[0.24,0.84,0.92,0.77,-0.89,-0.49,0.55,0.29,0.82,0.17,0.66,0.7,-0.35,-0.23,-0.18,0.63,0.37,0.62,-0.79,-0.47],\n", - "[0.59,0.62,-0.31,-0.32,-0.5,0.02,0.61,0.71,-0.64,-0.23,0.09,-0.91,-0.59,0.88,-0.53,0.57,0.74,0.73,-0.43,-0.55],\n", - "[-0.93,0.97,0.7,0.38,0.07,0.05,0.6,0.56,-0.31,0.4,-0.35,-0.4,0.87,0.83,0.06,-0.86,0.57,-0.25,0.36,-0.33],\n", - "[-0.63,-0.56,0.64,0.88,-0.41,0.81,-0.2,-0.62,-0.82,-0.97,-0.42,0.66,-0.43,-0.64,0.74,-0.14,-0.09,-0.17,-0.47,0.92],\n", - "[0.58,0.53,0.39,-0.41,-0.26,0.62,0.75,0.97,0.56,-0.72,0.29,0.2,0.82,0.61,0.74,0.19,-0.75,-0.58,-0.48,0.53],\n", - "[0.59,1,-0.75,0.91,0.81,0.81,-0.11,0.86,0.75,0.42,-0.18,-0.98,-0.81,-0.9,0.11,-0.44,0.05,0.63,0.99,0.99],\n", - "[0.53,0.64,0.47,-0.78,-0.38,-0.5,-0.98,0.35,-0.27,-0.01,0.23,-0.55,0.05,0.89,0.53,-0.62,0.45,-0.4,0.15,0.46],\n", - "[-0.12,-0.7,-0.26,0.05,0.6,0.07,-0.67,-0.47,-0.32,-0.12,-0.26,0.88,-0.23,-0.36,-0.14,-0.48,0.51,0.74,-0.36,0.84],\n", - "[-0.56,0.26,-0.53,0.61,-0.18,-0.81,-0.28,-0.01,-0.47,-0.06,0.51,-0.72,0.61,-0.25,-0.34,0.79,-0.81,-0.42,0.65,0.79],\n", - "[-0.06,0.58,0.99,-0.95,0.28,-0.74,0.6,0.17,-0.4,-0.69,-0.37,-0.69,0.44,0.93,-0.94,-0.97,0.01,-0.35,0.91,0.47]\n", - "]\n", - "\n", - "\n", - "\n", - "181->166\n", - "\n", - "\n", - "15:w15:[20, 20]:[\n", - "[-0.74,-0.47,-0.56,0.04,-0.88,-1,-0.26,-0.18,-0.48,0.63,-0.6,-0.4,0.54,-0.6,-0.59,-0.04,-0.18,-0.51,0.79,-0.63],\n", - "[-0.53,0.89,0.75,-0.08,-0.36,0.6,0.12,0.29,-0.22,0.79,-0.06,-0.07,0.31,-0.3,0.02,0.76,-0.21,0.93,-0.06,0.17],\n", - "[-0.26,0.3,0.81,-0.51,-0.96,-0.77,0.67,-0.11,-0.63,-0.37,-0.17,-0.17,0.75,0.81,-0.14,0.37,0.4,-0.74,0.43,-0.95],\n", - "[-0.91,-0.03,0.96,0.94,0.53,-0.02,-0.59,0.34,-0.99,-0.81,0.74,0.54,-0.54,-0.54,-0.3,-0.01,0.81,-0.49,0.11,-0.69],\n", - "[-0.76,-0.79,0.17,0.67,0.18,-0.54,-0.21,-0.7,0.69,0.25,0.38,0.02,0.85,-0.11,-0.67,0.04,-0.74,-0.79,0.94,0.67],\n", - "[-0.47,0.87,0.16,0.99,0.18,-0.96,0.63,-0.13,-0.51,0.98,0.02,-0.37,-0.52,-0.75,-0.33,-0.25,-0.77,0.17,0.69,0.22],\n", - "[-0.25,0.73,-0.63,-0.52,-0.35,-0.35,-0.8,0.86,0.07,-0.91,0.67,-0.67,-0.7,0.66,-0.58,-0.9,0.14,0.7,0.6,0.92],\n", - "[-0.05,0.96,-0.65,0.58,1,0.2,0.33,0.6,-0.85,-0.34,-0.12,-0.04,-0.81,0.66,0.78,-0.73,0.99,0.25,-0.6,-0.33],\n", - "[0.03,0.34,0.39,0.63,0.76,-0.49,-0.94,0.7,-0.25,0.07,0.19,0.17,-0.29,-0.84,-0.01,-0.01,1,0.2,-0.36,-0.24],\n", - "[0.54,0.91,0.25,-0.37,0.67,0.28,-0.67,-0.69,0.93,0.66,0.01,-0.57,-0.49,-0.73,-0.78,-0.69,0.71,-0.22,-0.48,-0.99],\n", - "[-0.54,0.46,-0.72,0.67,0.94,-0.05,-0.58,0.08,-0.95,-0.03,0.35,0.67,-0.54,0.56,-0.93,0.02,0.03,0.05,-0.92,0.59],\n", - "[0.33,0.34,-0.72,0.37,0.41,0.86,-0.37,0.63,0.62,0.91,0.25,0.39,0.15,-0.9,-0.82,0.72,0.88,-0.88,0.04,0.35],\n", - "[-0.42,0.16,0.79,0.06,0.01,0.89,-0.61,-0.42,-0.08,-0.86,-0.22,0.68,-0.42,0.27,-0.84,-0.96,0.94,0.37,0,0.98],\n", - "[-0.55,-0.43,-0.65,0.48,0.04,-0.15,-0.68,-0.35,0.82,-0.79,0.86,0.81,0.9,-0.83,-0.35,0.8,-0.63,0.13,-0.12,0.81],\n", - "[0.97,0.4,-0.17,-0.29,-0.27,0.46,-0.73,-0.12,0.89,-0.56,-0.33,-0.95,-0.98,0.32,-0.32,0.91,0.06,0.43,0.44,0.03],\n", - "[-0.45,-0.75,-0.26,-0.15,0.73,-0.47,-0.7,-0.71,-0.58,-0.56,-0.78,-0.74,-0.58,-0.18,-0.28,0.9,0.27,0.05,-0.6,-0.69],\n", - "[-0.61,0.68,0.47,-0.56,-0.2,0.76,-0.58,0.48,-0.21,0.1,0.24,0.48,-0.7,-0.92,0.11,0.53,0.17,0.69,-0.44,-0.66],\n", - "[0.81,0.3,0.93,0.92,-0.76,0.54,-0.08,-0.46,-0.7,-0.25,0.47,0.24,-0.81,0.75,0.79,0.76,-0.74,-0.8,-0.74,0.86],\n", - "[-0.22,-0.04,-0.43,0.85,0.2,0.08,0.25,-0.59,0.31,0.44,0.02,0.56,-0.45,0.4,0.72,0.2,0.8,-0.75,-0.69,0.51],\n", - "[0.24,0.97,0.02,-0.81,-0.57,-0.01,0.13,0.31,0.69,0.83,-0.05,0.78,0.49,-0.42,-0.91,0.04,-0.74,0.53,-1,-0.15]\n", - "]\n", - "\n", - "\n", - "\n", - "184\n", - "\n", - "RuleNeuron:184:lstm2_out__right(T) :- lstm2_out__i(T), lstm2_out__n(T). [transformation=identity, combination=elproduct]\n", - "val: [0.01,-0.02,0,0,-0.02,-0.01,-0.02,-0,0.01,-0.01,-0.02,0,-0.02,-0.01,0.02,0,0.01,-0.02,-0.02,0.01]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "184->178\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "184->181\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "187\n", - "\n", - "RuleNeuron:187:lstm2_out__c(T) :- lstm2_out__left(T), lstm2_out__right(T). [transformation=identity]\n", - "val: [0.02,-0.02,0,0,-0.02,-0.01,-0.03,-0,0.01,-0.01,-0.03,0,-0.03,-0.01,0.02,0,0.01,-0.03,-0.02,0.01]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum\n", - "\n", - "\n", - "\n", - "187->175\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "187->184\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "191\n", - "\n", - "RuleNeuron:191:lstm2_out(T) :- lstm2_out__o(T), tanh(lstm2_out__c(T)). [transformation=identity, combination=elproduct]\n", - "val: [0.01,-0.01,0,0,-0.01,-0.01,-0.02,-0,0,-0.01,-0.02,0,-0.01,-0,0.01,0,0,-0.01,-0.01,0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "191->169\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "191->187\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "194\n", - "\n", - "WeightedRuleNeuron:194:lstm2_out__o(T) :- {20, 20} lstm1_out(T), {20, 20} lstm2_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.49,0.5,0.51,0.5,0.52,0.48,0.51,0.5,0.49,0.51,0.5,0.5,0.51,0.51,0.48,0.51,0.51,0.5,0.49,0.52]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "194->109\n", - "\n", - "\n", - "12:w12:[20, 20]:[\n", - "[-0.47,0.05,0.77,-0.82,0.69,-0.87,0.27,0.97,0.08,0.55,-0.66,-0.91,0.88,-0.82,0.07,-0.47,-0.05,0.09,0.5,-0.11],\n", - "[-0.67,0.38,-0.22,0.1,0.69,-0.92,0.38,0.12,-0.19,-0.69,-0.27,0.07,0.52,-0.33,-0.1,-0.14,-0.19,0.72,0.51,-0.04],\n", - "[-0.47,-0.54,-0.55,0.97,-0.46,-0.39,-0.27,-0.05,-0.2,0.54,-0.67,-0.03,-0.31,-0.83,-0.22,-0.36,0.4,-0.17,0.77,-0.14],\n", - "[0.3,0.87,0.14,-0.3,-0.21,-0.26,-0.62,-0.19,-0.64,-0.08,-0.15,0.46,0.42,0.3,0.8,0.09,0.39,-0.53,0.61,-0.9],\n", - "[-0.83,0.44,-0.56,0.14,0.6,-0.38,-0.71,-0.3,-0.6,0.85,-0.96,0.08,-0.04,-0.24,-0.94,0.5,-0.92,-0.99,0.89,0.68],\n", - "[-0.25,-0.83,-0.81,-0.13,0.12,-0.2,-0.28,-0.9,-0.13,-0.54,-0.42,0.26,0.11,-0.73,-0.33,0.21,0.55,0.89,-0.95,0.54],\n", - "[0.5,0.06,-0.35,0.01,0.72,0.93,0.6,-0.77,-0.2,0.95,-0.55,-0.03,-0.38,0.24,-0.56,-0.25,0.53,0.08,0.88,-0.37],\n", - "[-0.99,-0.54,0.28,-0.63,-0.52,-0.65,0.27,-0.57,-0.49,0.87,0.42,-0.02,-0.73,0.44,-0.42,0.07,-0.61,-0.56,0.26,-0.69],\n", - "[-0.5,-0.86,-0.32,-0.65,0.45,0.19,0.78,-0.29,0.61,-0.82,-0.26,-0.19,0.13,-0.47,0.55,0.14,-0.2,-0.55,-0.57,-0.87],\n", - "[-0.28,0.36,0.91,-0.47,0.92,0.79,0.99,-0.39,0.62,-0.27,0.74,0.12,0.93,-0.93,0.61,-0.42,-0.39,-0.27,0.58,-0.15],\n", - "[-0.36,-0.05,-0.59,0.26,0.08,-0.56,0.47,-0.41,0.42,0.89,0.36,-0.15,-0.92,-0.28,0.33,-0.87,-0.34,0.9,0.35,-0.85],\n", - "[0.35,-0.57,0.13,0.83,-0.49,0.53,-1,-0.14,-0.03,0.15,-0.25,-0.26,0.74,-0.42,0.18,0.62,-0.13,0.75,0.83,0.11],\n", - "[-0.66,0.15,0.8,0.93,0.93,0.01,-0.6,0.31,0.62,-0.65,0.17,-0.83,-0.25,-0.37,0.65,0.71,-0.6,0.45,-0.88,-0.26],\n", - "[-0.06,0.72,-0.4,-0.21,-0.86,-0.44,-0.69,0.51,0.97,-0.71,0.1,0.06,-0.51,-0.09,-0.88,0.68,0.43,0.71,0.86,-0.11],\n", - "[0.23,-0.34,-0.85,-0.67,-0.27,-0.32,-0.11,-0.06,-0.02,0.19,-0.97,-0.54,0.1,0.63,-0.82,-0.74,0.72,0.12,-0.41,-0.94],\n", - "[-0.06,0.22,-0.08,0.77,-0.53,0.71,0.82,-0.98,-0.67,-0.19,-0.94,0.16,-0.88,-0.99,0.83,0.24,-0.18,0.43,0.56,0.93],\n", - "[0.87,-0.53,-0.57,0.84,0.24,-0.66,-0.8,0.67,-0.8,-0.75,0.16,0.91,-0.34,-0.44,-0.88,0.17,0.82,-0.98,-0.42,0.53],\n", - "[-0.65,0.48,0.1,-0.42,-0.48,-0.46,0.83,-0.32,-0.32,0.08,0.7,0.13,-0.52,0.66,0.93,0.68,0.95,-0.44,0.3,0.72],\n", - "[-0.22,0.11,-0.02,0.71,-0.94,0.22,0.16,0.82,-0.22,-0.18,0.87,0.16,-0.01,0.28,0.44,-0.03,-0.87,-0.02,0.45,-0.59],\n", - "[-0.96,0.38,0.45,0.34,0.98,-0.37,-0.19,0.61,-0.29,0.58,0.01,-0.24,-0.73,-0.94,-0.26,0.1,-0.73,0.53,0.14,0.48]\n", - "]\n", - "\n", - "\n", - "\n", - "194->191\n", - "\n", - "\n", - "13:w13:[20, 20]:[\n", - "[-0.39,-0.29,-0.09,-0.68,0.85,0.76,-0.02,0.19,-0.47,-0.1,0.87,0.37,0.15,-0.5,-0.75,-0.45,0.85,-0.88,0.4,-0.55],\n", - "[-0.52,0.24,0.04,-0.3,0.75,0.11,0.31,-0.73,0.4,-0.71,0.29,-0.83,-0.73,0.53,0.31,0.49,0.5,-0.37,-0.28,0.34],\n", - "[-0.92,0.48,-0.14,0.44,0.73,-0.74,-0.39,0.21,-0.09,0.57,-0.19,0.13,-0.26,-0.69,0.18,0.51,0.68,-0.14,-0.37,0.89],\n", - "[0.65,-0.71,-0.06,0.74,0.78,-0.6,-0.98,0.39,-0.26,-0.55,-0.18,0.44,0.28,-0.36,0.91,0.72,-0.57,0.91,-0.7,-0.91],\n", - "[-0.22,0.45,-0.72,-0.93,0.11,0.42,0.62,-0.93,0.6,-0.99,0.14,0.64,0.12,-0.68,-0.97,0.95,0.91,-0.64,-0.02,-0.43],\n", - "[-0.48,0.74,0.24,-0.38,-0.3,-0.16,0.89,0.26,-0.19,-0.57,0.16,-0.43,0.96,-0.86,-0.08,-0,0.39,-0.59,0.87,0.68],\n", - "[-0.8,0.32,0.74,0.25,0.92,0.76,-0.94,0.01,-0.49,0.47,0.1,0.56,-0.83,-0.31,-0.14,-0.17,0.35,0.26,0.29,-0.69],\n", - "[0.74,0.34,-0.74,0.45,-0.67,0.25,-0.93,0.37,-0.79,-0.66,-0.92,0.96,-0.98,-0.92,-0.85,0.95,-0.05,0.56,0.62,-0.37],\n", - "[0.22,-0.64,-0.18,0.48,-0.67,0.28,0.06,-0,0.57,-0.47,-0.15,-0.92,-0.57,0.93,0.69,-0.72,-0.87,0.49,-0.93,0.27],\n", - "[-0.98,0.78,-0.1,0.33,-0.42,0.07,0.89,-0.52,0.06,0.81,-0.65,0.65,-0.22,-0.63,0.18,0.18,-0.79,0.95,0.88,0.73],\n", - "[-0.54,0.96,-0.3,-0,0.8,0.07,-0.26,0.16,0.54,-0.02,0.53,-0.27,-0.26,0.81,0.87,0.9,0.99,-0.95,0.67,0.63],\n", - "[-0.25,0.73,0.42,-0.96,-0.21,-0.88,0.36,-0.14,0.44,-0.25,-0.45,-0.48,0.7,0.24,-0.87,0.43,-0.5,0.49,-0.27,0.08],\n", - "[-0.37,-0.89,0.2,0.63,0.81,0.4,-0.74,-0.06,0.14,0.7,0.36,0.19,-0.57,-0.74,-0.47,0.48,0.96,-0.36,-0.27,-0.81],\n", - "[-0.6,0.03,-0.3,0.4,-0.97,-0.32,0.65,-0.02,0.34,-0.42,-0.87,-0.3,0.94,-0.08,-0.8,-0.9,-0.24,0.3,-0.59,-0.18],\n", - "[-0.68,-0.87,-0.67,0.5,0.06,0.05,0.42,-0.55,-0.11,-0.08,0.36,-0.03,0.07,0.55,0.91,0.18,-0.93,0.93,0.77,-0.78],\n", - "[-0.74,0.64,-0.62,0.68,0.47,-0.1,-0.14,0.38,-0.66,-0.55,0.67,0.44,-0.67,0.84,-0.54,0.4,0.39,-1,0.78,0.75],\n", - "[0.35,-0.97,0.16,-0.8,-0.67,0.17,0.82,-0.91,0.75,0.17,-0.92,-0.09,0.75,0.68,-0.88,-0.49,-0.82,0.34,1,0.16],\n", - "[-0.98,-0.07,-0.36,0.17,-0.8,0.14,0.53,-0.21,0.66,0.31,0.84,-0.38,-0.98,0.72,0.32,0.77,0.79,-0.98,-0.23,-0.01],\n", - "[-0.8,-0.12,0.01,-0.2,-0.16,0.72,0.87,-0.2,-0.96,-0.79,-0.27,0.53,-0.84,0.31,-0.83,-0.7,0.27,-0.79,0.11,0.59],\n", - "[0.81,-0.99,0.27,0.25,-0.23,0.49,-0.56,-0.47,-0.83,0.11,-0.11,0.08,-0.87,-0.97,0.92,-0.13,0.55,0.55,0.32,-0.91]\n", - "]\n", - "\n", - "\n", - "\n", - "197\n", - "\n", - "WeightedRuleNeuron:197:lstm2_out__f(T) :- {20, 20} lstm1_out(T), {20, 20} lstm2_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.49,0.49,0.5,0.48,0.51,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.49,0.52,0.5,0.49,0.52,0.49,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "197->109\n", - "\n", - "\n", - "10:w10:[20, 20]:[\n", - "[-0.75,-0,0.21,-0.16,0.61,0.09,0.59,-0.02,-0.51,-0.54,0.27,-0.79,0.91,-0.16,-0.75,-0.4,0.96,-0.84,0.12,-0.66],\n", - "[0.93,-0.85,-0.91,-0.17,-0.1,0.2,-0.86,-0.26,-0.12,-0.1,0.35,-0.69,-0.83,0.09,-0.04,-0.45,-0.45,-0.67,-0.84,0.99],\n", - "[0.13,0.6,0.13,-0.09,-0.54,0.13,0.43,0.39,-0.07,0.09,-0.47,-0.35,0.33,0.52,0.95,0.83,0.16,0.43,0.38,0.67],\n", - "[0.02,0.82,-0.23,0.17,-0.66,-0.56,-0.46,-0.11,0.75,-0.51,-0.13,0.29,0.12,-0.01,-0.9,-0.85,0.71,0.19,-0.62,0.94],\n", - "[0.72,-0.49,-0.12,0.99,-0.2,-0.11,-0.91,-0,0.5,0.26,0.89,0.62,-0.38,0.47,-0.26,0.86,0.13,-0.96,0.32,0.99],\n", - "[0.55,0.81,0.77,0.08,-0.71,-0.38,0.12,0.35,0.96,-0.96,0.86,0.83,-0.66,-0.36,0.77,0.24,0.33,0.69,-0.83,0.43],\n", - "[0.96,0.3,-0.72,0.15,0.47,-0.52,0.51,0.47,-0.89,-0.53,0.01,-0.12,-0.47,-0.3,0.33,0.67,0.37,-0.4,0.37,0.89],\n", - "[-0.05,-0.46,-0.24,0.2,-0.06,0.14,-0.03,0.54,0.11,-0.37,-0.34,0.53,0.5,0.7,0.29,-0.31,-0.14,-0.94,0.81,0.36],\n", - "[-0.91,-0.23,0.46,0.38,0.6,0.1,0.26,-0.04,-0.85,0.2,-0.79,-0.92,-0.88,0.9,0.01,0.36,-0.9,-0.75,-0.51,-0.54],\n", - "[0.9,0.37,0.93,-0.33,-0.59,-0.6,0.99,0.83,-0.97,-0.29,-0.83,-0.99,0.61,0.75,-0.02,-0.57,-0.44,-0.56,0.98,-0.64],\n", - "[-0.87,0.22,0.59,-0.3,0.28,-0.27,0.12,0.8,0.58,-0.66,0.12,-0.44,0.82,-0.29,0.83,0.41,0.17,0.85,-0.54,0.02],\n", - "[0.15,0.52,-0.44,-0.67,0.94,0.89,0.12,0.03,0.19,-0.99,-0.65,0.23,0.62,0.05,-0.01,-0.03,0.8,-0.42,-0.87,0.77],\n", - "[0.58,-0.94,0.5,-0.4,-0.28,0.17,0.87,-0.85,0.85,-0.77,-0.06,0.86,-0.26,0.99,-0.27,0.44,-0.32,0.63,0.95,0.57],\n", - "[-0.15,0.89,-0.87,0.84,0.44,-0.9,-0.16,0.06,-0.64,0.94,-0.87,-0.62,-0.47,-0.13,0.61,0.08,0.96,-0.79,0.77,-0.63],\n", - "[-0.97,0.15,0.5,0.86,0.67,-0.75,-0.16,0.06,0.67,-0.75,0.91,0.71,-0.48,-0.67,-0.2,-0.22,0.02,-0.26,-0.58,0.1],\n", - "[0.12,-0.19,0.22,0.12,0.36,-0.8,-0.11,-0.57,-0.19,0.88,-0.39,-0.73,0.16,0.9,-0.74,0.65,0.58,0.93,-0.19,-0.78],\n", - "[0.58,0.2,-0.94,-0.2,0.42,0.7,-0.08,-0.09,0.05,-0.15,-0.68,-0.33,-0.57,0.71,0.01,-0.07,-0.75,-0.01,-0.26,0.61],\n", - "[0.21,-0.04,-0.6,-0.02,0.48,-0.83,-0.99,-0.24,0.83,0.33,0.94,-0.32,-0.08,0.97,-0.08,-0.3,-0.95,-0.14,-0.98,0.82],\n", - "[-0.34,-0.28,0.34,-0.18,-0.96,0.24,-0.4,0.8,0.88,0.84,0.61,0.24,0.13,0.63,-0.76,0.27,0.38,0.23,0.41,0.87],\n", - "[-0.01,0.32,0.24,-0.35,0.54,-0.6,0.73,0.2,0.83,-0.89,-0.1,-0.28,-0,-0.44,0.91,-0.45,-0.18,-0.94,0.07,-0.21]\n", - "]\n", - "\n", - "\n", - "\n", - "197->191\n", - "\n", - "\n", - "11:w11:[20, 20]:[\n", - "[0.27,0.56,-0.52,0.25,-0.59,-0.36,-0.1,-0.31,0.33,0.71,-0.34,0.55,-0.23,0.54,-0.88,0.19,-0.37,0.54,-0.74,0.88],\n", - "[0.11,-0.46,-0.42,0.95,0.43,-0.54,-0.94,0.82,-0.05,0.5,-0.54,-0.38,0.94,-0.73,0.01,-0.05,0.07,-0.09,0.94,0.31],\n", - "[-0.76,0.76,0.67,0.55,-0.53,0.22,0.62,-0.03,0.28,0,0.15,-0.54,0.79,0.98,0.57,-0.14,-0.86,-0.18,0.75,0.07],\n", - "[-0.41,-0.49,0.28,-0.26,-0.21,0.05,0.98,0.71,-0.82,0.8,0.33,0.88,-0.63,0.83,-0.24,-0.5,0.98,-0.55,-0.81,0.58],\n", - "[0.48,-0.18,-0.37,0.74,-0.61,0.43,0.81,-0.74,0.32,0.42,0.92,0.2,0.17,0.26,-0.55,-0.14,0.28,0.95,-0.57,-0.91],\n", - "[0.12,0.63,0.54,-0.48,-0.27,0.84,0.78,-0.83,0.15,0.97,0.88,0.38,0.62,0.74,0.89,-0.08,0.64,0.89,-0.03,0.39],\n", - "[-0.24,-0.22,0.78,-0.93,0.02,0.14,0.83,0.06,0.25,-0.33,0.14,0.41,0.46,-0.04,-0.83,0.94,-0.06,-0.06,0.83,-0.73],\n", - "[0.84,0.45,0.1,0.8,-0.46,-0.84,-0.14,0.51,-0.44,0.84,0.34,0.33,-0.97,-0.79,-0.14,0.52,0.6,0.3,0.19,-0.68],\n", - "[0.15,0.57,-0.26,0.95,0.31,-0.82,0.04,0.09,0.67,0.69,0.43,0.64,-0.7,-0.71,0.83,-0.86,-0.98,0.36,-0.54,0.68],\n", - "[0.01,0.14,0.2,-0.04,0.09,-0.76,0.32,-0.14,-0.79,0.69,0.7,0.6,-0.16,0.37,-0.81,0.51,-0.92,0.68,0.85,0.35],\n", - "[0.49,-0.17,0.16,-0.94,-0.82,-0.14,0.43,-0.08,0.81,0.44,0.49,0.17,-0.18,0.37,-0.33,0.35,0.47,0.59,-0.03,-0.14],\n", - "[0.19,-0.42,0.9,0.4,-0.15,0.13,0.9,-0.41,0.88,-0.48,-0.77,0.27,-0.72,-0.54,-0.13,-0.55,0.94,-0.66,-0.24,0.63],\n", - "[0.15,0.63,0.84,-0.37,0.31,-0.15,-0.95,-0.21,0.89,-0.33,-0.38,-0.7,0.63,-0.4,-0.8,0.82,-0.87,-0.06,0.22,0.98],\n", - "[0.3,0.58,-0.63,0.7,0.11,0.78,0.33,0.66,-0.15,-0.04,0.02,0.13,-0.22,0.14,-0.65,0.53,0.44,0.95,-0.28,-0.85],\n", - "[-0.6,0.53,-0.94,-0.1,0.58,0.28,0.89,0.15,-0.66,-0.03,-0.46,-0.08,0.77,0.22,-0.82,0.51,-0.31,0.3,-0.55,0.68],\n", - "[0.39,-0.21,-0.06,-0.61,0.45,-0.5,0.08,-0.48,-0.68,0.05,-0.51,-0.17,0.03,-0.68,-0.57,-0.11,-0.29,-0.94,-0.43,-0.12],\n", - "[-0.41,0.07,0.79,0.69,-0.62,0.43,-0.17,0.78,-0.84,0.57,0.86,0.46,-0.74,0.04,-0.48,0.84,-0.03,0.39,-0.9,-0.69],\n", - "[-0.65,-0.07,0.43,0.25,-0.08,-0.69,0.21,-0.04,0.6,0.87,-0.81,0.34,-0.26,-0.13,0.53,-0.24,0.22,-0.49,-0.11,-0.53],\n", - "[-0.42,-0.6,-0.8,0.08,0.81,0.83,0.26,0.18,0.68,0.05,-0.18,0.04,-0.25,0.69,-0.72,0.73,-1,0.27,0.56,-0.19],\n", - "[0.35,0.1,0.5,0.69,0.16,-0.62,-0.55,-0.59,0.19,-0.79,-0.03,0.64,-0.24,-0.42,-0.31,-0.43,0.99,-0.51,-0.08,0.7]\n", - "]\n", - "\n", - "\n", - "\n", - "200\n", - "\n", - "RuleNeuron:200:lstm2_out__left(T) :- lstm2_out__f(T), lstm2_out__c(Z), @next(Z, T). [transformation=identity, combination=elproduct]\n", - "val: [0.01,-0.01,0,0,-0.01,-0.01,-0.02,-0,0,-0.01,-0.02,0,-0.01,-0,0.01,0,0,-0.01,-0.01,0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "200->187\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "200->197\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "203\n", - "\n", - "WeightedRuleNeuron:203:lstm2_out__i(T) :- {20, 20} lstm1_out(T), {20, 20} lstm2_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.51,0.49,0.5,0.5,0.53,0.51,0.5,0.49,0.5,0.51,0.5,0.51,0.48,0.5,0.49,0.5,0.49,0.51,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "203->109\n", - "\n", - "\n", - "8:w8:[20, 20]:[\n", - "[0.27,0.6,-0.82,0.69,-0.56,0.92,-0.02,0.9,0.98,0.9,0.49,-0.39,0.84,0.46,0.35,-0.67,0.41,-0.22,0.58,-0.26],\n", - "[-0.46,-0.96,-0.7,0.85,-0.16,-0.03,0.96,-0.02,-0.97,-0.51,0.4,0.14,0.43,0.28,0.67,-0.27,-0.88,0.3,-0.92,-0.75],\n", - "[0.89,-0.26,0.12,-0.45,0.42,-0.23,0.97,-0.29,-0.87,-0.7,0.58,-0.23,0.5,0.48,-0.76,-0.73,-0.41,-0.44,0.26,-0.46],\n", - "[0.19,0.29,0.3,-0.02,-0.97,0.99,0.62,-0.81,0.41,-0.77,-0.21,0.66,-0.61,0.2,-0.97,-0.77,-0.95,-0.48,0.96,-0.56],\n", - "[-0.61,-0.21,0.15,-0.07,-0.72,0.83,0.47,0.47,0.78,0.17,-0.6,-0.12,-0.26,-0.67,0.78,-0.54,-0.42,0.01,-0.15,0.29],\n", - "[-0.86,-0.58,-0.92,-0.78,-0.66,0.71,0.64,0.58,0.57,-0.21,0.95,0.7,-0.1,0.5,-0.36,-0.99,-0.84,-0.31,-0.34,-0.21],\n", - "[-0.48,0.52,-0.74,0.82,-0.76,-0.02,0.41,0.52,0.79,-0.39,-0.98,0.69,-0.14,-0,-0.67,-0.3,-0.55,-0.91,-0.97,0.94],\n", - "[-0.36,-0.32,-0.49,0.61,-0.67,-0.73,0.89,-0.58,0.88,0.97,0.98,0.13,-0.47,-0.13,0.99,-0.67,0.89,-0.93,-0.12,-0.48],\n", - "[0.2,-0.66,-0.01,-0.08,-0.54,0.47,0.65,0.41,-0.07,-0.01,0.39,-0.46,-0.54,-0.69,-0.38,-0.65,0.05,-0.91,0.74,0.79],\n", - "[0.48,0.09,0.71,-0.87,0.02,-0.23,-0.26,0.17,-0.54,-0.85,0.7,-0.45,0.14,0.91,-0.39,-0.76,-0.79,0.07,-0.23,0.75],\n", - "[0.61,0.89,0.85,0.52,-0.01,0.38,-0.34,0.67,0.46,-0.78,-0.92,0.08,-0.37,-0.38,0.36,0.88,0.12,0.24,0.81,-0.8],\n", - "[0.74,-0.7,0.5,-0.7,0.23,-0.46,-0.03,0.01,-0.61,-0.27,-0.62,-0.95,-0.4,0.53,-0.97,0.41,0.62,-0.84,0.91,0.98],\n", - "[-0.29,0.78,-0.58,0.26,0.45,-0.58,0.28,0.08,-0.29,0.5,0.31,-0.86,0.55,0.41,0.7,0.29,0.46,-0.24,-0.74,0.56],\n", - "[0.31,0.44,0.02,-0.96,0.43,-0.08,0.86,0.14,-0.57,0.07,-0.48,0.91,0.97,0.84,-0.06,-0.6,0.86,-0.65,-0.23,0.75],\n", - "[0.86,0.22,0.36,-0.23,0.14,0.93,0.8,0.78,0.65,-0.18,-0.99,-0.39,-0.07,0.18,0.4,0.83,-0.31,-0.18,-0.95,-0.43],\n", - "[0.83,-0.29,0.99,-0.63,-0.77,-0.98,-0.99,0.3,0.38,0.38,-0.31,-0.59,-0.28,-0.79,0.75,0.75,0.16,-0.79,-0.94,-0.01],\n", - "[0.51,0.49,-0.06,0.07,-0.07,-0.67,0.31,0.29,-0.84,0.66,-0.99,0.44,-0.17,-0.65,-0.81,0.13,-0.46,0.56,-0.75,0.61],\n", - "[0.5,0.25,-0.43,0.79,0.88,-0.35,-0.88,0.8,-0.87,0.22,-0.2,0.98,0.12,0.41,0.25,-0.49,0.3,0.55,-0.08,-0.96],\n", - "[0.49,-0.97,-0.72,-0.09,0.69,0,-0.2,0.88,-0.23,-0.02,0.12,-0.5,-0.53,-0.34,-0.78,-0.02,0.31,0.81,-0.37,-0.48],\n", - "[0.36,0.89,-0.91,-0.05,0.98,-0.04,0.38,0.96,0.39,-0.75,-0.28,0.83,0.29,0.69,0.43,0.57,0.17,-0.05,0.34,0.76]\n", - "]\n", - "\n", - "\n", - "\n", - "203->191\n", - "\n", - "\n", - "9:w9:[20, 20]:[\n", - "[-0.08,-0.9,0.89,0.73,-0.35,0.92,0.86,-0.84,0.7,0.28,0.12,-0.41,-0.54,-0.13,-0.99,0.04,-0.51,-0.42,0.34,0.32],\n", - "[0.54,0.4,0.8,0.55,-0.18,-0,0.51,0.89,-0.81,0.76,0.63,0.49,-0.18,-0.31,0.26,0.31,-0.66,-0.92,-0.29,0.77],\n", - "[-0.6,-0.32,-0.17,-0.44,0.25,-0.21,0.35,-0.35,-0.98,0.63,-0.47,0.24,0.1,0.06,-0.22,-0.38,-0.89,-0.12,0.95,0.29],\n", - "[-0.94,-0.15,-0.23,0.86,-0.52,0.54,0.38,0.62,-0.87,-0.38,-0.91,-0.8,0.78,0.87,-0.5,-0.94,-0.79,0.09,-0.08,-0.74],\n", - "[0.39,0.72,0.98,0.48,0.07,-0.84,0.12,-0.07,-0.4,0.32,-0.99,0.58,0.22,-0.53,-0.34,-0.14,-0.22,-0.94,-0.61,-0.97],\n", - "[0.66,0.57,0.08,0.97,-0.86,0.7,0.12,0.59,0.49,0.25,0.2,0.74,0.87,-0.4,0.2,0.81,0.11,0.02,0,0.33],\n", - "[-0.7,-0.66,0.69,-0,-0.64,0.31,0.95,0.21,-0.21,-0.61,0.49,-0.95,0.05,-0.76,-0.42,-0.55,0.1,0.22,0.05,-0.23],\n", - "[-0.71,-0.38,-0.37,-0.83,-0.24,-0.56,0.47,0.38,0.2,-0.8,0.71,-0.51,0.41,0.43,0.14,-0.29,-0.8,-0.54,-0.43,-0.57],\n", - "[0.53,0.97,-0.53,-0.09,-0.36,-0.71,0.94,0.02,0.48,0.37,0.73,0.05,-0.11,-0.3,-0.66,-0.85,0.34,-0.27,0.6,-0.44],\n", - "[0.67,0.89,-0.53,-0.59,-0.54,0.56,-0.39,0.81,0.67,0.98,-0.36,0.39,-0.81,-0.35,0.77,0.75,0.37,0.44,-0.91,-0.25],\n", - "[0.71,0.25,0.81,0.02,0.05,0.87,0.37,-0.39,0.45,0.58,-0.7,0.38,0.78,0.89,0.78,-0.67,0.01,0.64,-0.15,-0.2],\n", - "[-0.39,0.92,-0.95,0.4,-0.03,-0.71,0.02,-0.81,0.62,0.64,0.12,-0.35,-0.29,-0.18,0.89,-0.87,-0.02,-0.26,-0.58,-0.86],\n", - "[-0.54,0.62,-0.1,-0.64,-0.18,0.94,0.36,0.15,-0.3,0.45,-0.94,0.62,-0.03,0.89,-0.63,0.79,-0.69,-0.15,-0.89,0.5],\n", - "[-0.51,0.69,-0.59,-0.7,0.29,-0.92,0.62,-0.07,-0.18,0.88,-0.92,-0.23,-0.21,-0.71,-0.3,0.91,0.67,-0.31,-0.58,-0.58],\n", - "[-0.02,0.19,0.29,-0.86,-0.45,0.02,0.4,0.9,-0.37,-0.01,-0.3,-0.8,0.66,-0.6,0.23,0.03,-0.95,-0.12,0.43,0.85],\n", - "[-0.46,-0.83,0.56,-0.66,0.96,-0.27,-0.23,-0.81,-0.08,0.05,-0.44,0.73,0.17,0.37,-0.62,0.15,0.14,-0.43,-0.91,0.73],\n", - "[-0.34,0.16,-0.68,-0.82,0.74,-0.72,0.88,0.68,0.74,0.06,-0.33,-0.57,-0.54,0.49,0.96,0.27,0.08,0.11,-0.79,0.49],\n", - "[0.39,-0.83,0.16,-0.33,0.2,0.35,-0.64,0.9,0.3,-0.09,-0.92,-0.05,-0.84,0.18,-0.13,0.37,0.05,0.7,0.26,-0.28],\n", - "[-0.8,0.19,-0.7,0.1,-0.83,-0.92,0.95,0.86,0.89,-0.56,-0.02,0.14,0.26,0.76,0.18,0.29,-0.34,-0.75,-0.58,-0.95],\n", - "[0.51,0.09,0.65,-0.51,0.21,-0.62,0.69,-0.19,0.02,0.49,0,-0.01,0.71,0.83,-0.94,0.39,0.98,-0.7,-0.95,-0.82]\n", - "]\n", - "\n", - "\n", - "\n", - "206\n", - "\n", - "WeightedRuleNeuron:206:lstm2_out__n(T) :- {20, 20} lstm1_out(T), {20, 20} lstm2_out(Z), @next(Z, T). [transformation=tanh]\n", - "val: [0.04,-0.06,0.01,0.02,-0.03,-0.05,-0.05,0,0.02,-0.03,-0.08,0.01,-0.06,-0.02,0.05,-0.03,0,-0.06,-0.05,0.01]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Tanh\n", - "\n", - "\n", - "\n", - "206->109\n", - "\n", - "\n", - "14:w14:[20, 20]:[\n", - "[-0.47,0.48,-0.24,0.29,-0.5,-0.41,-0.41,-0.06,0.83,0.5,0.06,0.26,-0.83,-0.03,-0.29,0.88,0.44,-0.77,0.05,-0.03],\n", - "[-0.94,-0.46,0.73,-0.97,-0.27,-0.19,0.84,0.06,0.82,0.02,0.06,0.21,0.65,-0.05,0.2,-0.13,0.33,-0.37,0.67,0.54],\n", - "[-0.68,-0.28,-0.63,-0.22,-0.34,-0.55,0.73,-0.64,0.09,-0.73,-0.36,-0.7,0.3,-0.87,-0.64,-0.25,0.06,0.82,-0.37,-0.47],\n", - "[0,0.81,-0.2,-0.95,0.47,-0.35,0.83,0.68,-0,0.01,-0.07,0.73,-0.09,-0.66,-0.36,0.18,-0.45,-0.85,0.36,-0.8],\n", - "[0.78,-0.98,0.25,-0.9,-0.23,-0.43,0.95,-0.49,0.94,0.85,-0.4,-0.57,0.8,0.74,-0.91,0.87,-0.9,0.61,-0.06,-0.98],\n", - "[-0.98,-0.54,-0.49,0.86,0.68,0.5,-1,-0.45,-0.42,0.7,0.02,0.13,-0.18,0,0.49,-0.55,-0.23,0.12,-0.04,-0.34],\n", - "[-0.62,-0.94,-0.22,0.13,-0.31,0.55,0.62,-0.85,-0.75,0.29,0.59,0.39,0.67,-0.23,-0.45,-0.16,0.23,-0.45,0.5,0.49],\n", - "[0.72,0.75,-0.94,-0.8,0.57,0.53,0.07,-0.44,0.81,0.71,-0.32,-0.18,-0.62,-0.3,0.5,0.99,-0.05,-0.84,-0.13,0.89],\n", - "[-0.77,0.35,-0.56,0.47,0.03,0.2,0.02,-0.35,0.87,0.96,-0.93,0.16,-0.61,0.45,0.7,0.31,-0.33,-0.27,0.92,0.87],\n", - "[0.95,0.66,0.35,0.57,-0.56,-0.65,0.63,0.6,-0.97,-0.54,-0.56,0.41,0.46,-0.97,0.04,0.4,-0.4,0.39,-0.88,0.27],\n", - "[0.24,0.84,0.92,0.77,-0.89,-0.49,0.55,0.29,0.82,0.17,0.66,0.7,-0.35,-0.23,-0.18,0.63,0.37,0.62,-0.79,-0.47],\n", - "[0.59,0.62,-0.31,-0.32,-0.5,0.02,0.61,0.71,-0.64,-0.23,0.09,-0.91,-0.59,0.88,-0.53,0.57,0.74,0.73,-0.43,-0.55],\n", - "[-0.93,0.97,0.7,0.38,0.07,0.05,0.6,0.56,-0.31,0.4,-0.35,-0.4,0.87,0.83,0.06,-0.86,0.57,-0.25,0.36,-0.33],\n", - "[-0.63,-0.56,0.64,0.88,-0.41,0.81,-0.2,-0.62,-0.82,-0.97,-0.42,0.66,-0.43,-0.64,0.74,-0.14,-0.09,-0.17,-0.47,0.92],\n", - "[0.58,0.53,0.39,-0.41,-0.26,0.62,0.75,0.97,0.56,-0.72,0.29,0.2,0.82,0.61,0.74,0.19,-0.75,-0.58,-0.48,0.53],\n", - "[0.59,1,-0.75,0.91,0.81,0.81,-0.11,0.86,0.75,0.42,-0.18,-0.98,-0.81,-0.9,0.11,-0.44,0.05,0.63,0.99,0.99],\n", - "[0.53,0.64,0.47,-0.78,-0.38,-0.5,-0.98,0.35,-0.27,-0.01,0.23,-0.55,0.05,0.89,0.53,-0.62,0.45,-0.4,0.15,0.46],\n", - "[-0.12,-0.7,-0.26,0.05,0.6,0.07,-0.67,-0.47,-0.32,-0.12,-0.26,0.88,-0.23,-0.36,-0.14,-0.48,0.51,0.74,-0.36,0.84],\n", - "[-0.56,0.26,-0.53,0.61,-0.18,-0.81,-0.28,-0.01,-0.47,-0.06,0.51,-0.72,0.61,-0.25,-0.34,0.79,-0.81,-0.42,0.65,0.79],\n", - "[-0.06,0.58,0.99,-0.95,0.28,-0.74,0.6,0.17,-0.4,-0.69,-0.37,-0.69,0.44,0.93,-0.94,-0.97,0.01,-0.35,0.91,0.47]\n", - "]\n", - "\n", - "\n", - "\n", - "206->191\n", - "\n", - "\n", - "15:w15:[20, 20]:[\n", - "[-0.74,-0.47,-0.56,0.04,-0.88,-1,-0.26,-0.18,-0.48,0.63,-0.6,-0.4,0.54,-0.6,-0.59,-0.04,-0.18,-0.51,0.79,-0.63],\n", - "[-0.53,0.89,0.75,-0.08,-0.36,0.6,0.12,0.29,-0.22,0.79,-0.06,-0.07,0.31,-0.3,0.02,0.76,-0.21,0.93,-0.06,0.17],\n", - "[-0.26,0.3,0.81,-0.51,-0.96,-0.77,0.67,-0.11,-0.63,-0.37,-0.17,-0.17,0.75,0.81,-0.14,0.37,0.4,-0.74,0.43,-0.95],\n", - "[-0.91,-0.03,0.96,0.94,0.53,-0.02,-0.59,0.34,-0.99,-0.81,0.74,0.54,-0.54,-0.54,-0.3,-0.01,0.81,-0.49,0.11,-0.69],\n", - "[-0.76,-0.79,0.17,0.67,0.18,-0.54,-0.21,-0.7,0.69,0.25,0.38,0.02,0.85,-0.11,-0.67,0.04,-0.74,-0.79,0.94,0.67],\n", - "[-0.47,0.87,0.16,0.99,0.18,-0.96,0.63,-0.13,-0.51,0.98,0.02,-0.37,-0.52,-0.75,-0.33,-0.25,-0.77,0.17,0.69,0.22],\n", - "[-0.25,0.73,-0.63,-0.52,-0.35,-0.35,-0.8,0.86,0.07,-0.91,0.67,-0.67,-0.7,0.66,-0.58,-0.9,0.14,0.7,0.6,0.92],\n", - "[-0.05,0.96,-0.65,0.58,1,0.2,0.33,0.6,-0.85,-0.34,-0.12,-0.04,-0.81,0.66,0.78,-0.73,0.99,0.25,-0.6,-0.33],\n", - "[0.03,0.34,0.39,0.63,0.76,-0.49,-0.94,0.7,-0.25,0.07,0.19,0.17,-0.29,-0.84,-0.01,-0.01,1,0.2,-0.36,-0.24],\n", - "[0.54,0.91,0.25,-0.37,0.67,0.28,-0.67,-0.69,0.93,0.66,0.01,-0.57,-0.49,-0.73,-0.78,-0.69,0.71,-0.22,-0.48,-0.99],\n", - "[-0.54,0.46,-0.72,0.67,0.94,-0.05,-0.58,0.08,-0.95,-0.03,0.35,0.67,-0.54,0.56,-0.93,0.02,0.03,0.05,-0.92,0.59],\n", - "[0.33,0.34,-0.72,0.37,0.41,0.86,-0.37,0.63,0.62,0.91,0.25,0.39,0.15,-0.9,-0.82,0.72,0.88,-0.88,0.04,0.35],\n", - "[-0.42,0.16,0.79,0.06,0.01,0.89,-0.61,-0.42,-0.08,-0.86,-0.22,0.68,-0.42,0.27,-0.84,-0.96,0.94,0.37,0,0.98],\n", - "[-0.55,-0.43,-0.65,0.48,0.04,-0.15,-0.68,-0.35,0.82,-0.79,0.86,0.81,0.9,-0.83,-0.35,0.8,-0.63,0.13,-0.12,0.81],\n", - "[0.97,0.4,-0.17,-0.29,-0.27,0.46,-0.73,-0.12,0.89,-0.56,-0.33,-0.95,-0.98,0.32,-0.32,0.91,0.06,0.43,0.44,0.03],\n", - "[-0.45,-0.75,-0.26,-0.15,0.73,-0.47,-0.7,-0.71,-0.58,-0.56,-0.78,-0.74,-0.58,-0.18,-0.28,0.9,0.27,0.05,-0.6,-0.69],\n", - "[-0.61,0.68,0.47,-0.56,-0.2,0.76,-0.58,0.48,-0.21,0.1,0.24,0.48,-0.7,-0.92,0.11,0.53,0.17,0.69,-0.44,-0.66],\n", - "[0.81,0.3,0.93,0.92,-0.76,0.54,-0.08,-0.46,-0.7,-0.25,0.47,0.24,-0.81,0.75,0.79,0.76,-0.74,-0.8,-0.74,0.86],\n", - "[-0.22,-0.04,-0.43,0.85,0.2,0.08,0.25,-0.59,0.31,0.44,0.02,0.56,-0.45,0.4,0.72,0.2,0.8,-0.75,-0.69,0.51],\n", - "[0.24,0.97,0.02,-0.81,-0.57,-0.01,0.13,0.31,0.69,0.83,-0.05,0.78,0.49,-0.42,-0.91,0.04,-0.74,0.53,-1,-0.15]\n", - "]\n", - "\n", - "\n", - "\n", - "209\n", - "\n", - "RuleNeuron:209:lstm2_out__right(T) :- lstm2_out__i(T), lstm2_out__n(T). [transformation=identity, combination=elproduct]\n", - "val: [0.02,-0.03,0.01,0.01,-0.01,-0.02,-0.03,0,0.01,-0.02,-0.04,0.01,-0.03,-0.01,0.03,-0.01,0,-0.03,-0.03,0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "209->203\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "209->206\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "212\n", - "\n", - "RuleNeuron:212:lstm2_out__c(T) :- lstm2_out__left(T), lstm2_out__right(T). [transformation=identity]\n", - "val: [0.03,-0.04,0.01,0.01,-0.02,-0.03,-0.04,0,0.01,-0.02,-0.05,0.01,-0.04,-0.02,0.04,-0.01,0,-0.04,-0.04,0.01]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum\n", - "\n", - "\n", - "\n", - "212->200\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "212->209\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "216\n", - "\n", - "RuleNeuron:216:lstm2_out(T) :- lstm2_out__o(T), tanh(lstm2_out__c(T)). [transformation=identity, combination=elproduct]\n", - "val: [0.01,-0.02,0,0.01,-0.01,-0.01,-0.02,0,0.01,-0.01,-0.03,0,-0.02,-0.01,0.02,-0.01,0,-0.02,-0.02,0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "216->194\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "216->212\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "223\n", - "\n", - "WeightedRuleNeuron:223:lstm3_out__o(T) :- {20, 20} lstm2_out(T), {20, 20} lstm3_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "223->4\n", - "\n", - "\n", - "21:w21:[20, 20]:[\n", - "[-0.34,1,-0.36,0.04,-0.77,0.08,-0.31,-0.87,0.07,-0.75,-0.88,0.5,0.84,0.51,-0.77,-0.39,0.95,0.36,-0.75,-0.06],\n", - "[0.94,0.07,0.1,0.55,-0.99,-0.21,-0.57,0.96,0.44,-0.05,0.08,-0.55,-0.94,0.72,0.48,0.52,0.69,-0.61,0.55,0.49],\n", - "[0.5,0.58,0.06,-0.22,-0.53,-0.3,0.75,-0.05,0.63,-0.74,-0.18,-0.74,0.29,-1,-0.08,-0.01,-0.33,0.38,0.46,0.76],\n", - "[0.18,-0.64,-0.66,0.65,-0.84,-0.81,-0.83,-0.07,-0.51,0.16,-0.53,0.08,-0.64,-0.44,0.57,-0.2,0.15,0.26,-0.89,-0.38],\n", - "[0.44,0.61,0.73,0.18,0.91,-0.01,0.74,0.89,0.11,0.42,0.3,0.45,0.83,0.52,0.13,0.62,-0.64,0.48,-0.62,-0.67],\n", - "[0.65,0.42,-0.81,-0.85,0.07,-0.87,-0.92,0.33,0.36,0.19,-0.77,-0.6,-0.87,0.84,-0.44,-0.44,-0.53,-0.43,0.64,-0.6],\n", - "[-0.23,0.45,0.33,-0.67,-0.56,-0.8,-0.86,-0.05,-0.14,0.23,-0.33,-0.56,-0.35,0.65,-0.54,-0.31,0.77,0.04,0,0.59],\n", - "[-0.61,-0.67,-0.07,-0.04,-0.44,-0.78,0.97,0.67,0.44,-0.77,-0.51,0.91,-0.79,-0.41,0.14,-0.59,-0.68,-0.9,-0.26,0.2],\n", - "[0.46,-0.74,0.01,-0.72,-0.8,-0.03,0.71,0.63,-0.8,-0.47,0.37,0.9,-0.62,0.62,-0.33,-0.86,-0.24,-0.37,-0.48,0.05],\n", - "[-0.86,-0.14,0.77,0.55,-0.82,0.53,0.5,-0.56,-0.75,-0.64,0.02,-0.54,-0.23,-0.36,0.85,-0.71,0.66,0.78,0.12,0.52],\n", - "[0.06,0.61,-0.97,-0.98,0.68,-0.1,0.75,-0.21,0.4,0.99,0.93,0.94,-0.62,0.64,0,0.11,0.33,-0.99,0.76,-0.64],\n", - "[-0.8,0.59,-0.78,0.41,-0.95,-0.68,0.66,0.97,-0.22,-0.1,0.22,0.76,0.97,-0.26,0.32,0.86,-0.09,0.2,0.12,0.08],\n", - "[0.86,0.03,-0.08,-0.88,-0.53,0.48,-0.44,-0.51,-0.67,-0.22,0.78,-0.65,0.5,-0.56,0.76,0.91,0.4,-0.81,-0.02,0.97],\n", - "[-0.91,0.26,0.07,-0.79,0.8,-0.93,0.23,-0.27,0.31,0.51,0.53,0.31,-0.32,0.71,0.49,-0.7,0.86,-0.12,0.05,0.59],\n", - "[-0.52,-0.94,-0.64,-0.69,0.58,-0.44,-0.31,-0.96,-0.41,0.7,-0.05,-0.24,0.71,-0.94,0.8,0.46,0.6,-0.86,-0.91,0.11],\n", - "[0.79,-0.93,0.33,0.24,-0.33,0,0.7,-0.03,-0.84,-0.64,0.15,-0.17,-0.09,-0.47,-0.84,0.85,0.16,-0.79,0.69,0.66],\n", - "[-0.7,-0.05,-0.6,0.74,-0.32,0.38,-0.26,0.47,0.96,0.12,0.74,-0.12,-0.69,-0.86,-0.33,0.85,-0.68,0.27,0.02,0.82],\n", - "[-0.21,0.38,-0.99,-0.97,-0.82,-0.21,0.1,-0.17,0.96,-0.75,0.59,-0.51,0.9,0.26,0.36,-0.16,-0.22,0.4,-0.94,-0.23],\n", - "[-0.06,-0.45,0.18,-0.89,0.87,-0.68,-0.74,0.74,0.93,0.09,-0.94,0.59,0,0.75,-0.9,0.85,-0.57,0.34,-0.69,-0.7],\n", - "[-0.12,-0.32,-0.72,-0.31,0.7,-0.68,0.12,-0.31,-0.12,-0.99,0.84,-0.83,-0.56,0.5,-0.21,-0.74,-0.01,0.27,-0.41,0.88]\n", - "]\n", - "\n", - "\n", - "\n", - "223->141\n", - "\n", - "\n", - "20:w20:[20, 20]:[\n", - "[0.08,-0.47,0.4,0.1,0.97,-0.07,0.69,-0.43,0.15,0.43,-0.36,0.51,0.65,-0.34,-0.7,-0.26,-0.27,-0.07,0.88,-0.57],\n", - "[-0.6,-0.9,-0.6,-0.23,0.92,-0.98,0.52,0.36,0.74,-0.54,-0.85,-0.31,-0.09,-0.07,0.31,-0.1,-0.71,0.07,0.69,0.52],\n", - "[0.55,0.63,-0.65,0.84,0.23,-0.08,-0.81,-0.86,-0.87,-0.25,-0.15,0.97,-0.27,-0.04,-0.93,-0.37,0.07,-0.04,-0.34,-0.04],\n", - "[0.61,0.82,0.79,-0.54,-0.88,0.84,-0.68,-0.77,-0.19,-0.97,0.06,-0.13,0.95,0.19,-0.54,0.25,0.97,1,-0.27,-0.03],\n", - "[0.65,-0.17,-0.65,0.94,-0.24,-0.93,-0.83,0.7,0.98,-0.09,-0.47,0.87,0.22,-0.13,0.51,-0.66,-0.14,0.65,0.1,-0.17],\n", - "[0.68,0.22,-0.43,-0.39,-0.88,0.7,0.79,-0.57,0.62,-0.14,-0.26,-0.91,-0.16,-0.16,-0.59,-0.68,-0.02,-0.76,-0.37,0.03],\n", - "[0.23,0.42,-0.44,0.72,-0.69,0.72,-0.08,0.16,0.68,0.07,-0.48,-0.06,-0.68,0.24,-0.91,0.77,0.6,-0.79,-0.07,-0.75],\n", - "[-0.96,0.94,0.42,-0.5,0.86,0.86,-0.7,-0.79,0.35,0.83,-0.91,0.12,-0.55,-0.82,0.06,0.54,-0.98,0.91,0.7,-0.16],\n", - "[-0.95,-0.54,0.93,0.69,0.96,0.75,0.92,-0.61,-0.72,0.75,-0.42,0.3,0.89,-0.8,0.84,-0.62,-0.76,0.98,0.16,-0.51],\n", - "[-0.83,-0.08,0.51,-0.85,0.11,0.77,0.88,0.44,0.84,0.68,0.3,0.72,0.36,0.68,-0.99,1,-0.25,0.93,-0.3,0.96],\n", - "[0.58,-0.75,-0.47,0.83,0.85,-0.36,-0.07,-0.32,-0.85,-0.57,0.01,-0.04,0.35,0.99,-0.74,-0.33,0.97,-0.15,0.61,-0.05],\n", - "[0.05,0.8,0.23,-0.63,-0.51,-0.55,1,0.91,0.35,0.34,0.7,-0.65,-0.93,-0.98,0.1,-0.68,-0.71,-0.66,0.94,0.17],\n", - "[-0.87,-0.86,-0.69,0.9,-0.72,0.05,-0.24,-0.25,-0.04,0.89,0.24,0.73,-0.31,0.81,0.76,0.91,0.75,0.19,0.56,-0.48],\n", - "[-0.04,0.34,0.75,0.28,0.65,0.77,-0.02,-0.6,0.52,0,0.46,0.02,0.43,-0.16,0.97,-0.73,-0.85,0.29,0.57,0.69],\n", - "[0.59,0.43,-0.68,-0.19,-0.99,-0.19,0.15,-0.19,0.47,-0.64,0.05,-0.71,0.96,0.01,-0.45,-0.8,-0.86,0.01,0.75,-0.55],\n", - "[-0.53,0.96,-0.12,0.83,-0.3,-0.32,0.97,-0.69,-0.22,0.28,0.02,-0.76,0.81,-0.29,0.66,-0.63,0.42,-0.25,0.42,-0.78],\n", - "[-0.02,-0.7,-0.7,-0.99,-0.06,0.57,-0.41,-0.21,-0.53,0.07,-0.25,0.52,0.9,0.73,-0.6,0.23,0.31,-0.7,-0.65,-0.35],\n", - "[0.93,0.6,0.3,0.23,0.41,-0,-0.52,0.97,0.95,-0.76,-0.43,0.76,-0.88,0.6,0.95,-0.93,-0.69,0.26,0.43,0.85],\n", - "[-0.72,-0.47,-0.5,0.87,-0.89,-0.97,0.48,-0.06,0.1,0.43,-0.21,0.95,0.19,-0.99,0.1,-0.18,0.94,0.22,-0.9,0.31],\n", - "[0.21,-0.25,0.28,0.87,0.3,0.92,0.23,-0.57,-0.61,-0.25,-0.37,-0.2,-0.88,0.26,-0.34,0.96,0.13,-0.64,0.87,0.33]\n", - "]\n", - "\n", - "\n", - "\n", - "235\n", - "\n", - "WeightedRuleNeuron:235:lstm3_out__i(T) :- {20, 20} lstm2_out(T), {20, 20} lstm3_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "235->4\n", - "\n", - "\n", - "17:w17:[20, 20]:[\n", - "[-0.57,-0.37,-0.8,-0.7,-0.37,0.12,0.49,0.97,-0.37,-0.72,0.91,-0.17,-0.58,-0.47,-0.23,-0.36,-0.92,0.16,0.77,-0.5],\n", - "[-0.74,0.89,-0.1,-0.55,0.16,0.49,-0.1,0.22,-0.95,0.83,0.43,-0.88,0.64,0.25,0.86,0.93,-0.32,0.67,0.28,0.46],\n", - "[0.48,0.97,0.09,0.69,0.4,-0.66,-0.48,0.13,0.51,-0.33,0.01,0.45,0.47,-0.38,0.86,-0.64,-0.26,0.13,0.94,-0.87],\n", - "[0.09,-0.99,-0.88,-0.72,0.92,-0.55,-0.62,-0.36,0.49,-0.61,-0.4,-0.17,0.15,-0.09,-0.31,-0.55,-0.79,-0.09,0.48,-0.93],\n", - "[-0.89,-0.89,0.82,0.36,-0.51,0.11,-0.67,-0.11,0.73,-0.48,0.08,-0.87,0.63,0.83,-0.16,-0.09,-0.12,-0.83,0.51,0.54],\n", - "[-0.56,0.82,-0.37,0.45,0.91,-0.1,-0.92,-0.42,0.51,-0.26,0.61,-0.53,0.52,-0.23,-0.97,-0.66,0.46,0.32,-0.55,-0.57],\n", - "[1,0.9,0.64,0.29,0.92,-0.64,0.76,-0.16,-0.91,0.61,0.96,0.58,0.35,-0.47,-0.46,0.47,0.6,0.04,-0.58,-0.14],\n", - "[0.55,0.49,0.45,-0,-0.87,0.95,-0.92,-0.44,0.81,0.24,-0.43,-0.82,-0.15,0.11,-0.48,0.8,0.69,0.42,0.97,0.15],\n", - "[-0.23,0.89,-0.1,0.12,-0.71,-0.68,-0.94,0,0.55,0.38,-0.71,-0.86,-0.86,-0.74,-0.14,-0.71,0.12,0.82,-0.12,-0.87],\n", - "[-0.63,-0.63,-0.91,0.59,0.21,0.51,-0.93,-0.05,-0.58,-0.87,-0.59,0.73,-0.85,0.08,-0.96,-0.86,0.94,0.85,-0.36,0.82],\n", - "[-0.07,0.2,0.93,-0.45,0.07,-0.58,-0.75,0.76,-0.08,-0.23,-0.9,0.5,0.88,0.58,-0.12,-0.7,-0.77,-0.53,-0.47,-0.33],\n", - "[-0.99,0.91,0.38,-0.66,-0.83,0.82,-0.84,-0.24,-0.59,-0.2,0.03,0.53,0.68,0.7,0.85,0.3,0.13,-0.15,-0.69,-0.1],\n", - "[-0.7,0.23,-0.1,-0.68,-0.88,0.55,0.1,-0.43,-0.53,-0.68,-0.01,0.5,0.97,-0.98,-0.46,-0.82,-0.23,0.46,-0.82,-0.51],\n", - "[-0.45,-0.43,0.38,-0.71,0.87,0.23,-0.98,-0.1,0.51,0.71,-0.74,0.14,0.83,-0.92,0.31,-0.19,0.48,0.65,0.58,-0.78],\n", - "[0.5,0.85,-0.26,0.12,0.14,0.63,0.35,0.97,0.66,-0.42,-0.31,0.74,0.8,0.37,0.88,0.96,0.61,-0.01,0.56,0.93],\n", - "[0.76,-0.21,-0.42,-0.88,0.92,0.41,-0.99,0.64,0.05,0.36,-0.35,-0.9,-0.93,-0.73,0.27,0.32,0.92,-0.68,-0.89,-0.53],\n", - "[-0.41,-0.86,1,-0.96,-0.46,0.93,0.89,0.83,0.44,0.31,-0.09,-0.72,-0.73,-0.78,-0.75,-0.49,0.34,0.92,-0.41,-0.69],\n", - "[0.18,-0.8,0.55,0.98,-0.25,0.66,-0.37,0.52,-0.06,0.1,0.43,0.6,-0.88,0.04,-0.3,-0.41,-0.13,0.64,0.48,-0.22],\n", - "[0.25,-0.7,-0.49,-0.09,0.22,0.79,-0.52,0.34,-0.75,-0.24,-0.61,-0.95,-0.44,-0.91,-0.76,0.03,0.04,0.97,-0.58,-0.21],\n", - "[-0.73,0.98,-0.82,0.96,-0.15,-0.74,-0.3,0.47,0.96,0.2,0.61,0.99,0.59,0.1,0.93,-0.9,-0.99,-0.27,-0.04,0.81]\n", - "]\n", - "\n", - "\n", - "\n", - "235->141\n", - "\n", - "\n", - "16:w16:[20, 20]:[\n", - "[-0.84,0.69,0.84,-0.81,0.03,0.06,0.4,0.31,0.02,-0.77,0.12,0.29,-0.97,-0.73,0.24,-0.18,0.88,0.49,0.13,-0.84],\n", - "[0.53,0.35,0.43,0.72,-0.31,0.28,0.45,0.2,-0.21,-0.69,-0.01,-0.18,-0.81,-0.15,0.04,-0.47,0.71,-0.84,0.23,0.26],\n", - "[-0.03,-0.21,-0.64,-0.6,-0.16,0.45,0.39,-1,-0.27,0.69,0.13,-0.21,0.22,-0.87,0.46,0.07,0.33,0.85,-0.03,-0.47],\n", - "[-0.63,0.25,0.38,0.24,-0.62,-0.33,-0.7,-0.2,0.08,0.49,-0.28,0.29,-0.93,0.22,0.71,-0.51,-0.71,-0.95,-0.15,-0.6],\n", - "[-0.13,0.49,-0.77,0.24,0.5,0.9,-0.32,0.86,0.93,-0.28,0.82,-0.52,0.31,0.89,0.11,-0.23,-0.28,0.58,0.27,0.49],\n", - "[-0.65,-0.26,-0.93,-0.24,-0.32,0.3,0.95,0.79,-0.89,-0.59,-0.45,0.5,-0.38,0.81,0.82,0.7,0.27,-0.13,-0.69,-0.07],\n", - "[0.49,-0.24,-0.34,0.53,0.61,-0.27,-0.38,0.14,0.19,-0.46,-0.67,-0.72,-0.57,-0.18,-1,-0.33,-0.72,-0.74,0.95,0.1],\n", - "[-0.58,-0.82,0.01,0.22,-0.96,-0.27,-0.37,0.66,-0.11,-0.58,0.15,-0.36,-0.14,-0.57,-0.15,0.47,-0.61,0.07,0.11,-0.05],\n", - "[0.34,0.39,0.95,0.38,0.59,-0.14,-0.9,-0.5,0.08,-0.77,0.97,-0.66,0.35,-0.63,-0.24,0.93,-0.34,-0.93,-0.71,0.53],\n", - "[-0.98,-0.76,-0.71,-0.99,-0.33,0.52,-0.9,-0.91,0.88,0.27,0.23,1,-0.73,0.69,-0.56,0.44,0.49,0.52,0.7,0.75],\n", - "[-0.8,-0.76,0.72,0.52,0.31,0.85,0.41,0.06,0.77,-0.54,0.38,-0.22,0.99,0.98,-0.85,-0.54,0.68,0.4,-0.33,-0.53],\n", - "[0.82,0.51,-0.64,-0.91,0.15,-0.95,-0.72,0.9,0.79,-0,0.25,0.82,-0.72,-0.34,-0.32,-0.77,0.66,-0.55,0.06,-0.88],\n", - "[0.18,-0.03,0.96,-0.45,-0.44,0.43,-0.84,0.98,0.7,0.7,-0.98,-0.26,-0.51,0.1,-0.64,-0.65,-0.75,-0.97,-0.12,0.87],\n", - "[0.21,0.34,-0.86,-0.54,-0.36,-0.8,0.07,0.67,0.93,-0.57,-0.68,-0.7,-0.5,0.18,-0.37,-0.23,0.3,0.18,-0.21,0.81],\n", - "[-0.02,0.57,0.98,-0.7,-0.89,-0.74,0.72,0.04,-0.56,-0.6,0.93,0.54,0.13,0.44,0.1,-0.3,-0.07,0.35,0.26,0.16],\n", - "[0.51,-0.48,-0.18,-0.49,-0.12,0.82,-0.89,0.83,0.4,0.75,-0.77,0.65,-0.93,0.36,-0.55,-0.4,-0.19,-0.6,-0.58,0.87],\n", - "[0.24,-0.93,0.16,0.99,0.63,-0.62,0.06,-0.68,0.23,-0.6,0.73,0.09,0.9,-0.51,-0.85,0.96,0.3,0.8,0.23,-0.36],\n", - "[-0.98,0.26,0.47,0.35,-0.72,-0.13,0.57,-0.62,0.08,-0.11,0.05,-0.19,0.28,0.39,-0.35,0.64,-0.87,0.37,-0.54,0.78],\n", - "[-0.24,0.85,0.42,-0.52,0.34,-0.83,0.36,0.22,0.34,0.86,0.5,0.33,0.08,0.61,0.91,0.06,-0.51,-0.94,-0.65,0.43],\n", - "[-0.24,-0.63,-0.38,0.41,0.4,0.99,0.62,0.09,-0.58,-0.86,0.2,-0.38,-0.04,0.41,-0.27,-0.05,-0.92,-0.06,-0.03,0.78]\n", - "]\n", - "\n", - "\n", - "\n", - "238\n", - "\n", - "WeightedRuleNeuron:238:lstm3_out__n(T) :- {20, 20} lstm2_out(T), {20, 20} lstm3_out(Z), @next(Z, T). [transformation=tanh]\n", - "val: [-0,-0,0,0,0,-0,-0,0,-0,-0,0,0,-0,0.01,0,0,0,-0,0,-0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Tanh\n", - "\n", - "\n", - "\n", - "238->4\n", - "\n", - "\n", - "23:w23:[20, 20]:[\n", - "[-0.94,-0.68,0.32,0.7,-0.63,0.71,-0.19,0.17,0.41,-0.91,0.04,-0.25,0.03,-0.62,-0.84,0.23,-0.91,-0.5,-0.45,-0.77],\n", - "[-0.01,0.51,-0.34,0.82,-0.41,0.6,0.57,0.76,0.06,0.09,-0.87,-0.52,0.78,0.52,-0.15,-0.97,-0.67,0.33,-0.43,-0.07],\n", - "[-0.19,-0.46,0.82,0.22,-0.71,-0.64,0.4,0.43,0.44,0.73,0.91,-0.91,-0.38,-0.45,0.53,0.82,0.7,0.28,0.89,-0.06],\n", - "[0.96,-0.17,-0.52,0.09,-0.75,-0.63,0.56,-0.32,0.43,0.19,0.5,0.34,0.5,-0.21,0.01,-0.98,-0.02,0.61,0.84,0.03],\n", - "[0.86,0.14,0,0.89,0.26,-0.25,-0.8,-0.57,-0.87,-0.3,-0.57,-0.55,-0.68,0.6,0.46,0.06,0.36,-0.01,0.61,-0.54],\n", - "[-0.28,-0.62,0.7,0.19,-0.71,0,-0.69,0.52,0.34,-0.12,-0.32,-1,-0.38,0.08,-0.82,-0.99,-0.57,0.61,0.71,0.61],\n", - "[0.3,-0.45,0.65,-0.96,0.36,-0.07,0.19,-0.47,0.01,-0.04,0.76,0.51,-0.55,-0.58,-0.37,-0.24,-0.01,0.14,-0.81,0.35],\n", - "[0.34,-0.15,-0.56,0.06,-0.45,-0.12,-0.12,0.14,-0.57,0.82,0.41,0.62,-0.72,-0.61,0.35,-0.77,-0.33,-0.44,-0.86,0.72],\n", - "[-0.65,0.42,-0.59,-0.57,0.36,0.89,0.14,-0.51,0.83,-0.78,0.13,-0.44,-0.86,0.87,0.17,0.44,0.56,0.16,-0.15,0.72],\n", - "[0.94,-0.83,0.36,0.91,0.44,-0.88,-0.91,-0.19,0.97,-0.46,0.63,0.01,0.58,0.67,0.15,0.94,-0.42,0.62,-0.83,-0.79],\n", - "[0.36,-0.98,0.26,-0.61,-0.95,-0.71,0.22,0.28,-0.3,-0.34,-0.97,0.75,-0.91,-0.07,0.12,-0.41,-0.43,0.94,0.12,0.81],\n", - "[0.25,0.72,-0.17,0.46,-0.82,-0.72,-0.18,-0.04,0.9,-0.68,0.79,-0.22,-0.21,0.29,-0.46,0.58,-0.06,0.21,-0.57,-0.62],\n", - "[-0.4,0.65,-0.2,0.24,0.04,0.79,0.17,-0.67,0.55,-0.16,-0.22,0.06,0.53,-0.23,-0.8,0.93,0.71,-0.37,-0.96,0.25],\n", - "[-0.36,-0.23,-0.17,-0.68,-0.04,-0.52,0.44,-0.96,0.26,0.78,0.89,0.51,-0.44,0.69,-0.22,0.01,0.09,-0.72,-0.24,-0.16],\n", - "[0.5,0.54,0.45,-0.85,-0.96,0.75,0.51,-0.4,0.08,0.08,-0.72,-0.42,-0.07,-0.55,0.03,-0.13,0.65,-0.04,-0.92,0.15],\n", - "[0.45,-0.03,-0.81,-0.7,-0.98,-0.94,-0.84,-0.87,-0.8,-0.38,-0.77,0.81,-0.62,1,-0.05,-0.05,0.91,-0.87,-0.65,0.06],\n", - "[-0.72,0.5,0.37,-0.73,0.6,-0.29,0.59,0.59,0.54,-0.24,0.98,-0.49,0.9,0.87,-0.87,0.44,-0.4,-0.47,0.47,0.76],\n", - "[-0.87,0.19,0.94,-0.69,0.16,0.35,-0.51,-0.42,-0.59,0.31,0.06,-0.12,-0.32,-0.56,0.6,0.43,-0.52,-0.59,-0.66,-0.53],\n", - "[-0.43,-0.25,-0.84,0.18,0.43,-0.53,0.97,0.61,-0.38,-0.23,0.87,0.56,-0.01,-0.59,0.58,-0.24,-0.52,-0.61,0.08,-0.04],\n", - "[0.64,0.67,-0.76,-0.37,0.8,-0.37,-0.75,0.47,0.23,-0.87,0.3,-0.53,-0.15,0.4,0.83,0.42,-0.79,-0.09,0.17,-0.28]\n", - "]\n", - "\n", - "\n", - "\n", - "238->141\n", - "\n", - "\n", - "22:w22:[20, 20]:[\n", - "[-0.98,-0.36,-0.06,-0.54,-0.52,0.49,-0.48,0.79,-0.96,0.02,-0.42,-0.91,-0.26,0.68,-0.71,0.88,0.71,0.75,-0.19,0.23],\n", - "[0.76,0.01,0.73,-0.07,0.09,0.35,-0.03,0.97,-0.41,0.07,0.69,-0.05,-0.57,0.49,-0.97,-0.1,-0.57,0.31,-0.04,-0.46],\n", - "[-0.26,-0.97,0.22,-0.27,0.49,0.14,-0.32,-0.26,-0.04,0.17,-0.61,-0.16,0.44,-0.27,-0,0.44,0.67,-0.03,0.3,0.3],\n", - "[0.54,-0.06,0.31,0.13,0.8,0.01,0.29,0.1,0.14,0.47,0.96,-0.01,0.32,0.38,0.52,0.98,0.86,-0.31,0.62,0.05],\n", - "[0.21,-0.14,-0.75,0.95,0.01,0.2,0.2,-0.57,-0.04,-0.53,-0.73,-0.51,-0.48,0.54,0.74,-0.55,0.59,0.13,-0.96,-0.24],\n", - "[-0.92,0.19,0.85,0.65,0.66,-0.18,-0.32,0.86,0.63,-0.76,0.29,-0.71,0.25,-0.18,0.23,0.18,-0.04,0.83,0.26,-0.35],\n", - "[-0.85,0.41,0.4,0.19,-0.44,0.79,-0.03,0.6,-0.4,0.8,0.29,0.59,-0.52,-0.4,-0.93,-0.45,0.5,0.39,-0.77,-0.89],\n", - "[-0.71,0.01,0.44,-0.34,0.65,-0.66,0.18,0.97,-0.99,0.53,0.1,0.08,0.77,0.52,0.93,-0.77,0.93,-0.9,-0.53,0.2],\n", - "[0.98,-0.28,-0.51,-0.95,-0.62,0.08,0.01,0.76,0.48,-0.68,-0.08,-0.78,-0.72,0.06,0.33,-0.32,-0.98,-0.85,-0.98,0.72],\n", - "[-0.56,0.42,0.71,0.57,0.01,0.9,-0.98,0.88,0.64,-0.82,0.42,0.67,-0.23,0.77,-0.11,-0.15,-0.74,-0.77,-0.54,-0.75],\n", - "[-0.15,-0.47,-0.79,0.91,-0.34,0.33,-0.78,0.24,-0.16,-0.03,0.24,0.37,0.94,0.92,0.83,-0.92,0.7,-0.2,-0.62,-0.43],\n", - "[0.36,0.32,0.43,0.51,0.91,0.8,0.29,-0.27,0.08,0.25,-0.64,0.37,0.56,-0.4,-0.89,-0.85,0.33,-0.85,-0.61,-0.84],\n", - "[0.62,-0.36,-0.97,0.58,0.89,-0.53,-0.02,-0.65,-0.4,0.75,-0.54,-0.81,0.92,0.19,0.79,-0.87,-0.26,-0.97,-0.64,-0.1],\n", - "[-0.12,-0.16,-0.77,-0.21,-0.05,0.26,0.71,-0.34,0.74,-0.68,-0.42,0.27,0.78,0.7,0.76,0.33,-0.76,0.65,0.15,-0.93],\n", - "[-0.61,-0.61,0.16,0.79,-0.81,0.07,0.2,0.16,-0.98,0.83,0.44,0.9,0.84,0.63,0.04,-0.08,-0.13,-0.69,0.98,0.9],\n", - "[-0.88,-0.4,-0.99,0.94,0.83,0.47,0.04,0.04,-0.18,0.87,-0.14,0.62,-0,0.58,0.94,0.87,-0.37,0.72,-0.98,-0.74],\n", - "[0.47,0.14,0.82,-0.05,0.89,0.09,0.51,0.87,0.31,-0.85,0.29,0.59,0.1,0.77,-0.15,0.21,0.19,-0.32,0.32,-0.77],\n", - "[0.17,-0.51,0.76,0.2,0.53,0.21,-0.46,0.11,0.18,0.56,-0.02,0.44,0.77,0.25,-0.43,0.88,-0.24,-0.1,0.03,-0.12],\n", - "[-0.42,0.68,0.31,0.69,-0.97,0.13,0.22,-0.73,-0.57,-0.47,-0.15,0.6,-0.13,0.51,0.81,-0.44,-0.37,-0.63,0.14,0.63],\n", - "[0.71,0.54,-0.54,0.88,0.09,-0.83,-0.18,0.43,-0.4,-0.09,-0.01,-0.74,0.9,-0.14,0.4,0.92,0.6,0.57,-0.34,0.6]\n", - "]\n", - "\n", - "\n", - "\n", - "241\n", - "\n", - "RuleNeuron:241:lstm3_out__right(T) :- lstm3_out__i(T), lstm3_out__n(T). [transformation=identity, combination=elproduct]\n", - "val: [-0,-0,0,0,0,-0,-0,0,-0,-0,0,0,-0,0,0,0,0,-0,0,-0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "241->235\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "241->238\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "248\n", - "\n", - "RuleNeuron:248:lstm3_out(T) :- lstm3_out__o(T), tanh(lstm3_out__c(T)). [transformation=identity, combination=elproduct]\n", - "val: [-0,-0,0,0,0,-0,-0,0,-0,-0,0,0,-0,0,0,0,0,-0,0,-0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "248->223\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "248->241\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "251\n", - "\n", - "WeightedRuleNeuron:251:lstm3_out__o(T) :- {20, 20} lstm2_out(T), {20, 20} lstm3_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.5,0.5,0.51,0.5,0.5,0.5,0.5,0.5,0.5,0.49,0.5,0.5,0.5,0.49,0.5,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "251->166\n", - "\n", - "\n", - "20:w20:[20, 20]:[\n", - "[0.08,-0.47,0.4,0.1,0.97,-0.07,0.69,-0.43,0.15,0.43,-0.36,0.51,0.65,-0.34,-0.7,-0.26,-0.27,-0.07,0.88,-0.57],\n", - "[-0.6,-0.9,-0.6,-0.23,0.92,-0.98,0.52,0.36,0.74,-0.54,-0.85,-0.31,-0.09,-0.07,0.31,-0.1,-0.71,0.07,0.69,0.52],\n", - "[0.55,0.63,-0.65,0.84,0.23,-0.08,-0.81,-0.86,-0.87,-0.25,-0.15,0.97,-0.27,-0.04,-0.93,-0.37,0.07,-0.04,-0.34,-0.04],\n", - "[0.61,0.82,0.79,-0.54,-0.88,0.84,-0.68,-0.77,-0.19,-0.97,0.06,-0.13,0.95,0.19,-0.54,0.25,0.97,1,-0.27,-0.03],\n", - "[0.65,-0.17,-0.65,0.94,-0.24,-0.93,-0.83,0.7,0.98,-0.09,-0.47,0.87,0.22,-0.13,0.51,-0.66,-0.14,0.65,0.1,-0.17],\n", - "[0.68,0.22,-0.43,-0.39,-0.88,0.7,0.79,-0.57,0.62,-0.14,-0.26,-0.91,-0.16,-0.16,-0.59,-0.68,-0.02,-0.76,-0.37,0.03],\n", - "[0.23,0.42,-0.44,0.72,-0.69,0.72,-0.08,0.16,0.68,0.07,-0.48,-0.06,-0.68,0.24,-0.91,0.77,0.6,-0.79,-0.07,-0.75],\n", - "[-0.96,0.94,0.42,-0.5,0.86,0.86,-0.7,-0.79,0.35,0.83,-0.91,0.12,-0.55,-0.82,0.06,0.54,-0.98,0.91,0.7,-0.16],\n", - "[-0.95,-0.54,0.93,0.69,0.96,0.75,0.92,-0.61,-0.72,0.75,-0.42,0.3,0.89,-0.8,0.84,-0.62,-0.76,0.98,0.16,-0.51],\n", - "[-0.83,-0.08,0.51,-0.85,0.11,0.77,0.88,0.44,0.84,0.68,0.3,0.72,0.36,0.68,-0.99,1,-0.25,0.93,-0.3,0.96],\n", - "[0.58,-0.75,-0.47,0.83,0.85,-0.36,-0.07,-0.32,-0.85,-0.57,0.01,-0.04,0.35,0.99,-0.74,-0.33,0.97,-0.15,0.61,-0.05],\n", - "[0.05,0.8,0.23,-0.63,-0.51,-0.55,1,0.91,0.35,0.34,0.7,-0.65,-0.93,-0.98,0.1,-0.68,-0.71,-0.66,0.94,0.17],\n", - "[-0.87,-0.86,-0.69,0.9,-0.72,0.05,-0.24,-0.25,-0.04,0.89,0.24,0.73,-0.31,0.81,0.76,0.91,0.75,0.19,0.56,-0.48],\n", - "[-0.04,0.34,0.75,0.28,0.65,0.77,-0.02,-0.6,0.52,0,0.46,0.02,0.43,-0.16,0.97,-0.73,-0.85,0.29,0.57,0.69],\n", - "[0.59,0.43,-0.68,-0.19,-0.99,-0.19,0.15,-0.19,0.47,-0.64,0.05,-0.71,0.96,0.01,-0.45,-0.8,-0.86,0.01,0.75,-0.55],\n", - "[-0.53,0.96,-0.12,0.83,-0.3,-0.32,0.97,-0.69,-0.22,0.28,0.02,-0.76,0.81,-0.29,0.66,-0.63,0.42,-0.25,0.42,-0.78],\n", - "[-0.02,-0.7,-0.7,-0.99,-0.06,0.57,-0.41,-0.21,-0.53,0.07,-0.25,0.52,0.9,0.73,-0.6,0.23,0.31,-0.7,-0.65,-0.35],\n", - "[0.93,0.6,0.3,0.23,0.41,-0,-0.52,0.97,0.95,-0.76,-0.43,0.76,-0.88,0.6,0.95,-0.93,-0.69,0.26,0.43,0.85],\n", - "[-0.72,-0.47,-0.5,0.87,-0.89,-0.97,0.48,-0.06,0.1,0.43,-0.21,0.95,0.19,-0.99,0.1,-0.18,0.94,0.22,-0.9,0.31],\n", - "[0.21,-0.25,0.28,0.87,0.3,0.92,0.23,-0.57,-0.61,-0.25,-0.37,-0.2,-0.88,0.26,-0.34,0.96,0.13,-0.64,0.87,0.33]\n", - "]\n", - "\n", - "\n", - "\n", - "251->248\n", - "\n", - "\n", - "21:w21:[20, 20]:[\n", - "[-0.34,1,-0.36,0.04,-0.77,0.08,-0.31,-0.87,0.07,-0.75,-0.88,0.5,0.84,0.51,-0.77,-0.39,0.95,0.36,-0.75,-0.06],\n", - "[0.94,0.07,0.1,0.55,-0.99,-0.21,-0.57,0.96,0.44,-0.05,0.08,-0.55,-0.94,0.72,0.48,0.52,0.69,-0.61,0.55,0.49],\n", - "[0.5,0.58,0.06,-0.22,-0.53,-0.3,0.75,-0.05,0.63,-0.74,-0.18,-0.74,0.29,-1,-0.08,-0.01,-0.33,0.38,0.46,0.76],\n", - "[0.18,-0.64,-0.66,0.65,-0.84,-0.81,-0.83,-0.07,-0.51,0.16,-0.53,0.08,-0.64,-0.44,0.57,-0.2,0.15,0.26,-0.89,-0.38],\n", - "[0.44,0.61,0.73,0.18,0.91,-0.01,0.74,0.89,0.11,0.42,0.3,0.45,0.83,0.52,0.13,0.62,-0.64,0.48,-0.62,-0.67],\n", - "[0.65,0.42,-0.81,-0.85,0.07,-0.87,-0.92,0.33,0.36,0.19,-0.77,-0.6,-0.87,0.84,-0.44,-0.44,-0.53,-0.43,0.64,-0.6],\n", - "[-0.23,0.45,0.33,-0.67,-0.56,-0.8,-0.86,-0.05,-0.14,0.23,-0.33,-0.56,-0.35,0.65,-0.54,-0.31,0.77,0.04,0,0.59],\n", - "[-0.61,-0.67,-0.07,-0.04,-0.44,-0.78,0.97,0.67,0.44,-0.77,-0.51,0.91,-0.79,-0.41,0.14,-0.59,-0.68,-0.9,-0.26,0.2],\n", - "[0.46,-0.74,0.01,-0.72,-0.8,-0.03,0.71,0.63,-0.8,-0.47,0.37,0.9,-0.62,0.62,-0.33,-0.86,-0.24,-0.37,-0.48,0.05],\n", - "[-0.86,-0.14,0.77,0.55,-0.82,0.53,0.5,-0.56,-0.75,-0.64,0.02,-0.54,-0.23,-0.36,0.85,-0.71,0.66,0.78,0.12,0.52],\n", - "[0.06,0.61,-0.97,-0.98,0.68,-0.1,0.75,-0.21,0.4,0.99,0.93,0.94,-0.62,0.64,0,0.11,0.33,-0.99,0.76,-0.64],\n", - "[-0.8,0.59,-0.78,0.41,-0.95,-0.68,0.66,0.97,-0.22,-0.1,0.22,0.76,0.97,-0.26,0.32,0.86,-0.09,0.2,0.12,0.08],\n", - "[0.86,0.03,-0.08,-0.88,-0.53,0.48,-0.44,-0.51,-0.67,-0.22,0.78,-0.65,0.5,-0.56,0.76,0.91,0.4,-0.81,-0.02,0.97],\n", - "[-0.91,0.26,0.07,-0.79,0.8,-0.93,0.23,-0.27,0.31,0.51,0.53,0.31,-0.32,0.71,0.49,-0.7,0.86,-0.12,0.05,0.59],\n", - "[-0.52,-0.94,-0.64,-0.69,0.58,-0.44,-0.31,-0.96,-0.41,0.7,-0.05,-0.24,0.71,-0.94,0.8,0.46,0.6,-0.86,-0.91,0.11],\n", - "[0.79,-0.93,0.33,0.24,-0.33,0,0.7,-0.03,-0.84,-0.64,0.15,-0.17,-0.09,-0.47,-0.84,0.85,0.16,-0.79,0.69,0.66],\n", - "[-0.7,-0.05,-0.6,0.74,-0.32,0.38,-0.26,0.47,0.96,0.12,0.74,-0.12,-0.69,-0.86,-0.33,0.85,-0.68,0.27,0.02,0.82],\n", - "[-0.21,0.38,-0.99,-0.97,-0.82,-0.21,0.1,-0.17,0.96,-0.75,0.59,-0.51,0.9,0.26,0.36,-0.16,-0.22,0.4,-0.94,-0.23],\n", - "[-0.06,-0.45,0.18,-0.89,0.87,-0.68,-0.74,0.74,0.93,0.09,-0.94,0.59,0,0.75,-0.9,0.85,-0.57,0.34,-0.69,-0.7],\n", - "[-0.12,-0.32,-0.72,-0.31,0.7,-0.68,0.12,-0.31,-0.12,-0.99,0.84,-0.83,-0.56,0.5,-0.21,-0.74,-0.01,0.27,-0.41,0.88]\n", - "]\n", - "\n", - "\n", - "\n", - "254\n", - "\n", - "WeightedRuleNeuron:254:lstm3_out__f(T) :- {20, 20} lstm2_out(T), {20, 20} lstm3_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.49,0.5,0.5,0.5,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "254->166\n", - "\n", - "\n", - "18:w18:[20, 20]:[\n", - "[0.67,0.87,0.79,0.15,0.83,-0.69,0.82,-0.27,0.08,-0.53,0.14,-0.69,0.14,0.99,-0.98,-0.08,0.26,0.41,-0.22,0.17],\n", - "[0.93,0.81,-0.3,-0.88,0.68,0.61,0.6,-0.8,0.25,-1,-0.43,0.6,0.56,-0.19,0.98,-0.12,0.62,-0.12,0.34,-0.66],\n", - "[0.72,0.76,-0.82,-0.59,-0.87,0.71,-0.61,0.17,0.9,0.85,-0.35,-0.2,-0.92,-0.22,0.76,-0.01,0.69,-0.3,-0.03,-0.13],\n", - "[0.18,0.05,-0.36,0.68,0.29,-0.86,0.3,0.83,0.16,-0.54,0.17,-0.2,0.2,-0.84,-0.81,-0.96,0.9,-0.72,0.95,0.79],\n", - "[-0.79,-0.71,0.47,-0.31,0.83,0.73,-0.98,0.8,0.72,-0.06,0.14,-0.12,-0.35,0.12,-0.32,0.64,-0.32,0.24,-0.98,-0.18],\n", - "[0.34,-0.22,-0.12,0.21,-0.7,-0.23,-0.94,-0.2,-0.63,-0.01,-0.4,0.83,0.78,-0.14,-0.06,0.49,0.95,0.22,-0.01,0.77],\n", - "[0.13,0.81,-0.9,-0.71,-0.43,0.03,-0.38,0.48,0.51,-0.35,0.65,-0.73,-0.69,0.54,0.61,0.75,0.96,-0.12,-0.72,0.43],\n", - "[-0.59,0.24,0.98,-0.66,-0.92,0.73,-0.18,0.84,-0.52,0.95,0.11,0.39,-0.53,0.41,-0.63,0.25,0.86,0.43,0.53,0.96],\n", - "[0.54,-0.05,-0.3,-0.94,0.55,0.81,-0.13,0.11,0.44,0.98,-0.09,-0.18,-0.91,0.16,-0.59,-0.03,-0.21,-0.79,-0.03,-0.56],\n", - "[-0.02,-0.65,-0.14,-0.94,0.87,-0.29,0.15,0.25,0.91,0.64,-0.13,0.38,0.02,-0.95,0.04,0.9,-0.55,0.96,-1,-0.55],\n", - "[-0.05,-0.79,-0.99,0.48,0.38,-0.45,-0.57,-0,-0.5,-0.69,0.93,-0.23,0.43,-0.19,-0.65,-0.7,0.91,-0.73,0.05,-0.14],\n", - "[0.53,-0.14,-0.76,-0.1,0.74,0.1,-0.37,-0.75,0.37,-0.97,0.94,0.83,-0.22,0.3,-0.5,0.59,0.29,-0.5,-0.26,-0.06],\n", - "[-0.95,-0.12,0.95,-0.9,0.49,-0.73,0.45,-0.57,0.02,0.41,0.21,0.98,-0.58,-0.96,0.58,-0.62,-0.51,-0.2,-0.1,0.67],\n", - "[0.26,-0.3,0.63,0.15,-0.01,-0.87,0.67,0.47,0.67,0.33,-0.28,-0.43,-0.79,-0.82,-0.45,0.23,-0.59,0.46,-0.59,0.79],\n", - "[0.4,0.99,0.02,-0.41,0.08,-0.93,0.93,-0.33,0.13,-0.77,-0.54,-0.32,0.72,-0.39,-0.13,0.58,0.51,-0.4,0.05,-0.42],\n", - "[0.34,0.24,0.39,0.51,-0.95,-0.04,-0.46,-0.49,0.29,-0.54,-0.08,-0.61,0.37,-0.4,-0.75,-0.98,0.34,-0.37,-0.32,-0.57],\n", - "[0.09,-0.75,0.48,-0.84,0.56,-0.64,0.71,0.52,-0.21,-0.83,-0.18,0.06,-0.16,0.22,-0.08,-0.71,0.61,-0.88,-0.38,-0.42],\n", - "[0.93,-0.05,0.81,0.1,-0.69,0.31,0.75,-0.32,0.25,0.47,0.27,0.45,-0.63,0.27,-0.58,-0.77,0.94,-0.15,-0.77,0.01],\n", - "[-0.04,-0.2,-0.19,0.66,0.59,-0.33,-0.41,0.91,0.23,-0.28,-0.64,0.38,-0.8,-0.19,0.07,-0.62,0.99,-0.16,-0.36,-0.65],\n", - "[-0.45,-0.85,0.45,-0.55,0.08,0.68,0.38,0.62,0.86,-0.76,-0,-0.3,0.43,-0.59,0.82,-0.67,0.67,0.42,0.31,0.9]\n", - "]\n", - "\n", - "\n", - "\n", - "254->248\n", - "\n", - "\n", - "19:w19:[20, 20]:[\n", - "[0.3,-0.36,0.82,0.87,0.8,0.3,-0.49,0.45,-0.46,-0.68,0.93,-0.14,-0.34,-0.65,-0.97,0.42,0.46,-0.38,0.88,0.54],\n", - "[-0.66,0.31,0.86,0.29,0.11,-0.1,0.63,0.15,0.93,-0.8,-0.42,-0.1,-0.48,0.64,0.55,-0.88,-0.33,-0.29,-0.13,0.93],\n", - "[-0.23,0.5,-0.25,0.43,0.09,-0.34,-0.93,-0.59,-0.63,0.37,0.14,-0.95,-0.1,0.4,0.35,0.43,-0.02,-0.43,0.38,0.87],\n", - "[-0.23,0.78,0.97,0.71,-0.31,-0.01,0.17,0.92,0.73,-0.46,0.42,-0.87,0.1,0.49,-0.01,-0.66,-0.59,0.87,-0.34,-0.14],\n", - "[-0.66,0.2,-0.08,-0.03,-0.27,-0.19,-0.5,-0.8,0.24,0.1,-0.48,0.91,0.31,-0.61,0.15,-0.15,0.95,0.52,-0.62,0.95],\n", - "[-0.7,-0.88,-0.93,0.65,-0.03,0.93,-0.51,0.26,0.73,-0.29,0.15,0.39,0.06,0.68,-0.22,0.23,-0.78,0.71,0.24,-0.98],\n", - "[-0,0.06,-0.82,-0.16,0.12,0.84,-0.68,-0.36,-0.45,-0.73,-0.44,0.35,0.86,0.79,-1,-0.41,-0.32,-0.23,-0.29,0.92],\n", - "[-0.4,-0.83,0.86,-0.84,0.43,0.56,0.91,-0.25,0.03,0.54,0.28,-0.87,-0.62,0.81,0.18,0.39,-0.61,-0.97,0.24,0.76],\n", - "[0.13,0.52,-0.23,0.91,-0.58,-0.42,-0.25,0.29,0.82,-0.56,-0.12,0.43,0.53,-0.7,0.34,0.63,-0.96,-0.92,0.39,-0.18],\n", - "[-0.28,-0.08,0.4,-0.51,0.18,-0.95,-0.62,-0.02,-0.35,0.93,-0.31,-0.69,0.51,-0.54,-0.68,-0.34,0.35,-0.68,0.84,-0.56],\n", - "[-0.83,-0.38,-0.27,-0.99,-0.1,0.43,-0.89,0.93,-0.78,-0.14,0.02,-0.03,0.79,-0.92,0.37,-0.78,0.98,-0.27,0.63,0.68],\n", - "[1,0.21,-0.53,0.37,0.91,0.22,0.85,-0.01,-0.34,-0.67,-0.14,0.85,0.06,0.41,0.61,0.23,-0.57,-0.98,-0.14,-0.38],\n", - "[0.53,-0.77,-0.36,-0.62,0.27,0.72,-0.7,-0.42,0.84,-0.45,0.2,-0.25,-0.22,-0.82,0.88,0.08,-0.27,0.51,-0.2,-0.71],\n", - "[-0.09,-0.42,-0.86,0.58,0.28,0.1,-0.83,-0.24,0.14,-0.1,0.25,0.04,0.79,-0.97,0.86,0.7,-0.21,-0.51,0.47,-0.65],\n", - "[-0.61,0.92,0.7,0.6,0.78,-0.61,-0.3,0.18,-0.15,0.68,0.55,-0.9,0.48,0.34,0.8,0.63,0.49,-0.66,-0.42,0.65],\n", - "[0.79,-0.93,0.08,0.13,0.62,0.03,-0.58,0.49,0.18,-0.19,0.75,0.84,-0.48,0.43,-0.51,0.34,-0.02,0.61,-0.29,-0.88],\n", - "[-0.63,-0.64,0.91,-0.31,0.42,0.32,0.39,-0.81,0.08,-0.09,0.15,-0.96,0.84,0.98,-0.94,-0.3,0.41,-0.26,-0.04,0.02],\n", - "[0.55,0.37,0.41,0.26,0.47,-0.17,-0.37,-0.27,-0.3,0.42,0.63,-0.45,0.87,-0.49,-0.83,-0.77,-0.22,-0.62,0.99,-0.75],\n", - "[0.57,-0.1,0.53,0.47,-0.07,0.34,-0.94,0.3,-0.12,0.73,-0.66,0.75,-0.15,0.96,-0.51,0.71,-0.46,-0.49,-0.72,-0.25],\n", - "[0.36,0.19,0.62,0.48,-0.81,-0.96,-0.32,-0.63,-0.58,-0.37,-0.17,-0.14,-0.48,-0.53,-0.63,0.58,-0.21,0.32,0.81,-0.5]\n", - "]\n", - "\n", - "\n", - "\n", - "257\n", - "\n", - "RuleNeuron:257:lstm3_out__left(T) :- lstm3_out__f(T), lstm3_out__c(Z), @next(Z, T). [transformation=identity, combination=elproduct]\n", - "val: [-0,-0,0,0,0,-0,-0,0,-0,-0,0,0,-0,0,0,0,0,-0,0,-0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "257->241\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "257->254\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "260\n", - "\n", - "WeightedRuleNeuron:260:lstm3_out__i(T) :- {20, 20} lstm2_out(T), {20, 20} lstm3_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.5,0.5,0.51,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "260->166\n", - "\n", - "\n", - "16:w16:[20, 20]:[\n", - "[-0.84,0.69,0.84,-0.81,0.03,0.06,0.4,0.31,0.02,-0.77,0.12,0.29,-0.97,-0.73,0.24,-0.18,0.88,0.49,0.13,-0.84],\n", - "[0.53,0.35,0.43,0.72,-0.31,0.28,0.45,0.2,-0.21,-0.69,-0.01,-0.18,-0.81,-0.15,0.04,-0.47,0.71,-0.84,0.23,0.26],\n", - "[-0.03,-0.21,-0.64,-0.6,-0.16,0.45,0.39,-1,-0.27,0.69,0.13,-0.21,0.22,-0.87,0.46,0.07,0.33,0.85,-0.03,-0.47],\n", - "[-0.63,0.25,0.38,0.24,-0.62,-0.33,-0.7,-0.2,0.08,0.49,-0.28,0.29,-0.93,0.22,0.71,-0.51,-0.71,-0.95,-0.15,-0.6],\n", - "[-0.13,0.49,-0.77,0.24,0.5,0.9,-0.32,0.86,0.93,-0.28,0.82,-0.52,0.31,0.89,0.11,-0.23,-0.28,0.58,0.27,0.49],\n", - "[-0.65,-0.26,-0.93,-0.24,-0.32,0.3,0.95,0.79,-0.89,-0.59,-0.45,0.5,-0.38,0.81,0.82,0.7,0.27,-0.13,-0.69,-0.07],\n", - "[0.49,-0.24,-0.34,0.53,0.61,-0.27,-0.38,0.14,0.19,-0.46,-0.67,-0.72,-0.57,-0.18,-1,-0.33,-0.72,-0.74,0.95,0.1],\n", - "[-0.58,-0.82,0.01,0.22,-0.96,-0.27,-0.37,0.66,-0.11,-0.58,0.15,-0.36,-0.14,-0.57,-0.15,0.47,-0.61,0.07,0.11,-0.05],\n", - "[0.34,0.39,0.95,0.38,0.59,-0.14,-0.9,-0.5,0.08,-0.77,0.97,-0.66,0.35,-0.63,-0.24,0.93,-0.34,-0.93,-0.71,0.53],\n", - "[-0.98,-0.76,-0.71,-0.99,-0.33,0.52,-0.9,-0.91,0.88,0.27,0.23,1,-0.73,0.69,-0.56,0.44,0.49,0.52,0.7,0.75],\n", - "[-0.8,-0.76,0.72,0.52,0.31,0.85,0.41,0.06,0.77,-0.54,0.38,-0.22,0.99,0.98,-0.85,-0.54,0.68,0.4,-0.33,-0.53],\n", - "[0.82,0.51,-0.64,-0.91,0.15,-0.95,-0.72,0.9,0.79,-0,0.25,0.82,-0.72,-0.34,-0.32,-0.77,0.66,-0.55,0.06,-0.88],\n", - "[0.18,-0.03,0.96,-0.45,-0.44,0.43,-0.84,0.98,0.7,0.7,-0.98,-0.26,-0.51,0.1,-0.64,-0.65,-0.75,-0.97,-0.12,0.87],\n", - "[0.21,0.34,-0.86,-0.54,-0.36,-0.8,0.07,0.67,0.93,-0.57,-0.68,-0.7,-0.5,0.18,-0.37,-0.23,0.3,0.18,-0.21,0.81],\n", - "[-0.02,0.57,0.98,-0.7,-0.89,-0.74,0.72,0.04,-0.56,-0.6,0.93,0.54,0.13,0.44,0.1,-0.3,-0.07,0.35,0.26,0.16],\n", - "[0.51,-0.48,-0.18,-0.49,-0.12,0.82,-0.89,0.83,0.4,0.75,-0.77,0.65,-0.93,0.36,-0.55,-0.4,-0.19,-0.6,-0.58,0.87],\n", - "[0.24,-0.93,0.16,0.99,0.63,-0.62,0.06,-0.68,0.23,-0.6,0.73,0.09,0.9,-0.51,-0.85,0.96,0.3,0.8,0.23,-0.36],\n", - "[-0.98,0.26,0.47,0.35,-0.72,-0.13,0.57,-0.62,0.08,-0.11,0.05,-0.19,0.28,0.39,-0.35,0.64,-0.87,0.37,-0.54,0.78],\n", - "[-0.24,0.85,0.42,-0.52,0.34,-0.83,0.36,0.22,0.34,0.86,0.5,0.33,0.08,0.61,0.91,0.06,-0.51,-0.94,-0.65,0.43],\n", - "[-0.24,-0.63,-0.38,0.41,0.4,0.99,0.62,0.09,-0.58,-0.86,0.2,-0.38,-0.04,0.41,-0.27,-0.05,-0.92,-0.06,-0.03,0.78]\n", - "]\n", - "\n", - "\n", - "\n", - "260->248\n", - "\n", - "\n", - "17:w17:[20, 20]:[\n", - "[-0.57,-0.37,-0.8,-0.7,-0.37,0.12,0.49,0.97,-0.37,-0.72,0.91,-0.17,-0.58,-0.47,-0.23,-0.36,-0.92,0.16,0.77,-0.5],\n", - "[-0.74,0.89,-0.1,-0.55,0.16,0.49,-0.1,0.22,-0.95,0.83,0.43,-0.88,0.64,0.25,0.86,0.93,-0.32,0.67,0.28,0.46],\n", - "[0.48,0.97,0.09,0.69,0.4,-0.66,-0.48,0.13,0.51,-0.33,0.01,0.45,0.47,-0.38,0.86,-0.64,-0.26,0.13,0.94,-0.87],\n", - "[0.09,-0.99,-0.88,-0.72,0.92,-0.55,-0.62,-0.36,0.49,-0.61,-0.4,-0.17,0.15,-0.09,-0.31,-0.55,-0.79,-0.09,0.48,-0.93],\n", - "[-0.89,-0.89,0.82,0.36,-0.51,0.11,-0.67,-0.11,0.73,-0.48,0.08,-0.87,0.63,0.83,-0.16,-0.09,-0.12,-0.83,0.51,0.54],\n", - "[-0.56,0.82,-0.37,0.45,0.91,-0.1,-0.92,-0.42,0.51,-0.26,0.61,-0.53,0.52,-0.23,-0.97,-0.66,0.46,0.32,-0.55,-0.57],\n", - "[1,0.9,0.64,0.29,0.92,-0.64,0.76,-0.16,-0.91,0.61,0.96,0.58,0.35,-0.47,-0.46,0.47,0.6,0.04,-0.58,-0.14],\n", - "[0.55,0.49,0.45,-0,-0.87,0.95,-0.92,-0.44,0.81,0.24,-0.43,-0.82,-0.15,0.11,-0.48,0.8,0.69,0.42,0.97,0.15],\n", - "[-0.23,0.89,-0.1,0.12,-0.71,-0.68,-0.94,0,0.55,0.38,-0.71,-0.86,-0.86,-0.74,-0.14,-0.71,0.12,0.82,-0.12,-0.87],\n", - "[-0.63,-0.63,-0.91,0.59,0.21,0.51,-0.93,-0.05,-0.58,-0.87,-0.59,0.73,-0.85,0.08,-0.96,-0.86,0.94,0.85,-0.36,0.82],\n", - "[-0.07,0.2,0.93,-0.45,0.07,-0.58,-0.75,0.76,-0.08,-0.23,-0.9,0.5,0.88,0.58,-0.12,-0.7,-0.77,-0.53,-0.47,-0.33],\n", - "[-0.99,0.91,0.38,-0.66,-0.83,0.82,-0.84,-0.24,-0.59,-0.2,0.03,0.53,0.68,0.7,0.85,0.3,0.13,-0.15,-0.69,-0.1],\n", - "[-0.7,0.23,-0.1,-0.68,-0.88,0.55,0.1,-0.43,-0.53,-0.68,-0.01,0.5,0.97,-0.98,-0.46,-0.82,-0.23,0.46,-0.82,-0.51],\n", - "[-0.45,-0.43,0.38,-0.71,0.87,0.23,-0.98,-0.1,0.51,0.71,-0.74,0.14,0.83,-0.92,0.31,-0.19,0.48,0.65,0.58,-0.78],\n", - "[0.5,0.85,-0.26,0.12,0.14,0.63,0.35,0.97,0.66,-0.42,-0.31,0.74,0.8,0.37,0.88,0.96,0.61,-0.01,0.56,0.93],\n", - "[0.76,-0.21,-0.42,-0.88,0.92,0.41,-0.99,0.64,0.05,0.36,-0.35,-0.9,-0.93,-0.73,0.27,0.32,0.92,-0.68,-0.89,-0.53],\n", - "[-0.41,-0.86,1,-0.96,-0.46,0.93,0.89,0.83,0.44,0.31,-0.09,-0.72,-0.73,-0.78,-0.75,-0.49,0.34,0.92,-0.41,-0.69],\n", - "[0.18,-0.8,0.55,0.98,-0.25,0.66,-0.37,0.52,-0.06,0.1,0.43,0.6,-0.88,0.04,-0.3,-0.41,-0.13,0.64,0.48,-0.22],\n", - "[0.25,-0.7,-0.49,-0.09,0.22,0.79,-0.52,0.34,-0.75,-0.24,-0.61,-0.95,-0.44,-0.91,-0.76,0.03,0.04,0.97,-0.58,-0.21],\n", - "[-0.73,0.98,-0.82,0.96,-0.15,-0.74,-0.3,0.47,0.96,0.2,0.61,0.99,0.59,0.1,0.93,-0.9,-0.99,-0.27,-0.04,0.81]\n", - "]\n", - "\n", - "\n", - "\n", - "263\n", - "\n", - "WeightedRuleNeuron:263:lstm3_out__n(T) :- {20, 20} lstm2_out(T), {20, 20} lstm3_out(Z), @next(Z, T). [transformation=tanh]\n", - "val: [0,-0.01,0,0.01,0.01,-0.02,-0.01,-0,-0,-0,0.01,0,-0,0.02,0,0.02,0.01,-0.01,0.01,-0.02]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Tanh\n", - "\n", - "\n", - "\n", - "263->166\n", - "\n", - "\n", - "22:w22:[20, 20]:[\n", - "[-0.98,-0.36,-0.06,-0.54,-0.52,0.49,-0.48,0.79,-0.96,0.02,-0.42,-0.91,-0.26,0.68,-0.71,0.88,0.71,0.75,-0.19,0.23],\n", - "[0.76,0.01,0.73,-0.07,0.09,0.35,-0.03,0.97,-0.41,0.07,0.69,-0.05,-0.57,0.49,-0.97,-0.1,-0.57,0.31,-0.04,-0.46],\n", - "[-0.26,-0.97,0.22,-0.27,0.49,0.14,-0.32,-0.26,-0.04,0.17,-0.61,-0.16,0.44,-0.27,-0,0.44,0.67,-0.03,0.3,0.3],\n", - "[0.54,-0.06,0.31,0.13,0.8,0.01,0.29,0.1,0.14,0.47,0.96,-0.01,0.32,0.38,0.52,0.98,0.86,-0.31,0.62,0.05],\n", - "[0.21,-0.14,-0.75,0.95,0.01,0.2,0.2,-0.57,-0.04,-0.53,-0.73,-0.51,-0.48,0.54,0.74,-0.55,0.59,0.13,-0.96,-0.24],\n", - "[-0.92,0.19,0.85,0.65,0.66,-0.18,-0.32,0.86,0.63,-0.76,0.29,-0.71,0.25,-0.18,0.23,0.18,-0.04,0.83,0.26,-0.35],\n", - "[-0.85,0.41,0.4,0.19,-0.44,0.79,-0.03,0.6,-0.4,0.8,0.29,0.59,-0.52,-0.4,-0.93,-0.45,0.5,0.39,-0.77,-0.89],\n", - "[-0.71,0.01,0.44,-0.34,0.65,-0.66,0.18,0.97,-0.99,0.53,0.1,0.08,0.77,0.52,0.93,-0.77,0.93,-0.9,-0.53,0.2],\n", - "[0.98,-0.28,-0.51,-0.95,-0.62,0.08,0.01,0.76,0.48,-0.68,-0.08,-0.78,-0.72,0.06,0.33,-0.32,-0.98,-0.85,-0.98,0.72],\n", - "[-0.56,0.42,0.71,0.57,0.01,0.9,-0.98,0.88,0.64,-0.82,0.42,0.67,-0.23,0.77,-0.11,-0.15,-0.74,-0.77,-0.54,-0.75],\n", - "[-0.15,-0.47,-0.79,0.91,-0.34,0.33,-0.78,0.24,-0.16,-0.03,0.24,0.37,0.94,0.92,0.83,-0.92,0.7,-0.2,-0.62,-0.43],\n", - "[0.36,0.32,0.43,0.51,0.91,0.8,0.29,-0.27,0.08,0.25,-0.64,0.37,0.56,-0.4,-0.89,-0.85,0.33,-0.85,-0.61,-0.84],\n", - "[0.62,-0.36,-0.97,0.58,0.89,-0.53,-0.02,-0.65,-0.4,0.75,-0.54,-0.81,0.92,0.19,0.79,-0.87,-0.26,-0.97,-0.64,-0.1],\n", - "[-0.12,-0.16,-0.77,-0.21,-0.05,0.26,0.71,-0.34,0.74,-0.68,-0.42,0.27,0.78,0.7,0.76,0.33,-0.76,0.65,0.15,-0.93],\n", - "[-0.61,-0.61,0.16,0.79,-0.81,0.07,0.2,0.16,-0.98,0.83,0.44,0.9,0.84,0.63,0.04,-0.08,-0.13,-0.69,0.98,0.9],\n", - "[-0.88,-0.4,-0.99,0.94,0.83,0.47,0.04,0.04,-0.18,0.87,-0.14,0.62,-0,0.58,0.94,0.87,-0.37,0.72,-0.98,-0.74],\n", - "[0.47,0.14,0.82,-0.05,0.89,0.09,0.51,0.87,0.31,-0.85,0.29,0.59,0.1,0.77,-0.15,0.21,0.19,-0.32,0.32,-0.77],\n", - "[0.17,-0.51,0.76,0.2,0.53,0.21,-0.46,0.11,0.18,0.56,-0.02,0.44,0.77,0.25,-0.43,0.88,-0.24,-0.1,0.03,-0.12],\n", - "[-0.42,0.68,0.31,0.69,-0.97,0.13,0.22,-0.73,-0.57,-0.47,-0.15,0.6,-0.13,0.51,0.81,-0.44,-0.37,-0.63,0.14,0.63],\n", - "[0.71,0.54,-0.54,0.88,0.09,-0.83,-0.18,0.43,-0.4,-0.09,-0.01,-0.74,0.9,-0.14,0.4,0.92,0.6,0.57,-0.34,0.6]\n", - "]\n", - "\n", - "\n", - "\n", - "263->248\n", - "\n", - "\n", - "23:w23:[20, 20]:[\n", - "[-0.94,-0.68,0.32,0.7,-0.63,0.71,-0.19,0.17,0.41,-0.91,0.04,-0.25,0.03,-0.62,-0.84,0.23,-0.91,-0.5,-0.45,-0.77],\n", - "[-0.01,0.51,-0.34,0.82,-0.41,0.6,0.57,0.76,0.06,0.09,-0.87,-0.52,0.78,0.52,-0.15,-0.97,-0.67,0.33,-0.43,-0.07],\n", - "[-0.19,-0.46,0.82,0.22,-0.71,-0.64,0.4,0.43,0.44,0.73,0.91,-0.91,-0.38,-0.45,0.53,0.82,0.7,0.28,0.89,-0.06],\n", - "[0.96,-0.17,-0.52,0.09,-0.75,-0.63,0.56,-0.32,0.43,0.19,0.5,0.34,0.5,-0.21,0.01,-0.98,-0.02,0.61,0.84,0.03],\n", - "[0.86,0.14,0,0.89,0.26,-0.25,-0.8,-0.57,-0.87,-0.3,-0.57,-0.55,-0.68,0.6,0.46,0.06,0.36,-0.01,0.61,-0.54],\n", - "[-0.28,-0.62,0.7,0.19,-0.71,0,-0.69,0.52,0.34,-0.12,-0.32,-1,-0.38,0.08,-0.82,-0.99,-0.57,0.61,0.71,0.61],\n", - "[0.3,-0.45,0.65,-0.96,0.36,-0.07,0.19,-0.47,0.01,-0.04,0.76,0.51,-0.55,-0.58,-0.37,-0.24,-0.01,0.14,-0.81,0.35],\n", - "[0.34,-0.15,-0.56,0.06,-0.45,-0.12,-0.12,0.14,-0.57,0.82,0.41,0.62,-0.72,-0.61,0.35,-0.77,-0.33,-0.44,-0.86,0.72],\n", - "[-0.65,0.42,-0.59,-0.57,0.36,0.89,0.14,-0.51,0.83,-0.78,0.13,-0.44,-0.86,0.87,0.17,0.44,0.56,0.16,-0.15,0.72],\n", - "[0.94,-0.83,0.36,0.91,0.44,-0.88,-0.91,-0.19,0.97,-0.46,0.63,0.01,0.58,0.67,0.15,0.94,-0.42,0.62,-0.83,-0.79],\n", - "[0.36,-0.98,0.26,-0.61,-0.95,-0.71,0.22,0.28,-0.3,-0.34,-0.97,0.75,-0.91,-0.07,0.12,-0.41,-0.43,0.94,0.12,0.81],\n", - "[0.25,0.72,-0.17,0.46,-0.82,-0.72,-0.18,-0.04,0.9,-0.68,0.79,-0.22,-0.21,0.29,-0.46,0.58,-0.06,0.21,-0.57,-0.62],\n", - "[-0.4,0.65,-0.2,0.24,0.04,0.79,0.17,-0.67,0.55,-0.16,-0.22,0.06,0.53,-0.23,-0.8,0.93,0.71,-0.37,-0.96,0.25],\n", - "[-0.36,-0.23,-0.17,-0.68,-0.04,-0.52,0.44,-0.96,0.26,0.78,0.89,0.51,-0.44,0.69,-0.22,0.01,0.09,-0.72,-0.24,-0.16],\n", - "[0.5,0.54,0.45,-0.85,-0.96,0.75,0.51,-0.4,0.08,0.08,-0.72,-0.42,-0.07,-0.55,0.03,-0.13,0.65,-0.04,-0.92,0.15],\n", - "[0.45,-0.03,-0.81,-0.7,-0.98,-0.94,-0.84,-0.87,-0.8,-0.38,-0.77,0.81,-0.62,1,-0.05,-0.05,0.91,-0.87,-0.65,0.06],\n", - "[-0.72,0.5,0.37,-0.73,0.6,-0.29,0.59,0.59,0.54,-0.24,0.98,-0.49,0.9,0.87,-0.87,0.44,-0.4,-0.47,0.47,0.76],\n", - "[-0.87,0.19,0.94,-0.69,0.16,0.35,-0.51,-0.42,-0.59,0.31,0.06,-0.12,-0.32,-0.56,0.6,0.43,-0.52,-0.59,-0.66,-0.53],\n", - "[-0.43,-0.25,-0.84,0.18,0.43,-0.53,0.97,0.61,-0.38,-0.23,0.87,0.56,-0.01,-0.59,0.58,-0.24,-0.52,-0.61,0.08,-0.04],\n", - "[0.64,0.67,-0.76,-0.37,0.8,-0.37,-0.75,0.47,0.23,-0.87,0.3,-0.53,-0.15,0.4,0.83,0.42,-0.79,-0.09,0.17,-0.28]\n", - "]\n", - "\n", - "\n", - "\n", - "266\n", - "\n", - "RuleNeuron:266:lstm3_out__right(T) :- lstm3_out__i(T), lstm3_out__n(T). [transformation=identity, combination=elproduct]\n", - "val: [0,-0.01,0,0.01,0,-0.01,-0.01,-0,-0,-0,0,0,-0,0.01,0,0.01,0,-0,0,-0.01]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "266->260\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "266->263\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "269\n", - "\n", - "RuleNeuron:269:lstm3_out__c(T) :- lstm3_out__left(T), lstm3_out__right(T). [transformation=identity]\n", - "val: [0,-0.01,0,0.01,0,-0.01,-0.01,0,-0,-0,0,0,-0,0.01,0,0.01,0,-0,0.01,-0.01]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum\n", - "\n", - "\n", - "\n", - "269->257\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "269->266\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "273\n", - "\n", - "RuleNeuron:273:lstm3_out(T) :- lstm3_out__o(T), tanh(lstm3_out__c(T)). [transformation=identity, combination=elproduct]\n", - "val: [0,-0,0,0,0,-0,-0,0,-0,-0,0,0,-0,0.01,0,0,0,-0,0,-0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "273->251\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "273->269\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "276\n", - "\n", - "WeightedRuleNeuron:276:lstm3_out__o(T) :- {20, 20} lstm2_out(T), {20, 20} lstm3_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.49,0.5,0.5,0.5,0.51,0.5,0.5,0.5,0.5,0.5,0.5,0.49,0.49,0.5,0.5,0.49,0.5,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "276->191\n", - "\n", - "\n", - "20:w20:[20, 20]:[\n", - "[0.08,-0.47,0.4,0.1,0.97,-0.07,0.69,-0.43,0.15,0.43,-0.36,0.51,0.65,-0.34,-0.7,-0.26,-0.27,-0.07,0.88,-0.57],\n", - "[-0.6,-0.9,-0.6,-0.23,0.92,-0.98,0.52,0.36,0.74,-0.54,-0.85,-0.31,-0.09,-0.07,0.31,-0.1,-0.71,0.07,0.69,0.52],\n", - "[0.55,0.63,-0.65,0.84,0.23,-0.08,-0.81,-0.86,-0.87,-0.25,-0.15,0.97,-0.27,-0.04,-0.93,-0.37,0.07,-0.04,-0.34,-0.04],\n", - "[0.61,0.82,0.79,-0.54,-0.88,0.84,-0.68,-0.77,-0.19,-0.97,0.06,-0.13,0.95,0.19,-0.54,0.25,0.97,1,-0.27,-0.03],\n", - "[0.65,-0.17,-0.65,0.94,-0.24,-0.93,-0.83,0.7,0.98,-0.09,-0.47,0.87,0.22,-0.13,0.51,-0.66,-0.14,0.65,0.1,-0.17],\n", - "[0.68,0.22,-0.43,-0.39,-0.88,0.7,0.79,-0.57,0.62,-0.14,-0.26,-0.91,-0.16,-0.16,-0.59,-0.68,-0.02,-0.76,-0.37,0.03],\n", - "[0.23,0.42,-0.44,0.72,-0.69,0.72,-0.08,0.16,0.68,0.07,-0.48,-0.06,-0.68,0.24,-0.91,0.77,0.6,-0.79,-0.07,-0.75],\n", - "[-0.96,0.94,0.42,-0.5,0.86,0.86,-0.7,-0.79,0.35,0.83,-0.91,0.12,-0.55,-0.82,0.06,0.54,-0.98,0.91,0.7,-0.16],\n", - "[-0.95,-0.54,0.93,0.69,0.96,0.75,0.92,-0.61,-0.72,0.75,-0.42,0.3,0.89,-0.8,0.84,-0.62,-0.76,0.98,0.16,-0.51],\n", - "[-0.83,-0.08,0.51,-0.85,0.11,0.77,0.88,0.44,0.84,0.68,0.3,0.72,0.36,0.68,-0.99,1,-0.25,0.93,-0.3,0.96],\n", - "[0.58,-0.75,-0.47,0.83,0.85,-0.36,-0.07,-0.32,-0.85,-0.57,0.01,-0.04,0.35,0.99,-0.74,-0.33,0.97,-0.15,0.61,-0.05],\n", - "[0.05,0.8,0.23,-0.63,-0.51,-0.55,1,0.91,0.35,0.34,0.7,-0.65,-0.93,-0.98,0.1,-0.68,-0.71,-0.66,0.94,0.17],\n", - "[-0.87,-0.86,-0.69,0.9,-0.72,0.05,-0.24,-0.25,-0.04,0.89,0.24,0.73,-0.31,0.81,0.76,0.91,0.75,0.19,0.56,-0.48],\n", - "[-0.04,0.34,0.75,0.28,0.65,0.77,-0.02,-0.6,0.52,0,0.46,0.02,0.43,-0.16,0.97,-0.73,-0.85,0.29,0.57,0.69],\n", - "[0.59,0.43,-0.68,-0.19,-0.99,-0.19,0.15,-0.19,0.47,-0.64,0.05,-0.71,0.96,0.01,-0.45,-0.8,-0.86,0.01,0.75,-0.55],\n", - "[-0.53,0.96,-0.12,0.83,-0.3,-0.32,0.97,-0.69,-0.22,0.28,0.02,-0.76,0.81,-0.29,0.66,-0.63,0.42,-0.25,0.42,-0.78],\n", - "[-0.02,-0.7,-0.7,-0.99,-0.06,0.57,-0.41,-0.21,-0.53,0.07,-0.25,0.52,0.9,0.73,-0.6,0.23,0.31,-0.7,-0.65,-0.35],\n", - "[0.93,0.6,0.3,0.23,0.41,-0,-0.52,0.97,0.95,-0.76,-0.43,0.76,-0.88,0.6,0.95,-0.93,-0.69,0.26,0.43,0.85],\n", - "[-0.72,-0.47,-0.5,0.87,-0.89,-0.97,0.48,-0.06,0.1,0.43,-0.21,0.95,0.19,-0.99,0.1,-0.18,0.94,0.22,-0.9,0.31],\n", - "[0.21,-0.25,0.28,0.87,0.3,0.92,0.23,-0.57,-0.61,-0.25,-0.37,-0.2,-0.88,0.26,-0.34,0.96,0.13,-0.64,0.87,0.33]\n", - "]\n", - "\n", - "\n", - "\n", - "276->273\n", - "\n", - "\n", - "21:w21:[20, 20]:[\n", - "[-0.34,1,-0.36,0.04,-0.77,0.08,-0.31,-0.87,0.07,-0.75,-0.88,0.5,0.84,0.51,-0.77,-0.39,0.95,0.36,-0.75,-0.06],\n", - "[0.94,0.07,0.1,0.55,-0.99,-0.21,-0.57,0.96,0.44,-0.05,0.08,-0.55,-0.94,0.72,0.48,0.52,0.69,-0.61,0.55,0.49],\n", - "[0.5,0.58,0.06,-0.22,-0.53,-0.3,0.75,-0.05,0.63,-0.74,-0.18,-0.74,0.29,-1,-0.08,-0.01,-0.33,0.38,0.46,0.76],\n", - "[0.18,-0.64,-0.66,0.65,-0.84,-0.81,-0.83,-0.07,-0.51,0.16,-0.53,0.08,-0.64,-0.44,0.57,-0.2,0.15,0.26,-0.89,-0.38],\n", - "[0.44,0.61,0.73,0.18,0.91,-0.01,0.74,0.89,0.11,0.42,0.3,0.45,0.83,0.52,0.13,0.62,-0.64,0.48,-0.62,-0.67],\n", - "[0.65,0.42,-0.81,-0.85,0.07,-0.87,-0.92,0.33,0.36,0.19,-0.77,-0.6,-0.87,0.84,-0.44,-0.44,-0.53,-0.43,0.64,-0.6],\n", - "[-0.23,0.45,0.33,-0.67,-0.56,-0.8,-0.86,-0.05,-0.14,0.23,-0.33,-0.56,-0.35,0.65,-0.54,-0.31,0.77,0.04,0,0.59],\n", - "[-0.61,-0.67,-0.07,-0.04,-0.44,-0.78,0.97,0.67,0.44,-0.77,-0.51,0.91,-0.79,-0.41,0.14,-0.59,-0.68,-0.9,-0.26,0.2],\n", - "[0.46,-0.74,0.01,-0.72,-0.8,-0.03,0.71,0.63,-0.8,-0.47,0.37,0.9,-0.62,0.62,-0.33,-0.86,-0.24,-0.37,-0.48,0.05],\n", - "[-0.86,-0.14,0.77,0.55,-0.82,0.53,0.5,-0.56,-0.75,-0.64,0.02,-0.54,-0.23,-0.36,0.85,-0.71,0.66,0.78,0.12,0.52],\n", - "[0.06,0.61,-0.97,-0.98,0.68,-0.1,0.75,-0.21,0.4,0.99,0.93,0.94,-0.62,0.64,0,0.11,0.33,-0.99,0.76,-0.64],\n", - "[-0.8,0.59,-0.78,0.41,-0.95,-0.68,0.66,0.97,-0.22,-0.1,0.22,0.76,0.97,-0.26,0.32,0.86,-0.09,0.2,0.12,0.08],\n", - "[0.86,0.03,-0.08,-0.88,-0.53,0.48,-0.44,-0.51,-0.67,-0.22,0.78,-0.65,0.5,-0.56,0.76,0.91,0.4,-0.81,-0.02,0.97],\n", - "[-0.91,0.26,0.07,-0.79,0.8,-0.93,0.23,-0.27,0.31,0.51,0.53,0.31,-0.32,0.71,0.49,-0.7,0.86,-0.12,0.05,0.59],\n", - "[-0.52,-0.94,-0.64,-0.69,0.58,-0.44,-0.31,-0.96,-0.41,0.7,-0.05,-0.24,0.71,-0.94,0.8,0.46,0.6,-0.86,-0.91,0.11],\n", - "[0.79,-0.93,0.33,0.24,-0.33,0,0.7,-0.03,-0.84,-0.64,0.15,-0.17,-0.09,-0.47,-0.84,0.85,0.16,-0.79,0.69,0.66],\n", - "[-0.7,-0.05,-0.6,0.74,-0.32,0.38,-0.26,0.47,0.96,0.12,0.74,-0.12,-0.69,-0.86,-0.33,0.85,-0.68,0.27,0.02,0.82],\n", - "[-0.21,0.38,-0.99,-0.97,-0.82,-0.21,0.1,-0.17,0.96,-0.75,0.59,-0.51,0.9,0.26,0.36,-0.16,-0.22,0.4,-0.94,-0.23],\n", - "[-0.06,-0.45,0.18,-0.89,0.87,-0.68,-0.74,0.74,0.93,0.09,-0.94,0.59,0,0.75,-0.9,0.85,-0.57,0.34,-0.69,-0.7],\n", - "[-0.12,-0.32,-0.72,-0.31,0.7,-0.68,0.12,-0.31,-0.12,-0.99,0.84,-0.83,-0.56,0.5,-0.21,-0.74,-0.01,0.27,-0.41,0.88]\n", - "]\n", - "\n", - "\n", - "\n", - "279\n", - "\n", - "WeightedRuleNeuron:279:lstm3_out__f(T) :- {20, 20} lstm2_out(T), {20, 20} lstm3_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.51,0.5,0.5,0.5,0.49,0.49,0.49,0.5,0.49,0.5,0.5,0.51,0.5,0.49,0.5,0.5,0.49,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "279->191\n", - "\n", - "\n", - "18:w18:[20, 20]:[\n", - "[0.67,0.87,0.79,0.15,0.83,-0.69,0.82,-0.27,0.08,-0.53,0.14,-0.69,0.14,0.99,-0.98,-0.08,0.26,0.41,-0.22,0.17],\n", - "[0.93,0.81,-0.3,-0.88,0.68,0.61,0.6,-0.8,0.25,-1,-0.43,0.6,0.56,-0.19,0.98,-0.12,0.62,-0.12,0.34,-0.66],\n", - "[0.72,0.76,-0.82,-0.59,-0.87,0.71,-0.61,0.17,0.9,0.85,-0.35,-0.2,-0.92,-0.22,0.76,-0.01,0.69,-0.3,-0.03,-0.13],\n", - "[0.18,0.05,-0.36,0.68,0.29,-0.86,0.3,0.83,0.16,-0.54,0.17,-0.2,0.2,-0.84,-0.81,-0.96,0.9,-0.72,0.95,0.79],\n", - "[-0.79,-0.71,0.47,-0.31,0.83,0.73,-0.98,0.8,0.72,-0.06,0.14,-0.12,-0.35,0.12,-0.32,0.64,-0.32,0.24,-0.98,-0.18],\n", - "[0.34,-0.22,-0.12,0.21,-0.7,-0.23,-0.94,-0.2,-0.63,-0.01,-0.4,0.83,0.78,-0.14,-0.06,0.49,0.95,0.22,-0.01,0.77],\n", - "[0.13,0.81,-0.9,-0.71,-0.43,0.03,-0.38,0.48,0.51,-0.35,0.65,-0.73,-0.69,0.54,0.61,0.75,0.96,-0.12,-0.72,0.43],\n", - "[-0.59,0.24,0.98,-0.66,-0.92,0.73,-0.18,0.84,-0.52,0.95,0.11,0.39,-0.53,0.41,-0.63,0.25,0.86,0.43,0.53,0.96],\n", - "[0.54,-0.05,-0.3,-0.94,0.55,0.81,-0.13,0.11,0.44,0.98,-0.09,-0.18,-0.91,0.16,-0.59,-0.03,-0.21,-0.79,-0.03,-0.56],\n", - "[-0.02,-0.65,-0.14,-0.94,0.87,-0.29,0.15,0.25,0.91,0.64,-0.13,0.38,0.02,-0.95,0.04,0.9,-0.55,0.96,-1,-0.55],\n", - "[-0.05,-0.79,-0.99,0.48,0.38,-0.45,-0.57,-0,-0.5,-0.69,0.93,-0.23,0.43,-0.19,-0.65,-0.7,0.91,-0.73,0.05,-0.14],\n", - "[0.53,-0.14,-0.76,-0.1,0.74,0.1,-0.37,-0.75,0.37,-0.97,0.94,0.83,-0.22,0.3,-0.5,0.59,0.29,-0.5,-0.26,-0.06],\n", - "[-0.95,-0.12,0.95,-0.9,0.49,-0.73,0.45,-0.57,0.02,0.41,0.21,0.98,-0.58,-0.96,0.58,-0.62,-0.51,-0.2,-0.1,0.67],\n", - "[0.26,-0.3,0.63,0.15,-0.01,-0.87,0.67,0.47,0.67,0.33,-0.28,-0.43,-0.79,-0.82,-0.45,0.23,-0.59,0.46,-0.59,0.79],\n", - "[0.4,0.99,0.02,-0.41,0.08,-0.93,0.93,-0.33,0.13,-0.77,-0.54,-0.32,0.72,-0.39,-0.13,0.58,0.51,-0.4,0.05,-0.42],\n", - "[0.34,0.24,0.39,0.51,-0.95,-0.04,-0.46,-0.49,0.29,-0.54,-0.08,-0.61,0.37,-0.4,-0.75,-0.98,0.34,-0.37,-0.32,-0.57],\n", - "[0.09,-0.75,0.48,-0.84,0.56,-0.64,0.71,0.52,-0.21,-0.83,-0.18,0.06,-0.16,0.22,-0.08,-0.71,0.61,-0.88,-0.38,-0.42],\n", - "[0.93,-0.05,0.81,0.1,-0.69,0.31,0.75,-0.32,0.25,0.47,0.27,0.45,-0.63,0.27,-0.58,-0.77,0.94,-0.15,-0.77,0.01],\n", - "[-0.04,-0.2,-0.19,0.66,0.59,-0.33,-0.41,0.91,0.23,-0.28,-0.64,0.38,-0.8,-0.19,0.07,-0.62,0.99,-0.16,-0.36,-0.65],\n", - "[-0.45,-0.85,0.45,-0.55,0.08,0.68,0.38,0.62,0.86,-0.76,-0,-0.3,0.43,-0.59,0.82,-0.67,0.67,0.42,0.31,0.9]\n", - "]\n", - "\n", - "\n", - "\n", - "279->273\n", - "\n", - "\n", - "19:w19:[20, 20]:[\n", - "[0.3,-0.36,0.82,0.87,0.8,0.3,-0.49,0.45,-0.46,-0.68,0.93,-0.14,-0.34,-0.65,-0.97,0.42,0.46,-0.38,0.88,0.54],\n", - "[-0.66,0.31,0.86,0.29,0.11,-0.1,0.63,0.15,0.93,-0.8,-0.42,-0.1,-0.48,0.64,0.55,-0.88,-0.33,-0.29,-0.13,0.93],\n", - "[-0.23,0.5,-0.25,0.43,0.09,-0.34,-0.93,-0.59,-0.63,0.37,0.14,-0.95,-0.1,0.4,0.35,0.43,-0.02,-0.43,0.38,0.87],\n", - "[-0.23,0.78,0.97,0.71,-0.31,-0.01,0.17,0.92,0.73,-0.46,0.42,-0.87,0.1,0.49,-0.01,-0.66,-0.59,0.87,-0.34,-0.14],\n", - "[-0.66,0.2,-0.08,-0.03,-0.27,-0.19,-0.5,-0.8,0.24,0.1,-0.48,0.91,0.31,-0.61,0.15,-0.15,0.95,0.52,-0.62,0.95],\n", - "[-0.7,-0.88,-0.93,0.65,-0.03,0.93,-0.51,0.26,0.73,-0.29,0.15,0.39,0.06,0.68,-0.22,0.23,-0.78,0.71,0.24,-0.98],\n", - "[-0,0.06,-0.82,-0.16,0.12,0.84,-0.68,-0.36,-0.45,-0.73,-0.44,0.35,0.86,0.79,-1,-0.41,-0.32,-0.23,-0.29,0.92],\n", - "[-0.4,-0.83,0.86,-0.84,0.43,0.56,0.91,-0.25,0.03,0.54,0.28,-0.87,-0.62,0.81,0.18,0.39,-0.61,-0.97,0.24,0.76],\n", - "[0.13,0.52,-0.23,0.91,-0.58,-0.42,-0.25,0.29,0.82,-0.56,-0.12,0.43,0.53,-0.7,0.34,0.63,-0.96,-0.92,0.39,-0.18],\n", - "[-0.28,-0.08,0.4,-0.51,0.18,-0.95,-0.62,-0.02,-0.35,0.93,-0.31,-0.69,0.51,-0.54,-0.68,-0.34,0.35,-0.68,0.84,-0.56],\n", - "[-0.83,-0.38,-0.27,-0.99,-0.1,0.43,-0.89,0.93,-0.78,-0.14,0.02,-0.03,0.79,-0.92,0.37,-0.78,0.98,-0.27,0.63,0.68],\n", - "[1,0.21,-0.53,0.37,0.91,0.22,0.85,-0.01,-0.34,-0.67,-0.14,0.85,0.06,0.41,0.61,0.23,-0.57,-0.98,-0.14,-0.38],\n", - "[0.53,-0.77,-0.36,-0.62,0.27,0.72,-0.7,-0.42,0.84,-0.45,0.2,-0.25,-0.22,-0.82,0.88,0.08,-0.27,0.51,-0.2,-0.71],\n", - "[-0.09,-0.42,-0.86,0.58,0.28,0.1,-0.83,-0.24,0.14,-0.1,0.25,0.04,0.79,-0.97,0.86,0.7,-0.21,-0.51,0.47,-0.65],\n", - "[-0.61,0.92,0.7,0.6,0.78,-0.61,-0.3,0.18,-0.15,0.68,0.55,-0.9,0.48,0.34,0.8,0.63,0.49,-0.66,-0.42,0.65],\n", - "[0.79,-0.93,0.08,0.13,0.62,0.03,-0.58,0.49,0.18,-0.19,0.75,0.84,-0.48,0.43,-0.51,0.34,-0.02,0.61,-0.29,-0.88],\n", - "[-0.63,-0.64,0.91,-0.31,0.42,0.32,0.39,-0.81,0.08,-0.09,0.15,-0.96,0.84,0.98,-0.94,-0.3,0.41,-0.26,-0.04,0.02],\n", - "[0.55,0.37,0.41,0.26,0.47,-0.17,-0.37,-0.27,-0.3,0.42,0.63,-0.45,0.87,-0.49,-0.83,-0.77,-0.22,-0.62,0.99,-0.75],\n", - "[0.57,-0.1,0.53,0.47,-0.07,0.34,-0.94,0.3,-0.12,0.73,-0.66,0.75,-0.15,0.96,-0.51,0.71,-0.46,-0.49,-0.72,-0.25],\n", - "[0.36,0.19,0.62,0.48,-0.81,-0.96,-0.32,-0.63,-0.58,-0.37,-0.17,-0.14,-0.48,-0.53,-0.63,0.58,-0.21,0.32,0.81,-0.5]\n", - "]\n", - "\n", - "\n", - "\n", - "282\n", - "\n", - "RuleNeuron:282:lstm3_out__left(T) :- lstm3_out__f(T), lstm3_out__c(Z), @next(Z, T). [transformation=identity, combination=elproduct]\n", - "val: [0,-0,0,0,0,-0,-0,0,-0,-0,0,0,-0,0.01,0,0,0,-0,0,-0]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "282->269\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "282->279\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "285\n", - "\n", - "WeightedRuleNeuron:285:lstm3_out__i(T) :- {20, 20} lstm2_out(T), {20, 20} lstm3_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.5,0.49,0.5,0.51,0.49,0.5,0.5,0.51,0.49,0.49,0.5,0.49,0.51,0.51,0.5,0.5,0.51,0.49]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "285->191\n", - "\n", - "\n", - "16:w16:[20, 20]:[\n", - "[-0.84,0.69,0.84,-0.81,0.03,0.06,0.4,0.31,0.02,-0.77,0.12,0.29,-0.97,-0.73,0.24,-0.18,0.88,0.49,0.13,-0.84],\n", - "[0.53,0.35,0.43,0.72,-0.31,0.28,0.45,0.2,-0.21,-0.69,-0.01,-0.18,-0.81,-0.15,0.04,-0.47,0.71,-0.84,0.23,0.26],\n", - "[-0.03,-0.21,-0.64,-0.6,-0.16,0.45,0.39,-1,-0.27,0.69,0.13,-0.21,0.22,-0.87,0.46,0.07,0.33,0.85,-0.03,-0.47],\n", - "[-0.63,0.25,0.38,0.24,-0.62,-0.33,-0.7,-0.2,0.08,0.49,-0.28,0.29,-0.93,0.22,0.71,-0.51,-0.71,-0.95,-0.15,-0.6],\n", - "[-0.13,0.49,-0.77,0.24,0.5,0.9,-0.32,0.86,0.93,-0.28,0.82,-0.52,0.31,0.89,0.11,-0.23,-0.28,0.58,0.27,0.49],\n", - "[-0.65,-0.26,-0.93,-0.24,-0.32,0.3,0.95,0.79,-0.89,-0.59,-0.45,0.5,-0.38,0.81,0.82,0.7,0.27,-0.13,-0.69,-0.07],\n", - "[0.49,-0.24,-0.34,0.53,0.61,-0.27,-0.38,0.14,0.19,-0.46,-0.67,-0.72,-0.57,-0.18,-1,-0.33,-0.72,-0.74,0.95,0.1],\n", - "[-0.58,-0.82,0.01,0.22,-0.96,-0.27,-0.37,0.66,-0.11,-0.58,0.15,-0.36,-0.14,-0.57,-0.15,0.47,-0.61,0.07,0.11,-0.05],\n", - "[0.34,0.39,0.95,0.38,0.59,-0.14,-0.9,-0.5,0.08,-0.77,0.97,-0.66,0.35,-0.63,-0.24,0.93,-0.34,-0.93,-0.71,0.53],\n", - "[-0.98,-0.76,-0.71,-0.99,-0.33,0.52,-0.9,-0.91,0.88,0.27,0.23,1,-0.73,0.69,-0.56,0.44,0.49,0.52,0.7,0.75],\n", - "[-0.8,-0.76,0.72,0.52,0.31,0.85,0.41,0.06,0.77,-0.54,0.38,-0.22,0.99,0.98,-0.85,-0.54,0.68,0.4,-0.33,-0.53],\n", - "[0.82,0.51,-0.64,-0.91,0.15,-0.95,-0.72,0.9,0.79,-0,0.25,0.82,-0.72,-0.34,-0.32,-0.77,0.66,-0.55,0.06,-0.88],\n", - "[0.18,-0.03,0.96,-0.45,-0.44,0.43,-0.84,0.98,0.7,0.7,-0.98,-0.26,-0.51,0.1,-0.64,-0.65,-0.75,-0.97,-0.12,0.87],\n", - "[0.21,0.34,-0.86,-0.54,-0.36,-0.8,0.07,0.67,0.93,-0.57,-0.68,-0.7,-0.5,0.18,-0.37,-0.23,0.3,0.18,-0.21,0.81],\n", - "[-0.02,0.57,0.98,-0.7,-0.89,-0.74,0.72,0.04,-0.56,-0.6,0.93,0.54,0.13,0.44,0.1,-0.3,-0.07,0.35,0.26,0.16],\n", - "[0.51,-0.48,-0.18,-0.49,-0.12,0.82,-0.89,0.83,0.4,0.75,-0.77,0.65,-0.93,0.36,-0.55,-0.4,-0.19,-0.6,-0.58,0.87],\n", - "[0.24,-0.93,0.16,0.99,0.63,-0.62,0.06,-0.68,0.23,-0.6,0.73,0.09,0.9,-0.51,-0.85,0.96,0.3,0.8,0.23,-0.36],\n", - "[-0.98,0.26,0.47,0.35,-0.72,-0.13,0.57,-0.62,0.08,-0.11,0.05,-0.19,0.28,0.39,-0.35,0.64,-0.87,0.37,-0.54,0.78],\n", - "[-0.24,0.85,0.42,-0.52,0.34,-0.83,0.36,0.22,0.34,0.86,0.5,0.33,0.08,0.61,0.91,0.06,-0.51,-0.94,-0.65,0.43],\n", - "[-0.24,-0.63,-0.38,0.41,0.4,0.99,0.62,0.09,-0.58,-0.86,0.2,-0.38,-0.04,0.41,-0.27,-0.05,-0.92,-0.06,-0.03,0.78]\n", - "]\n", - "\n", - "\n", - "\n", - "285->273\n", - "\n", - "\n", - "17:w17:[20, 20]:[\n", - "[-0.57,-0.37,-0.8,-0.7,-0.37,0.12,0.49,0.97,-0.37,-0.72,0.91,-0.17,-0.58,-0.47,-0.23,-0.36,-0.92,0.16,0.77,-0.5],\n", - "[-0.74,0.89,-0.1,-0.55,0.16,0.49,-0.1,0.22,-0.95,0.83,0.43,-0.88,0.64,0.25,0.86,0.93,-0.32,0.67,0.28,0.46],\n", - "[0.48,0.97,0.09,0.69,0.4,-0.66,-0.48,0.13,0.51,-0.33,0.01,0.45,0.47,-0.38,0.86,-0.64,-0.26,0.13,0.94,-0.87],\n", - "[0.09,-0.99,-0.88,-0.72,0.92,-0.55,-0.62,-0.36,0.49,-0.61,-0.4,-0.17,0.15,-0.09,-0.31,-0.55,-0.79,-0.09,0.48,-0.93],\n", - "[-0.89,-0.89,0.82,0.36,-0.51,0.11,-0.67,-0.11,0.73,-0.48,0.08,-0.87,0.63,0.83,-0.16,-0.09,-0.12,-0.83,0.51,0.54],\n", - "[-0.56,0.82,-0.37,0.45,0.91,-0.1,-0.92,-0.42,0.51,-0.26,0.61,-0.53,0.52,-0.23,-0.97,-0.66,0.46,0.32,-0.55,-0.57],\n", - "[1,0.9,0.64,0.29,0.92,-0.64,0.76,-0.16,-0.91,0.61,0.96,0.58,0.35,-0.47,-0.46,0.47,0.6,0.04,-0.58,-0.14],\n", - "[0.55,0.49,0.45,-0,-0.87,0.95,-0.92,-0.44,0.81,0.24,-0.43,-0.82,-0.15,0.11,-0.48,0.8,0.69,0.42,0.97,0.15],\n", - "[-0.23,0.89,-0.1,0.12,-0.71,-0.68,-0.94,0,0.55,0.38,-0.71,-0.86,-0.86,-0.74,-0.14,-0.71,0.12,0.82,-0.12,-0.87],\n", - "[-0.63,-0.63,-0.91,0.59,0.21,0.51,-0.93,-0.05,-0.58,-0.87,-0.59,0.73,-0.85,0.08,-0.96,-0.86,0.94,0.85,-0.36,0.82],\n", - "[-0.07,0.2,0.93,-0.45,0.07,-0.58,-0.75,0.76,-0.08,-0.23,-0.9,0.5,0.88,0.58,-0.12,-0.7,-0.77,-0.53,-0.47,-0.33],\n", - "[-0.99,0.91,0.38,-0.66,-0.83,0.82,-0.84,-0.24,-0.59,-0.2,0.03,0.53,0.68,0.7,0.85,0.3,0.13,-0.15,-0.69,-0.1],\n", - "[-0.7,0.23,-0.1,-0.68,-0.88,0.55,0.1,-0.43,-0.53,-0.68,-0.01,0.5,0.97,-0.98,-0.46,-0.82,-0.23,0.46,-0.82,-0.51],\n", - "[-0.45,-0.43,0.38,-0.71,0.87,0.23,-0.98,-0.1,0.51,0.71,-0.74,0.14,0.83,-0.92,0.31,-0.19,0.48,0.65,0.58,-0.78],\n", - "[0.5,0.85,-0.26,0.12,0.14,0.63,0.35,0.97,0.66,-0.42,-0.31,0.74,0.8,0.37,0.88,0.96,0.61,-0.01,0.56,0.93],\n", - "[0.76,-0.21,-0.42,-0.88,0.92,0.41,-0.99,0.64,0.05,0.36,-0.35,-0.9,-0.93,-0.73,0.27,0.32,0.92,-0.68,-0.89,-0.53],\n", - "[-0.41,-0.86,1,-0.96,-0.46,0.93,0.89,0.83,0.44,0.31,-0.09,-0.72,-0.73,-0.78,-0.75,-0.49,0.34,0.92,-0.41,-0.69],\n", - "[0.18,-0.8,0.55,0.98,-0.25,0.66,-0.37,0.52,-0.06,0.1,0.43,0.6,-0.88,0.04,-0.3,-0.41,-0.13,0.64,0.48,-0.22],\n", - "[0.25,-0.7,-0.49,-0.09,0.22,0.79,-0.52,0.34,-0.75,-0.24,-0.61,-0.95,-0.44,-0.91,-0.76,0.03,0.04,0.97,-0.58,-0.21],\n", - "[-0.73,0.98,-0.82,0.96,-0.15,-0.74,-0.3,0.47,0.96,0.2,0.61,0.99,0.59,0.1,0.93,-0.9,-0.99,-0.27,-0.04,0.81]\n", - "]\n", - "\n", - "\n", - "\n", - "288\n", - "\n", - "WeightedRuleNeuron:288:lstm3_out__n(T) :- {20, 20} lstm2_out(T), {20, 20} lstm3_out(Z), @next(Z, T). [transformation=tanh]\n", - "val: [0.01,-0.03,-0,0.02,0.01,-0.04,-0.04,0,-0.01,-0,0.01,0.01,-0.01,0.04,0.01,0.05,0.01,-0,0.01,-0.04]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Tanh\n", - "\n", - "\n", - "\n", - "288->191\n", - "\n", - "\n", - "22:w22:[20, 20]:[\n", - "[-0.98,-0.36,-0.06,-0.54,-0.52,0.49,-0.48,0.79,-0.96,0.02,-0.42,-0.91,-0.26,0.68,-0.71,0.88,0.71,0.75,-0.19,0.23],\n", - "[0.76,0.01,0.73,-0.07,0.09,0.35,-0.03,0.97,-0.41,0.07,0.69,-0.05,-0.57,0.49,-0.97,-0.1,-0.57,0.31,-0.04,-0.46],\n", - "[-0.26,-0.97,0.22,-0.27,0.49,0.14,-0.32,-0.26,-0.04,0.17,-0.61,-0.16,0.44,-0.27,-0,0.44,0.67,-0.03,0.3,0.3],\n", - "[0.54,-0.06,0.31,0.13,0.8,0.01,0.29,0.1,0.14,0.47,0.96,-0.01,0.32,0.38,0.52,0.98,0.86,-0.31,0.62,0.05],\n", - "[0.21,-0.14,-0.75,0.95,0.01,0.2,0.2,-0.57,-0.04,-0.53,-0.73,-0.51,-0.48,0.54,0.74,-0.55,0.59,0.13,-0.96,-0.24],\n", - "[-0.92,0.19,0.85,0.65,0.66,-0.18,-0.32,0.86,0.63,-0.76,0.29,-0.71,0.25,-0.18,0.23,0.18,-0.04,0.83,0.26,-0.35],\n", - "[-0.85,0.41,0.4,0.19,-0.44,0.79,-0.03,0.6,-0.4,0.8,0.29,0.59,-0.52,-0.4,-0.93,-0.45,0.5,0.39,-0.77,-0.89],\n", - "[-0.71,0.01,0.44,-0.34,0.65,-0.66,0.18,0.97,-0.99,0.53,0.1,0.08,0.77,0.52,0.93,-0.77,0.93,-0.9,-0.53,0.2],\n", - "[0.98,-0.28,-0.51,-0.95,-0.62,0.08,0.01,0.76,0.48,-0.68,-0.08,-0.78,-0.72,0.06,0.33,-0.32,-0.98,-0.85,-0.98,0.72],\n", - "[-0.56,0.42,0.71,0.57,0.01,0.9,-0.98,0.88,0.64,-0.82,0.42,0.67,-0.23,0.77,-0.11,-0.15,-0.74,-0.77,-0.54,-0.75],\n", - "[-0.15,-0.47,-0.79,0.91,-0.34,0.33,-0.78,0.24,-0.16,-0.03,0.24,0.37,0.94,0.92,0.83,-0.92,0.7,-0.2,-0.62,-0.43],\n", - "[0.36,0.32,0.43,0.51,0.91,0.8,0.29,-0.27,0.08,0.25,-0.64,0.37,0.56,-0.4,-0.89,-0.85,0.33,-0.85,-0.61,-0.84],\n", - "[0.62,-0.36,-0.97,0.58,0.89,-0.53,-0.02,-0.65,-0.4,0.75,-0.54,-0.81,0.92,0.19,0.79,-0.87,-0.26,-0.97,-0.64,-0.1],\n", - "[-0.12,-0.16,-0.77,-0.21,-0.05,0.26,0.71,-0.34,0.74,-0.68,-0.42,0.27,0.78,0.7,0.76,0.33,-0.76,0.65,0.15,-0.93],\n", - "[-0.61,-0.61,0.16,0.79,-0.81,0.07,0.2,0.16,-0.98,0.83,0.44,0.9,0.84,0.63,0.04,-0.08,-0.13,-0.69,0.98,0.9],\n", - "[-0.88,-0.4,-0.99,0.94,0.83,0.47,0.04,0.04,-0.18,0.87,-0.14,0.62,-0,0.58,0.94,0.87,-0.37,0.72,-0.98,-0.74],\n", - "[0.47,0.14,0.82,-0.05,0.89,0.09,0.51,0.87,0.31,-0.85,0.29,0.59,0.1,0.77,-0.15,0.21,0.19,-0.32,0.32,-0.77],\n", - "[0.17,-0.51,0.76,0.2,0.53,0.21,-0.46,0.11,0.18,0.56,-0.02,0.44,0.77,0.25,-0.43,0.88,-0.24,-0.1,0.03,-0.12],\n", - "[-0.42,0.68,0.31,0.69,-0.97,0.13,0.22,-0.73,-0.57,-0.47,-0.15,0.6,-0.13,0.51,0.81,-0.44,-0.37,-0.63,0.14,0.63],\n", - "[0.71,0.54,-0.54,0.88,0.09,-0.83,-0.18,0.43,-0.4,-0.09,-0.01,-0.74,0.9,-0.14,0.4,0.92,0.6,0.57,-0.34,0.6]\n", - "]\n", - "\n", - "\n", - "\n", - "288->273\n", - "\n", - "\n", - "23:w23:[20, 20]:[\n", - "[-0.94,-0.68,0.32,0.7,-0.63,0.71,-0.19,0.17,0.41,-0.91,0.04,-0.25,0.03,-0.62,-0.84,0.23,-0.91,-0.5,-0.45,-0.77],\n", - "[-0.01,0.51,-0.34,0.82,-0.41,0.6,0.57,0.76,0.06,0.09,-0.87,-0.52,0.78,0.52,-0.15,-0.97,-0.67,0.33,-0.43,-0.07],\n", - "[-0.19,-0.46,0.82,0.22,-0.71,-0.64,0.4,0.43,0.44,0.73,0.91,-0.91,-0.38,-0.45,0.53,0.82,0.7,0.28,0.89,-0.06],\n", - "[0.96,-0.17,-0.52,0.09,-0.75,-0.63,0.56,-0.32,0.43,0.19,0.5,0.34,0.5,-0.21,0.01,-0.98,-0.02,0.61,0.84,0.03],\n", - "[0.86,0.14,0,0.89,0.26,-0.25,-0.8,-0.57,-0.87,-0.3,-0.57,-0.55,-0.68,0.6,0.46,0.06,0.36,-0.01,0.61,-0.54],\n", - "[-0.28,-0.62,0.7,0.19,-0.71,0,-0.69,0.52,0.34,-0.12,-0.32,-1,-0.38,0.08,-0.82,-0.99,-0.57,0.61,0.71,0.61],\n", - "[0.3,-0.45,0.65,-0.96,0.36,-0.07,0.19,-0.47,0.01,-0.04,0.76,0.51,-0.55,-0.58,-0.37,-0.24,-0.01,0.14,-0.81,0.35],\n", - "[0.34,-0.15,-0.56,0.06,-0.45,-0.12,-0.12,0.14,-0.57,0.82,0.41,0.62,-0.72,-0.61,0.35,-0.77,-0.33,-0.44,-0.86,0.72],\n", - "[-0.65,0.42,-0.59,-0.57,0.36,0.89,0.14,-0.51,0.83,-0.78,0.13,-0.44,-0.86,0.87,0.17,0.44,0.56,0.16,-0.15,0.72],\n", - "[0.94,-0.83,0.36,0.91,0.44,-0.88,-0.91,-0.19,0.97,-0.46,0.63,0.01,0.58,0.67,0.15,0.94,-0.42,0.62,-0.83,-0.79],\n", - "[0.36,-0.98,0.26,-0.61,-0.95,-0.71,0.22,0.28,-0.3,-0.34,-0.97,0.75,-0.91,-0.07,0.12,-0.41,-0.43,0.94,0.12,0.81],\n", - "[0.25,0.72,-0.17,0.46,-0.82,-0.72,-0.18,-0.04,0.9,-0.68,0.79,-0.22,-0.21,0.29,-0.46,0.58,-0.06,0.21,-0.57,-0.62],\n", - "[-0.4,0.65,-0.2,0.24,0.04,0.79,0.17,-0.67,0.55,-0.16,-0.22,0.06,0.53,-0.23,-0.8,0.93,0.71,-0.37,-0.96,0.25],\n", - "[-0.36,-0.23,-0.17,-0.68,-0.04,-0.52,0.44,-0.96,0.26,0.78,0.89,0.51,-0.44,0.69,-0.22,0.01,0.09,-0.72,-0.24,-0.16],\n", - "[0.5,0.54,0.45,-0.85,-0.96,0.75,0.51,-0.4,0.08,0.08,-0.72,-0.42,-0.07,-0.55,0.03,-0.13,0.65,-0.04,-0.92,0.15],\n", - "[0.45,-0.03,-0.81,-0.7,-0.98,-0.94,-0.84,-0.87,-0.8,-0.38,-0.77,0.81,-0.62,1,-0.05,-0.05,0.91,-0.87,-0.65,0.06],\n", - "[-0.72,0.5,0.37,-0.73,0.6,-0.29,0.59,0.59,0.54,-0.24,0.98,-0.49,0.9,0.87,-0.87,0.44,-0.4,-0.47,0.47,0.76],\n", - "[-0.87,0.19,0.94,-0.69,0.16,0.35,-0.51,-0.42,-0.59,0.31,0.06,-0.12,-0.32,-0.56,0.6,0.43,-0.52,-0.59,-0.66,-0.53],\n", - "[-0.43,-0.25,-0.84,0.18,0.43,-0.53,0.97,0.61,-0.38,-0.23,0.87,0.56,-0.01,-0.59,0.58,-0.24,-0.52,-0.61,0.08,-0.04],\n", - "[0.64,0.67,-0.76,-0.37,0.8,-0.37,-0.75,0.47,0.23,-0.87,0.3,-0.53,-0.15,0.4,0.83,0.42,-0.79,-0.09,0.17,-0.28]\n", - "]\n", - "\n", - "\n", - "\n", - "291\n", - "\n", - "RuleNeuron:291:lstm3_out__right(T) :- lstm3_out__i(T), lstm3_out__n(T). [transformation=identity, combination=elproduct]\n", - "val: [0.01,-0.01,-0,0.01,0.01,-0.02,-0.02,0,-0,-0,0.01,0.01,-0,0.02,0,0.03,0,-0,0.01,-0.02]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "291->285\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "291->288\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "294\n", - "\n", - "RuleNeuron:294:lstm3_out__c(T) :- lstm3_out__left(T), lstm3_out__right(T). [transformation=identity]\n", - "val: [0.01,-0.02,0,0.01,0.01,-0.02,-0.02,0,-0,-0,0.01,0.01,-0.01,0.03,0,0.03,0.01,-0,0.01,-0.02]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum\n", - "\n", - "\n", - "\n", - "294->282\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "294->291\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "298\n", - "\n", - "RuleNeuron:298:lstm3_out(T) :- lstm3_out__o(T), tanh(lstm3_out__c(T)). [transformation=identity, combination=elproduct]\n", - "val: [0,-0.01,0,0.01,0,-0.01,-0.01,0,-0,-0,0,0,-0,0.01,0,0.02,0,-0,0,-0.01]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "298->276\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "298->294\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "301\n", - "\n", - "WeightedRuleNeuron:301:lstm3_out__o(T) :- {20, 20} lstm2_out(T), {20, 20} lstm3_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.49,0.49,0.5,0.51,0.5,0.52,0.48,0.51,0.49,0.5,0.5,0.5,0.47,0.48,0.5,0.5,0.49,0.5,0.5,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "301->216\n", - "\n", - "\n", - "20:w20:[20, 20]:[\n", - "[0.08,-0.47,0.4,0.1,0.97,-0.07,0.69,-0.43,0.15,0.43,-0.36,0.51,0.65,-0.34,-0.7,-0.26,-0.27,-0.07,0.88,-0.57],\n", - "[-0.6,-0.9,-0.6,-0.23,0.92,-0.98,0.52,0.36,0.74,-0.54,-0.85,-0.31,-0.09,-0.07,0.31,-0.1,-0.71,0.07,0.69,0.52],\n", - "[0.55,0.63,-0.65,0.84,0.23,-0.08,-0.81,-0.86,-0.87,-0.25,-0.15,0.97,-0.27,-0.04,-0.93,-0.37,0.07,-0.04,-0.34,-0.04],\n", - "[0.61,0.82,0.79,-0.54,-0.88,0.84,-0.68,-0.77,-0.19,-0.97,0.06,-0.13,0.95,0.19,-0.54,0.25,0.97,1,-0.27,-0.03],\n", - "[0.65,-0.17,-0.65,0.94,-0.24,-0.93,-0.83,0.7,0.98,-0.09,-0.47,0.87,0.22,-0.13,0.51,-0.66,-0.14,0.65,0.1,-0.17],\n", - "[0.68,0.22,-0.43,-0.39,-0.88,0.7,0.79,-0.57,0.62,-0.14,-0.26,-0.91,-0.16,-0.16,-0.59,-0.68,-0.02,-0.76,-0.37,0.03],\n", - "[0.23,0.42,-0.44,0.72,-0.69,0.72,-0.08,0.16,0.68,0.07,-0.48,-0.06,-0.68,0.24,-0.91,0.77,0.6,-0.79,-0.07,-0.75],\n", - "[-0.96,0.94,0.42,-0.5,0.86,0.86,-0.7,-0.79,0.35,0.83,-0.91,0.12,-0.55,-0.82,0.06,0.54,-0.98,0.91,0.7,-0.16],\n", - "[-0.95,-0.54,0.93,0.69,0.96,0.75,0.92,-0.61,-0.72,0.75,-0.42,0.3,0.89,-0.8,0.84,-0.62,-0.76,0.98,0.16,-0.51],\n", - "[-0.83,-0.08,0.51,-0.85,0.11,0.77,0.88,0.44,0.84,0.68,0.3,0.72,0.36,0.68,-0.99,1,-0.25,0.93,-0.3,0.96],\n", - "[0.58,-0.75,-0.47,0.83,0.85,-0.36,-0.07,-0.32,-0.85,-0.57,0.01,-0.04,0.35,0.99,-0.74,-0.33,0.97,-0.15,0.61,-0.05],\n", - "[0.05,0.8,0.23,-0.63,-0.51,-0.55,1,0.91,0.35,0.34,0.7,-0.65,-0.93,-0.98,0.1,-0.68,-0.71,-0.66,0.94,0.17],\n", - "[-0.87,-0.86,-0.69,0.9,-0.72,0.05,-0.24,-0.25,-0.04,0.89,0.24,0.73,-0.31,0.81,0.76,0.91,0.75,0.19,0.56,-0.48],\n", - "[-0.04,0.34,0.75,0.28,0.65,0.77,-0.02,-0.6,0.52,0,0.46,0.02,0.43,-0.16,0.97,-0.73,-0.85,0.29,0.57,0.69],\n", - "[0.59,0.43,-0.68,-0.19,-0.99,-0.19,0.15,-0.19,0.47,-0.64,0.05,-0.71,0.96,0.01,-0.45,-0.8,-0.86,0.01,0.75,-0.55],\n", - "[-0.53,0.96,-0.12,0.83,-0.3,-0.32,0.97,-0.69,-0.22,0.28,0.02,-0.76,0.81,-0.29,0.66,-0.63,0.42,-0.25,0.42,-0.78],\n", - "[-0.02,-0.7,-0.7,-0.99,-0.06,0.57,-0.41,-0.21,-0.53,0.07,-0.25,0.52,0.9,0.73,-0.6,0.23,0.31,-0.7,-0.65,-0.35],\n", - "[0.93,0.6,0.3,0.23,0.41,-0,-0.52,0.97,0.95,-0.76,-0.43,0.76,-0.88,0.6,0.95,-0.93,-0.69,0.26,0.43,0.85],\n", - "[-0.72,-0.47,-0.5,0.87,-0.89,-0.97,0.48,-0.06,0.1,0.43,-0.21,0.95,0.19,-0.99,0.1,-0.18,0.94,0.22,-0.9,0.31],\n", - "[0.21,-0.25,0.28,0.87,0.3,0.92,0.23,-0.57,-0.61,-0.25,-0.37,-0.2,-0.88,0.26,-0.34,0.96,0.13,-0.64,0.87,0.33]\n", - "]\n", - "\n", - "\n", - "\n", - "301->298\n", - "\n", - "\n", - "21:w21:[20, 20]:[\n", - "[-0.34,1,-0.36,0.04,-0.77,0.08,-0.31,-0.87,0.07,-0.75,-0.88,0.5,0.84,0.51,-0.77,-0.39,0.95,0.36,-0.75,-0.06],\n", - "[0.94,0.07,0.1,0.55,-0.99,-0.21,-0.57,0.96,0.44,-0.05,0.08,-0.55,-0.94,0.72,0.48,0.52,0.69,-0.61,0.55,0.49],\n", - "[0.5,0.58,0.06,-0.22,-0.53,-0.3,0.75,-0.05,0.63,-0.74,-0.18,-0.74,0.29,-1,-0.08,-0.01,-0.33,0.38,0.46,0.76],\n", - "[0.18,-0.64,-0.66,0.65,-0.84,-0.81,-0.83,-0.07,-0.51,0.16,-0.53,0.08,-0.64,-0.44,0.57,-0.2,0.15,0.26,-0.89,-0.38],\n", - "[0.44,0.61,0.73,0.18,0.91,-0.01,0.74,0.89,0.11,0.42,0.3,0.45,0.83,0.52,0.13,0.62,-0.64,0.48,-0.62,-0.67],\n", - "[0.65,0.42,-0.81,-0.85,0.07,-0.87,-0.92,0.33,0.36,0.19,-0.77,-0.6,-0.87,0.84,-0.44,-0.44,-0.53,-0.43,0.64,-0.6],\n", - "[-0.23,0.45,0.33,-0.67,-0.56,-0.8,-0.86,-0.05,-0.14,0.23,-0.33,-0.56,-0.35,0.65,-0.54,-0.31,0.77,0.04,0,0.59],\n", - "[-0.61,-0.67,-0.07,-0.04,-0.44,-0.78,0.97,0.67,0.44,-0.77,-0.51,0.91,-0.79,-0.41,0.14,-0.59,-0.68,-0.9,-0.26,0.2],\n", - "[0.46,-0.74,0.01,-0.72,-0.8,-0.03,0.71,0.63,-0.8,-0.47,0.37,0.9,-0.62,0.62,-0.33,-0.86,-0.24,-0.37,-0.48,0.05],\n", - "[-0.86,-0.14,0.77,0.55,-0.82,0.53,0.5,-0.56,-0.75,-0.64,0.02,-0.54,-0.23,-0.36,0.85,-0.71,0.66,0.78,0.12,0.52],\n", - "[0.06,0.61,-0.97,-0.98,0.68,-0.1,0.75,-0.21,0.4,0.99,0.93,0.94,-0.62,0.64,0,0.11,0.33,-0.99,0.76,-0.64],\n", - "[-0.8,0.59,-0.78,0.41,-0.95,-0.68,0.66,0.97,-0.22,-0.1,0.22,0.76,0.97,-0.26,0.32,0.86,-0.09,0.2,0.12,0.08],\n", - "[0.86,0.03,-0.08,-0.88,-0.53,0.48,-0.44,-0.51,-0.67,-0.22,0.78,-0.65,0.5,-0.56,0.76,0.91,0.4,-0.81,-0.02,0.97],\n", - "[-0.91,0.26,0.07,-0.79,0.8,-0.93,0.23,-0.27,0.31,0.51,0.53,0.31,-0.32,0.71,0.49,-0.7,0.86,-0.12,0.05,0.59],\n", - "[-0.52,-0.94,-0.64,-0.69,0.58,-0.44,-0.31,-0.96,-0.41,0.7,-0.05,-0.24,0.71,-0.94,0.8,0.46,0.6,-0.86,-0.91,0.11],\n", - "[0.79,-0.93,0.33,0.24,-0.33,0,0.7,-0.03,-0.84,-0.64,0.15,-0.17,-0.09,-0.47,-0.84,0.85,0.16,-0.79,0.69,0.66],\n", - "[-0.7,-0.05,-0.6,0.74,-0.32,0.38,-0.26,0.47,0.96,0.12,0.74,-0.12,-0.69,-0.86,-0.33,0.85,-0.68,0.27,0.02,0.82],\n", - "[-0.21,0.38,-0.99,-0.97,-0.82,-0.21,0.1,-0.17,0.96,-0.75,0.59,-0.51,0.9,0.26,0.36,-0.16,-0.22,0.4,-0.94,-0.23],\n", - "[-0.06,-0.45,0.18,-0.89,0.87,-0.68,-0.74,0.74,0.93,0.09,-0.94,0.59,0,0.75,-0.9,0.85,-0.57,0.34,-0.69,-0.7],\n", - "[-0.12,-0.32,-0.72,-0.31,0.7,-0.68,0.12,-0.31,-0.12,-0.99,0.84,-0.83,-0.56,0.5,-0.21,-0.74,-0.01,0.27,-0.41,0.88]\n", - "]\n", - "\n", - "\n", - "\n", - "304\n", - "\n", - "WeightedRuleNeuron:304:lstm3_out__f(T) :- {20, 20} lstm2_out(T), {20, 20} lstm3_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.51,0.5,0.51,0.5,0.49,0.49,0.49,0.5,0.48,0.51,0.5,0.51,0.5,0.47,0.51,0.49,0.48,0.51,0.5]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "304->216\n", - "\n", - "\n", - "18:w18:[20, 20]:[\n", - "[0.67,0.87,0.79,0.15,0.83,-0.69,0.82,-0.27,0.08,-0.53,0.14,-0.69,0.14,0.99,-0.98,-0.08,0.26,0.41,-0.22,0.17],\n", - "[0.93,0.81,-0.3,-0.88,0.68,0.61,0.6,-0.8,0.25,-1,-0.43,0.6,0.56,-0.19,0.98,-0.12,0.62,-0.12,0.34,-0.66],\n", - "[0.72,0.76,-0.82,-0.59,-0.87,0.71,-0.61,0.17,0.9,0.85,-0.35,-0.2,-0.92,-0.22,0.76,-0.01,0.69,-0.3,-0.03,-0.13],\n", - "[0.18,0.05,-0.36,0.68,0.29,-0.86,0.3,0.83,0.16,-0.54,0.17,-0.2,0.2,-0.84,-0.81,-0.96,0.9,-0.72,0.95,0.79],\n", - "[-0.79,-0.71,0.47,-0.31,0.83,0.73,-0.98,0.8,0.72,-0.06,0.14,-0.12,-0.35,0.12,-0.32,0.64,-0.32,0.24,-0.98,-0.18],\n", - "[0.34,-0.22,-0.12,0.21,-0.7,-0.23,-0.94,-0.2,-0.63,-0.01,-0.4,0.83,0.78,-0.14,-0.06,0.49,0.95,0.22,-0.01,0.77],\n", - "[0.13,0.81,-0.9,-0.71,-0.43,0.03,-0.38,0.48,0.51,-0.35,0.65,-0.73,-0.69,0.54,0.61,0.75,0.96,-0.12,-0.72,0.43],\n", - "[-0.59,0.24,0.98,-0.66,-0.92,0.73,-0.18,0.84,-0.52,0.95,0.11,0.39,-0.53,0.41,-0.63,0.25,0.86,0.43,0.53,0.96],\n", - "[0.54,-0.05,-0.3,-0.94,0.55,0.81,-0.13,0.11,0.44,0.98,-0.09,-0.18,-0.91,0.16,-0.59,-0.03,-0.21,-0.79,-0.03,-0.56],\n", - "[-0.02,-0.65,-0.14,-0.94,0.87,-0.29,0.15,0.25,0.91,0.64,-0.13,0.38,0.02,-0.95,0.04,0.9,-0.55,0.96,-1,-0.55],\n", - "[-0.05,-0.79,-0.99,0.48,0.38,-0.45,-0.57,-0,-0.5,-0.69,0.93,-0.23,0.43,-0.19,-0.65,-0.7,0.91,-0.73,0.05,-0.14],\n", - "[0.53,-0.14,-0.76,-0.1,0.74,0.1,-0.37,-0.75,0.37,-0.97,0.94,0.83,-0.22,0.3,-0.5,0.59,0.29,-0.5,-0.26,-0.06],\n", - "[-0.95,-0.12,0.95,-0.9,0.49,-0.73,0.45,-0.57,0.02,0.41,0.21,0.98,-0.58,-0.96,0.58,-0.62,-0.51,-0.2,-0.1,0.67],\n", - "[0.26,-0.3,0.63,0.15,-0.01,-0.87,0.67,0.47,0.67,0.33,-0.28,-0.43,-0.79,-0.82,-0.45,0.23,-0.59,0.46,-0.59,0.79],\n", - "[0.4,0.99,0.02,-0.41,0.08,-0.93,0.93,-0.33,0.13,-0.77,-0.54,-0.32,0.72,-0.39,-0.13,0.58,0.51,-0.4,0.05,-0.42],\n", - "[0.34,0.24,0.39,0.51,-0.95,-0.04,-0.46,-0.49,0.29,-0.54,-0.08,-0.61,0.37,-0.4,-0.75,-0.98,0.34,-0.37,-0.32,-0.57],\n", - "[0.09,-0.75,0.48,-0.84,0.56,-0.64,0.71,0.52,-0.21,-0.83,-0.18,0.06,-0.16,0.22,-0.08,-0.71,0.61,-0.88,-0.38,-0.42],\n", - "[0.93,-0.05,0.81,0.1,-0.69,0.31,0.75,-0.32,0.25,0.47,0.27,0.45,-0.63,0.27,-0.58,-0.77,0.94,-0.15,-0.77,0.01],\n", - "[-0.04,-0.2,-0.19,0.66,0.59,-0.33,-0.41,0.91,0.23,-0.28,-0.64,0.38,-0.8,-0.19,0.07,-0.62,0.99,-0.16,-0.36,-0.65],\n", - "[-0.45,-0.85,0.45,-0.55,0.08,0.68,0.38,0.62,0.86,-0.76,-0,-0.3,0.43,-0.59,0.82,-0.67,0.67,0.42,0.31,0.9]\n", - "]\n", - "\n", - "\n", - "\n", - "304->298\n", - "\n", - "\n", - "19:w19:[20, 20]:[\n", - "[0.3,-0.36,0.82,0.87,0.8,0.3,-0.49,0.45,-0.46,-0.68,0.93,-0.14,-0.34,-0.65,-0.97,0.42,0.46,-0.38,0.88,0.54],\n", - "[-0.66,0.31,0.86,0.29,0.11,-0.1,0.63,0.15,0.93,-0.8,-0.42,-0.1,-0.48,0.64,0.55,-0.88,-0.33,-0.29,-0.13,0.93],\n", - "[-0.23,0.5,-0.25,0.43,0.09,-0.34,-0.93,-0.59,-0.63,0.37,0.14,-0.95,-0.1,0.4,0.35,0.43,-0.02,-0.43,0.38,0.87],\n", - "[-0.23,0.78,0.97,0.71,-0.31,-0.01,0.17,0.92,0.73,-0.46,0.42,-0.87,0.1,0.49,-0.01,-0.66,-0.59,0.87,-0.34,-0.14],\n", - "[-0.66,0.2,-0.08,-0.03,-0.27,-0.19,-0.5,-0.8,0.24,0.1,-0.48,0.91,0.31,-0.61,0.15,-0.15,0.95,0.52,-0.62,0.95],\n", - "[-0.7,-0.88,-0.93,0.65,-0.03,0.93,-0.51,0.26,0.73,-0.29,0.15,0.39,0.06,0.68,-0.22,0.23,-0.78,0.71,0.24,-0.98],\n", - "[-0,0.06,-0.82,-0.16,0.12,0.84,-0.68,-0.36,-0.45,-0.73,-0.44,0.35,0.86,0.79,-1,-0.41,-0.32,-0.23,-0.29,0.92],\n", - "[-0.4,-0.83,0.86,-0.84,0.43,0.56,0.91,-0.25,0.03,0.54,0.28,-0.87,-0.62,0.81,0.18,0.39,-0.61,-0.97,0.24,0.76],\n", - "[0.13,0.52,-0.23,0.91,-0.58,-0.42,-0.25,0.29,0.82,-0.56,-0.12,0.43,0.53,-0.7,0.34,0.63,-0.96,-0.92,0.39,-0.18],\n", - "[-0.28,-0.08,0.4,-0.51,0.18,-0.95,-0.62,-0.02,-0.35,0.93,-0.31,-0.69,0.51,-0.54,-0.68,-0.34,0.35,-0.68,0.84,-0.56],\n", - "[-0.83,-0.38,-0.27,-0.99,-0.1,0.43,-0.89,0.93,-0.78,-0.14,0.02,-0.03,0.79,-0.92,0.37,-0.78,0.98,-0.27,0.63,0.68],\n", - "[1,0.21,-0.53,0.37,0.91,0.22,0.85,-0.01,-0.34,-0.67,-0.14,0.85,0.06,0.41,0.61,0.23,-0.57,-0.98,-0.14,-0.38],\n", - "[0.53,-0.77,-0.36,-0.62,0.27,0.72,-0.7,-0.42,0.84,-0.45,0.2,-0.25,-0.22,-0.82,0.88,0.08,-0.27,0.51,-0.2,-0.71],\n", - "[-0.09,-0.42,-0.86,0.58,0.28,0.1,-0.83,-0.24,0.14,-0.1,0.25,0.04,0.79,-0.97,0.86,0.7,-0.21,-0.51,0.47,-0.65],\n", - "[-0.61,0.92,0.7,0.6,0.78,-0.61,-0.3,0.18,-0.15,0.68,0.55,-0.9,0.48,0.34,0.8,0.63,0.49,-0.66,-0.42,0.65],\n", - "[0.79,-0.93,0.08,0.13,0.62,0.03,-0.58,0.49,0.18,-0.19,0.75,0.84,-0.48,0.43,-0.51,0.34,-0.02,0.61,-0.29,-0.88],\n", - "[-0.63,-0.64,0.91,-0.31,0.42,0.32,0.39,-0.81,0.08,-0.09,0.15,-0.96,0.84,0.98,-0.94,-0.3,0.41,-0.26,-0.04,0.02],\n", - "[0.55,0.37,0.41,0.26,0.47,-0.17,-0.37,-0.27,-0.3,0.42,0.63,-0.45,0.87,-0.49,-0.83,-0.77,-0.22,-0.62,0.99,-0.75],\n", - "[0.57,-0.1,0.53,0.47,-0.07,0.34,-0.94,0.3,-0.12,0.73,-0.66,0.75,-0.15,0.96,-0.51,0.71,-0.46,-0.49,-0.72,-0.25],\n", - "[0.36,0.19,0.62,0.48,-0.81,-0.96,-0.32,-0.63,-0.58,-0.37,-0.17,-0.14,-0.48,-0.53,-0.63,0.58,-0.21,0.32,0.81,-0.5]\n", - "]\n", - "\n", - "\n", - "\n", - "307\n", - "\n", - "RuleNeuron:307:lstm3_out__left(T) :- lstm3_out__f(T), lstm3_out__c(Z), @next(Z, T). [transformation=identity, combination=elproduct]\n", - "val: [0,-0.01,0,0.01,0,-0.01,-0.01,0,-0,-0,0,0,-0,0.01,0,0.02,0,-0,0,-0.01]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "307->294\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "307->304\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "310\n", - "\n", - "WeightedRuleNeuron:310:lstm3_out__i(T) :- {20, 20} lstm2_out(T), {20, 20} lstm3_out(Z), @next(Z, T). [transformation=sigmoid]\n", - "val: [0.5,0.5,0.49,0.49,0.49,0.52,0.49,0.5,0.49,0.51,0.48,0.49,0.51,0.49,0.51,0.51,0.5,0.51,0.51,0.49]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Sigmoid\n", - "\n", - "\n", - "\n", - "310->216\n", - "\n", - "\n", - "16:w16:[20, 20]:[\n", - "[-0.84,0.69,0.84,-0.81,0.03,0.06,0.4,0.31,0.02,-0.77,0.12,0.29,-0.97,-0.73,0.24,-0.18,0.88,0.49,0.13,-0.84],\n", - "[0.53,0.35,0.43,0.72,-0.31,0.28,0.45,0.2,-0.21,-0.69,-0.01,-0.18,-0.81,-0.15,0.04,-0.47,0.71,-0.84,0.23,0.26],\n", - "[-0.03,-0.21,-0.64,-0.6,-0.16,0.45,0.39,-1,-0.27,0.69,0.13,-0.21,0.22,-0.87,0.46,0.07,0.33,0.85,-0.03,-0.47],\n", - "[-0.63,0.25,0.38,0.24,-0.62,-0.33,-0.7,-0.2,0.08,0.49,-0.28,0.29,-0.93,0.22,0.71,-0.51,-0.71,-0.95,-0.15,-0.6],\n", - "[-0.13,0.49,-0.77,0.24,0.5,0.9,-0.32,0.86,0.93,-0.28,0.82,-0.52,0.31,0.89,0.11,-0.23,-0.28,0.58,0.27,0.49],\n", - "[-0.65,-0.26,-0.93,-0.24,-0.32,0.3,0.95,0.79,-0.89,-0.59,-0.45,0.5,-0.38,0.81,0.82,0.7,0.27,-0.13,-0.69,-0.07],\n", - "[0.49,-0.24,-0.34,0.53,0.61,-0.27,-0.38,0.14,0.19,-0.46,-0.67,-0.72,-0.57,-0.18,-1,-0.33,-0.72,-0.74,0.95,0.1],\n", - "[-0.58,-0.82,0.01,0.22,-0.96,-0.27,-0.37,0.66,-0.11,-0.58,0.15,-0.36,-0.14,-0.57,-0.15,0.47,-0.61,0.07,0.11,-0.05],\n", - "[0.34,0.39,0.95,0.38,0.59,-0.14,-0.9,-0.5,0.08,-0.77,0.97,-0.66,0.35,-0.63,-0.24,0.93,-0.34,-0.93,-0.71,0.53],\n", - "[-0.98,-0.76,-0.71,-0.99,-0.33,0.52,-0.9,-0.91,0.88,0.27,0.23,1,-0.73,0.69,-0.56,0.44,0.49,0.52,0.7,0.75],\n", - "[-0.8,-0.76,0.72,0.52,0.31,0.85,0.41,0.06,0.77,-0.54,0.38,-0.22,0.99,0.98,-0.85,-0.54,0.68,0.4,-0.33,-0.53],\n", - "[0.82,0.51,-0.64,-0.91,0.15,-0.95,-0.72,0.9,0.79,-0,0.25,0.82,-0.72,-0.34,-0.32,-0.77,0.66,-0.55,0.06,-0.88],\n", - "[0.18,-0.03,0.96,-0.45,-0.44,0.43,-0.84,0.98,0.7,0.7,-0.98,-0.26,-0.51,0.1,-0.64,-0.65,-0.75,-0.97,-0.12,0.87],\n", - "[0.21,0.34,-0.86,-0.54,-0.36,-0.8,0.07,0.67,0.93,-0.57,-0.68,-0.7,-0.5,0.18,-0.37,-0.23,0.3,0.18,-0.21,0.81],\n", - "[-0.02,0.57,0.98,-0.7,-0.89,-0.74,0.72,0.04,-0.56,-0.6,0.93,0.54,0.13,0.44,0.1,-0.3,-0.07,0.35,0.26,0.16],\n", - "[0.51,-0.48,-0.18,-0.49,-0.12,0.82,-0.89,0.83,0.4,0.75,-0.77,0.65,-0.93,0.36,-0.55,-0.4,-0.19,-0.6,-0.58,0.87],\n", - "[0.24,-0.93,0.16,0.99,0.63,-0.62,0.06,-0.68,0.23,-0.6,0.73,0.09,0.9,-0.51,-0.85,0.96,0.3,0.8,0.23,-0.36],\n", - "[-0.98,0.26,0.47,0.35,-0.72,-0.13,0.57,-0.62,0.08,-0.11,0.05,-0.19,0.28,0.39,-0.35,0.64,-0.87,0.37,-0.54,0.78],\n", - "[-0.24,0.85,0.42,-0.52,0.34,-0.83,0.36,0.22,0.34,0.86,0.5,0.33,0.08,0.61,0.91,0.06,-0.51,-0.94,-0.65,0.43],\n", - "[-0.24,-0.63,-0.38,0.41,0.4,0.99,0.62,0.09,-0.58,-0.86,0.2,-0.38,-0.04,0.41,-0.27,-0.05,-0.92,-0.06,-0.03,0.78]\n", - "]\n", - "\n", - "\n", - "\n", - "310->298\n", - "\n", - "\n", - "17:w17:[20, 20]:[\n", - "[-0.57,-0.37,-0.8,-0.7,-0.37,0.12,0.49,0.97,-0.37,-0.72,0.91,-0.17,-0.58,-0.47,-0.23,-0.36,-0.92,0.16,0.77,-0.5],\n", - "[-0.74,0.89,-0.1,-0.55,0.16,0.49,-0.1,0.22,-0.95,0.83,0.43,-0.88,0.64,0.25,0.86,0.93,-0.32,0.67,0.28,0.46],\n", - "[0.48,0.97,0.09,0.69,0.4,-0.66,-0.48,0.13,0.51,-0.33,0.01,0.45,0.47,-0.38,0.86,-0.64,-0.26,0.13,0.94,-0.87],\n", - "[0.09,-0.99,-0.88,-0.72,0.92,-0.55,-0.62,-0.36,0.49,-0.61,-0.4,-0.17,0.15,-0.09,-0.31,-0.55,-0.79,-0.09,0.48,-0.93],\n", - "[-0.89,-0.89,0.82,0.36,-0.51,0.11,-0.67,-0.11,0.73,-0.48,0.08,-0.87,0.63,0.83,-0.16,-0.09,-0.12,-0.83,0.51,0.54],\n", - "[-0.56,0.82,-0.37,0.45,0.91,-0.1,-0.92,-0.42,0.51,-0.26,0.61,-0.53,0.52,-0.23,-0.97,-0.66,0.46,0.32,-0.55,-0.57],\n", - "[1,0.9,0.64,0.29,0.92,-0.64,0.76,-0.16,-0.91,0.61,0.96,0.58,0.35,-0.47,-0.46,0.47,0.6,0.04,-0.58,-0.14],\n", - "[0.55,0.49,0.45,-0,-0.87,0.95,-0.92,-0.44,0.81,0.24,-0.43,-0.82,-0.15,0.11,-0.48,0.8,0.69,0.42,0.97,0.15],\n", - "[-0.23,0.89,-0.1,0.12,-0.71,-0.68,-0.94,0,0.55,0.38,-0.71,-0.86,-0.86,-0.74,-0.14,-0.71,0.12,0.82,-0.12,-0.87],\n", - "[-0.63,-0.63,-0.91,0.59,0.21,0.51,-0.93,-0.05,-0.58,-0.87,-0.59,0.73,-0.85,0.08,-0.96,-0.86,0.94,0.85,-0.36,0.82],\n", - "[-0.07,0.2,0.93,-0.45,0.07,-0.58,-0.75,0.76,-0.08,-0.23,-0.9,0.5,0.88,0.58,-0.12,-0.7,-0.77,-0.53,-0.47,-0.33],\n", - "[-0.99,0.91,0.38,-0.66,-0.83,0.82,-0.84,-0.24,-0.59,-0.2,0.03,0.53,0.68,0.7,0.85,0.3,0.13,-0.15,-0.69,-0.1],\n", - "[-0.7,0.23,-0.1,-0.68,-0.88,0.55,0.1,-0.43,-0.53,-0.68,-0.01,0.5,0.97,-0.98,-0.46,-0.82,-0.23,0.46,-0.82,-0.51],\n", - "[-0.45,-0.43,0.38,-0.71,0.87,0.23,-0.98,-0.1,0.51,0.71,-0.74,0.14,0.83,-0.92,0.31,-0.19,0.48,0.65,0.58,-0.78],\n", - "[0.5,0.85,-0.26,0.12,0.14,0.63,0.35,0.97,0.66,-0.42,-0.31,0.74,0.8,0.37,0.88,0.96,0.61,-0.01,0.56,0.93],\n", - "[0.76,-0.21,-0.42,-0.88,0.92,0.41,-0.99,0.64,0.05,0.36,-0.35,-0.9,-0.93,-0.73,0.27,0.32,0.92,-0.68,-0.89,-0.53],\n", - "[-0.41,-0.86,1,-0.96,-0.46,0.93,0.89,0.83,0.44,0.31,-0.09,-0.72,-0.73,-0.78,-0.75,-0.49,0.34,0.92,-0.41,-0.69],\n", - "[0.18,-0.8,0.55,0.98,-0.25,0.66,-0.37,0.52,-0.06,0.1,0.43,0.6,-0.88,0.04,-0.3,-0.41,-0.13,0.64,0.48,-0.22],\n", - "[0.25,-0.7,-0.49,-0.09,0.22,0.79,-0.52,0.34,-0.75,-0.24,-0.61,-0.95,-0.44,-0.91,-0.76,0.03,0.04,0.97,-0.58,-0.21],\n", - "[-0.73,0.98,-0.82,0.96,-0.15,-0.74,-0.3,0.47,0.96,0.2,0.61,0.99,0.59,0.1,0.93,-0.9,-0.99,-0.27,-0.04,0.81]\n", - "]\n", - "\n", - "\n", - "\n", - "313\n", - "\n", - "WeightedRuleNeuron:313:lstm3_out__n(T) :- {20, 20} lstm2_out(T), {20, 20} lstm3_out(Z), @next(Z, T). [transformation=tanh]\n", - "val: [0.03,-0.04,-0.02,0.02,0.01,-0.06,-0.09,0.01,-0.01,0,0,0.04,-0.01,0.05,0.01,0.11,0.01,0.01,-0,-0.07]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum+Tanh\n", - "\n", - "\n", - "\n", - "313->216\n", - "\n", - "\n", - "22:w22:[20, 20]:[\n", - "[-0.98,-0.36,-0.06,-0.54,-0.52,0.49,-0.48,0.79,-0.96,0.02,-0.42,-0.91,-0.26,0.68,-0.71,0.88,0.71,0.75,-0.19,0.23],\n", - "[0.76,0.01,0.73,-0.07,0.09,0.35,-0.03,0.97,-0.41,0.07,0.69,-0.05,-0.57,0.49,-0.97,-0.1,-0.57,0.31,-0.04,-0.46],\n", - "[-0.26,-0.97,0.22,-0.27,0.49,0.14,-0.32,-0.26,-0.04,0.17,-0.61,-0.16,0.44,-0.27,-0,0.44,0.67,-0.03,0.3,0.3],\n", - "[0.54,-0.06,0.31,0.13,0.8,0.01,0.29,0.1,0.14,0.47,0.96,-0.01,0.32,0.38,0.52,0.98,0.86,-0.31,0.62,0.05],\n", - "[0.21,-0.14,-0.75,0.95,0.01,0.2,0.2,-0.57,-0.04,-0.53,-0.73,-0.51,-0.48,0.54,0.74,-0.55,0.59,0.13,-0.96,-0.24],\n", - "[-0.92,0.19,0.85,0.65,0.66,-0.18,-0.32,0.86,0.63,-0.76,0.29,-0.71,0.25,-0.18,0.23,0.18,-0.04,0.83,0.26,-0.35],\n", - "[-0.85,0.41,0.4,0.19,-0.44,0.79,-0.03,0.6,-0.4,0.8,0.29,0.59,-0.52,-0.4,-0.93,-0.45,0.5,0.39,-0.77,-0.89],\n", - "[-0.71,0.01,0.44,-0.34,0.65,-0.66,0.18,0.97,-0.99,0.53,0.1,0.08,0.77,0.52,0.93,-0.77,0.93,-0.9,-0.53,0.2],\n", - "[0.98,-0.28,-0.51,-0.95,-0.62,0.08,0.01,0.76,0.48,-0.68,-0.08,-0.78,-0.72,0.06,0.33,-0.32,-0.98,-0.85,-0.98,0.72],\n", - "[-0.56,0.42,0.71,0.57,0.01,0.9,-0.98,0.88,0.64,-0.82,0.42,0.67,-0.23,0.77,-0.11,-0.15,-0.74,-0.77,-0.54,-0.75],\n", - "[-0.15,-0.47,-0.79,0.91,-0.34,0.33,-0.78,0.24,-0.16,-0.03,0.24,0.37,0.94,0.92,0.83,-0.92,0.7,-0.2,-0.62,-0.43],\n", - "[0.36,0.32,0.43,0.51,0.91,0.8,0.29,-0.27,0.08,0.25,-0.64,0.37,0.56,-0.4,-0.89,-0.85,0.33,-0.85,-0.61,-0.84],\n", - "[0.62,-0.36,-0.97,0.58,0.89,-0.53,-0.02,-0.65,-0.4,0.75,-0.54,-0.81,0.92,0.19,0.79,-0.87,-0.26,-0.97,-0.64,-0.1],\n", - "[-0.12,-0.16,-0.77,-0.21,-0.05,0.26,0.71,-0.34,0.74,-0.68,-0.42,0.27,0.78,0.7,0.76,0.33,-0.76,0.65,0.15,-0.93],\n", - "[-0.61,-0.61,0.16,0.79,-0.81,0.07,0.2,0.16,-0.98,0.83,0.44,0.9,0.84,0.63,0.04,-0.08,-0.13,-0.69,0.98,0.9],\n", - "[-0.88,-0.4,-0.99,0.94,0.83,0.47,0.04,0.04,-0.18,0.87,-0.14,0.62,-0,0.58,0.94,0.87,-0.37,0.72,-0.98,-0.74],\n", - "[0.47,0.14,0.82,-0.05,0.89,0.09,0.51,0.87,0.31,-0.85,0.29,0.59,0.1,0.77,-0.15,0.21,0.19,-0.32,0.32,-0.77],\n", - "[0.17,-0.51,0.76,0.2,0.53,0.21,-0.46,0.11,0.18,0.56,-0.02,0.44,0.77,0.25,-0.43,0.88,-0.24,-0.1,0.03,-0.12],\n", - "[-0.42,0.68,0.31,0.69,-0.97,0.13,0.22,-0.73,-0.57,-0.47,-0.15,0.6,-0.13,0.51,0.81,-0.44,-0.37,-0.63,0.14,0.63],\n", - "[0.71,0.54,-0.54,0.88,0.09,-0.83,-0.18,0.43,-0.4,-0.09,-0.01,-0.74,0.9,-0.14,0.4,0.92,0.6,0.57,-0.34,0.6]\n", - "]\n", - "\n", - "\n", - "\n", - "313->298\n", - "\n", - "\n", - "23:w23:[20, 20]:[\n", - "[-0.94,-0.68,0.32,0.7,-0.63,0.71,-0.19,0.17,0.41,-0.91,0.04,-0.25,0.03,-0.62,-0.84,0.23,-0.91,-0.5,-0.45,-0.77],\n", - "[-0.01,0.51,-0.34,0.82,-0.41,0.6,0.57,0.76,0.06,0.09,-0.87,-0.52,0.78,0.52,-0.15,-0.97,-0.67,0.33,-0.43,-0.07],\n", - "[-0.19,-0.46,0.82,0.22,-0.71,-0.64,0.4,0.43,0.44,0.73,0.91,-0.91,-0.38,-0.45,0.53,0.82,0.7,0.28,0.89,-0.06],\n", - "[0.96,-0.17,-0.52,0.09,-0.75,-0.63,0.56,-0.32,0.43,0.19,0.5,0.34,0.5,-0.21,0.01,-0.98,-0.02,0.61,0.84,0.03],\n", - "[0.86,0.14,0,0.89,0.26,-0.25,-0.8,-0.57,-0.87,-0.3,-0.57,-0.55,-0.68,0.6,0.46,0.06,0.36,-0.01,0.61,-0.54],\n", - "[-0.28,-0.62,0.7,0.19,-0.71,0,-0.69,0.52,0.34,-0.12,-0.32,-1,-0.38,0.08,-0.82,-0.99,-0.57,0.61,0.71,0.61],\n", - "[0.3,-0.45,0.65,-0.96,0.36,-0.07,0.19,-0.47,0.01,-0.04,0.76,0.51,-0.55,-0.58,-0.37,-0.24,-0.01,0.14,-0.81,0.35],\n", - "[0.34,-0.15,-0.56,0.06,-0.45,-0.12,-0.12,0.14,-0.57,0.82,0.41,0.62,-0.72,-0.61,0.35,-0.77,-0.33,-0.44,-0.86,0.72],\n", - "[-0.65,0.42,-0.59,-0.57,0.36,0.89,0.14,-0.51,0.83,-0.78,0.13,-0.44,-0.86,0.87,0.17,0.44,0.56,0.16,-0.15,0.72],\n", - "[0.94,-0.83,0.36,0.91,0.44,-0.88,-0.91,-0.19,0.97,-0.46,0.63,0.01,0.58,0.67,0.15,0.94,-0.42,0.62,-0.83,-0.79],\n", - "[0.36,-0.98,0.26,-0.61,-0.95,-0.71,0.22,0.28,-0.3,-0.34,-0.97,0.75,-0.91,-0.07,0.12,-0.41,-0.43,0.94,0.12,0.81],\n", - "[0.25,0.72,-0.17,0.46,-0.82,-0.72,-0.18,-0.04,0.9,-0.68,0.79,-0.22,-0.21,0.29,-0.46,0.58,-0.06,0.21,-0.57,-0.62],\n", - "[-0.4,0.65,-0.2,0.24,0.04,0.79,0.17,-0.67,0.55,-0.16,-0.22,0.06,0.53,-0.23,-0.8,0.93,0.71,-0.37,-0.96,0.25],\n", - "[-0.36,-0.23,-0.17,-0.68,-0.04,-0.52,0.44,-0.96,0.26,0.78,0.89,0.51,-0.44,0.69,-0.22,0.01,0.09,-0.72,-0.24,-0.16],\n", - "[0.5,0.54,0.45,-0.85,-0.96,0.75,0.51,-0.4,0.08,0.08,-0.72,-0.42,-0.07,-0.55,0.03,-0.13,0.65,-0.04,-0.92,0.15],\n", - "[0.45,-0.03,-0.81,-0.7,-0.98,-0.94,-0.84,-0.87,-0.8,-0.38,-0.77,0.81,-0.62,1,-0.05,-0.05,0.91,-0.87,-0.65,0.06],\n", - "[-0.72,0.5,0.37,-0.73,0.6,-0.29,0.59,0.59,0.54,-0.24,0.98,-0.49,0.9,0.87,-0.87,0.44,-0.4,-0.47,0.47,0.76],\n", - "[-0.87,0.19,0.94,-0.69,0.16,0.35,-0.51,-0.42,-0.59,0.31,0.06,-0.12,-0.32,-0.56,0.6,0.43,-0.52,-0.59,-0.66,-0.53],\n", - "[-0.43,-0.25,-0.84,0.18,0.43,-0.53,0.97,0.61,-0.38,-0.23,0.87,0.56,-0.01,-0.59,0.58,-0.24,-0.52,-0.61,0.08,-0.04],\n", - "[0.64,0.67,-0.76,-0.37,0.8,-0.37,-0.75,0.47,0.23,-0.87,0.3,-0.53,-0.15,0.4,0.83,0.42,-0.79,-0.09,0.17,-0.28]\n", - "]\n", - "\n", - "\n", - "\n", - "316\n", - "\n", - "RuleNeuron:316:lstm3_out__right(T) :- lstm3_out__i(T), lstm3_out__n(T). [transformation=identity, combination=elproduct]\n", - "val: [0.02,-0.02,-0.01,0.01,0.01,-0.03,-0.04,0,-0,0,0,0.02,-0,0.02,0,0.06,0.01,0.01,-0,-0.03]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "316->310\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "316->313\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "319\n", - "\n", - "RuleNeuron:319:lstm3_out__c(T) :- lstm3_out__left(T), lstm3_out__right(T). [transformation=identity]\n", - "val: [0.02,-0.03,-0.01,0.02,0.01,-0.04,-0.05,0,-0.01,-0,0,0.02,-0.01,0.04,0.01,0.07,0.01,0,0,-0.04]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: Sum\n", - "\n", - "\n", - "\n", - "319->307\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "319->316\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "325\n", - "\n", - "RuleNeuron:325:lstm3_out(T) :- lstm3_out__o(T), tanh(lstm3_out__c(T)). [transformation=identity, combination=elproduct]\n", - "val: [0.01,-0.01,-0,0.01,0.01,-0.02,-0.03,0,-0,-0,0,0.01,-0,0.02,0,0.04,0,0,0,-0.02]\n", - "grad: [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]\n", - "dim: [20, 1]\n", - "fcn: ElementProduct\n", - "\n", - "\n", - "\n", - "325->301\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "325->319\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "328\n", - "\n", - "WeightedRuleNeuron:328:predict :- {1, 20} lstm3_out(X). [transformation=sigmoid]\n", - "val: 0.5\n", - "grad: 0\n", - "dim: []\n", - "fcn: +Sigmoid\n", - "\n", - "\n", - "\n", - "328->325\n", - "\n", - "\n", - "24:w24:[1, 20]:[-0.9,-0.85,-0.99,-0.48,-0.69,0.15,-0.96,0.37,0.75,-0.44,-0.54,-0.2,0.73,-0.9,-0.24,0.51,-0.45,0.16,-0.38,0.32]\n", - "\n", - "\n", - "\n", - "329\n", - "\n", - "WeightedRuleNeuron:329:predict :- {1, 20} lstm3_out(X). [transformation=sigmoid]\n", - "val: 0.5\n", - "grad: 0\n", - "dim: []\n", - "fcn: +Sigmoid\n", - "\n", - "\n", - "\n", - "329->248\n", - "\n", - "\n", - "24:w24:[1, 20]:[-0.9,-0.85,-0.99,-0.48,-0.69,0.15,-0.96,0.37,0.75,-0.44,-0.54,-0.2,0.73,-0.9,-0.24,0.51,-0.45,0.16,-0.38,0.32]\n", - "\n", - "\n", - "\n", - "330\n", - "\n", - "WeightedRuleNeuron:330:predict :- {1, 20} lstm3_out(X). [transformation=sigmoid]\n", - "val: 0.5\n", - "grad: 0\n", - "dim: []\n", - "fcn: +Sigmoid\n", - "\n", - "\n", - "\n", - "330->273\n", - "\n", - "\n", - "24:w24:[1, 20]:[-0.9,-0.85,-0.99,-0.48,-0.69,0.15,-0.96,0.37,0.75,-0.44,-0.54,-0.2,0.73,-0.9,-0.24,0.51,-0.45,0.16,-0.38,0.32]\n", - "\n", - "\n", - "\n", - "331\n", - "\n", - "WeightedRuleNeuron:331:predict :- {1, 20} lstm3_out(X). [transformation=sigmoid]\n", - "val: 0.5\n", - "grad: 0\n", - "dim: []\n", - "fcn: +Sigmoid\n", - "\n", - "\n", - "\n", - "331->298\n", - "\n", - "\n", - "24:w24:[1, 20]:[-0.9,-0.85,-0.99,-0.48,-0.69,0.15,-0.96,0.37,0.75,-0.44,-0.54,-0.2,0.73,-0.9,-0.24,0.51,-0.45,0.16,-0.38,0.32]\n", - "\n", - "\n", - "\n", - "332\n", - "\n", - "WeightedRuleNeuron:332:predict :- {1, 20} lstm3_out(X). [transformation=sigmoid]\n", - "val: 0.5\n", - "grad: 0\n", - "dim: []\n", - "fcn: +Sigmoid\n", - "\n", - "\n", - "\n", - "332->4\n", - "\n", - "\n", - "24:w24:[1, 20]:[-0.9,-0.85,-0.99,-0.48,-0.69,0.15,-0.96,0.37,0.75,-0.44,-0.54,-0.2,0.73,-0.9,-0.24,0.51,-0.45,0.16,-0.38,0.32]\n", - "\n", - "\n", - "\n", - "333\n", - "\n", - "AggregationNeuron:333:predict :- {1, 20} lstm3_out(X). [transformation=sigmoid]\n", - "val: 0.5\n", - "grad: 0\n", - "dim: []\n", - "fcn: Average\n", - "\n", - "\n", - "\n", - "333->328\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "333->329\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "333->330\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "333->331\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "333->332\n", - "\n", - "\n", - "\n", - "\n", - "\n", - "334\n", - "\n", - "\n", - "\n", - "AtomNeuron:334:predict\n", - "val: 0.5\n", - "grad: 0\n", - "dim: []\n", - "fcn: +Identity\n", - "\n", - "\n", - "\n", - "334->333\n", - "\n", - "\n", - "\n", - "\n", - "" - ], - "text/plain": [ - "" - ] - }, - "execution_count": 10, - "metadata": {}, - "output_type": "execute_result" - } - ], - "source": [ - "# Display sample\n", - "built_dataset.samples[0].draw(img_type=\"svg\")" - ] - }, - { - "cell_type": "code", - "execution_count": 11, "id": "77a350e4", "metadata": {}, "outputs": [ @@ -4013,318 +211,318 @@ "name": "stdout", "output_type": "stream", "text": [ - "Epoch 0 Loss: 6.083160490000402 Mean loss: 0.06083160490000403\n", - "Epoch 1 Loss: 2.2840650084958822 Mean loss: 0.022840650084958822\n", - "Epoch 2 Loss: 1.4858856718724218 Mean loss: 0.014858856718724217\n", - "Epoch 3 Loss: 1.1878093438444064 Mean loss: 0.011878093438444063\n", - "Epoch 4 Loss: 1.0768520829935488 Mean loss: 0.010768520829935489\n", - "Epoch 5 Loss: 1.0123350387217116 Mean loss: 0.010123350387217116\n", - "Epoch 6 Loss: 0.9656553714956269 Mean loss: 0.009656553714956268\n", - "Epoch 7 Loss: 0.9324729977460346 Mean loss: 0.009324729977460346\n", - "Epoch 8 Loss: 0.9022516771087317 Mean loss: 0.009022516771087317\n", - "Epoch 9 Loss: 0.8790322979991856 Mean loss: 0.008790322979991856\n", - "Epoch 10 Loss: 0.8534107807542335 Mean loss: 0.008534107807542334\n", - "Epoch 11 Loss: 0.8324739065199404 Mean loss: 0.008324739065199404\n", - "Epoch 12 Loss: 0.807513596480301 Mean loss: 0.008075135964803009\n", - "Epoch 13 Loss: 0.7868600054668282 Mean loss: 0.007868600054668282\n", - "Epoch 14 Loss: 0.7617469762882513 Mean loss: 0.007617469762882512\n", - "Epoch 15 Loss: 0.7385103699472664 Mean loss: 0.0073851036994726635\n", - "Epoch 16 Loss: 0.7092846216249051 Mean loss: 0.0070928462162490515\n", - "Epoch 17 Loss: 0.6804915967261108 Mean loss: 0.006804915967261108\n", - "Epoch 18 Loss: 0.6464659471068225 Mean loss: 0.0064646594710682246\n", - "Epoch 19 Loss: 0.6116479375864817 Mean loss: 0.006116479375864817\n", - "Epoch 20 Loss: 0.5738761583275848 Mean loss: 0.005738761583275848\n", - "Epoch 21 Loss: 0.537949396048625 Mean loss: 0.00537949396048625\n", - "Epoch 22 Loss: 0.5018689123831259 Mean loss: 0.005018689123831259\n", - "Epoch 23 Loss: 0.47481070581329715 Mean loss: 0.004748107058132971\n", - "Epoch 24 Loss: 0.4471177682202233 Mean loss: 0.004471177682202233\n", - "Epoch 25 Loss: 0.4270975208381949 Mean loss: 0.0042709752083819495\n", - "Epoch 26 Loss: 0.4065414722428367 Mean loss: 0.004065414722428367\n", - "Epoch 27 Loss: 0.39059723308889716 Mean loss: 0.0039059723308889713\n", - "Epoch 28 Loss: 0.37502081585425073 Mean loss: 0.0037502081585425073\n", - "Epoch 29 Loss: 0.3616113501473641 Mean loss: 0.003616113501473641\n", - "Epoch 30 Loss: 0.3484865189725933 Mean loss: 0.003484865189725933\n", - "Epoch 31 Loss: 0.3363718305391395 Mean loss: 0.003363718305391395\n", - "Epoch 32 Loss: 0.32543222524656923 Mean loss: 0.0032543222524656925\n", - "Epoch 33 Loss: 0.3160366080328809 Mean loss: 0.003160366080328809\n", - "Epoch 34 Loss: 0.3079445370904387 Mean loss: 0.0030794453709043867\n", - "Epoch 35 Loss: 0.30088264406091886 Mean loss: 0.0030088264406091885\n", - "Epoch 36 Loss: 0.2945482716354868 Mean loss: 0.002945482716354868\n", - "Epoch 37 Loss: 0.2887927930553754 Mean loss: 0.002887927930553754\n", - "Epoch 38 Loss: 0.28355691395717453 Mean loss: 0.0028355691395717452\n", - "Epoch 39 Loss: 0.2788925951147424 Mean loss: 0.002788925951147424\n", - "Epoch 40 Loss: 0.2748177983640656 Mean loss: 0.002748177983640656\n", - "Epoch 41 Loss: 0.2712331222157517 Mean loss: 0.002712331222157517\n", - "Epoch 42 Loss: 0.267947591461536 Mean loss: 0.00267947591461536\n", - "Epoch 43 Loss: 0.26482637296627226 Mean loss: 0.0026482637296627227\n", - "Epoch 44 Loss: 0.2618278690600767 Mean loss: 0.002618278690600767\n", - "Epoch 45 Loss: 0.25895268297222657 Mean loss: 0.0025895268297222657\n", - "Epoch 46 Loss: 0.25620322431117626 Mean loss: 0.0025620322431117626\n", - "Epoch 47 Loss: 0.25357954706711766 Mean loss: 0.0025357954706711765\n", - "Epoch 48 Loss: 0.2510801694547265 Mean loss: 0.002510801694547265\n", - "Epoch 49 Loss: 0.24870342733602796 Mean loss: 0.0024870342733602797\n", - "Epoch 50 Loss: 0.24644654234583854 Mean loss: 0.0024644654234583854\n", - "Epoch 51 Loss: 0.24430506909732894 Mean loss: 0.0024430506909732894\n", - "Epoch 52 Loss: 0.24227211094403014 Mean loss: 0.0024227211094403015\n", - "Epoch 53 Loss: 0.24033816531522745 Mean loss: 0.0024033816531522745\n", - "Epoch 54 Loss: 0.2384912918492733 Mean loss: 0.002384912918492733\n", - "Epoch 55 Loss: 0.23671779926293116 Mean loss: 0.0023671779926293115\n", - "Epoch 56 Loss: 0.23500316361053725 Mean loss: 0.0023500316361053726\n", - "Epoch 57 Loss: 0.23333313790789803 Mean loss: 0.00233333137907898\n", - "Epoch 58 Loss: 0.23169491912773862 Mean loss: 0.0023169491912773863\n", - "Epoch 59 Loss: 0.2300783620904086 Mean loss: 0.002300783620904086\n", - "Epoch 60 Loss: 0.22847710697762427 Mean loss: 0.0022847710697762426\n", - "Epoch 61 Loss: 0.22688941172872537 Mean loss: 0.0022688941172872537\n", - "Epoch 62 Loss: 0.22531838589796252 Mean loss: 0.002253183858979625\n", - "Epoch 63 Loss: 0.22377141023167735 Mean loss: 0.0022377141023167736\n", - "Epoch 64 Loss: 0.2222587473180738 Mean loss: 0.002222587473180738\n", - "Epoch 65 Loss: 0.22079165523409952 Mean loss: 0.0022079165523409953\n", - "Epoch 66 Loss: 0.21938051101686143 Mean loss: 0.0021938051101686143\n", - "Epoch 67 Loss: 0.21803342482280744 Mean loss: 0.0021803342482280743\n", - "Epoch 68 Loss: 0.21675558896861644 Mean loss: 0.0021675558896861644\n", - "Epoch 69 Loss: 0.21554931657259155 Mean loss: 0.0021554931657259153\n", - "Epoch 70 Loss: 0.2144145327695956 Mean loss: 0.002144145327695956\n", - "Epoch 71 Loss: 0.21334944095874855 Mean loss: 0.0021334944095874856\n", - "Epoch 72 Loss: 0.21235115342080232 Mean loss: 0.002123511534208023\n", - "Epoch 73 Loss: 0.21141617601708299 Mean loss: 0.0021141617601708297\n", - "Epoch 74 Loss: 0.2105407210866146 Mean loss: 0.002105407210866146\n", - "Epoch 75 Loss: 0.2097208728094562 Mean loss: 0.002097208728094562\n", - "Epoch 76 Loss: 0.20895264943451458 Mean loss: 0.002089526494345146\n", - "Epoch 77 Loss: 0.20823200756652255 Mean loss: 0.0020823200756652253\n", - "Epoch 78 Loss: 0.20755482491472854 Mean loss: 0.0020755482491472852\n", - "Epoch 79 Loss: 0.206916886222528 Mean loss: 0.00206916886222528\n", - "Epoch 80 Loss: 0.2063138856816492 Mean loss: 0.0020631388568164923\n", - "Epoch 81 Loss: 0.20574144958590843 Mean loss: 0.0020574144958590843\n", - "Epoch 82 Loss: 0.20519517607491758 Mean loss: 0.0020519517607491757\n", - "Epoch 83 Loss: 0.204670684603382 Mean loss: 0.00204670684603382\n", - "Epoch 84 Loss: 0.2041636662069527 Mean loss: 0.002041636662069527\n", - "Epoch 85 Loss: 0.20366992626157315 Mean loss: 0.0020366992626157316\n", - "Epoch 86 Loss: 0.20318541338629365 Mean loss: 0.0020318541338629364\n", - "Epoch 87 Loss: 0.2027062307098171 Mean loss: 0.002027062307098171\n", - "Epoch 88 Loss: 0.20222862818392587 Mean loss: 0.002022286281839259\n", - "Epoch 89 Loss: 0.20174897644495168 Mean loss: 0.0020174897644495166\n", - "Epoch 90 Loss: 0.2012637238085716 Mean loss: 0.002012637238085716\n", - "Epoch 91 Loss: 0.20076933836761576 Mean loss: 0.0020076933836761577\n", - "Epoch 92 Loss: 0.2002622370197211 Mean loss: 0.002002622370197211\n", - "Epoch 93 Loss: 0.19973870293313342 Mean loss: 0.001997387029331334\n", - "Epoch 94 Loss: 0.1991947927612296 Mean loss: 0.001991947927612296\n", - "Epoch 95 Loss: 0.19862623512855923 Mean loss: 0.001986262351285592\n", - "Epoch 96 Loss: 0.198028322871328 Mean loss: 0.00198028322871328\n", - "Epoch 97 Loss: 0.19739580352459032 Mean loss: 0.001973958035245903\n", - "Epoch 98 Loss: 0.19672277588089948 Mean loss: 0.001967227758808995\n", - "Epoch 99 Loss: 0.1960026052328529 Mean loss: 0.001960026052328529\n", - "Epoch 100 Loss: 0.19522787562176763 Mean loss: 0.0019522787562176764\n", - "Epoch 101 Loss: 0.1943904010918067 Mean loss: 0.001943904010918067\n", - "Epoch 102 Loss: 0.19348130776081687 Mean loss: 0.0019348130776081687\n", - "Epoch 103 Loss: 0.19249113081415486 Mean loss: 0.0019249113081415486\n", - "Epoch 104 Loss: 0.1914096122305827 Mean loss: 0.001914096122305827\n", - "Epoch 105 Loss: 0.1902246016097717 Mean loss: 0.001902246016097717\n", - "Epoch 106 Loss: 0.18893570169672097 Mean loss: 0.0018893570169672097\n", - "Epoch 107 Loss: 0.18816446384864324 Mean loss: 0.0018816446384864324\n", - "Epoch 108 Loss: 0.19580734117764276 Mean loss: 0.0019580734117764275\n", - "Epoch 109 Loss: 0.19450393596976412 Mean loss: 0.0019450393596976412\n", - "Epoch 110 Loss: 0.19598815927830862 Mean loss: 0.001959881592783086\n", - "Epoch 111 Loss: 0.19019711408727544 Mean loss: 0.0019019711408727543\n", - "Epoch 112 Loss: 0.18578404152295358 Mean loss: 0.0018578404152295358\n", - "Epoch 113 Loss: 0.18577666774590582 Mean loss: 0.001857766677459058\n", - "Epoch 114 Loss: 0.1826241482441486 Mean loss: 0.001826241482441486\n", - "Epoch 115 Loss: 0.17754927472814208 Mean loss: 0.0017754927472814208\n", - "Epoch 116 Loss: 0.1788830011488663 Mean loss: 0.001788830011488663\n", - "Epoch 117 Loss: 0.17670998431936757 Mean loss: 0.0017670998431936757\n", - "Epoch 118 Loss: 0.16991395976844317 Mean loss: 0.0016991395976844318\n", - "Epoch 119 Loss: 0.17219193468624594 Mean loss: 0.0017219193468624595\n", - "Epoch 120 Loss: 0.16966548782351834 Mean loss: 0.0016966548782351835\n", - "Epoch 121 Loss: 0.16158054237288072 Mean loss: 0.0016158054237288071\n", - "Epoch 122 Loss: 0.16435186349478598 Mean loss: 0.0016435186349478598\n", - "Epoch 123 Loss: 0.15660026507801217 Mean loss: 0.0015660026507801218\n", - "Epoch 124 Loss: 0.15362313662330648 Mean loss: 0.0015362313662330648\n", - "Epoch 125 Loss: 0.1507173054258388 Mean loss: 0.001507173054258388\n" + "Epoch 0 Loss: 5.715762719156138 Mean loss: 0.05715762719156138\n", + "Epoch 1 Loss: 2.9277588745454284 Mean loss: 0.029277588745454285\n", + "Epoch 2 Loss: 1.7611229763634915 Mean loss: 0.017611229763634915\n", + "Epoch 3 Loss: 1.3051838788499823 Mean loss: 0.013051838788499824\n", + "Epoch 4 Loss: 1.1662684966567458 Mean loss: 0.011662684966567458\n", + "Epoch 5 Loss: 1.0855330067647546 Mean loss: 0.010855330067647545\n", + "Epoch 6 Loss: 1.0328900527402585 Mean loss: 0.010328900527402585\n", + "Epoch 7 Loss: 0.9959791472418418 Mean loss: 0.009959791472418418\n", + "Epoch 8 Loss: 0.9673639566131574 Mean loss: 0.009673639566131575\n", + "Epoch 9 Loss: 0.943190286218842 Mean loss: 0.00943190286218842\n", + "Epoch 10 Loss: 0.9214379508033702 Mean loss: 0.009214379508033702\n", + "Epoch 11 Loss: 0.9009343752858782 Mean loss: 0.009009343752858782\n", + "Epoch 12 Loss: 0.8807454481776903 Mean loss: 0.008807454481776902\n", + "Epoch 13 Loss: 0.859680635661283 Mean loss: 0.00859680635661283\n", + "Epoch 14 Loss: 0.8356006684848919 Mean loss: 0.00835600668484892\n", + "Epoch 15 Loss: 0.8046631387132461 Mean loss: 0.008046631387132462\n", + "Epoch 16 Loss: 0.7675871476228703 Mean loss: 0.007675871476228703\n", + "Epoch 17 Loss: 0.7434498832759233 Mean loss: 0.007434498832759233\n", + "Epoch 18 Loss: 0.7243190964729866 Mean loss: 0.007243190964729866\n", + "Epoch 19 Loss: 0.6991885489431245 Mean loss: 0.006991885489431245\n", + "Epoch 20 Loss: 0.6718486755730448 Mean loss: 0.006718486755730449\n", + "Epoch 21 Loss: 0.6454120936761021 Mean loss: 0.006454120936761021\n", + "Epoch 22 Loss: 0.6214004080215063 Mean loss: 0.006214004080215063\n", + "Epoch 23 Loss: 0.5994502912021199 Mean loss: 0.005994502912021199\n", + "Epoch 24 Loss: 0.5788167114021876 Mean loss: 0.005788167114021876\n", + "Epoch 25 Loss: 0.559247764033065 Mean loss: 0.00559247764033065\n", + "Epoch 26 Loss: 0.5407149614886022 Mean loss: 0.005407149614886022\n", + "Epoch 27 Loss: 0.523187610005583 Mean loss: 0.00523187610005583\n", + "Epoch 28 Loss: 0.5066117804763391 Mean loss: 0.0050661178047633915\n", + "Epoch 29 Loss: 0.49093151162512083 Mean loss: 0.004909315116251208\n", + "Epoch 30 Loss: 0.4760956277288028 Mean loss: 0.004760956277288028\n", + "Epoch 31 Loss: 0.462055536282987 Mean loss: 0.0046205553628298705\n", + "Epoch 32 Loss: 0.4487627612360135 Mean loss: 0.004487627612360135\n", + "Epoch 33 Loss: 0.4361689158412638 Mean loss: 0.004361689158412638\n", + "Epoch 34 Loss: 0.42422780636932467 Mean loss: 0.004242278063693246\n", + "Epoch 35 Loss: 0.4128990141009483 Mean loss: 0.0041289901410094835\n", + "Epoch 36 Loss: 0.40215289805012805 Mean loss: 0.004021528980501281\n", + "Epoch 37 Loss: 0.39197695098588625 Mean loss: 0.003919769509858863\n", + "Epoch 38 Loss: 0.3823822264113285 Mean loss: 0.0038238222641132852\n", + "Epoch 39 Loss: 0.3734067425037506 Mean loss: 0.0037340674250375063\n", + "Epoch 40 Loss: 0.3651111425593625 Mean loss: 0.003651111425593625\n", + "Epoch 41 Loss: 0.3575608198891939 Mean loss: 0.003575608198891939\n", + "Epoch 42 Loss: 0.3507916982689369 Mean loss: 0.003507916982689369\n", + "Epoch 43 Loss: 0.3447695631277147 Mean loss: 0.0034476956312771466\n", + "Epoch 44 Loss: 0.3393658030860818 Mean loss: 0.0033936580308608177\n", + "Epoch 45 Loss: 0.33436783168888845 Mean loss: 0.0033436783168888846\n", + "Epoch 46 Loss: 0.32952082876959177 Mean loss: 0.003295208287695918\n", + "Epoch 47 Loss: 0.32457992685019393 Mean loss: 0.003245799268501939\n", + "Epoch 48 Loss: 0.3193575261274225 Mean loss: 0.003193575261274225\n", + "Epoch 49 Loss: 0.31375871698299845 Mean loss: 0.0031375871698299846\n", + "Epoch 50 Loss: 0.3077914154585002 Mean loss: 0.003077914154585002\n", + "Epoch 51 Loss: 0.3015426679018456 Mean loss: 0.003015426679018456\n", + "Epoch 52 Loss: 0.2951366275174731 Mean loss: 0.002951366275174731\n", + "Epoch 53 Loss: 0.28869962741651306 Mean loss: 0.0028869962741651304\n", + "Epoch 54 Loss: 0.28234283400110505 Mean loss: 0.0028234283400110504\n", + "Epoch 55 Loss: 0.2761565324619596 Mean loss: 0.002761565324619596\n", + "Epoch 56 Loss: 0.27020668960665245 Mean loss: 0.0027020668960665243\n", + "Epoch 57 Loss: 0.264530374534784 Mean loss: 0.00264530374534784\n", + "Epoch 58 Loss: 0.2591341783043994 Mean loss: 0.002591341783043994\n", + "Epoch 59 Loss: 0.25400252751993224 Mean loss: 0.0025400252751993225\n", + "Epoch 60 Loss: 0.24911779579969673 Mean loss: 0.002491177957996967\n", + "Epoch 61 Loss: 0.24448547252984704 Mean loss: 0.0024448547252984704\n", + "Epoch 62 Loss: 0.24015198612730096 Mean loss: 0.0024015198612730095\n", + "Epoch 63 Loss: 0.23620183102347447 Mean loss: 0.0023620183102347446\n", + "Epoch 64 Loss: 0.23272635130473782 Mean loss: 0.002327263513047378\n", + "Epoch 65 Loss: 0.2297750273360357 Mean loss: 0.002297750273360357\n", + "Epoch 66 Loss: 0.2273225342833131 Mean loss: 0.002273225342833131\n", + "Epoch 67 Loss: 0.22527972711373448 Mean loss: 0.0022527972711373447\n", + "Epoch 68 Loss: 0.22353749966425981 Mean loss: 0.0022353749966425984\n", + "Epoch 69 Loss: 0.22200551402122554 Mean loss: 0.0022200551402122553\n", + "Epoch 70 Loss: 0.2206242104845903 Mean loss: 0.002206242104845903\n", + "Epoch 71 Loss: 0.21935768770290523 Mean loss: 0.0021935768770290523\n", + "Epoch 72 Loss: 0.21818299290192464 Mean loss: 0.0021818299290192462\n", + "Epoch 73 Loss: 0.2170832747880538 Mean loss: 0.002170832747880538\n", + "Epoch 74 Loss: 0.21604487825180313 Mean loss: 0.0021604487825180314\n", + "Epoch 75 Loss: 0.21505641862342847 Mean loss: 0.0021505641862342847\n", + "Epoch 76 Loss: 0.2141084583896326 Mean loss: 0.002141084583896326\n", + "Epoch 77 Loss: 0.21319325515469714 Mean loss: 0.0021319325515469713\n", + "Epoch 78 Loss: 0.2123044903394485 Mean loss: 0.0021230449033944852\n", + "Epoch 79 Loss: 0.21143701857387207 Mean loss: 0.002114370185738721\n", + "Epoch 80 Loss: 0.21058667971229067 Mean loss: 0.002105866797122907\n", + "Epoch 81 Loss: 0.2097501846081654 Mean loss: 0.0020975018460816543\n", + "Epoch 82 Loss: 0.20892505656191063 Mean loss: 0.002089250565619106\n", + "Epoch 83 Loss: 0.2081095940721001 Mean loss: 0.002081095940721001\n", + "Epoch 84 Loss: 0.20730281999808942 Mean loss: 0.002073028199980894\n", + "Epoch 85 Loss: 0.2065043945897857 Mean loss: 0.002065043945897857\n", + "Epoch 86 Loss: 0.20571448831191463 Mean loss: 0.0020571448831191463\n", + "Epoch 87 Loss: 0.20493362729760697 Mean loss: 0.00204933627297607\n", + "Epoch 88 Loss: 0.20416253390350145 Mean loss: 0.0020416253390350145\n", + "Epoch 89 Loss: 0.20340198527569672 Mean loss: 0.0020340198527569672\n", + "Epoch 90 Loss: 0.2026527059499982 Mean loss: 0.002026527059499982\n", + "Epoch 91 Loss: 0.2019153004166202 Mean loss: 0.002019153004166202\n", + "Epoch 92 Loss: 0.20119022243571869 Mean loss: 0.0020119022243571868\n", + "Epoch 93 Loss: 0.20047777225339308 Mean loss: 0.002004777722533931\n", + "Epoch 94 Loss: 0.1997781112405046 Mean loss: 0.001997781112405046\n", + "Epoch 95 Loss: 0.19909128485788063 Mean loss: 0.0019909128485788062\n", + "Epoch 96 Loss: 0.1984172476750606 Mean loss: 0.001984172476750606\n", + "Epoch 97 Loss: 0.19775588707180503 Mean loss: 0.00197755887071805\n", + "Epoch 98 Loss: 0.19710704443349975 Mean loss: 0.0019710704443349973\n", + "Epoch 99 Loss: 0.19647053388450983 Mean loss: 0.0019647053388450984\n", + "Epoch 100 Loss: 0.19584615904077207 Mean loss: 0.001958461590407721\n", + "Epoch 101 Loss: 0.19523372821404827 Mean loss: 0.0019523372821404828\n", + "Epoch 102 Loss: 0.19463306826017948 Mean loss: 0.0019463306826017948\n", + "Epoch 103 Loss: 0.19404403702570655 Mean loss: 0.0019404403702570655\n", + "Epoch 104 Loss: 0.19346653418987497 Mean loss: 0.0019346653418987497\n", + "Epoch 105 Loss: 0.19290051021733362 Mean loss: 0.0019290051021733361\n", + "Epoch 106 Loss: 0.1923459730863331 Mean loss: 0.0019234597308633312\n", + "Epoch 107 Loss: 0.1918029923903704 Mean loss: 0.001918029923903704\n", + "Epoch 108 Loss: 0.1912717002958099 Mean loss: 0.001912717002958099\n", + "Epoch 109 Loss: 0.1907522886591839 Mean loss: 0.001907522886591839\n", + "Epoch 110 Loss: 0.19024500136249248 Mean loss: 0.0019024500136249247\n", + "Epoch 111 Loss: 0.18975012061617288 Mean loss: 0.0018975012061617288\n", + "Epoch 112 Loss: 0.18926794561879703 Mean loss: 0.0018926794561879704\n", + "Epoch 113 Loss: 0.18879876157425932 Mean loss: 0.0018879876157425932\n", + "Epoch 114 Loss: 0.18834279669555534 Mean loss: 0.0018834279669555534\n", + "Epoch 115 Loss: 0.18790016453769684 Mean loss: 0.0018790016453769684\n", + "Epoch 116 Loss: 0.1874707888983872 Mean loss: 0.001874707888983872\n", + "Epoch 117 Loss: 0.1870543087323777 Mean loss: 0.0018705430873237768\n", + "Epoch 118 Loss: 0.18664996120448468 Mean loss: 0.0018664996120448469\n", + "Epoch 119 Loss: 0.18625644233648053 Mean loss: 0.0018625644233648054\n", + "Epoch 120 Loss: 0.18587174685836352 Mean loss: 0.0018587174685836352\n", + "Epoch 121 Loss: 0.18549299196910551 Mean loss: 0.001854929919691055\n", + "Epoch 122 Loss: 0.1851162337331781 Mean loss: 0.001851162337331781\n", + "Epoch 123 Loss: 0.18473628953972737 Mean loss: 0.0018473628953972736\n", + "Epoch 124 Loss: 0.18434658486947458 Mean loss: 0.0018434658486947459\n", + "Epoch 125 Loss: 0.1839390466178675 Mean loss: 0.001839390466178675\n", + "Epoch 126 Loss: 0.18350406735446725 Mean loss: 0.0018350406735446725\n", + "Epoch 127 Loss: 0.1830305646723058 Mean loss: 0.001830305646723058\n", + "Epoch 128 Loss: 0.1825061592355457 Mean loss: 0.001825061592355457\n", + "Epoch 129 Loss: 0.1819175022152493 Mean loss: 0.001819175022152493\n", + "Epoch 130 Loss: 0.1812508170923714 Mean loss: 0.001812508170923714\n", + "Epoch 131 Loss: 0.1804928184945482 Mean loss: 0.001804928184945482\n", + "Epoch 132 Loss: 0.179632380834296 Mean loss: 0.0017963238083429601\n", + "Epoch 133 Loss: 0.17866366499831146 Mean loss: 0.0017866366499831145\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Epoch 126 Loss: 0.14895697766636043 Mean loss: 0.0014895697766636044\n", - "Epoch 127 Loss: 0.14882017999683764 Mean loss: 0.0014882017999683765\n", - "Epoch 128 Loss: 0.15011007733618287 Mean loss: 0.0015011007733618286\n", - "Epoch 129 Loss: 0.1452656301834221 Mean loss: 0.001452656301834221\n", - "Epoch 130 Loss: 0.1473628173633166 Mean loss: 0.001473628173633166\n", - "Epoch 131 Loss: 0.14376499782897112 Mean loss: 0.0014376499782897112\n", - "Epoch 132 Loss: 0.1437787506164816 Mean loss: 0.001437787506164816\n", - "Epoch 133 Loss: 0.15047429676666596 Mean loss: 0.0015047429676666596\n", - "Epoch 134 Loss: 0.13399597816249426 Mean loss: 0.0013399597816249427\n", - "Epoch 135 Loss: 0.14216185125057132 Mean loss: 0.0014216185125057131\n", - "Epoch 136 Loss: 0.1374803619545433 Mean loss: 0.001374803619545433\n", - "Epoch 137 Loss: 0.13166273404984574 Mean loss: 0.0013166273404984575\n", - "Epoch 138 Loss: 0.13153480516568078 Mean loss: 0.0013153480516568078\n", - "Epoch 139 Loss: 0.1286446283226079 Mean loss: 0.001286446283226079\n", - "Epoch 140 Loss: 0.12786214645100585 Mean loss: 0.0012786214645100586\n", - "Epoch 141 Loss: 0.12862788937600045 Mean loss: 0.0012862788937600044\n", - "Epoch 142 Loss: 0.1276703336722236 Mean loss: 0.0012767033367222362\n", - "Epoch 143 Loss: 0.127294822300196 Mean loss: 0.00127294822300196\n", - "Epoch 144 Loss: 0.1255729539929333 Mean loss: 0.0012557295399293331\n", - "Epoch 145 Loss: 0.12449152453277823 Mean loss: 0.0012449152453277823\n", - "Epoch 146 Loss: 0.1241520881888694 Mean loss: 0.001241520881888694\n", - "Epoch 147 Loss: 0.1220946003392575 Mean loss: 0.001220946003392575\n", - "Epoch 148 Loss: 0.1386100768602433 Mean loss: 0.001386100768602433\n", - "Epoch 149 Loss: 0.1229360935889032 Mean loss: 0.001229360935889032\n", - "Epoch 150 Loss: 0.1199826774558212 Mean loss: 0.001199826774558212\n", - "Epoch 151 Loss: 0.1223851355092127 Mean loss: 0.0012238513550921271\n", - "Epoch 152 Loss: 0.11552160302678739 Mean loss: 0.0011552160302678739\n", - "Epoch 153 Loss: 0.1161009745508707 Mean loss: 0.001161009745508707\n", - "Epoch 154 Loss: 0.12080764589711339 Mean loss: 0.0012080764589711339\n", - "Epoch 155 Loss: 0.12319141996494747 Mean loss: 0.0012319141996494747\n", - "Epoch 156 Loss: 0.1158836811700347 Mean loss: 0.001158836811700347\n", - "Epoch 157 Loss: 0.1153594899823644 Mean loss: 0.001153594899823644\n", - "Epoch 158 Loss: 0.11476300608361631 Mean loss: 0.001147630060836163\n", - "Epoch 159 Loss: 0.11689415593979512 Mean loss: 0.001168941559397951\n", - "Epoch 160 Loss: 0.12683857040256535 Mean loss: 0.0012683857040256534\n", - "Epoch 161 Loss: 0.11129082249090212 Mean loss: 0.0011129082249090212\n", - "Epoch 162 Loss: 0.11960465663717192 Mean loss: 0.0011960465663717191\n", - "Epoch 163 Loss: 0.10512041441231634 Mean loss: 0.0010512041441231635\n", - "Epoch 164 Loss: 0.11538800816221598 Mean loss: 0.0011538800816221597\n", - "Epoch 165 Loss: 0.11731346268335903 Mean loss: 0.0011731346268335904\n", - "Epoch 166 Loss: 0.11516251003893982 Mean loss: 0.0011516251003893983\n", - "Epoch 167 Loss: 0.12985611424529997 Mean loss: 0.0012985611424529996\n", - "Epoch 168 Loss: 0.1013678255016612 Mean loss: 0.0010136782550166118\n", - "Epoch 169 Loss: 0.11977371861657109 Mean loss: 0.0011977371861657108\n", - "Epoch 170 Loss: 0.10990479788583829 Mean loss: 0.001099047978858383\n", - "Epoch 171 Loss: 0.10729259917806595 Mean loss: 0.0010729259917806595\n", - "Epoch 172 Loss: 0.11742794851810652 Mean loss: 0.001174279485181065\n", - "Epoch 173 Loss: 0.10034025938796261 Mean loss: 0.001003402593879626\n", - "Epoch 174 Loss: 0.11841517654673434 Mean loss: 0.0011841517654673434\n", - "Epoch 175 Loss: 0.10567052249284892 Mean loss: 0.0010567052249284893\n", - "Epoch 176 Loss: 0.10665246694340459 Mean loss: 0.0010665246694340459\n", - "Epoch 177 Loss: 0.11198522948406578 Mean loss: 0.0011198522948406577\n", - "Epoch 178 Loss: 0.1029199972950369 Mean loss: 0.0010291999729503689\n", - "Epoch 179 Loss: 0.1220893882054049 Mean loss: 0.001220893882054049\n", - "Epoch 180 Loss: 0.10523872677439694 Mean loss: 0.0010523872677439695\n", - "Epoch 181 Loss: 0.11786081831993529 Mean loss: 0.0011786081831993529\n", - "Epoch 182 Loss: 0.10205034063082091 Mean loss: 0.0010205034063082092\n", - "Epoch 183 Loss: 0.11248286207461641 Mean loss: 0.0011248286207461641\n", - "Epoch 184 Loss: 0.09721189210986972 Mean loss: 0.0009721189210986972\n", - "Epoch 185 Loss: 0.11654154671085364 Mean loss: 0.0011654154671085363\n", - "Epoch 186 Loss: 0.10178745603789904 Mean loss: 0.0010178745603789904\n", - "Epoch 187 Loss: 0.10461227897362944 Mean loss: 0.0010461227897362944\n", - "Epoch 188 Loss: 0.10680712355881301 Mean loss: 0.00106807123558813\n", - "Epoch 189 Loss: 0.0987064391373997 Mean loss: 0.000987064391373997\n", - "Epoch 190 Loss: 0.1082486093232308 Mean loss: 0.001082486093232308\n", - "Epoch 191 Loss: 0.10169815085834212 Mean loss: 0.0010169815085834212\n", - "Epoch 192 Loss: 0.12690476250915858 Mean loss: 0.0012690476250915858\n", - "Epoch 193 Loss: 0.11895071603544563 Mean loss: 0.0011895071603544563\n", - "Epoch 194 Loss: 0.1252819051938504 Mean loss: 0.0012528190519385041\n", - "Epoch 195 Loss: 0.11930809673156229 Mean loss: 0.0011930809673156228\n", - "Epoch 196 Loss: 0.1021465660340279 Mean loss: 0.001021465660340279\n", - "Epoch 197 Loss: 0.1042080034115485 Mean loss: 0.0010420800341154849\n", - "Epoch 198 Loss: 0.09730490637930274 Mean loss: 0.0009730490637930274\n", - "Epoch 199 Loss: 0.10960531626045962 Mean loss: 0.0010960531626045963\n", - "Epoch 200 Loss: 0.09932577426342126 Mean loss: 0.0009932577426342126\n", - "Epoch 201 Loss: 0.11358885362828214 Mean loss: 0.0011358885362828214\n", - "Epoch 202 Loss: 0.10063435613182822 Mean loss: 0.0010063435613182823\n", - "Epoch 203 Loss: 0.10894849033688014 Mean loss: 0.0010894849033688014\n", - "Epoch 204 Loss: 0.1036726490302544 Mean loss: 0.001036726490302544\n", - "Epoch 205 Loss: 0.10189171000229726 Mean loss: 0.0010189171000229726\n", - "Epoch 206 Loss: 0.10731966998567467 Mean loss: 0.0010731966998567468\n", - "Epoch 207 Loss: 0.09836754939826567 Mean loss: 0.0009836754939826567\n", - "Epoch 208 Loss: 0.115603464983346 Mean loss: 0.00115603464983346\n", - "Epoch 209 Loss: 0.1009500571614545 Mean loss: 0.001009500571614545\n", - "Epoch 210 Loss: 0.1213533513147023 Mean loss: 0.001213533513147023\n", - "Epoch 211 Loss: 0.1001639038618272 Mean loss: 0.0010016390386182721\n", - "Epoch 212 Loss: 0.11246077738658453 Mean loss: 0.0011246077738658452\n", - "Epoch 213 Loss: 0.10676459906063875 Mean loss: 0.0010676459906063874\n", - "Epoch 214 Loss: 0.11501875218568439 Mean loss: 0.0011501875218568438\n", - "Epoch 215 Loss: 0.11205418595556486 Mean loss: 0.0011205418595556487\n", - "Epoch 216 Loss: 0.09569108439226329 Mean loss: 0.0009569108439226329\n", - "Epoch 217 Loss: 0.10156188681893441 Mean loss: 0.001015618868189344\n", - "Epoch 218 Loss: 0.09422111132922581 Mean loss: 0.000942211113292258\n", - "Epoch 219 Loss: 0.09789595305217748 Mean loss: 0.0009789595305217747\n", - "Epoch 220 Loss: 0.09393177034407546 Mean loss: 0.0009393177034407546\n", - "Epoch 221 Loss: 0.10518339501590486 Mean loss: 0.0010518339501590485\n", - "Epoch 222 Loss: 0.09807845552939332 Mean loss: 0.0009807845552939332\n", - "Epoch 223 Loss: 0.10880643701705353 Mean loss: 0.0010880643701705354\n", - "Epoch 224 Loss: 0.1020052866041747 Mean loss: 0.001020052866041747\n", - "Epoch 225 Loss: 0.10901518415386335 Mean loss: 0.0010901518415386336\n", - "Epoch 226 Loss: 0.10223555076419197 Mean loss: 0.0010223555076419198\n", - "Epoch 227 Loss: 0.1091480621155891 Mean loss: 0.001091480621155891\n", - "Epoch 228 Loss: 0.1003731089883593 Mean loss: 0.001003731089883593\n", - "Epoch 229 Loss: 0.10825074000352475 Mean loss: 0.0010825074000352475\n", - "Epoch 230 Loss: 0.09794348918057849 Mean loss: 0.000979434891805785\n", - "Epoch 231 Loss: 0.10449357070418688 Mean loss: 0.001044935707041869\n", - "Epoch 232 Loss: 0.09778684574190152 Mean loss: 0.0009778684574190152\n", - "Epoch 233 Loss: 0.10646128196840761 Mean loss: 0.0010646128196840762\n", - "Epoch 234 Loss: 0.10197685273767754 Mean loss: 0.0010197685273767754\n", - "Epoch 235 Loss: 0.09574424243710178 Mean loss: 0.0009574424243710178\n", - "Epoch 236 Loss: 0.09799143209907854 Mean loss: 0.0009799143209907853\n", - "Epoch 237 Loss: 0.09739274446061073 Mean loss: 0.0009739274446061073\n", - "Epoch 238 Loss: 0.09834212265231346 Mean loss: 0.0009834212265231346\n", - "Epoch 239 Loss: 0.09628509296203167 Mean loss: 0.0009628509296203166\n", - "Epoch 240 Loss: 0.0954098455837166 Mean loss: 0.000954098455837166\n", - "Epoch 241 Loss: 0.09944375177783467 Mean loss: 0.0009944375177783466\n", - "Epoch 242 Loss: 0.09840629812152626 Mean loss: 0.0009840629812152626\n", - "Epoch 243 Loss: 0.0942411837881239 Mean loss: 0.0009424118378812391\n", - "Epoch 244 Loss: 0.0897834051870545 Mean loss: 0.0008978340518705449\n", - "Epoch 245 Loss: 0.09773485187185174 Mean loss: 0.0009773485187185174\n", - "Epoch 246 Loss: 0.09111532666492525 Mean loss: 0.0009111532666492525\n", - "Epoch 247 Loss: 0.09689982037886531 Mean loss: 0.0009689982037886532\n", - "Epoch 248 Loss: 0.08986459250564661 Mean loss: 0.0008986459250564661\n", - "Epoch 249 Loss: 0.09596942526637334 Mean loss: 0.0009596942526637333\n", - "Epoch 250 Loss: 0.09171480532400897 Mean loss: 0.0009171480532400896\n", - "Epoch 251 Loss: 0.0946808402910875 Mean loss: 0.0009468084029108751\n", - "Epoch 252 Loss: 0.09149382783573408 Mean loss: 0.0009149382783573408\n", - "Epoch 253 Loss: 0.09387105855534868 Mean loss: 0.0009387105855534868\n" + "Epoch 134 Loss: 0.17759167189052794 Mean loss: 0.0017759167189052793\n", + "Epoch 135 Loss: 0.17644053981762586 Mean loss: 0.0017644053981762586\n", + "Epoch 136 Loss: 0.17526154450490783 Mean loss: 0.0017526154450490782\n", + "Epoch 137 Loss: 0.17413095922905666 Mean loss: 0.0017413095922905665\n", + "Epoch 138 Loss: 0.17312538813821396 Mean loss: 0.0017312538813821397\n", + "Epoch 139 Loss: 0.17228097052651925 Mean loss: 0.0017228097052651924\n", + "Epoch 140 Loss: 0.17157358319005347 Mean loss: 0.0017157358319005347\n", + "Epoch 141 Loss: 0.17094400200536386 Mean loss: 0.0017094400200536387\n", + "Epoch 142 Loss: 0.17033847752046316 Mean loss: 0.0017033847752046317\n", + "Epoch 143 Loss: 0.16972759209255095 Mean loss: 0.0016972759209255095\n", + "Epoch 144 Loss: 0.16910260939603503 Mean loss: 0.0016910260939603503\n", + "Epoch 145 Loss: 0.16846511472376482 Mean loss: 0.0016846511472376482\n", + "Epoch 146 Loss: 0.16781930253814165 Mean loss: 0.0016781930253814165\n", + "Epoch 147 Loss: 0.16716849950688698 Mean loss: 0.0016716849950688699\n", + "Epoch 148 Loss: 0.16651447146704423 Mean loss: 0.0016651447146704423\n", + "Epoch 149 Loss: 0.16585787738378577 Mean loss: 0.0016585787738378577\n", + "Epoch 150 Loss: 0.16519889638400126 Mean loss: 0.0016519889638400126\n", + "Epoch 151 Loss: 0.164537652974168 Mean loss: 0.00164537652974168\n", + "Epoch 152 Loss: 0.16387440079540924 Mean loss: 0.0016387440079540924\n", + "Epoch 153 Loss: 0.16320954644242483 Mean loss: 0.0016320954644242483\n", + "Epoch 154 Loss: 0.16254360375306082 Mean loss: 0.0016254360375306082\n", + "Epoch 155 Loss: 0.1618771372928104 Mean loss: 0.001618771372928104\n", + "Epoch 156 Loss: 0.16121072106841697 Mean loss: 0.0016121072106841698\n", + "Epoch 157 Loss: 0.16054491718718786 Mean loss: 0.0016054491718718786\n", + "Epoch 158 Loss: 0.1598802697070902 Mean loss: 0.0015988026970709022\n", + "Epoch 159 Loss: 0.15921730689127644 Mean loss: 0.0015921730689127645\n", + "Epoch 160 Loss: 0.1585565464665272 Mean loss: 0.001585565464665272\n", + "Epoch 161 Loss: 0.15789850066384373 Mean loss: 0.0015789850066384372\n", + "Epoch 162 Loss: 0.1572436796184536 Mean loss: 0.0015724367961845358\n", + "Epoch 163 Loss: 0.1565925928197653 Mean loss: 0.0015659259281976532\n", + "Epoch 164 Loss: 0.15594574880428771 Mean loss: 0.0015594574880428771\n", + "Epoch 165 Loss: 0.1553036533580766 Mean loss: 0.0015530365335807662\n", + "Epoch 166 Loss: 0.15466680630710566 Mean loss: 0.0015466680630710566\n", + "Epoch 167 Loss: 0.15403569666069408 Mean loss: 0.0015403569666069409\n", + "Epoch 168 Loss: 0.15341079555173948 Mean loss: 0.0015341079555173949\n", + "Epoch 169 Loss: 0.15279254619508775 Mean loss: 0.0015279254619508776\n", + "Epoch 170 Loss: 0.152181350057888 Mean loss: 0.00152181350057888\n", + "Epoch 171 Loss: 0.15157754867778553 Mean loss: 0.0015157754867778552\n", + "Epoch 172 Loss: 0.1509814011174741 Mean loss: 0.001509814011174741\n", + "Epoch 173 Loss: 0.15039305790689453 Mean loss: 0.0015039305790689452\n", + "Epoch 174 Loss: 0.14981253344516474 Mean loss: 0.0014981253344516473\n", + "Epoch 175 Loss: 0.1492396800902562 Mean loss: 0.0014923968009025622\n", + "Epoch 176 Loss: 0.148674168326758 Mean loss: 0.00148674168326758\n", + "Epoch 177 Loss: 0.14811547809846778 Mean loss: 0.001481154780984678\n", + "Epoch 178 Loss: 0.14756290612133063 Mean loss: 0.0014756290612133063\n", + "Epoch 179 Loss: 0.1470155922601608 Mean loss: 0.001470155922601608\n", + "Epoch 180 Loss: 0.14647256467358702 Mean loss: 0.0014647256467358702\n", + "Epoch 181 Loss: 0.1459327988843325 Mean loss: 0.001459327988843325\n", + "Epoch 182 Loss: 0.1453952815238505 Mean loss: 0.001453952815238505\n", + "Epoch 183 Loss: 0.14485906705454016 Mean loss: 0.0014485906705454016\n", + "Epoch 184 Loss: 0.14432331679880647 Mean loss: 0.0014432331679880648\n", + "Epoch 185 Loss: 0.1437873143319819 Mean loss: 0.0014378731433198192\n", + "Epoch 186 Loss: 0.14325045822760757 Mean loss: 0.0014325045822760757\n", + "Epoch 187 Loss: 0.14271223958009574 Mean loss: 0.0014271223958009574\n", + "Epoch 188 Loss: 0.142172215095839 Mean loss: 0.00142172215095839\n", + "Epoch 189 Loss: 0.14162998572891852 Mean loss: 0.0014162998572891852\n", + "Epoch 190 Loss: 0.1410851867096637 Mean loss: 0.001410851867096637\n", + "Epoch 191 Loss: 0.14053748961010812 Mean loss: 0.0014053748961010812\n", + "Epoch 192 Loss: 0.1399866131207283 Mean loss: 0.001399866131207283\n", + "Epoch 193 Loss: 0.13943233770875316 Mean loss: 0.0013943233770875316\n", + "Epoch 194 Loss: 0.138874520071301 Mean loss: 0.0013887452007130098\n", + "Epoch 195 Loss: 0.1383131051086693 Mean loss: 0.001383131051086693\n", + "Epoch 196 Loss: 0.1377481347851872 Mean loss: 0.001377481347851872\n", + "Epoch 197 Loss: 0.1371797540956103 Mean loss: 0.0013717975409561032\n", + "Epoch 198 Loss: 0.1366082145149286 Mean loss: 0.001366082145149286\n", + "Epoch 199 Loss: 0.13603387522869817 Mean loss: 0.0013603387522869816\n", + "Epoch 200 Loss: 0.13545720246809503 Mean loss: 0.0013545720246809503\n", + "Epoch 201 Loss: 0.1348787674640386 Mean loss: 0.001348787674640386\n", + "Epoch 202 Loss: 0.13429924375944688 Mean loss: 0.0013429924375944688\n", + "Epoch 203 Loss: 0.13371940476736877 Mean loss: 0.0013371940476736876\n", + "Epoch 204 Loss: 0.13314012252599722 Mean loss: 0.0013314012252599722\n", + "Epoch 205 Loss: 0.13256236860606466 Mean loss: 0.0013256236860606466\n", + "Epoch 206 Loss: 0.1319872180481535 Mean loss: 0.001319872180481535\n", + "Epoch 207 Loss: 0.13141585698591668 Mean loss: 0.0013141585698591667\n", + "Epoch 208 Loss: 0.13084959425786313 Mean loss: 0.0013084959425786313\n", + "Epoch 209 Loss: 0.13028987699785982 Mean loss: 0.0013028987699785983\n", + "Epoch 210 Loss: 0.12973831018878595 Mean loss: 0.0012973831018878595\n", + "Epoch 211 Loss: 0.12919668058149178 Mean loss: 0.0012919668058149177\n", + "Epoch 212 Loss: 0.1286669859368506 Mean loss: 0.001286669859368506\n", + "Epoch 213 Loss: 0.12815147064077437 Mean loss: 0.0012815147064077436\n", + "Epoch 214 Loss: 0.1276526681117191 Mean loss: 0.0012765266811171911\n", + "Epoch 215 Loss: 0.127173449966066 Mean loss: 0.00127173449966066\n", + "Epoch 216 Loss: 0.1267170832312286 Mean loss: 0.001267170832312286\n", + "Epoch 217 Loss: 0.12628729942230912 Mean loss: 0.0012628729942230912\n", + "Epoch 218 Loss: 0.12588837722837865 Mean loss: 0.0012588837722837865\n", + "Epoch 219 Loss: 0.1255252265375812 Mean loss: 0.001255252265375812\n", + "Epoch 220 Loss: 0.12520343908486295 Mean loss: 0.0012520343908486295\n", + "Epoch 221 Loss: 0.12492925759980228 Mean loss: 0.0012492925759980229\n", + "Epoch 222 Loss: 0.12470941293888725 Mean loss: 0.0012470941293888725\n", + "Epoch 223 Loss: 0.12455074901198011 Mean loss: 0.0012455074901198012\n", + "Epoch 224 Loss: 0.12445949085728787 Mean loss: 0.0012445949085728787\n", + "Epoch 225 Loss: 0.1244400346345009 Mean loss: 0.001244400346345009\n", + "Epoch 226 Loss: 0.12449335348984245 Mean loss: 0.0012449335348984245\n", + "Epoch 227 Loss: 0.12461531922387385 Mean loss: 0.0012461531922387384\n", + "Epoch 228 Loss: 0.12479515266177997 Mean loss: 0.0012479515266177997\n", + "Epoch 229 Loss: 0.12501418220702823 Mean loss: 0.0012501418220702823\n", + "Epoch 230 Loss: 0.12524600603326538 Mean loss: 0.0012524600603326537\n", + "Epoch 231 Loss: 0.12545964105885957 Mean loss: 0.0012545964105885956\n", + "Epoch 232 Loss: 0.12562439178514492 Mean loss: 0.0012562439178514491\n", + "Epoch 233 Loss: 0.12571627467655216 Mean loss: 0.0012571627467655216\n", + "Epoch 234 Loss: 0.12572740417497183 Mean loss: 0.0012572740417497183\n", + "Epoch 235 Loss: 0.12566421319041446 Mean loss: 0.0012566421319041446\n", + "Epoch 236 Loss: 0.12554022012398025 Mean loss: 0.0012554022012398025\n", + "Epoch 237 Loss: 0.12538186130044465 Mean loss: 0.0012538186130044465\n", + "Epoch 238 Loss: 0.12521932708144687 Mean loss: 0.0012521932708144687\n", + "Epoch 239 Loss: 0.12507113536810444 Mean loss: 0.0012507113536810444\n", + "Epoch 240 Loss: 0.12492832782433191 Mean loss: 0.0012492832782433191\n", + "Epoch 241 Loss: 0.12478197982103939 Mean loss: 0.0012478197982103939\n", + "Epoch 242 Loss: 0.12466294046424488 Mean loss: 0.0012466294046424489\n", + "Epoch 243 Loss: 0.12461052487921535 Mean loss: 0.0012461052487921535\n", + "Epoch 244 Loss: 0.12457305264758035 Mean loss: 0.0012457305264758035\n", + "Epoch 245 Loss: 0.12431741980872443 Mean loss: 0.0012431741980872444\n", + "Epoch 246 Loss: 0.12381713439807265 Mean loss: 0.0012381713439807264\n", + "Epoch 247 Loss: 0.12344432917940093 Mean loss: 0.0012344432917940093\n", + "Epoch 248 Loss: 0.12254093431244248 Mean loss: 0.0012254093431244247\n", + "Epoch 249 Loss: 0.12374290458646707 Mean loss: 0.0012374290458646707\n", + "Epoch 250 Loss: 0.1313600041513636 Mean loss: 0.001313600041513636\n", + "Epoch 251 Loss: 0.1396210379131796 Mean loss: 0.001396210379131796\n", + "Epoch 252 Loss: 0.12953141655998612 Mean loss: 0.0012953141655998612\n", + "Epoch 253 Loss: 0.12193954632208612 Mean loss: 0.0012193954632208613\n", + "Epoch 254 Loss: 0.11511606108934311 Mean loss: 0.0011511606108934311\n", + "Epoch 255 Loss: 0.13523000040869948 Mean loss: 0.0013523000040869947\n", + "Epoch 256 Loss: 0.14015283485335012 Mean loss: 0.001401528348533501\n", + "Epoch 257 Loss: 0.19466152203985948 Mean loss: 0.0019466152203985948\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Epoch 254 Loss: 0.09434323371034338 Mean loss: 0.0009434323371034338\n", - "Epoch 255 Loss: 0.09450034617171188 Mean loss: 0.0009450034617171188\n", - "Epoch 256 Loss: 0.12156915306071207 Mean loss: 0.0012156915306071208\n", - "Epoch 257 Loss: 0.1392795370284876 Mean loss: 0.0013927953702848762\n", - "Epoch 258 Loss: 0.08929249353623421 Mean loss: 0.0008929249353623421\n", - "Epoch 259 Loss: 0.09398243261584227 Mean loss: 0.0009398243261584227\n", - "Epoch 260 Loss: 0.08239302578103064 Mean loss: 0.0008239302578103064\n", - "Epoch 261 Loss: 0.08789133305047872 Mean loss: 0.0008789133305047872\n", - "Epoch 262 Loss: 0.08317340457395869 Mean loss: 0.0008317340457395869\n", - "Epoch 263 Loss: 0.08239071055252592 Mean loss: 0.0008239071055252592\n", - "Epoch 264 Loss: 0.08302280720106538 Mean loss: 0.0008302280720106538\n", - "Epoch 265 Loss: 0.08362600675293391 Mean loss: 0.0008362600675293391\n", - "Epoch 266 Loss: 0.08375331041873296 Mean loss: 0.0008375331041873297\n", - "Epoch 267 Loss: 0.08372338607674298 Mean loss: 0.0008372338607674298\n", - "Epoch 268 Loss: 0.0836066704860989 Mean loss: 0.000836066704860989\n", - "Epoch 269 Loss: 0.08333407508201558 Mean loss: 0.0008333407508201558\n", - "Epoch 270 Loss: 0.08341751949550281 Mean loss: 0.0008341751949550282\n", - "Epoch 271 Loss: 0.08397763642994897 Mean loss: 0.0008397763642994897\n", - "Epoch 272 Loss: 0.08405642086856399 Mean loss: 0.0008405642086856399\n", - "Epoch 273 Loss: 0.08349180637854459 Mean loss: 0.0008349180637854458\n", - "Epoch 274 Loss: 0.08348449112454358 Mean loss: 0.0008348449112454358\n", - "Epoch 275 Loss: 0.08423419134359587 Mean loss: 0.0008423419134359587\n", - "Epoch 276 Loss: 0.08424321019536178 Mean loss: 0.0008424321019536178\n", - "Epoch 277 Loss: 0.0828091234227086 Mean loss: 0.0008280912342270861\n", - "Epoch 278 Loss: 0.08211766184943146 Mean loss: 0.0008211766184943146\n", - "Epoch 279 Loss: 0.08781183618847502 Mean loss: 0.0008781183618847501\n", - "Epoch 280 Loss: 0.08463019298563688 Mean loss: 0.0008463019298563688\n", - "Epoch 281 Loss: 0.08194688822453519 Mean loss: 0.0008194688822453519\n", - "Epoch 282 Loss: 0.08200901937313357 Mean loss: 0.0008200901937313357\n", - "Epoch 283 Loss: 0.08399670991547481 Mean loss: 0.0008399670991547481\n", - "Epoch 284 Loss: 0.09225427238230037 Mean loss: 0.0009225427238230037\n", - "Epoch 285 Loss: 0.10908554145905743 Mean loss: 0.0010908554145905743\n", - "Epoch 286 Loss: 0.09534500026763959 Mean loss: 0.0009534500026763959\n", - "Epoch 287 Loss: 0.08448496928843417 Mean loss: 0.0008448496928843418\n", - "Epoch 288 Loss: 0.07501055803040634 Mean loss: 0.0007501055803040633\n", - "Epoch 289 Loss: 0.07750343929131569 Mean loss: 0.0007750343929131568\n", - "Epoch 290 Loss: 0.07832753695685768 Mean loss: 0.0007832753695685767\n", - "Epoch 291 Loss: 0.08001351938269927 Mean loss: 0.0008001351938269926\n", - "Epoch 292 Loss: 0.07829180069822811 Mean loss: 0.0007829180069822811\n", - "Epoch 293 Loss: 0.09550637483870658 Mean loss: 0.0009550637483870658\n", - "Epoch 294 Loss: 0.13488568872335843 Mean loss: 0.0013488568872335844\n", - "Epoch 295 Loss: 0.20717020653058996 Mean loss: 0.0020717020653058995\n", - "Epoch 296 Loss: 0.094329752895172 Mean loss: 0.0009432975289517199\n", - "Epoch 297 Loss: 0.07303287083098921 Mean loss: 0.0007303287083098921\n", - "Epoch 298 Loss: 0.068320019350551 Mean loss: 0.00068320019350551\n", - "Epoch 299 Loss: 0.0690655598732786 Mean loss: 0.000690655598732786\n" + "Epoch 258 Loss: 0.2386396544888986 Mean loss: 0.0023863965448889858\n", + "Epoch 259 Loss: 0.31201955073719995 Mean loss: 0.0031201955073719995\n", + "Epoch 260 Loss: 0.16216760835810654 Mean loss: 0.0016216760835810653\n", + "Epoch 261 Loss: 0.16648864338415775 Mean loss: 0.0016648864338415774\n", + "Epoch 262 Loss: 0.15540133652128238 Mean loss: 0.0015540133652128239\n", + "Epoch 263 Loss: 0.1493506538961854 Mean loss: 0.001493506538961854\n", + "Epoch 264 Loss: 0.139459332727619 Mean loss: 0.00139459332727619\n", + "Epoch 265 Loss: 0.1334717124180713 Mean loss: 0.0013347171241807128\n", + "Epoch 266 Loss: 0.11855875458958151 Mean loss: 0.0011855875458958152\n", + "Epoch 267 Loss: 0.11435600589535014 Mean loss: 0.0011435600589535014\n", + "Epoch 268 Loss: 0.1123635784064095 Mean loss: 0.001123635784064095\n", + "Epoch 269 Loss: 0.11118093040935359 Mean loss: 0.001111809304093536\n", + "Epoch 270 Loss: 0.11083272283506919 Mean loss: 0.001108327228350692\n", + "Epoch 271 Loss: 0.11096433694544319 Mean loss: 0.001109643369454432\n", + "Epoch 272 Loss: 0.11124282560667885 Mean loss: 0.0011124282560667886\n", + "Epoch 273 Loss: 0.11153044763219729 Mean loss: 0.0011153044763219729\n", + "Epoch 274 Loss: 0.11183847474497101 Mean loss: 0.00111838474744971\n", + "Epoch 275 Loss: 0.11206685988234781 Mean loss: 0.001120668598823478\n", + "Epoch 276 Loss: 0.11237973601835687 Mean loss: 0.0011237973601835687\n", + "Epoch 277 Loss: 0.11256486950344263 Mean loss: 0.0011256486950344264\n", + "Epoch 278 Loss: 0.11292920593137093 Mean loss: 0.0011292920593137094\n", + "Epoch 279 Loss: 0.11303735848023178 Mean loss: 0.0011303735848023177\n", + "Epoch 280 Loss: 0.11360464250177574 Mean loss: 0.0011360464250177575\n", + "Epoch 281 Loss: 0.11347393460645638 Mean loss: 0.0011347393460645638\n", + "Epoch 282 Loss: 0.1145471192345662 Mean loss: 0.001145471192345662\n", + "Epoch 283 Loss: 0.11389528270422605 Mean loss: 0.0011389528270422606\n", + "Epoch 284 Loss: 0.11569802798554582 Mean loss: 0.0011569802798554582\n", + "Epoch 285 Loss: 0.11458081657180909 Mean loss: 0.0011458081657180908\n", + "Epoch 286 Loss: 0.11640913599212238 Mean loss: 0.0011640913599212238\n", + "Epoch 287 Loss: 0.11590571841147927 Mean loss: 0.0011590571841147927\n", + "Epoch 288 Loss: 0.11645015069177882 Mean loss: 0.0011645015069177882\n", + "Epoch 289 Loss: 0.11906325723716964 Mean loss: 0.0011906325723716963\n", + "Epoch 290 Loss: 0.11968328788485184 Mean loss: 0.0011968328788485184\n", + "Epoch 291 Loss: 0.11872443331823831 Mean loss: 0.0011872443331823832\n", + "Epoch 292 Loss: 0.13143560921834582 Mean loss: 0.0013143560921834582\n", + "Epoch 293 Loss: 0.1302986078298137 Mean loss: 0.001302986078298137\n", + "Epoch 294 Loss: 0.11141804744501499 Mean loss: 0.00111418047445015\n", + "Epoch 295 Loss: 0.11860539878338315 Mean loss: 0.0011860539878338315\n", + "Epoch 296 Loss: 0.11598736596689051 Mean loss: 0.001159873659668905\n", + "Epoch 297 Loss: 0.12085387699001898 Mean loss: 0.00120853876990019\n", + "Epoch 298 Loss: 0.12358426447291998 Mean loss: 0.0012358426447291997\n", + "Epoch 299 Loss: 0.10966639497011618 Mean loss: 0.0010966639497011618\n" ] } ], @@ -4347,23 +545,23 @@ }, { "cell_type": "code", - "execution_count": 12, + "execution_count": 11, "id": "2248561a", "metadata": {}, "outputs": [ { "data": { "text/plain": [ - "[]" + "[]" ] }, - "execution_count": 12, + "execution_count": 11, "metadata": {}, "output_type": "execute_result" }, { "data": { - "image/png": "iVBORw0KGgoAAAANSUhEUgAAAXgAAAEGCAYAAABvtY4XAAAAOXRFWHRTb2Z0d2FyZQBNYXRwbG90bGliIHZlcnNpb24zLjUuMSwgaHR0cHM6Ly9tYXRwbG90bGliLm9yZy/YYfK9AAAACXBIWXMAAAsTAAALEwEAmpwYAAAdCElEQVR4nO3de3hcd33n8fd3rpJGF0fWxHFiO3IuhITcSIUJSUohPHRD4CGB0jRAWZby1DRcCssuu7Dtw5btPk+hu/Ck2VLaAEkhhAALpFy2QAKEhFvsyMExdtw4TuI4cXyRfNN1NJqZ7/5xjq4jyZLs49Ecf17Po2fOnBnN+R4f+TO/+Z3f/I65OyIiEj+JWhcgIiLRUMCLiMSUAl5EJKYU8CIiMaWAFxGJqVStC5iso6PDOzs7a12GiEjd2LRpU6+752d6bEkFfGdnJ93d3bUuQ0SkbpjZs7M9pi4aEZGYUsCLiMSUAl5EJKYU8CIiMaWAFxGJKQW8iEhMKeBFRGIqFgF/20+e5MEdPbUuQ0RkSYk04M1smZl908z+zcy2m9krotjO5372FL/c2RvFS4uI1K2ov8n6d8AP3f0tZpYBmqLYSMKgUtGFS0REJoss4M2sDXgl8B8A3L0IFKPYVsIM5buIyFRRdtGsBXqAO83sN2b2BTPLTX+Sma03s24z6+7pWVw/uhlUdOlBEZEpogz4FHAF8Dl3fykwCHx0+pPc/XZ373L3rnx+xgnRjimRMHRtWRGRqaIM+OeB5919Q3j/mwSBf8Kpi0ZEpFpkAe/u+4DnzOyCcNVrgMej2FZCXTQiIlWiHkXzAeDucATN08C7otiIqQUvIlIl0oB3981AV5TbgKAFrz54EZGpYvFN1oQZZTXhRUSmiE3AK99FRKaKR8An1EUjIjJdPALeTKNoRESmiVHA17oKEZGlJRYBr6kKRESqxSLgE2Yo30VEpopJwKsFLyIyXUwCXidZRUSmi0XAa6oCEZFqsQh4TVUgIlItJgGvqQpERKaLR8An1EUjIjJdPAJeo2hERKrEJOA1Dl5EZLqYBLxa8CIi08Ui4E3j4EVEqsQi4IMWfK2rEBFZWmIS8KZx8CIi08Qm4NWCFxGZKhYBr+mCRUSqxSLg1YIXEakWk4CHihJeRGSKWAR8MqFhkiIi06WifHEz2wX0A2Wg5O5dEW1HXTQiItNEGvChV7t7b5Qb0HTBIiLVYtFFoys6iYhUizrgHbjPzDaZ2fqZnmBm682s28y6e3p6FrURjaIREakWdcBf4+5XAK8D3mdmr5z+BHe/3d273L0rn88vaiMaBy8iUi3SgHf3PeHtAeBeYF0U29F0wSIi1SILeDPLmVnL2DLw+8DWKLal6YJFRKpFOYpmBXCvmY1t56vu/sMoNqSTrCIi1SILeHd/GrgsqtefzMyoVE7GlkRE6kdMhkmqi0ZEZLpYBLymKhARqRaLgNdUBSIi1WIR8JqqQESkWkwCXi14EZHpYhLwOskqIjJdLAI+GCapgBcRmSwWAa+pCkREqsUk4NVFIyIyXTwCPqGTrCIi08Ui4M2grBa8iMgUsQj4pJnGwYuITBOLgNc4eBGRajEJeJ1kFRGZLhYBb+EwSXXTiIhMiEXAJ4KLimgsvIjIJDEJ+OBW3TQiIhPiEfBhwutEq4jIhFgEvKkFLyJSJRYBrz54EZFqMQn44FYteBGRCTEJ+CDhNV2BiMiEWAW8V2pciIjIEhKTgA9u1UUjIjIh8oA3s6SZ/cbMvh/VNiaGSSrgRUTGnIwW/AeB7VFuwEzj4EVEpos04M1sFfB64AtRbmesi0Zz0YiITIi6BX8r8F+AWU9/mtl6M+s2s+6enp5FbSShFryISJXIAt7M3gAccPdNcz3P3W939y5378rn84valk6yiohUi7IFfzXwRjPbBXwNuNbMvhLFhib64BXwIiJjIgt4d/+Yu69y907gZuCn7v7HUWxLUxWIiFSL1Tj4sjrhRUTGpU7GRtz9Z8DPonr9hLpoRESqxKMFr/ngRUSqxCPgNQ5eRKRKTAJeLXgRkeliEvDBrfrgRUQmxCLgNQ5eRKRaLAJe4+BFRKrFJOCDW7XgRUQmxCTgdZJVRGS6WAS8qQUvIlIlFgE/3oJXE15EZFwsAj6pb7KKiFSJRcCri0ZEpFosAl6TjYmIVItVwCvfRUQmxCTgg1u14EVEJsQi4E3j4EVEqswr4M0sZ2aJcPlFZvZGM0tHW9r8qQUvIlJtvi34h4AGMzsLuA94B/DPURW1UBN98Ap4EZEx8w14c/ch4M3AP7j7HwIvia6shZn4olONCxERWULmHfBm9grg7cD/C9cloylp4cbGwZfVghcRGTffgP8Q8DHgXnffZmbnAA9EVtUCqYtGRKRaaj5PcvcHgQcBwpOtve7+51EWthCaqkBEpNp8R9F81cxazSwHbAUeN7OPRFva/GkUjYhItfl20Vzk7n3AjcAPgLUEI2lmZWYNZrbRzB4zs21m9onjK3XObQFqwYuITDbfgE+H495vBL7r7qPAseJ0BLjW3S8DLgeuM7MrF1voXMZa8OqDFxGZMN+A/ydgF5ADHjKzs4G+uX7BAwPh3XT4E0kCa7IxEZFq8wp4d7/N3c9y9+vD4H4WePWxfs/Mkma2GTgA3O/uG46v3JlpHLyISLX5nmRtM7PPmFl3+PNpgtb8nNy97O6XA6uAdWZ28QyvvX7sdXt6ehZaf/gawa1a8CIiE+bbRXMH0A/cFP70AXfOdyPufoRg3Px1Mzx2u7t3uXtXPp+f70tOkUhoumARkenmNQ4eONfd/2DS/U+EXS+zMrM8MOruR8ysEXgt8KnFlTk3DZMUEak23xb8sJldM3bHzK4Gho/xOyuBB8xsC/AIQR/89xdX5tzG+uA1VYGIyIT5tuD/DPiymbWF9w8D75zrF9x9C/DS46ht3ib64E/G1kRE6sN8pyp4DLjMzFrD+31m9iFgS4S1zVtSc9GIiFRZ0BWd3L0v/EYrwIcjqGdRJoZJKuBFRMYczyX77IRVcZwSmqpARKTK8QT8kolTC/dCo2hERCbM2QdvZv3MHOQGNEZS0SJMzAdf40JERJaQOQPe3VtOViHHQ+PgRUSqHU8XzZKhPngRkWqxCHjNRSMiUi0WAa9rsoqIVItVwJc1XbCIyLiYBHxwqy4aEZEJsQh4M8NMXTQiIpPFIuAh6KbRKBoRkQkxCnh10YiITBabgDe14EVEpohNwCfUBy8iMkWMAt7URSMiMknMAr7WVYiILB2xCfhU0iiW9E0nEZExsQn4xnSSwmi51mWIiCwZsQr4YQW8iMi42AR8QzpJYVRdNCIiY2IT8I0ZddGIiEwWn4BXF42IyBSRBbyZrTazB8zscTPbZmYfjGpbEHTRDBcV8CIiY+a8JutxKgH/yd0fNbMWYJOZ3e/uj0exMXXRiIhMFVkL3t33uvuj4XI/sB04K6rtNaYT6qIREZnkpPTBm1kn8FJgwwyPrTezbjPr7unpWfQ21AcvIjJV5AFvZs3At4APuXvf9Mfd/XZ373L3rnw+v+jtNGTUBy8iMlmkAW9maYJwv9vdvx3lthrTSUZKFSqakEZEBIh2FI0BXwS2u/tnotrOmMZ0EoBCSa14ERGItgV/NfAO4Foz2xz+XB/VxhrCgFc3jYhIILJhku7+C8Ciev3pxlrwOtEqIhKIzTdZGzJhF40CXkQEiFHAj/fBa8IxEREghgGvLhoRkUB8Aj4T7IpOsoqIBGIT8A1qwYuITBGbgJ/og1fAi4hAnAI+o3HwIiKTxSfg1UUjIjJFbAJeffAiIlPFJuCzqQRmUFAXjYgIEKOANzPNCS8iMklsAh4gl03RXyjVugwRkSUhVgF/ekuWA/0jtS5DRGRJiFXAr2htYN/RQq3LEBFZEmIX8Af6FfAiIhCzgD+jtYHegSLFkmaUFBGJVcCvaM0CqBUvIkLcAr6tAYD9fTrRKiISr4BvGQt4teBFRGIV8Ge0KeBFRMbEKuBPa0qTSSbYp4AXEYlXwJsZaztybNvTV+tSRERqLlYBD/C753ew8ZlDDBU1ZYGInNpiF/CvfFGeYrnChqcP1boUEZGaiizgzewOMztgZluj2sZM1q1tpyGd4P7t+0/mZkVElpwoW/D/DFwX4evPqCGd5A2Xnsm3H32e3gGNhxeRU1dkAe/uDwE16Se55VXnMlKq8E8PPlWLzYuILAk174M3s/Vm1m1m3T09PSfkNc/NN3PT76zmjl/u4rHnjpyQ1xQRqTc1D3h3v93du9y9K5/Pn7DX/W+vv5B8c5b33LWJZw8OnrDXFRGpFzUP+Ki0Naa5810vo1Aq87bPb+C5Q0O1LklE5KSKbcADXLiyla+8++X0F0Z50z/8kl/u7K11SSIiJ02UwyTvAX4NXGBmz5vZu6Pa1lwuPquNb91yFcuaMvzxFzdw6493UK54LUoRETmpohxF81Z3X+nuaXdf5e5fjGpbx3L+iha+876redPlZ3Hrj5/knXdspEfXbhWRmIt1F81kuWyKT990GZ988yU8susQr7/t5zz89MFalyUiEplTJuAhmIzs5nVr+Jf3XU0um+Jtn3+Yzz6wk4q6bEQkhk6pgB9z4cpWvveBa3j9pWfyv370BOvv2kRhtFzrskRETqhTMuABmrMpbrv5cj7+hov48fb9/NlXFPIiEi+nbMBD0GXzJ9es5ZNvvoQHd/SoJS8isXJKB/yYm9et4VNvvpSfP9nDe+7axGi5UuuSRESOmwI+dNPLVvM3bwpa8h//zlbcdeJVROpbqtYFLCU3r1vD84eH+fsHdtK5PMd7fu/cWpckIrJoCvhpPvzaF7Hr4CB/84N/Y017E6+7ZGWtSxIRWRQF/DSJhPG///Ay9h4t8B+/sZmVyxq5fPWyWpclIrJg6oOfQUM6ye3v+B3yLVnededGfrP7cK1LEhFZMAX8LJY3Z7nrT15OS0Oat37+Ye7btq/WJYmILIgCfg6dHTm+dctVvGhFC+vv2sSHv75Zk5SJSN1QwB9DviXLN97zCt7/6vP43pYXuPbTP+POXz6jL0SJyJJnS2m8d1dXl3d3d9e6jFk91TPAf//ONn6xs5eO5izvurqTm7pWk2/J1ro0ETlFmdkmd++a8TEF/MK4O79++iCf+9lT/PzJXlIJ4zUXns5NXau5+rwOGtLJWpcoIqeQuQJewyQXyMy46twOrjq3g50H+vlG9/N8a9Pz/GjbfhrSCdatXc7lq5dx2ao2LlnVRr45i5nVumwROQWpBX8CFEsVfvVULw/u6OFXOw/y5IF+xqaYb8mmOLujic7lOc5e3sSK1gbyzVnyLRM/TRm9z4rI4qgFH7FMKsGrLjidV11wOgCDIyW2vdDH1j1HefbgILsODvHbPUf5wdZ9M14PtimTJN+SZVljmtbwpy38aW2YWM5lkzRnUzRlUuSyyfHbxnRSnxJEpIoCPgK5bIp1a9tZt7Z9yvpyxTk0WKSnf4TegRF6+kfoCW97B0Y4MjTK0eFR9hwZpm84WB4tH/sTlhnkMimaMkly2fA2k6IpG9w2pJNkUkYmmSCdTJBJTdwG64xMKhneBo+lkwlSyYnfmWk5lbTx56bD5VTC5v1mU6k4ZujNSU4pzx0aor9Q4qIzWyPflgL+JEombLxbZj7cncJohaNh2A+MlBgqlhgcKQe3xTJDI5NvJz02UubwYJHnDw8zXCwzWq5QLFcYLYW383jjWKzJYT/2hjH+ZpBIkE4ZhrFjfz/litPSkKK1Mc3KtgaW57LksknOXNZIUyZJJpmgsyNHS0OaVMJYm8/R2pCOrHaRKLk77737UQ70F3j4Y6+JvHGjgF/CzIzGTJLGTJIz2hpO6Gu7+3jQF0sViqUKo+UKI6UKpUqFUtnH3xBKlWC5VHZGy5XwJ1gulSsUpy2Xpj1ntDzxeuOvUXHe9vI1NKaT9BdKHB0e5YUjw2zf18dAoUTPwAgznR5KWDCVxLLGNBeubKU9l6G9OUNLNkUiYaQSRsKMXDbFaU0ZegdGcOD3zs/TM1BgRWsDK1obSCeP/RUQd2ewWKY5q/8m9cbdOTw0SnsuU+tSpnh09xF+u+coAE/s7+fFZ0Tbitdf7inKzMimkmRTwBIcxl8YLVOqOCOjZXbsH6BQKlMqO7/dc5SBQon9fQWe6R1k2wt9HBosUlzARVrMYHkuS2tjiuZsCjPDCD5hNaaDN9SmTJIXjgzzyK7D/O75HZx/egvFcplz880kE8a+owUODhQ5J5+jKZuiUnEGiyU27z7CyzqDrrlLVrVxZKjIvqMFKg4/3LaPa198Ou1NGbLpBC85s5Xte/tZ3d5ET/8IyQRctmoZLxwpkMsm6X72MKWyc0ZbFsM4J58DYKRU4UDfCFedu5yh0TIJg1QiQV9hlId29HDftv3kW7IcGiry0tXLeOu6NTRlkpQrzr6+AsuaMjz+Qh+XrmrjQN8IZ7Q1kEoYR4dHaUgn+fXTvaw+rYnvb9nLGy8/kyNDRZqzaZ7uGaAxk+TKc5ZXDQce6247OFgcf9POZZMMFcvksinKZSebTpBNBW+sY58gM6kT/13LXb2DfOSbj/Ho7iPc86dXVnWVRmVgpETCmHXQRKXi3PrjHTRlgn+Xn+/ojTzgNYpG6p67M1p2Ku6UK07Znf5CiSNDRVob0uzrK7Btz1HWLG/iQN8ILxwtcKCvQH8h6NZyZ/x3C6NlhoplhkfLJM24+rwOHnqyh97+ERIJo79QAoI3g7bGNIcGi1NqybdkZ53Ooq0xzdHh0ePe34RBxYMayhUnmTCyqQRDxeDb1ctzGQqjZdqbMzx3aDh4A0gmaM6mODRYJJUwShUffx0zSCcSFMsVsqkEI6W53yw7mrNkksbKZY1U3GlIJdn83BEaM8mqfw9gfHtNmSQJC7rthotliuUKaztytOcyjJYrVBzSCSOVNIbDY9Cey7CitYGKw8GBEU5ryjAwUqKjOcuhwRGWNWUYLpZpaUhRKFUojJbp3nUIBzLJBNl0gpd1tvOyznYGR0rkW7Kkk4nxYw7gUHWBn7G7jk8sj68LpJNGUybFSKlM967DfO2R3STNeOPlZ/LmK1axtiPHQKHEwEjwN/bVjbu5Z+Nu/vrGi/nyr3bRlE3xtT+9ksbM8X13pmZfdDKz64C/A5LAF9z9k3M9XwEvS5m70zsQBFh7LkMyYRweLI6HZSqZoLUhxZMHBmjOptjwzEHObGvk/BUtjJYrLM9luOvhZ+lcnqOtKc3ug0OcdVojjz13hOXNGXKZFHuPFji9JUvvYJEVLVlaG9MMjgQhcXCgSF9hlJFShQtWtLB9Xx/tTRn6CqP0DZdY095EJpXg7S9fQyrsgnp092Huf3w/xVKF3oERLj6zjT1HhrlsdRvb9/ZzdvimVxgts6wpw7MHB7n6vA6e7hnk5ee084sne3nRGS2UyhXOyTdzeLDIPRt305AOPuFk0wkGCiVefEYrpYqzpr0Jx2lrTDNQKJHLpugdGKExneRA/wgVD7romjLBp6cn9vdzdHiUbCqBmVGuBN2F2VTwKerwUJG9RwuUK06+JcvhoSK5TPCa+ZYG+gvBp46BQonGTJJsKkG+Jcv/uOFidh8a4s/v+Q0Jg8NDx//GOpdMMsF1F59BNpXge1teoDA685vku69Zy1++/kLu3rCbv/yXrTRnU5zekqWtKc297716UduuScCbWRLYAbwWeB54BHiruz8+2+8o4EXkRCtXnGd6B1meC87JlN1JhN1yE+c4bXx5bNXYCdDJzzMmTooWy8GnvUwqwdntufGWeH9hlAd39HBosEhzNjV+Dufs5TkuOKNl/Pc3PnOI7295gYODRdqbMvz1jRcvav9qNQ5+HbDT3Z8Oi/gacAMwa8CLiJxoyYRx3unNAJx2Ek66tjSkecOlZx7zeTMNpT7RopxN8izguUn3nw/XTWFm682s28y6e3p6IixHROTUUvPpgt39dnfvcveufD5f63JERGIjyoDfA6yedH9VuE5ERE6CKAP+EeB8M1trZhngZuC7EW5PREQmiewkq7uXzOz9wI8Ihkne4e7botqeiIhMFek3Wd39X4F/jXIbIiIys5qfZBURkWgo4EVEYmpJzUVjZj3As4v89Q6g9wSWU0val6UnLvsB2pelarH7cra7zzjGfEkF/PEws+7Zvq5bb7QvS09c9gO0L0tVFPuiLhoRkZhSwIuIxFScAv72WhdwAmlflp647AdoX5aqE74vsemDFxGRqeLUghcRkUkU8CIiMVX3AW9m15nZE2a208w+Wut6FsrMdpnZb81ss5l1h+vazex+M3syvD2t1nXOxMzuMLMDZrZ10roZa7fAbeFx2mJmV9Su8mqz7Mtfmdme8NhsNrPrJz32sXBfnjCzf1ebqmdmZqvN7AEze9zMtpnZB8P1dXds5tiXujs2ZtZgZhvN7LFwXz4Rrl9rZhvCmr8eTs6ImWXD+zvDxzsXvFF3r9sfgknMngLOATLAY8BFta5rgfuwC+iYtu5vgY+Gyx8FPlXrOmep/ZXAFcDWY9UOXA/8gOAKaFcCG2pd/zz25a+A/zzDcy8K/9aywNrwbzBZ632YVN9K4IpwuYXg0pkX1eOxmWNf6u7YhP++zeFyGtgQ/nt/A7g5XP+PwC3h8nuBfwyXbwa+vtBt1nsLfvyygO5eBMYuC1jvbgC+FC5/CbixdqXMzt0fAg5NWz1b7TcAX/bAw8AyM1t5Ugqdh1n2ZTY3AF9z9xF3fwbYSfC3uCS4+153fzRc7ge2E1xNre6OzRz7Mpsle2zCf9+B8G46/HHgWuCb4frpx2XseH0TeI3ZxFVk56PeA35elwVc4hy4z8w2mdn6cN0Kd98bLu8DVtSmtEWZrfZ6PVbvD7st7pjUVVY3+xJ+rH8pQWuxro/NtH2BOjw2ZpY0s83AAeB+gk8YR9y9FD5lcr3j+xI+fhRYvpDt1XvAx8E17n4F8DrgfWb2yskPevD5rC7HstZz7aHPAecClwN7gU/XtJoFMrNm4FvAh9y9b/Jj9XZsZtiXujw27l5298sJrnC3DnhxlNur94Cv+8sCuvue8PYAcC/BQd8/9hE5vD1QuwoXbLba6+5Yufv+8D9kBfg8Ex/1l/y+mFmaIBDvdvdvh6vr8tjMtC/1fGwA3P0I8ADwCoIusbFrc0yud3xfwsfbgIML2U69B3xdXxbQzHJm1jK2DPw+sJVgH94ZPu2dwHdqU+GizFb7d4F/H47YuBI4Oqm7YEma1g/9JoJjA8G+3ByOclgLnA9sPNn1zSbsp/0isN3dPzPpobo7NrPtSz0eGzPLm9mycLkReC3BOYUHgLeET5t+XMaO11uAn4afvOav1meWT8CZ6esJzqw/BfxFretZYO3nEJzxfwzYNlY/QT/bT4AngR8D7bWudZb67yH4eDxK0Hf47tlqJxhB8NnwOP0W6Kp1/fPYl7vCWreE/9lWTnr+X4T78gTwulrXP21friHoftkCbA5/rq/HYzPHvtTdsQEuBX4T1rwV+Hi4/hyCN6GdwP8FsuH6hvD+zvDxcxa6TU1VICISU/XeRSMiIrNQwIuIxJQCXkQkphTwIiIxpYAXEYkpBbycUsysPGkGws12AmcgNbPOybNRitRa6thPEYmVYQ++Ki4Se2rBizA+L//fWjA3/0YzOy9c32lmPw0ntfqJma0J168ws3vDub0fM7OrwpdKmtnnw/m+7wu/sShSEwp4OdU0Tuui+aNJjx1190uAvwduDdf9H+BL7n4pcDdwW7j+NuBBd7+MYB75beH684HPuvtLgCPAH0S6NyJz0DdZ5ZRiZgPu3jzD+l3Ate7+dDi51T53X25mvQRfgx8N1+919w4z6wFWufvIpNfoBO539/PD+/8VSLv7/zwJuyZSRS14kQk+y/JCjExaLqPzXFJDCniRCX806fbX4fKvCGYpBXg78PNw+SfALTB+EYe2k1WkyHypdSGnmsbwijpjfujuY0MlTzOzLQSt8LeG6z4A3GlmHwF6gHeF6z8I3G5m7yZoqd9CMBulyJKhPngRxvvgu9y9t9a1iJwo6qIREYkpteBFRGJKLXgRkZhSwIuIxJQCXkQkphTwIiIxpYAXEYmp/w8Eir9xWebpuAAAAABJRU5ErkJggg==\n", + "image/png": "iVBORw0KGgoAAAANSUhEUgAAAXgAAAEGCAYAAABvtY4XAAAAOXRFWHRTb2Z0d2FyZQBNYXRwbG90bGliIHZlcnNpb24zLjUuMSwgaHR0cHM6Ly9tYXRwbG90bGliLm9yZy/YYfK9AAAACXBIWXMAAAsTAAALEwEAmpwYAAAaUklEQVR4nO3de5Ad5Xnn8d9zLjPSjO5oJIRGaCQuAckgkKewMVpqjZ1wWRuI7Y1xHIcYstq1TRZqd53Y5a2sXUnVblIVypGD45WBWI7tGK8NNsv6AuFqfBEegSRLwkjCEkigywhdZnSb23n2j+5z5sycmdHMSK0+553vpzQ1ffpc+mm19Ov3vP12t7m7AADhyaRdAAAgGQQ8AASKgAeAQBHwABAoAh4AApVLu4Bys2fP9paWlrTLAICasW7dugPu3jTUc1UV8C0tLWpra0u7DACoGWb22nDP0UUDAIEi4AEgUAQ8AASKgAeAQBHwABAoAh4AAkXAA0Cgggj4VU9u07Nb29MuAwCqShAB/4/PvKrntxHwAFAuiIDPmFTgviUAMEAYAZ8x9ZHwADBAEAGfzZi49SAADBREwGfM1EfAA8AA4QR8Ie0qAKC6BBHw2YzoogGAQYII+KgFT8ADQLlwAp4WPAAMEETAR6No0q4CAKpLEAGfMdFFAwCDhBHwGbpoAGCwIAI+a5zoBACDBRHwjKIBgEphBHyGE50AYLAgAp4TnQCgUhABzzh4AKgUTsDTBw8AAwQR8JzoBACVggh4TnQCgEqBBDx98AAwWBABzx2dAKBSLskPN7Odkjol9UnqdffWJJbDQVYAqJRowMfe7e4HklxAdC2aJJcAALUnjC4a40QnABgs6YB3SY+b2TozWznUC8xspZm1mVlbe3v7uBZCFw0AVEo64Fe4+3JJN0r6lJldO/gF7r7a3VvdvbWpqWlcC4muRUPAA0C5RAPe3d+If++X9Iikq5JYTnS54CQ+GQBqV2IBb2aNZja1OC3p9yRtSmJZmYwYBw8AgyQ5imaupEfMrLicb7n7j5NYUMZMBbpoAGCAxALe3X8raVlSn18uY6YCLXgAGCCMYZLckxUAKgQR8FEXTdpVAEB1CSTgRRcNAAwSRMBnGQcPABWCCPhMxkS+A8BAYQQ8XTQAUCGIgM9yLRoAqBBEwEddNAQ8AJQLI+A5kxUAKgQR8JzoBACVggj46FIFaVcBANUlkIAXXTQAMEgQAU8XDQBUCiLgM/ENP7gvKwD0CybgJdEPDwBlggj4bLwWnOwEAP2CCPhMptiCJ+ABoCiMgDcCHgAGCyLgs3HA00UDAP2CCPj+LpqUCwGAKhJGwEf5zslOAFAmiIDPxgnPyU4A0C+IgOcgKwBUCivgCykXAgBVJIiAL53oRAseAEqCCPj+FjwBDwBFYQU8LXgAKEk84M0sa2YvmdljSS2jNIqGFjwAlJyNFvzdkl5OcgFWHAdPvgNASaIBb2bNkv6dpPuTXE6Wi40BQIWkW/BflPTnkoYdwGhmK82szcza2tvbx7WQLH3wAFAhsYA3s/dJ2u/u60Z6nbuvdvdWd29tamoa77Ik0QcPAOWSbMFfI+lmM9sp6duSrjOzbySxoFIXDSc6AUBJYgHv7p9192Z3b5F0m6Sn3P2PklhW8UQnumgAoF8Q4+BLXTQEPACU5M7GQtz9GUnPJPX5Wc5kBYAKQbTgs9zwAwAqBBHwxROdGEUDAP2CCHjGwQNApTACnjNZAaBCEAHPiU4AUCmIgKcFDwCVwgh4btkHABWCCPjSKBpa8ABQEkTA91+LhoAHgKKwAp58B4CSIAI+QxcNAFQIJODpogGAwYIIeIZJAkClIAI+w4lOAFAhjICnBQ8AFYII+P6LjaVcCABUkSACPsPlggGgQhgBTxcNAFQII+AZJgkAFYII+GzpptspFwIAVSSIgM/Ea0ELHgD6hRHw3LIPACoEEfDFM1m5Fg0A9Asi4IstePIdAPqNKuDNrNHMMvH0xWZ2s5nlky1t9BgHDwCVRtuCf07SJDObL+lxSR+T9LWkihqrUhcNAQ8AJaMNeHP345I+IOnL7v7vJS0d8Q1mk8zsBTPbYGabzewLp1vsCMuSmeT00QBAyagD3syulvRRSf8vnpc9xXu6JF3n7sskXSHpBjN757iqHIWMGQdZAaBMbpSvu0fSZyU94u6bzWyxpKdHeoNHzemj8cN8/JNYAmfN1FdI6tMBoPaMKuDd/VlJz0pSfLD1gLv/51O9z8yyktZJulDSfe6+dojXrJS0UpLOP//80Vc+SCZDFw0AlBvtKJpvmdk0M2uUtEnSFjP79Kne5+597n6FpGZJV5nZ24Z4zWp3b3X31qampjGW3y9jxkFWACgz2j74Je7eIelWST+StEjRSJpRcffDirp0bhhjfaOWpQ8eAAYYbcDn43Hvt0p61N17dIr+dDNrMrMZ8fRkSb8r6TfjL/UUBeYy6qETHgBKRnuQ9X9L2ilpg6TnzGyhpI5TvGeepDVxP3xG0nfc/bHxFnoqddmMenppwQNA0WgPsq6StKps1mtm9u5TvGejpCtPo7YxyeeMFjwAlBntQdbpZnavmbXFP38nqTHh2sYkn82oi4AHgJLR9sE/KKlT0h/EPx2S/imposYj6qIh4AGgaLR98Be4+wfLHn/BzNYnUM+41XGQFQAGGG0L/oSZrSg+MLNrJJ1IpqTxyWcz6ibgAaBktC34/yTp62Y2PX58SNLtyZQ0PvmsMYoGAMqMdhTNBknLzGxa/LjDzO6RtDHB2sakLpdVx4metMsAgKoxpjs6uXtHfEarJP2XBOoZt7qsqZuDrABQcjq37LMzVsUZkM9ykBUAyp1OwFdVhzejaABgoBH74M2sU0MHuUmanEhF45TPZuiiAYAyIwa8u089W4WcrmiYZFV9qQCAVJ1OF01VqaeLBgAGCCbg84yiAYABAgp4WvAAUC6YgK/LZdRbcBW4bR8ASAoo4PPZaFW4Hg0ARIIJ+Lo44OmmAYBIOAGfKwY8XTQAIAUU8KUuGkbSAICkoAI+ujQOXTQAEAkm4ItdNBxkBYBIOAFPFw0ADBBMwOcZRQMAAwQT8P2jaAh4AJACCvj+UTQMkwQAKaCAr8tFo2g4yAoAkcQC3swWmNnTZrbFzDab2d1JLUsq64PnICsASDrFDT9OU6+k/+ruL5rZVEnrzOwJd9+SxMLogweAgRJrwbv7Hnd/MZ7ulPSypPlJLY+LjQHAQGelD97MWiRdKWntEM+tNLM2M2trb28f9zIYBw8AAyUe8GY2RdL3JN3j7h2Dn3f31e7e6u6tTU1N414OFxsDgIESDXgzyysK92+6+8NJLqt/mGRfkosBgJqR5Cgak/SApJfd/d6kllPUf7ExWvAAICXbgr9G0sckXWdm6+Ofm5JaGBcbA4CBEhsm6e7PS7KkPn+wfIaDrABQLpgzWTMZUy5jjIMHgFgwAS9F3TQEPABEggr4fDajLrpoAEBSYAE/pT6no129aZcBAFUhqICfNjmvjhM9aZcBAFUhqICfPjmnjhO04AFACizgp03Kq+MkLXgAkEILeLpoAKAkrICflNcRAh4AJIUW8JNzOtbdp17GwgNAWAE/fXJektR5kgOtABBUwE+bFAU8B1oBILSAj1vw9MMDQGgBPym6OCZj4QEgtICfTBcNABQFFfDFg6yMhQeAwAKePngA6BdUwDfWZZUxumgAQAos4M1M0yfndeg4AQ8AQQW8JC1umqJX9namXQYApC64gL+8ebo2v3mEyxUAmPCCC/hlzTN0sqegrfuOpl0KAKQquIC/vHm6JGnj7sPpFgIAKQsu4FvOadS0STm1vXYo7VIAIFXBBXwmY3rvpXP1k817dbKnL+1yACA1wQW8JN1y5Xx1nuzVM6/sT7sUAEhNYgFvZg+a2X4z25TUMoZzzQXnqGlqvdb8/DW5+9lePABUhSRb8F+TdEOCnz+sXDajT/7bC/SL376lZ7a2p1ECAKQusYB39+ckHUzq80/lo+9YqJZzGvS5h3+tt452pVUGAKQmyD54SarLZfSljyzXgWPd+g9fb1Mn16cBMMGkHvBmttLM2sysrb39zHanXNY8XX//4Su0cfcR3bb6l9p18PgZ/XwAqGapB7y7r3b3VndvbWpqOuOff+Nl8/TVP27V6weP6+Z/eF7PbztwxpcBANUo9YA/G959yRw9etcKNU2t18ceXKt7H3+Fa9UACF6SwyT/RdIvJP2Ome02szuTWtZoLJrdqO9/6hp9aHmzVj21XX94/1rtPXIyzZIAIFFWTePEW1tbva2tLfHlPPzibv3372/SpHxWX/zwFbr24jPfNQQAZ4OZrXP31qGemxBdNIN9YHmzHr1rheZMrdefrmmjXx5AkCZkwEvShXOm6KGVV2txU6M+8Y112nHgWNolAcAZNWEDXpKmN+R1/+2tymZNd3ztV4Q8gKBM6ICXpOaZDXrg9lYdPt6t9636qR761etplwQAZ8SED3hJevvCWfq/f7ZCyxbM0F9879f668e2qFConoPPADAeBHyseWaD/vnOd+hP3tWi+5/foU9+80Wd6OZ68gBqFwFfJpsxff7mpfrL9y3RT7bs1W1f/aXaO7lQGYDaRMAP4Y4Vi/SVP3q7Xtnbod//8s+0fX9n2iUBwJgR8MO4fum5emjl1TrZ06db7/u5vvOrXdw8BEBNIeBHsGzBDP3grhV62/xp+vPvbdSda9q0v4PLGwCoDQT8KcyfMVnf+tN36n+8f4l+tv2A3nvvs3rg+R3q4WJlAKocAT8KmYzp49cs0o/u/jdatmCG/uqxLbr+i8/pqd/so9sGQNUi4MdgcdMUff2Oq/TA7a1yl+74Wptu/fLP9eTLBD2A6jMhryZ5JnT3FvTddbv15We2a/ehE7p03jR9/F0tev+y8zS5Lpt2eQAmiJGuJknAn6aevoJ+sP5NrX7uVW3dd1TTJuX0obcv0AeWz9fS86bJzNIuEUDACPizwN31wo6D+sba1/XjTXvU0+e6oKlRNy+br/dcOoewB5AIAv4sO3SsWz/atFffX/+GXthxUJI0Z2q9rr24Sa0LZ2rZghm6aM4U5bIcAgFwegj4FO3vPKlnX2nXM1vb9fy2AzpyokeSNDmf1SXzpmrJvGlact40LZk3TZecO43+ewBjQsBXCXfXzreOa8Ouw1q/67Be3tOhLXs61HmyV5KUsejesZfOm6bLm6drWfMMvW3+dDXW51KuHKhN63cd1j3ffknf+Y9Xa860SWmXk4iRAp7kOIvMTItmN2rR7EbdeuV8SVHo7z50Qlv2dGjLm1Hgv/T6YT22cY+kKPQvnjtVy5pn6PIFUehfPHeq6nJ07wCn8sDzO7TzreNau+Og3r/svLTLOesI+JSZmRbMatCCWQ26fum5pfkHjnZp4+7D2rDriDbsPqzHt+zVQ227JEVXvVw4q0GLm6bowjlTdEFToxbMalDT1Ho1Ta3X1PocB3Qx4R081q2fbNorSdr05hECHtVj9pR6XXfJXF13yVxJ/S399bsOa+u+Tr3aflTb9x/Vs1v3q6dvYDdbfS6jmQ11aqjLqqE+q4a6nBrqsmqMf0+uy6oum1FdLqP6XFZ1ueJ0/+/idF02q/p8pvT6Aa/LZpXLmvLZjPJZY6eCqrLqyW3q7ivonMY6bXmzI+1yUkHA14jyln653r6CXj94XHuOnFR7Z5f2d57U/o4udZzs0bHuPp3o7tOxrl69dbRbu7qP63h3n0709Km7t6Du3oJ6z+Cdq7IZUy4TBX4ua8plouAvPs5n4vnZjPIZK9s5ZIZ8X+VnFN8XPz9gOqN8rv/58uWVns9mSjur+kE7t2yGnVNIXnr9kNb8Yqduv3qhTvYU9PiWvXL3CdcIIeBrXC6b0eKmKVrcNGVc7+8reCnsu3r71NVbUHdfQV090e/i/OJris91xc/19hXU01dQT5+rt1BQb5+Xpnv6vP/5QjTd2+cDpo/29qqnOL8v2uGUT/fE7+/t8zO6MxoslzHV5Sp3NMV55TuV0k5phJ1Yaac1xE4nnx24Y8plTXXxDmzwjin6nP73lNc5YOfGN6gSd9f//OFvNHtKvT59wyV6+MXdeqhtl1b8zdP64PL5uvqC2Wptman8BBimTMBPcNmMaXLcbSPl0y5nRO7evwOIdya98c6jp7dQtlOpfL64o+nqjX66y36X78SKO5XiZ/TvpPp3YN19BR3v7o1f68Ps5Mp3UGdnpFqu+K2o7JtSXWlHVLYTymYqvk2VP1e+8xhqxzXUTsVMMpmK+xiL50XTpvjPsIbbOWVsiM82kyk6TvXTbQf0xqETuubC2XrPpXPUebJXq597VS++flh/dctSTanP6d2/M0dXtexRn7u+9PR2rXpqu2Y05NW6cKY++o6Fam2ZqamTRv9vv/jvsK8Q/463e+lx6XdB7lH9P912QK+2H9XlzTN002Xz1HGiRwV37Tp4QlPqc7qsefqolz8WDJMEEube/5++tCMYsOMZuFMZdsdVGPTa0jeegrrj95TvoHoLBXX3DrfTqfw2FX3TGvQtrLe/zmq8D33LOQ1qntmgF3YeVHdvdAnvxU2NuvFt5+qe915c0Uo/dKxbL+w8qCe27NPPth/QniPR/R0a6rKalM+WuhkLPjjAi4/H//cwdVKuNCR6sHcsmqU1d1ylSfmxnwfDMEkgRWbx8YSsxvUfuFoUCpXfalyu+I+KbUWXl01HO7jxtCML8ftKn1H2WS5pSn1O582YLEk61tWrtTve0ltHu3XzFeepPjf03/PMxjpdv/RcXb/0XHX3FvTky/v0+sHj2t/ZFX+Di3Zw2YyVwj4bf3vpf1w2v/g4Wzk/k4m+abikxbMbtfS8aXp0w5t6df9RnTs9qvv8WQ3auq9TW/d1JvJvI9EWvJndIOnvJWUl3e/u/2uk19OCB4CxGakFn9hRBjPLSrpP0o2Slkj6iJktSWp5AICBkjyMfJWk7e7+W3fvlvRtSbckuDwAQJkkA36+pF1lj3fH8wYws5Vm1mZmbe3t7QmWAwATS+oDQd19tbu3untrU1NT2uUAQDCSDPg3JC0oe9wczwMAnAVJBvyvJF1kZovMrE7SbZIeTXB5AIAyiY2Dd/deM7tL0k8UDZN80N03J7U8AMBAiZ7o5O4/lPTDJJcBABhaVV2qwMzaJb02zrfPlnTgDJaTJtal+oSyHhLrUq3Guy4L3X3IESpVFfCnw8zahjubq9awLtUnlPWQWJdqlcS6pD5MEgCQDAIeAAIVUsCvTruAM4h1qT6hrIfEulSrM74uwfTBAwAGCqkFDwAoQ8ADQKBqPuDN7AYze8XMtpvZZ9KuZ6zMbKeZ/drM1ptZWzxvlpk9YWbb4t8z065zKGb2oJntN7NNZfOGrN0iq+LttNHMlqdXeaVh1uXzZvZGvG3Wm9lNZc99Nl6XV8zs+nSqHpqZLTCzp81si5ltNrO74/k1t21GWJea2zZmNsnMXjCzDfG6fCGev8jM1sY1PxRf2kVmVh8/3h4/3zLmhUa3wKrNH0WXQHhV0mJJdZI2SFqSdl1jXIedkmYPmve3kj4TT39G0t+kXecwtV8rabmkTaeqXdJNkn6k6N7L75S0Nu36R7Eun5f034Z47ZL431q9pEXxv8Fs2utQVt88Scvj6amStsY119y2GWFdam7bxH+/U+LpvKS18d/3dyTdFs//iqRPxNOflPSVePo2SQ+NdZm13oIP9aYit0haE0+vkXRreqUMz92fk3Rw0Ozhar9F0tc98ktJM8xs3lkpdBSGWZfh3CLp2+7e5e47JG1X9G+xKrj7Hnd/MZ7ulPSyonsx1Ny2GWFdhlO12yb++z0aP8zHPy7pOknfjecP3i7F7fVdSe8xMxvLMms94Ed1U5Eq55IeN7N1ZrYynjfX3ffE03slzU2ntHEZrvZa3VZ3xd0WD5Z1ldXMusRf669U1Fqs6W0zaF2kGtw2ZpY1s/WS9kt6QtE3jMPu3hu/pLze0rrEzx+RdM5YllfrAR+CFe6+XNG9az9lZteWP+nR97OaHMtay7XH/lHSBZKukLRH0t+lWs0YmdkUSd+TdI+7d5Q/V2vbZoh1qclt4+597n6FovtjXCXpkiSXV+sBX/M3FXH3N+Lf+yU9omij7yt+RY5/70+vwjEbrvaa21buvi/+D1mQ9FX1f9Wv+nUxs7yiQPymuz8cz67JbTPUutTytpEkdz8s6WlJVyvqEite2be83tK6xM9Pl/TWWJZT6wFf0zcVMbNGM5tanJb0e5I2KVqH2+OX3S7pB+lUOC7D1f6opD+OR2y8U9KRsu6CqjSoH/r3FW0bKVqX2+JRDoskXSTphbNd33DiftoHJL3s7veWPVVz22a4danFbWNmTWY2I56eLOl3FR1TeFrSh+KXDd4uxe31IUlPxd+8Ri/tI8tn4Mj0TYqOrL8q6XNp1zPG2hcrOuK/QdLmYv2K+tmelLRN0r9KmpV2rcPU/y+Kvh73KOo7vHO42hWNILgv3k6/ltSadv2jWJd/jmvdGP9nm1f2+s/F6/KKpBvTrn/QuqxQ1P2yUdL6+OemWtw2I6xLzW0bSZdLeimueZOkv4znL1a0E9ou6f9Iqo/nT4ofb4+fXzzWZXKpAgAIVK130QAAhkHAA0CgCHgACBQBDwCBIuABIFAEPCYUM+sruwLhejuDVyA1s5byq1ECacud+iVAUE54dKo4EDxa8IBK1+X/W4uuzf+CmV0Yz28xs6fii1o9aWbnx/Pnmtkj8bW9N5jZu+KPyprZV+PrfT8en7EIpIKAx0QzeVAXzYfLnjvi7pdJ+gdJX4znfUnSGne/XNI3Ja2K56+S9Ky7L1N0HfnN8fyLJN3n7kslHZb0wUTXBhgBZ7JiQjGzo+4+ZYj5OyVd5+6/jS9utdfdzzGzA4pOg++J5+9x99lm1i6p2d27yj6jRdIT7n5R/PgvJOXd/a/PwqoBFWjBA/18mOmx6Cqb7hPHuZAiAh7o9+Gy37+Ip3+u6CqlkvRRST+Np5+U9AmpdBOH6WerSGC0aF1gopkc31Gn6MfuXhwqOdPMNipqhX8knvdnkv7JzD4tqV3Sx+P5d0tabWZ3Kmqpf0LR1SiBqkEfPKBSH3yrux9IuxbgTKGLBgACRQseAAJFCx4AAkXAA0CgCHgACBQBDwCBIuABIFD/H+MtaIuhHIYOAAAAAElFTkSuQmCC\n", "text/plain": [ "
" ] @@ -4385,7 +583,7 @@ }, { "cell_type": "code", - "execution_count": 13, + "execution_count": 12, "id": "1e2bf200", "metadata": {}, "outputs": [], @@ -4404,13 +602,13 @@ }, { "cell_type": "code", - "execution_count": 14, + "execution_count": 13, "id": "e4e0025d", "metadata": {}, "outputs": [ { "data": { - "image/png": "iVBORw0KGgoAAAANSUhEUgAAAXcAAAD4CAYAAAAXUaZHAAAAOXRFWHRTb2Z0d2FyZQBNYXRwbG90bGliIHZlcnNpb24zLjUuMSwgaHR0cHM6Ly9tYXRwbG90bGliLm9yZy/YYfK9AAAACXBIWXMAAAsTAAALEwEAmpwYAABY7UlEQVR4nO29d3xcV5n//z7Ti6SZUZdVLPcS9zjF6YnTSCcECC0BsoQFFsgCC4S6lGWX3y4Q4LtAAlkSkhBKAqT33h07cdy7ZUuyumak6RrNnN8f5440stpIGlmSc96v133NnXPvPffMtfWZZ57znOcRUko0Go1Gc3xhmuoBaDQajSb3aHHXaDSa4xAt7hqNRnMcosVdo9FojkO0uGs0Gs1xiGWqBwBQXFwsa2trp3oYGo1m9271umjR1I5DkxWbNm1ql1KWDHVsWoh7bW0tGzdunOphaDSac85Rr88/P5Wj0GSJEOLQcMe0W0aj0WiOQ6aF5a7RaKYJ3/rWVI9AkyO0uGs0mn7OP3+qR6DJEdoto9Fo+tm8WW2aGY+23DUaTT833aRe9YTqjEdb7hqNRnMcosVdo9FojkO0uGs0Gk0Oeac+wKZDnVM9DC3uGo1Gk0v+87GdfO6et0mlprZWhp5Q1Wg0/fzoR1M9ghlPIJKguTvGhrpOTp1bNGXj0OKu0Wj6Oe20qR7BjKcrmgDgwXeOTKm4a7eMRqPp59VX1aYZN92GuD+6tYme3tSUjUOLu0aj6ecb31CbZlz0JlOEe5KsrPIQiCR4eV/blI0lK3EXQniFEPcJIXYJIXYKIdYJIQqFEE8JIfYarz7jXCGE+IUQYp8QYosQYs3kfgSNRqOZHnTHegG4dEUFHqeVBzcfmbKxZGu5/xx4XEq5GFgJ7AS+DjwjpVwAPGO8B3gPsMDYbgR+ndMRazQazTQl7ZIpzrNz7qISXjvQMWVjGVXchRAe4CzgdgApZY+UMgBcCdxpnHYncJWxfyXwB6l4HfAKISpyPG6NRqOZdqQnUwscVkry7XRHe6dsLNlY7nOANuD3Qoi3hRC/E0K4gTIpZZNxTjNQZuxXAvUZ1zcYbRqNRnNc0x0zxN1pJc9uJZpIkkhOzaRqNqGQFmAN8Hkp5RtCiJ/T74IBQEophRBjitgXQtyIcttQU1Mzlks1Gs1kccstUz2CGU3aUvc4reQ7lLyG4714XbZjPpZsLPcGoEFK+Ybx/j6U2Lek3S3Ga6txvBGozri+ymgbgJTyNinlWinl2pKSIUsAajSaY82qVWrTjIs+t4zT0ifuwdjUuGZGFXcpZTNQL4RIV8xdD+wAHgSuN9quBx4w9h8ErjOiZk4FujLcNxqNZjrz9NNq04yLtFsm03JPtx1rsl2h+nngHiGEDTgAfAL1xfAXIcQNwCHgA8a5jwKXAPuAiHGuRqOZCfzwh+pVV2QaF13RBBaTwGk1k++wAhCaIss9K3GXUm4G1g5xaP0Q50rgcxMblkaj0cw8uqMJPE4rQgjy7NPcLaPRaDSa7OiKJihwKos97ZYJxbW4azQazYymO9ZLgSHqeX0TqlPjc9firtFoNDmiO8NyLzB87sEpstx1yl+NRtPPrbdO9QhmNN3RBJU+JwB2iwmrWUyZz12Lu0aj6WfRotHP0QxLd0xNqAJ9k6raLaPRaKaehx5Sm2bMSCnVhKrhjgHId1indyikRqN5l/CTn6jXyy+f2nHMQGKJFImk7LPcAcNy19EyGo1GM2PJTD2QJt9hmbIJVS3uGo1GkwMyUw+kyXdYteWu0Wg0x4JUSvLQO0eI9ORWdLszcrmnyXfoCVWNRqM5Jjy05Qifv/dtntzektN++90yA8V9qlao6glVjUbTz113TfUIJpXeZIqfP70XgECkJ6d9D+WWSU+oSikRQuT0fqOhxV2j0fRTXT36OTOYB985woH2MJD7nC9dkbRbJnNC1UoyJYklUjht5pzebzS0W0aj0fTz5z+r7TikN5ni58/sZWlFAXaLKecTnd1Gf5lumanML6PFXaPR9PPrX6vtOOSNg50c6ojwL+fNJ99h7RPjXNEdTeCymbGa+2U1bcVPRTikFneNRvOuoC0YB2BReT4FkxDF0hVNDPC3A1Naak+Lu0ajeVfgNyZQfS6bEaKYa7fMwNQDAHl2IzOkdstoNBrN5OCPJBAiXd/UOimWe+bqVMgo2KEtd41Go5kcuiI9FDismE1iUnK+dEd7B7llprLUng6F1Gg0/dx331SPYNLwRxJ4Xf0l8HIeChlNsLgif0DbVBbs0OKu0Wj6KS6e6hFMGv5ID16XDch9zhcpJe2hOMV59gHtbruKbdc+d41GM7XccYfajkMCkQS+oyz3ZErmpO9QvJd4b4oit21Au8VswmUz62gZjUYzxRzH4u6P9ODrs9yNic4cuUs6QioS52jLPX0vPaGq0Wg0k0RXhs+9zxeeI3dJe0jF0BfnDxb3PLuFYFy7ZTQajSbnJJIpgvFevM6Blnuu3CVpcT/aLaPuNTU53bW4azSa456AkdTL50773NOWe67EXbllSoaw3CdjwVQ2aHHXaDTHPen0vulombw+n3tu3TKFQ1ruU5PTXYdCajSafh59dKpHMCn405Z7RrQM5NYt43VZByQNS6MWTB17n7sWd41G04/LNdUjmBT6LPejfO65ygzZEeoZMlJG3Uv73DUazVTzq1+p7Tgj7XOfzGiZ4rzBLhlQXySRniS9yVRO7pUtWYm7EKJOCLFVCLFZCLHRaCsUQjwlhNhrvPqMdiGE+IUQYp8QYosQYs1kfgCNRpND/vIXtU0Rv33xAP/zxO6c99uXEdLwidstJqxmkdMJ1aJhLPfSfAcALUbK4WPFWCz3c6WUq6SUa433XweekVIuAJ4x3gO8B1hgbDcCx2fmf41Gk1PC8V5ueXoPj25rynnf/kgCq1ngNkrdCSFymhmyPRSnZBhxr/Q5ATgSiObkXtkyEbfMlcCdxv6dwFUZ7X+QitcBrxCiYgL30Wg07wIeeucI4Z7kpKzmDBh5ZTKLVOcqRDGWSBKM9Q7rlqn0Kst9uoq7BJ4UQmwSQtxotJVJKdNfsc1AmbFfCdRnXNtgtA1ACHGjEGKjEGJjW1vbOIau0WiOJ+7dcBjIfeFqUD537xBVknLxRdIZVi6f4dwys7zKcm/wH1txzzZa5gwpZaMQohR4SgixK/OglFIKIcaUgUdKeRtwG8DatWtzk71Ho9HMSLY1dvFOQxel+XZag3GSKYnZJEa/MEsy88qkyVVO977UA8OIu8tmweeyTk/LXUrZaLy2An8HTgZa0u4W47XVOL0RqM64vMpo02g0053nn1fbMea+N/ayynKIm+fVMUc05dx6D2TklUmjimRP3OfeL+5Du2VAWe+N003chRBuIUR+eh+4ENgGPAhcb5x2PfCAsf8gcJ0RNXMq0JXhvtFoNJpBXLn9Jv5huZn37voy/8/6i5wv+hnKcs+Vz719hIyQaWZ5ncfccs/GLVMG/N2YiLAAf5RSPi6EeBP4ixDiBuAQ8AHj/EeBS4B9QAT4RM5HrdFoJof/+R/1+pWvHNPblvceYb/nFJwFhcyuf46GHIq7lHJIy70gR9Eyw7plOg9C9xGoPZ1Kr5NX97UjpRwwqTuZjCruUsoDwMoh2juA9UO0S+BzORmdRqM5tjz8sHo9huIeivfiIUSrZwEUVzOr4TGi3Z1Q4clJ/9FEkp5kqi+vTJp0zpeJCm57sAe3zYzTCLPs4/Gb4fCr8LVDVHqdhHuSqs7qUV8yk4VeoarRaKaUts4ALhHHkleE8NUA0Os/lLP+j84rkybfYSElIdyTnFD/HeH44EiZ3h6oewliXRA41Bfrfiz97lrcNRrNlNLZoWIxHPnFWAuVuEt//UiXjAl/eGBGyDTptL8TDYccMvVAw5vQE1L7zdv6wiG1uGs0mncNXR0tALi8JdiKagEwBxty1n9gBMsdJp5fxh44yCzXUXlj9j8LwgwIaNnGrClYyKSzQmo0mn6czmN+y0iXWsSY7ysFXxlRacOaQ3H3Z+ZyT0ShbTd0NZBvUVOJE8oMmUrxy9BNvJR3PXBmf/v+Z6HqJIh0QPNWit12bBaTFneNRjNFPPbYMb9lrKsdALe3GGmzcFAW4QgfyVn/aXEvju6D/7oIkiq6pXbdD4G5E7LcG1tbqSRGuSnQ3xjphCNvwzk3Q9tOaHwLk0kwy+OgQbtlNBrNdCIQ6eHlvSqUL9ckwh0ACGchJpOgWZTijuVO3Fu6Y5hNAm+4Tgn7xT8GTzVFTS8BWRbsqN8A/+8kiIcGNN/1/FYAFnoynsvBFwAJ886DsmUQOASxrmMe667FXaPR9PODH6jNoLU7xqfv2shJ//E0H739DV470JHzW6bCfrXjKgSgzVxKQbwlZ/03dcUoy7djjgVUw5LLYf56XEdewUJvduJ+5G1o36Pi1g3qOyO8sPWAGrqM9J+7/1lweGDWaihfodpatlPpddJ4DPPLaHHXaDT9PPOM2gye2NHCE9tbuGS5SuzaFIjl/JamuJ+EsIJVVYHyW8vIT/qVfzwHtHTHKPM4IC3uTh/MW4+pJ8RqsS87t0ysW70mwn1Nv3x2LwXCGGO8u//ctj1K1M0WKF+m2oyImdZgnHjvxEIvs0WLu0ajGZb6zgg2i4kfXqVEKr0aM5dY4wFiFg8YC4m6bUaC2a7cTKo2dcWo8Dgg6gezDaxOmHMWUpg5z7q1L6vjiMS71GuPstA7QnHuf6uRKxbnqfZYhrjHusDpVfv5FeAqguYtVBrhkC1dx6ZohxZ3jUYzLPWdEap8TvLsFhxWU87FPRzvJS8VJGHtX40acs5SO4HDE+5fSklzV4yyAgdEA+Dwqi8RpxdRtZazLVtp6c7i10havHuU5b6/LUwyJVlTZsSkxIMZ53Yptwyoe5Utg5ZtFOerWPiOsBZ3jUYzxdT7IyzypBBv/YGf2G4j0pVbn3trMI5XhEg6fX1tMZch7l0TX8gUjPcS6Ukqyz0WUC6ZNPPWszi1n2igddjr+4gPdMsc7lQWfIk1PvA4GOLu7X9fvhxad1LoVF8E6eidyUaLu0aj6aeoSG0Gvo63+UXjB+GhL3Bp8lmKOjfl9Hat3TG8hDBliG6vu5wkJghMXNxbupRVrix3f7+7BGD+ekxIqgNvjN7RUZb74c4IJgFes2H1py33ZEJ9ATgy8uJ4qqA3RrFZ+ec7QlrcNRrNseb++9UGdEUTLEnswCp74L23ApCMdOX0di2G5W7J6/9CcTkdtEpfTiz3JkPcKzzOfrdMmlmrSWGiJHZw9I7Slrnhc6/vjFDhcWJJGKLeG1P5ZNJfApn3caooIK9Jnastd41GM6XUd0aoFm30WD0w/3wARCy34t7aFcVHCEdBcV9bnt1CvSxG5sBybzb86eUFabeMt/+gyUzcko+rt5voaMnD0p/byBdzuDNCTaFroK89HuyPyMm03I0QT3dvN1azoDOc21z1w6HFXaPR9HPzzWoDGvwRakQrSU812AsAMPd0k0rlbiFTZ1c3dpHAlt8v7vkOC405Eve0W6a0wA7RroE+dyBhL8QnQrQGR5lU7QuFVJZ7n7hnRsnEu/q/BDLF3bDcRdSPz2XrS2Q22Whx12g0/bz2mtqA+s4oVaINS9FcsNhImBzkESYQzZ3lGTYmM4WrX3Tz7BZaZCEi3AITXBHb1B2j0G3DYUaJb6a7BJBOH16CtHSPEsES7/e5R3uStAXjVBc6B06kxoNDi3v6s0U7KXTb6NRuGY1GM5XUd4SoNrVhNTI19toKKCBCRw7DIWPdKq9M2roFlYrXL/MQyZ6+Cczx0tIVM1wyhuhmumUAs7todMs9meiz2OkJU+9X+9VHu2Vi3SNa7kQ6teWu0Wimnu72Bmz0gm82ANLuoUCEacuhuPeGjNDKDHdJnsOCH2NxUGQCoZdS0hIIU55ewASDLHdrfhE+MYrlningPWEOdyhxVz73bnCX9J83lLg7PCr9r7bcNRrNdEB21qkdXy0AwumhgEhfQeic3CM6MK8MKLdMQBriHu0cf+dv/IbfBD5FeYFdRcrAIJ+7Lb8YH6NY7pmTyIlIX4x734RqQaU6Fh/GchdC3TeixD1tuUspiSUmLxWBFneNRtNPVRVUVZFKSewhY0LTWwuAxeWlQIRpD+bGcu/pTWFPpN0l/aKb77Dgl/nqTWT84p5s3k4VrdS64hAzvkSOcssIVxFO0YM/0D24gzSZfvWeMIc7I7htZgrdNuWK8VQZ5xmWuzCDzT2wD1chRDvxuW0EogmSKUlXNMHibz/OXa/VjfszjoTO567RaPq5+24A2rpjVKRakGaB8FYDYHEXUiAiOVs+H4j24MNIoescOKHa55ZJW/bjIN7dhguotXZC1HCtHOWWSf9iiHaNsEo1HRFjtimfe2eE6kKXKqqdabnHuvpTDxxdcNtZqCz3SitSqjUEDYbvvrTAMe7POBLactdoNIOo74xQbWoj7iwFiyr+LBwevCJCezA3bhl/OIFHhEiaHSqZl0Gew0IgF5Z7SE3WVtLe/yVxlFsmPdkZD47g209b7vnlfW6ZmkKXmmjtjSqfu9nWb7lnumTSuAoh6sfnVvllOsM9A907k4AWd41G089NN8FNN1Hvj1AtWpGGSwYAh4c8IrSPFhOeJZ1hZbn32r0D2t02CwEMt8ZEfO6GoJcmWzPS/Q68V9pyT4Xbh+8n7UfPn4XsCQ1ewGTPV1va5z6UuKctd0Pc/ZF+ca/W4q7RaCadzZth82Ya/VGqRSu24jn9xxwezKQIh3KzStUf6cErQsijrGmzSeCwO4iZ8yZkuVti6tqCRLOaULU4+36F9GFY7vaeruFXqcb6LXcZDxPvTVFT5MqYPC1Qi7xGtNx9yufu6rfc6zsjFLlt5NknxzuuxV2jmem07YaO/Tnt0t8dolz4MRfW9jcaohUPjd8PnklnuAePCGPKiJRJk2e3EDYVjN9yT/VP1tpCjYNTD6RxqZw2PhEcMmImlkji9xtWfX4F0oi7rx7Kco+NYrn3xiiyqapPabfMZFntoMVdo5nZxENwx2Xw6Fdy2m0qUI8J2RfjDvSJVm/Yn5Naqv5wDz6CWNxFg47lOSwETfnjt9xjAUykABCBemW5H+1vh742LyFah4gC+v7DO/j7azuRVpf65ZKMYSLF4vL8fl+8PV89m3RumeF87oDPpCaQ0+I+Wf520OKu0cxYfvrkbnb/48cQboVgc077tgWNMEgjxh3oEy1HMkR4tERbWeCPJPCJMCb30JZ7l8gft+Ue7FQ1WHvMbpVd8uiMkGksNpLWPApFcFDRjtZgjPs2NuBMhUlY8vrCG6vcUq167bPcC7LzuQOOni5cNjNtwThHAjEt7hqNZiDJlOTPL2ymasdtqmGkCcGxsHAhLFyIO2KUuPMOttxzFeseD7ZTJLr648QzyE9HzIzTct9dVwdAtHiZWuXa3Ti0WwbAVYhXhGg9apXq71+pI5FKkS8iRIQbbEqIV1fYVBhk2hdvN3zukQ6VpmCoL5G068nwu28/0kUyJbW4azSagRwJRPlncT92enjFdjoy0gGp1MQ7vu02uO028mLNJDGrGqBp0uJOJCfl9ko631Kun9mnDzqWZ7fQKcc2oSqlJGlkrDx0WP3ycM0+UR30HxzaLQOYXIUUiRAtGT73YCzB3a8f4pJlFZRY4/hTDnpMSohXlKRL66VztxuWe/rX0wiWezpiZmujmg+YFj53IYRZCPG2EOJh4/0cIcQbQoh9Qog/CyFsRrvdeL/POF47SWPXaN61HOqIcJX5FXYUns8z4bkImewP95sgqZSkoLeDiK0QTBkSYVikBSI3KQhqQ5tJYIXKEwcdy7NbaE+6oSeoimCMRKQTeuN8+q5N/PPdqlJUa/MRAKxVqweN/2iEq4gSc5jmrn5x/9OGeoKxXv757HmU2uK09tipD6mFSUuKjxJ3e74SeIx5iCHFvT8zpM9tI5ZQX8Q1RdNA3IEvAjsz3v8Y+JmUcj7gB24w2m8A/Eb7z4zzNBpNDmlobsYnQtQuOxVvsWFdTyTJVpobb6T7M/9CCX5ijpKBxxwqp7uH3CQPWxTfyiHXCYPDE4ECp5XmhCF8o61S/e158MKP2dnczVM7WnhqRwtBv/K5U7Gy/7wR3DKFphCN/mhf02sHOlhUls/yKg8+U5S2hJ23m9VnXuAzZDMeBJMVLA4l8GlGmFAl4qfQZQXAahbKdz9JZCXuQogq4FLgd8Z7AZwH3GeccidwlbF/pfEe4/h643yNRpMjupoOAJBXNhent1Q15sLvvmcP7YeaKBUBel1lA4+ZrUirW2WGnKjPPdbNwtR+jnjWDHl4ltdJS28WC5kSMeVyad/bV5v0K399hwLZTdJkg6L5YDIs7WEsd5yFeGSQxkC/uNd3RphtWNUuGSYoXTy2R0W6FNsSfZ8Be75KNWAUM1H3GULcLXawugesUq3yuTCbJk8as7XcbwG+CqSdekVAQErZa7xvAIwEC1QC9QDG8S7j/AEIIW4UQmwUQmxsa2sb3+g1mncpPe2q7qfJW4PTq0S4N5Sbv6NOq5MSEYC8skHHhMNDqTVOa/fEVqkm6l7FLCT+kpOGPF7lc2ak/R1B3INNACRDbUR6kqyq9tIVTVBIUMWwm8xQMEudO4zPHVchzlSItu4wPb0ppJRqha7hD7cmQoSFi6aIEmKRzu0eD/b9mhlV3I37EO2k0FjINJn+dshC3IUQlwGtUsqclj2XUt4mpVwrpVxbUlIy+gUajaafdPFobw0FheUABDuactK13+KgWHRjLigffNDhocgSHTImfCz07H+ZhDQTKx/sbwcl7n35ZUay3A1xTwVV4q8Pn1zDyXMKqXJEMbuN0n2eGvU6nFvGmOz0yBBNXVHaQnFiiRTVPickE4jeKLY8H2EMF0q6gEi8u98d48hC3I20v2nLvabQOfR5OSKbda+nA1cIIS4BHEAB8HPAK4SwGNZ5FdBonN8IVAMNQggL4AFy4AzUaDSgwiDdkSMkrHas7mK8pUpow/4WhrFNx0TYqSxUu69y8EGH8kGPWnN0FMThV9ki5+IpKBjyeJXXhV9mYbl3q4lTEVHiXpxv4/8+fhL2OyXYDT+3txoOMbxbxvCHe4Xyu9utZsCY7DTCHX2+YiJ+Y26gT9yDYDeEfDSfe/o+0U6K+sR9ii13KeXNUsoqKWUtcC3wrJTyI8BzwDXGadcDDxj7DxrvMY4/K3OxnE2j0QAqDLKCViLOWSAE5YUeQtJBfKS0tdmyahU9lUrsXEWzBh93ePCI8KCY8DGRiOJofYcNqcV4DRfF0RQ4Lf0JxUay3A1xt/QEsdNDcZ6dPLsFa9zfl1oAj0pZPNKEKoCPIA3+aF8q3mqfS9VdBWZXVpC0GGKcFvdYhuWedssMlcs9jZE8rChPfUlMubiPwNeALwkh9qF86rcb7bcDRUb7l4CvT2yIGo0mk7qOMJWinZThbqjwOOiU+SNnNsyWW25BLp8LgNVTMfi4w4NbhmkPxftiysdM2y5Mspd3UvP6siQejRCCQq+XhLBm5XMHKKK7TziJdPSLe80pKl4/f4jPA31umSJTkAZ/hHojW2OVr99yXz63ihe/eak6v8/n3p3hc0+7Z4bI5Z7GsNxPnO3jh1ct47zFg+c0csmY0pFJKZ8Hnjf2DwAnD3FODHh/Dsam0WiGoK4jwqWiDXvxWYAqKH1QeHBHcrNKVYSNMMK8oX3uzmSIlISOcJzS/HGE8rXuAmCvrOzLkjgUVYVuurrzKc7CcgcoFl3K5ZHsVekG0uI+/3z48q7h+zAs99nOOA3+KFazieI8O06buS+WXTg85DntKuyxxygwEg8O9rkP55IB9SUSDWAmxUdPnT38eTlCr1DVaGYYjS1tFIoQjuLavraI1atcERPlox/FemgbKQTklQ4+7vBg6w0hSI3fNdO2k6SwUifL8Rox30NR5XPSmcpDjma525TAVttCOKxmYzGX7Bf30TAs9xpHnIZA1EjoZUx2plMMpMXb5oaeCEhpTKim2zMs9+FwFapxZdZknUS0uGs0M4xwixEG6avpa0vYC3H2BibeeUMDTkKETB4wDyG8Dg+CFG5io8e673safn+psqQzad1Jm2M2bocdq3l4CaryOWlP5ZEMj2S5N0HFCgBm2w2LOr2Ya4hUwkNic4PZxix7hEZ/dEAYZJ8Qp0Xc6lZumd4YpHr7LXezBYzMkcPSl4Lg2MSXaHHXaGYYSf8htZOR1CvlKqIg1aUsygmSZw4Rtg1j9Wbklxk1YmbXI3DoZQgcGtjeuotG6+xh/e1p0rHuvaFh3E2plLLcy5W4V9qOFvcsLXchwFVEiVmFQjZ1xdRkKmTkjzFE2+ZWbpmjLXpQXwAjiXs6NDNXSd5GQYu7RjODSKYk9lA6Y2N1X7vJXYKdBPHIxH7ypwCvKUjUPszak77MkJHR3TKtRrYS/8H+tngIug5zQFQPGymTpsrnIiDzEcP53CPtkEpA4RzCuCg3G4I7VnEHcHjxiQgpqZ5x9dFumbSFbnMpt0yX8W+QOS9xyo2wfITpxrS452huZDS0uGs0M4iOUJwy2aaW1rv7feJ2j9pvbzky3KVZ0W12UGLqotc1hL8d+sIJK+zxkRcySdkv7p0Z4t62G4DdqcpRLfdKr7LcrT3D/CJJT6YWzKKDAkrEBMTd6SWfUN/bPss91gW2vH4Xlc2tQiFbd6j3pUv6+zjzy7D0iuHv4Ta+MMPHZkW+FneNZgbRGoxTJdqIuWYNyNjo9ikLsrN1YqtUO05aRzFdyCFSDwB9lnu1s2dkt0ywuT9Lpb+uv71NCf7W+KwRI2UAvC4rEVMBJpkcehIynXogr4LWVAG+lHG/sfrcAZw+nMlg39t+n3tgoKvF6oZEWH1xWZzgm0PWuLRbRqOZ8UzWur2W7hhVoo1kQfWAdk+REuOQf2Li3n39h7CKJOaCYWLCDaGrdMRGttzTli1ioLi37gSLg+2xQgrdw0fKgIp1N6erNA2Vztiw3LssxbRJD/lJI1oo0qlE2DqG5f0OL9aeLoRQBborPEaIZ6xr4MpWm6vfci9dPDAl8mhYbOr5actdo5mZPLa1iVN+9Axd0UTO+24NxqkU7ZgLB8ZJF5eqVAGRwMRWqUY6VBYRm3cYcS+oBKePU3o3jexzT7tkqk5Cdh5gd3Owrz1VtJBQj+zLsTIS9ry0uA9juQsTbdJDu/TgSnSmP8TYXDIATi8i1kVZvoNZXgeWdBRPNDBwZWs6FLJ1B5QuHds9QFnv2nLXaGYmm+sDtAbjPLurJed9t/sDlIhu7EUDxd1hpP1NdE9M3OP3/A4Ad9Hg0neASl27+qMsD72ECDYN/wuldaeaE6haS7KjjotueYENBzuhbRetTuXKmF+SN+p4XB5DpKOBQccOHtxH0lVKeyRJu/Rg6wlAMgFNW6Bo7qh9D8Dpg54QJ5Q7WVKeEQEzlFsm3AahloH+9mxxl2jLXaOZqRwxKvo8ujW3RasBYn4jl4r3qLwvtjx6sE7YKkyh8qbkFQ+RVybN2k9ilkmu5lkCkWF+nbTuUOLnq8WSjFBCF397dTt0N/JGqBSvy8rZi0bPBlvgU37qSPfA2PC69jD1dftokj7aQ3HaMQS4bRe0bofaM0f/sJkYrpefX1nLTz+4qr99kFvGDdIoDj4ey92tLXeNZsbSZBR9eGFPG6F47yhnj41kl+FTPzo1gBCEzF7MsbEVlN784wt58Tdf6G8wClHYj/7yyKRwLq1lZ/JhyzO0BkKDj6dSSmRLl/ZNONaIFo7seh2AR1p8XLlyFnaLedTxeYvUF0C3f6AgbjjYSbnoZF+sgPZQD+3SEPftf1evc84ete8BGK6XPBkmz56RlWWQWyYj2de4xL1Eh0JqNDOVpq4Ys4tc9PSmeHZXDjI1ZhIyXD35g6NZYjYfjp4xFJROpVgY2czqpr/w9Dv72dMSpDXPRiRlHyhiQ+Bfeh3lwk9y16ODDwYOqVWcpUugUIn7UkcnZ8lNJLDwcmIx7ztxGLfPURQVGxPFXQMF8Y2DnZQLP3U9Hl7c00an8KoD2+5XoYuzVmXVfx99NU4D/W3JXlXDNdMtYzNcSQ4v5A+Re2c03MVqTiCVHPu1Y0SLu0aTQ5IpSXN3jEuWV7AkL0Les99UpeByhDWSFvfBE569jkLyU11Ee7ITjlAwgEvEyRdRXv3Hbdx054tcYn4Dc2D0VLS2JRcTkXasDa8PPthmJOkqXQreGlIITvV1c6ltMy8nT6CytJjllSOs5MygvLiIXmkiFhzoltlysFEtpKKQF/e2kUyHGfrroGbd0KkTRiLtesmMyulbnertb7O6+j/beKqHuktApkavC5sDtLhrNDmkNRgjmZJUep3cXPQ853X9jVjdEAI4DpIpiaunnaQw9+cpycRdQrHopqkrOvjYEHQ2H+7bvyr1FO8J3ofXFMZefOmo15Z6XHTIAlJDJfVKh0GWLCJlstEki1gTe52KVBNPp07kfSdWkW1Z5dICJ924SIQCfW1HAlF6Aso95S6qVOubMhZ0MeesrPoeQNr1kim66f2h3DJl43DJQEYKgsmfVNXirtHkkCMBZaVXFVg5pftJABoOjJBudgx0hOOUECBmLx4yvtrsqaQMP02BSFb9BdtUqb6O6gtZIfbzOevDsPRK+M7/jnqt226hSxQgokMkweo8oOYEHAW0BuMcSpVSEVErU084+wNjSndrNgnCpnxkhui+WddJKQEA5tTOU+PJ9/Rb1XPGOJkKQ7tl0uGXQ7llxhMpAxmrVCff767FXaPJIWmreV5wA/ao8rdHWg/kpO/W7jil+Ek4h04N4CiuxiqSdLY0ZNVftFPFtMdO/jxYHJhkL5z7razHE7V6sQ41gRtu70sXXO+PcEga8wMVq/jwhesGTlhmQcycjyneH+f+xsFOamzKZbJyySIASvLsSjgdnr5EYmMiLeCZbpn0fqZbxlcLJgtUnzr2e0DGKtXJt9zH9pQ1Gs2INBmWe/n++5CuYtrCSWTmCs0J0BqMUSECkL9oyOP5pbUAhNsPA2tG7a+3S4VVemYvh3O/qZJwXfdFdfCxx0a9PukoxBGuH9AW6emls+EwQZHHElCVjaTxZbToklH7HHKctgJs6QReqEiZTxTGwQ+V1XP40Mk2zlxQAmKVyqtuGj0KZxBmq7LKB7hlAuo10y1TsghubgTrOIqUwDG13LW4azQ55EhXlEpbGMu+xxEn30jrxhdwhBpHvzALWrvjrBQBLJ6hwxRtPhWB0uuvH/L40YhgM2HpIK/AB6cb4ZDRh7Iej3AXkR/sJpWSmEyCPS1BPnP3Ju4ItbFTFlLYHaO+M8p2WYsUZsSSy7PuOxPp8OAIH0FKSWe4h32tIZYsiECXFVyF/OfVxkKn5X+YWMpjp290twyMX9jByHcjtM9do5lpNAViXON6G5FKwOqPEHFX4euZWKbGNG2BEEUiiMM3TAy6xwgv7M7uftZIC52mMSTXOgpbQQl5IkarX4ngvz+4HX8kQbklSIcs4PUDHTT4I+xwnYT40s5xT0KanT7yCdER7mFzfQCAamsQ8soGR6yMJ4IljcM7ultmopjMSuCPQay7FneNJoc0dUVZZT6ocpuULgXvbIqln1A4POG+o371C8AyVOFqAFcRCWHFHs4ueZgz3k63tXjc43H7lC+9qakRKSXbj3Rz+RIP1lSMkMXLa/s7qO+MqgyLQ8TlZ4s1vxAPYZr8Ud5p6MIkoFD6J9TnkDi9g90yZtvYEpBlwzFKQaDFXaPJIY2BGPNSdVB2AgiBo6QWk5A01O2ecN+JgJHOYKjC1QBC0G0txd2TXU6bgt52oo5h8rZngadIfcm0tx6huTtGVzTBykK1wtVbXMlrBzpoCESo8o0eNz8SzvxCbCJJc6efLQ0BFpTmYw63Dv8cxovDM9gt4/BM7NfAULhLdLSMRjOTiPcm6QxFqYgfhLJlABRWLQCgvX7PhPsXIUPcR7BY465ySlLto6Y9kKkURanOwUU5LrtMbVlQWKLEtaujmZ1NasJzcb7KFFlZVc2hjggN/mh/VaNxkudVvy787S1saehiRZUHQs2TYLn7BrtlcumSSeMu1tEyGs1MoqUrzmzRgjUVU5Y7UFajIlvCLRMPh7REjFQGQ6xOTZPMm0WF/zWaAlEWlOUPe153oAOP6Bm8hP4rX8l6PNZ8FfkR8bfS0qRS+tY6VYz9/No58HoQKZmw5Z7nUeK+53AjnWEHq2a5YHtH7i33odwymZEyueIYpf3VlrtGkyOOdEVZLIxVn4a4W72VJLCQ6jw0wpWjk0pJnD3tpDD1h9MNgdlbRRl+jvhH9vGnV6daR0oQNhpGzvSeYBu7moNUep24epQ41tbU4HWpFABVvolZ7sKpolV2HVRRQKsN1086lj5nOLzQG+tPF5F2y+Qad4n6VdDbk/u+M9DirtHkiKauKItNh5HCBCWLVaPJTKelbFA8eDb0JlMqtC/cQXsoTonsJGYrHDGO21lcg1Uk8beOHH4ZbFcLnVyFRyXwOucctWWDw0sKgQy3s6upmyUV+X3uBlNeCafMUZE41RO03NOuEXtvN1azYL7LyEQ5nsRdI5FepZp2zUymWwb6ywFOElrcNZoccSQQY4k4jCycNyDCIuKupLCnmUQylXVfj29rZtV3HmT3rR9F/vc8fnXHHZSKgAr/G4H80hogvZBpIFJKXj/QQU9viminEveC0uyyMw6J2UKPpQB7T4AD7WEWlxcowbK6webmylWVLCjNY5Z3gtEmhmvEQ5glFQXYooa/epRnMd779LlmJsstc4wKZWtx12hyxJFAlKXmekzlywce8M6mSrRyqCP7cMhd+w9wr/m7LGp+mKi0cX773awpjOMqqhzxOotP1VYdaiHTMxu3seH/vsS9z2zoywtfWFaT9ZiGIuEopFAESaYki9OWu2GZXrK8gqe+dDY2ywRlxrCePSKsJlOD6YnlXEfLqPsQDahfTJPmlklb7pPrd9firtHkiI7Odqpo7fO3p3GUzKFIBDl4JPuye1WH/s5yUx0vrf0l9zqv5QzTVryh/aMLWoESf3HUQqbItkdY88ilfMHyD4re/G9EsJkgTpx5ExMv4SrCh5pMXVxeYIj76BWWxoQhsAVEWFHlNXLai4GZIHNBplumJ6QqLk2GW6ZkMbz3VigZZ/KxLNHirtHkCFuHEctuhEGm8c6aD0Bnw/6s+8oPHiBgLuTMy67jhpt+APYCSPaMHiGSXsgUyVjI1LQF130fpjVVwG7PmVyQeB57+zb8pjEWkR4CW0ExhaIbu8VEbZFrgOWeM0xmktZ8SqxR1s0tUpa7uxjMOQ72y3TLpFMPTEq0TCGsvBYKho96ygWjirsQwiGE2CCEeEcIsV0I8T2jfY4Q4g0hxD4hxJ+FEDaj3W6832ccr53UT6DRTAOklPhCRiz7UZa7s1SJe7R1b9Z9lfQcJuCqVQ0OD6z9hNofLbZbCIK2UvJ6WvuKV7ftfg2AR074KbUf+glWkWQNOwlahxD3D3xAbVlizS+hSIRYWJaPxWxSIX65FnfA7PLx0ZUetdo11JL7MEgY6JZJL2aaDLfMMSIbyz0OnCelXAmsAi4WQpwK/Bj4mZRyPuAHbjDOvwHwG+0/M87TaKYFb9Z14g/3QE8kp6XOOsI9zEsdpseS15/jJU2xWshk7cxO3Fu7Y8ylkR7P3P7GUz8H1aeoKkOjEHdVUCrbaQ2qBUUdh3cSlxauOucU7OWL2OtV+c5jQ61O/exn1ZYlwlVEoQhy8QllRmRPe+7dMqBENm1Nh1pyv4ApfQ+EcstMRl6ZY8yo4i4V6Sq4VmOTwHnAfUb7ncBVxv6VxnuM4+tFtmVXNJpJJN6b5CO/fYOv3P0S8tfr4LGv5azvBn+UeeII0YK5g5erOwrwW8soCu/Lrq/GerwijKk0I7Vvfhnc8OSgXwVDYfZWUSE62deq/myF/wD1lDG7WC1qcp9zEwDJo1enAkQiassWdzEWevncaWVKfFOJyRF3p7ffmg5OkuVuMoOjYPLdMseIrHzuQgizEGIz0Ao8BewHAlLK9BrnBiA9jV8J1AMYx7uAQb//hBA3CiE2CiE2trVN/lJcjaa+M0JPMsW59f+L8NdB0zs567vRH2WuqQlZtGDI490FC6hNHqIrmhi1r656VaYuv2p8WRRdJbMpw8+BZhXSlxc6RKu1SrlNgKqV57Fr5c3MufAzgy++5BK1ZX0z40870tG/6nLSLPcApFIQbp0cyx2Upf4ucssgpUxKKVcBVcDJwOKJ3lhKeZuUcq2Ucm1JyST8Z9BojuJge4R1pu181PIMMWykclREA6C5rZ0K0YmjYug/jWTJUuaJI9S1jF4YOdGiyvIVzV42yplD465eiVUkCdZvhVSK4sQRwnkZpe2EYPF7v07JnOXDd5ItA8S9bWBbLklb7pEOSPVOjuUOKmKmq+Hd4ZbJREoZAJ4D1gFeIUR6uroKSC+JawSqAYzjHmByl2JpNFlwsC3If1huJ5o/m98lL8UUboV4aPQLsyDeoiZTHeVDV0lyVS3HKpK0Hdo+al9m/37i2LD4xheDLmatAsDS8g6xznrs9EDhvHH1NSpDifukWO5e5SqpN4qNHz2vkSsWXwqHX4U3bweEilKaoWQTLVMihPAa+07gAmAnSuSvMU67HnjA2H/QeI9x/FkpJ1IeRaPJDc0tzcw1NeNc9ykSRYaFHZhYzpc0pvRk6TBumcI5qwCIN2wdtS9P6CAt1qohi2BnhW8OUZObou4dtNUpF4+rYuH4+hoNl1Hs41iIeyIMT35bPeMFF+T+HgBnfgUWvgc69yv/+3j/DaYB2Yy8AnhOCLEFeBN4Skr5MPA14EtCiH0on/rtxvm3A0VG+5eAr+d+2BrN2Im0HVQ73hrspcqSlZ25KV7tDh4khYDCuUMet5Utphczlo5dI/YjpaQscZgud+34B2My0VmwhHm9+2k+qH4pFNWMz38/KpmWezpXymS5ZQD8B+GiH6map5OByQTv+61aYJQ/gaRq04BRVwFIKbcAq4doP4Dyvx/dHgPen5PRad59HHpVWWfX/QPsw6esHQ/J9JJ8TzW+SjPsgVDzfvLHu1Bw3zPg9CJnraEwdpguewW+4eprWmw0W6vxBocOh3zrsJ9ApIdVFU4qZSvtvivHOShFb+lylvjv5sXDW4lKG7Nrs3TLfPzjY7uRvQBMViXsPWFlYVtsYx3u6KQnNuefDwsvzH3/mdjz4YYnIKMo90xE53PXTC823AaNG6HuZVj0npx1G4734o42qUBeTzWzKwVd0kWkaS/j+gpJJeH+G6CgCv91zzJbHiGcPwffCJcE8uZR2bkFKSWZ0cEbDnbysdvfIJFM8YN1Jj4iJNayicUsOGefiH3P71nS/TKNpgrm27O0dMcq7kIoS333Y2qiczJcMgDly6FovrLajwUOz4yOlAGdfkAzneiJwJ4n1P7BF3PadV1HmFmig6TJDu5i5pflcViWkuw4OL4Om95R8dAtW2k7sJk5oonkKJOWiaIlVIk22jr64wu2NXZxwx1vssoT5urCOja8rj63t3pieUcK558EQJVoJ+Cszv7C9na1jYUll6nJzlAb1J4xtmuzpXQJfH4TlAw9Ya0ZjLbcNdOHvU9CIgLOwpyL+8H2MJWijd78SsxCUJJn521TBeWhwalxs+LAc8aOwLXpVtwiTlfZyMJjr1wO+6Bl32ZKi88H4NsPbCPfJrjb+T9Y23eC4dEoqR19sdJIWEoWEsGJiyjxgjnZX3iNESPx/PPZX3PpT9SmmVZoy10zfdjxD1WC7NTPQMu2nJYiq2sPUynasRSq8EIhBBFXFb540/jSEBx4XiUIqz2DykP/AKCgamRru3DuKgCi9ZsBNXm6tyXEv1W8rYT9rH+jeekneWv2DTjcE3QJmEwccaicNpaSoSN4NMc3Wtw10wPDJXO4/Hy++Kbhua57KWfdH2gPU23qxJwROy59tVjoRXaPXLVoyLEefh3mnoM84WpMxkLtvFkji3tp1UKaZSEFR9Tn6gz3kIhHuLDldzBrDZz7Tco/8DPWfOKnYxvPMHT7lPVfUKVdGe9GtLhrpgeGS+b7BxbycHsZSWteTl0zDW1+igiAp9//7ChTlm33kewSeqXpOfAyJHu45WAV6x/z0CtNRHCOWLgawGQ2sdF+CrWB1yER43BnhBvMj+GOt8KFPxick2aCFCy7iKDIo2bR2pz2q5kZaHHXTAvk2/fgNxXycmIhScw0+07MqbjH243FShniXlylFva0H96d/Til5KG//5G4tPD3jtmsXTqfxtKzEZWrsxLnuqKzcMgY1L3EkZY2brQ8TGj2+kmZiJx/+tXkf7cRtzf3KXg10x89oaqZejoPwr6n+EPve/n2FSv43kM72GFfSWX9C9DVCJ6RS8uNRiDSQ168WU1WZixbr56zgIQ0E2nJLlsjQFswztLoJpq9q3j2pvdgNglI3Asyu/qo8eozCB1x4Nr5MAV+L14RJn5u7rJTTpjPDJFMTDMj0Za7ZurZ+H+kELzquYwPn1zD3GI3bySUy4TmLRPufldzkEphTM56+y33cm8eTRTDGBKIHT6wmyWmw8h565WwgyqGbXNndX1NqY8XUytI7XqMFfV3s1Esw157Stb3n3Q++EG1aWY8Wtw1U0siinz7Lp6Wa1myaDFCCOaV5vFalzGp2pG9VT0cO450Uyk6kIi+GqOgImY6bJW4wtmHQ6Z2PAhA/qr3jmssc0vcPJ1cgyXSgqe3nce8Hx5XP5NGfb3aNDMeLe6aqWX73xFRP3ckLuCM+co3PL8kjx0BM9JVlBNx39nUzVxrJyK/YlBOkkjBHMoT9chUdm6V0sYn2c1sCmvGt4J0TnEez6VWkcLELjGXrorTx9XPpPGxj6lNM+PR4q6ZOhJReOH/o905hw2cwClzVYbBeaV5SAnR/DnQkX1R6eHY0dTNPJt/gEsmjaV0EW5itDbVjd5RsIWa8Fbedp/JeIuL+VxWUs4i/lT5Df419ilqirJz52g0Y0WLu2bqeOkn4D/IL2yfYlW1j3yHsqrnl+QB0G6vnrDlnkim2NsSooL2IXOAe2pULHjj3mGqMkkJe59Sed93PYwJSdOs8aebFUIwt8TNLa2r2SlnM7vINe6+NJqR0OKuGTvhjv4ak+OlbQ+8fAs9S6/h7tZaTp/fH643t8SNEHBYVECwaUIFNfa3hUgke/H0tAwIg0xTNX8FAN1GabtB7HsG7rkGfnUqva/fxoFUOfnVE6tgNKfY3Ve8uqZQi7tmctDirhkbyV64/QJ46Kbx95FKwcM3gc3FK/P+lZSkz98O4LCaqfa52NljFHDuHL9rZmdTN8tEnVpF6ps96HhecTVhnMj2YRYy7XkcrC6wOrF07OKJ1EksKJ9YdZ65xf2umNnaLaOZJHScuyZrnt7Rwsru5ynp3A+WYfKWZ8OGW+HQK3DF/+PpQxKXzcyqau+AU+aX5rGpo4hPgXLNVKwc1612Nvr5D9vvke4SxNKrBp8gBG32GvKCQxTtkBK59wkafSdT+PF7ePux3/PrN8t4vDRvXGNJM6dYXZ9vt+BzTVLRifHy5S9P9Qg0OUKLuyYr6trD3HjXmzzvNfKedI0zXK59Hzz9PVhwIbFlH+Lhh57l3MWl2CwDf0TOL83j3n0e9T90ApOqlfv+yAqxHy6+vb8k3FHEPHOZ1bKBULyXPHvGn0TbbkTgML9KXIDt6UNIy7kkbQ1UeCbwxYZyywDUFLnGPTE7aVx++VSPQJMjtFtGkxW3vXSAteyiJroLihdBvFtVo8+S5q4YHaE4PPQFsNjh8l/w+PYWuqIJPnLy4ELQ80rcBHst9OZXjn1StScMb/4O+fCXuCbwf+zOOwWWvW/Y0y2li6kUHeytbxrQLvc+CcDzqVXc+VodT+9sZX5Z/oQFubZY+dmn5WTq7t1q08x4tLhrRqU1GOO+TfV8xvoQHbKA1Bn/qg4Eslv8E+1Jcu7/PM+1/3EHHHqFN6pvgIIK/rjhMLVFLk6dO7jm5iLDrx1wzs5a3F/Z185n7t5EYsPt8MiXkVvv4+3kPLas+vcR8774ZquImSP7tw1oD255hJ2pGj7xnjMoybPTGIiyYIIuGQCXzcKlKyo4f0nZhPvKOZ/+tNo0Mx4t7ppRueOVOj7Ko5xr2sxtvZfSajcmJrN0zRzqDBNNJPnarHdIYuKzW+fzo0d3suFgJ9eeXIPJNFh4l1TkYzObOESFEncpR73PUztaeGxbM02v/RXKlvPNhQ/x0cQ3WbJk5MIXhUY4ZLBhZ39jrAt3y5u8alrDR0+dzXcuVwWmF5Xlpq7r/354DVevGRyaqdHkCu1z14xIMJbg8Ov38wvL3fhnX8xtuy/ljLiPcoBAduJe1x5BkOKs2HOI+etZ2buA2148gNUsuObEoQXObjGzZFYBW2IlnBjrUgWY3SNnN2wMRCmmi6rQVp50f5x732zgX86dz7LKkQtfiKJ5pDBBR3/ETMe2pygiiX3JxThtZi5dXoHtYyZOnTf4V4ZGMx3RlrtmRP72ylb+S/6CWPFyTO+7DYmJ7QErWJzZW+4dYU4Wu7GFj2BaeS2/+sgaLlhaxnXrainOsw973epqL6/5vepNFq6ZRn+UT5fvwiQkP61fxDmLSvjXCxaOPkCLnYC9goLQARJJlYagdcvTRKSdM865GFCLjy48oZwCxzSLbtFohkGLu2ZY4r1Jjrz6J/JEDNfVv8RT4KGswM6elpBayh84lFU/dR0RPuh4FWx5sOgSHFYzv71uLd++bOmI162q9rK714h1z0Lcj3RFOSu1gZ78GlavPZ2ff3B1f+bGUUh45zOHRnY3BwEoaHmTbaaFzC71ZnW9RjPd0OKuGZYH3j7CeYkXiBTM64szX1iWz57WIHhrsnbLNHYEuIjXYcnlYMs+QmRVtZdGWUxKmFXO9xEIx3tJRLqZH9yIbdkV/Of7VuAZQwy5Y/aJLBQN7Ny3H6J+KuL7OeI9cfqFKk423/qW2jQzHi3uxwtSwqY74HcXqKX9EySVktz//OucYtqF88Rr+6JNFpbls681RMpTnbVbxtP6Fm4ZVuI+BmYXuch3OfFbysCfIe6b7oQtfxkwyXokEOU809uYZQIWXzqm+wAUrHkfZiEROx8itPcVTEhkzWlj7mfGc/75atPMeLS4Hw/Eg3DfJ+ChL0LjJrj7alXBaAJsaexiVeAZAMTya/raF5blEUuk6LKWq0nOnvDIQ+tNckL0TZLCDHPOGtMYhBCsrPZSJ0uh01hBmkrB41+Hv30K7vtkX46b5rZWvmr5M7H82VA99uIXouwEjlhrmN/2JJ07niMuLVQsmWbpeI8FmzerTTPj0eJ+PPDST2DHA7D+u/CpZ9XioruvVqKfLYko1G/oe7unJchV5leIlZ0IhXP72hcYoYD10ohcGcU10+CPcrbpHTqL1oB97GGEq6q97IgXI9Numa7DkIjA7DPUZ/7NmdCwkYrXf8As0U7oPb8Ek3nM90EIGiouYkXvdtwHH2eLnMcJtdMwDn2yuekmtWlmPFrcZzpSwra/wbz1cOaXYNYquPpWaNsFB57Pvp8nvqkSgu15AoDEvhdYYjqMddUHBpyWXsSzN25UShrFNdPUcIAlpsP01J6X/VgyWFXt5VCqDBELQKQTWnepA+u/A598XH3+2y9kfsPfuTV1Jb7FY/t1kIl5+XsxCUlRvIG9zhV9KYg1mpmIFveZTtM7Kmpl6ZX9bVUnq9fupqGvOZpQG2y+R+0/+AVo38cle79DvakS85qPDjg132Gl0uvkrS5jpebRq1RTKdj9ODz6VQi3k9r7LADupReN9ZMBsKLKy2FpRMz4D6ovLYCSRVB9MvzzS7Dsfex0n8xf3B/OOjpmKOYvO5ndKRV3HymfRnVNNZpxMKq4CyGqhRDPCSF2CCG2CyG+aLQXCiGeEkLsNV59RrsQQvxCCLFPCLFFCLFmsj/Eu5od/wBhHjiJ6CoCkxW6s/S7b7gNeuNw9e8g3Aa3nok72c1dVf8O9sHL7dfW+nisTiJN1oHi3r4P/vdkuPeDKvPjPddQ1vAEbdKLp3b1uD5eodtG0GXknuk0xD2/Apxe1eb0wvt+y3fzv0+pb2KpeD1OK8871xOUTgoWvAv97Zrjimws917gy1LKpcCpwOeEEEuBrwPPSCkXAM8Y7wHeAywwthuBX+d81BqFlMrvPPfsgRkPTSYlgMEsLPd4SIn74kthxfvhjH+FRITvJz6Go2roNLsXLi2nM5ok7p410C3z1p3qV8T7bocP3gNNW1jU/SrvOE5EmMb/IzGvfL7aMSx3WbKYZ3e1cNkvX+Jr920B1OrUKq9z3PdIs3fuJzgzfgsnzK0c/WSNZhoz6l+clLJJSvmWsR8EdgKVwJXAncZpdwJXGftXAn+QitcBrxCiItcD14Bs3gKdBwjOu5RoT3LgwYJZ0H1k9E7evgtiATj9i+r9ud/k4Puf5O7keuYNkyTr7EUl2MwmmigeaLk3bFTx8MuvgSWXwRW/JImJvYXrx/cBDeZWltAifaTa90Pbbl70F/LJOzaypznE/W810B6K09wdY1YOxP2K1VWsXDg3ZzlkZhw/+pHaNDOeMZlTQohaYDXwBlAmpUybhs1AOrSgEsicZWsw2o7u60YhxEYhxMa2traxjlsTD7H7L9+hV5o460E36/7rGYKxRP/xgiws90QMXvk51Jym/NcAJhPbk9WAYP4w4p5nt3Da/CI2RCuRTVtUpE0yAUfehqqT+s7rXfEhTuz5LcHZExP3JeUF1MkykgdegESEx1t9fOjkav7yz+voTUnufLWOZErmRNzPWljCnZ88GYv5XTodddppatPMeLL+HyyEyAPuB26SUnZnHpNSSmD0tH0Dr7lNSrlWSrm2pKRkLJdqGjbCr9ex0P8Cf3Z/iOvPP5FAJMFjW5v7z8k3LPeRsim+fZf6AjjnawOa97WGEALmFg+f3vaCpWU8ElmCSMbh0KvQsg16o2wzLeIL977N6u8/yarvP0Ug5aR2gqXkFlfkczhVijWkfonsTlXx/rXVrKzyMLfEzZ2v1gFQ6Zu4uL/refVVtWlmPFmJuxDCihL2e6SUfzOaW9LuFuO11WhvBDIrEVcZbZrxsuuRgS6WR75Eb28v749/h/hp/8YX1y9gbrGb+zY19J9TUKHiwYcrZN0bh5d/RmfRGr65uRCZ8SWwvy1MpdeJ0zZ8vPj5S8rYkFpMUlhh/7PqCwe48VnBi3vbWL+kjA+sreYz58zjwhMmFi8+tziPBtHfR7NtNisqPQghuGLlLLpjvQBUeidWIUkDfOMbatPMeLKJlhHA7cBOKeVPMw49CFxv7F8PPJDRfp0RNXMq0JXhvtGMlZ4w/Okj8JzhB436oWkLuyquYpNcxBkLihFC8L4Tq9hQ18nhjog6r2CWej3aNfPa/8JfrlerO7sb+X7wcu7ZUM9rBzr6TtnXGhrWJZOmrMDB4uoytlmWwv7noH4DAXMRJm8VG75xPv/z/pV85/KlfO3ixXhdtgk9ApvFRCy/FoA2UcgJ82r63CZXrJzVd14u3DIazfFCNpb76cDHgPOEEJuN7RLgv4ALhBB7gfON9wCPAgeAfcBvgc/mftjvIjr2AVJZx1LC4dcByTOxhZTk2/sWFb13dSVCwP1vGdZ7viF6R4dDvnwL7Hsa9j5FR+lp/KN7IWaT4DcvqOX9yZTkQFuI+SWjVxw6f0kpj0aWQut25P5n2JScx2nziwfVQ80F9tJ5AOzsreSM+f153eeW5LG80oPXZcVl0+UJNJo0o/41SClfBoZbGTJopszwv39uguPSpGk3Ckh0N0Lbbqh7GWm286fGUs5YVNyXtXCW18np84r529sNfHH9AkwFRoBS5kKmeAjCrXDet+GML/H5373BLE+ED55Uw8+e3sP2I13k263Ee1OjWu4A5y4u5atPLQfuRUQ62JC4kHWTVMyisGoR1ME+WclZ8wcW7fju5Utp8Ecn5b4azUzlXRoScAzpjUPLDhWPPp5kXpl5zPc/A4deIVK6iqYInH6UyL13dSX1nVG2HelSce4w0C2Tzr9eOIfdrWFePdDJx9bV8vHTanHbzHzvoR3c/HcVN56NuC+tKKAzbyFBsxeAt1ILWDd35GpJ42VOTTXfTVzPY87LmFcycIJ2bW0hV63WcekaTSb6d+xkkojCz1dByIhimXsuXPePrC9PpSSm9r0qd7rZDtv/AU3vsGv2PwFw+vyBVvKZC5Swvra/gxVV88BVPHAiNp18y1fL7S8fwG4xce1J1XhcVj566mxuffEA5QUOvrh+ASfO9o06PiEE5ywp44V3lnGxeI1g4TLKPZMzqbmkPJ/rkxfxvgVV774c68eSW26Z6hFocoQW98nkwAtK2M/7NoRa1ErQzgMDsiwOR0cozjn//Twv+3biKVoAxQvhDbXY989tNcwrcVPhGTiBWFrgYF6Jm9cOdPDps+epiJlMcffXAbA1WsRfN23nE6fNwedWk51funAhl66o4IRZnjHlZzl3USnff/P93C/WsXrFrNEvGCcl+Xa+cuFCzl/6LszUeCxZtWqqR6DJEdotM5nsfgRs+XDa59WyfmFWhSay4O3DAYLxBLauA1C8AOar6Y1eLDzqr+Jblw5dom7dvCLePNipaoEWVEJwoLhLewFfe/gwpfl2/vWCBX2H7BYzK6q8Y068dfr8YlpNZTyXWj1p/nZQvxL+5bwFLC6fWP4YzSg8/bTaNDMeLe6TRDKZRO5+nNT888FiV6GJCy9W2Rd7e0a9fmtjF2X4ccooAdds5OzTSAgrb6Xm8cP3n8y5i0uHvG7d3GLCPUm2Nhp+98wJVf9B2m2V7GgO8u+Xn5CTlLZuu4VT5qq8NqfOLRzlbM2054c/VJtmxqPFfRJo7Y5x3Q9+hQi3ctPmWdz1Wp06sPaTKuvirodH7WNbYxerXSotw/PtHh7b3c13eq6jadUXR5w8TAvsa/s71BdKpF1N6gLdTft4M5DP+UtKuXhZ+cQ+ZAafO3c+X1i/gNJ8vYhIo5kuaHGfBJ7f08bpvRtICjPNJWfy/57bR7w3CfPOU5Oj79w7ah9bG7tYX6qyPNy118b3HtrOO6Xv5dIrPzTidUV5dhaV5fP6gY4BETO/enY39lAD0jeHn1+7OqeTkqfOLeJLFyzMWX8ajWbiaHGfBF7Z187Flrcw1Z7Ov1y6lpbuOA+8fUSl4p1zFjS+NWLOl9buGK3BOEusLSTMTjYFnLQG4/zo6uVZJbRaN6+IjXV+Em4l7t2th/njU69jF71cdOapuO16Hl2jOd7R4p5jpJQc2LuDuTQgFl3CmQuKOWFWAb95cT+plITyFcpVEmweto+tjSofTFWyAVPxAjxOG9evq2VVtTerMZw6t4hoIsnOsIoH37tvN1VCpf6xFM6Z2AfUaDQzAi3uOWZ3S5CaqFEKrmYdQgg+ffY8DrSFeWpnixJ3gOYtAy/sCcM7f4Zkgq2NXQgBBeE6zCULePGr5/Kdy4aOjhmKMxYUY7eYePCQFWx5sOdJFtmM3DFa3DUjceutatPMeN694r7ht3DrWfD2PSoXeY54eW87y00HkWYblCpBvmRZOeUFDv7xdiOUnaBOPFrcN/4e/n4j3Hste+ubOa0whKmrHooW4HFaMY0hRDHPbmH9klL+sd1PavV1rOx6hvfk7weTBQqqcvVRNccjixapTTPjeXeKe7AFnvquytvywGfhV+sgHsxJ16/u7+Ak+yFE6VKwqAVCFrOJU+YW8tZhP9KeD7450Lx14IUNb4LVDfuf5ZuHPskfwp9RYjz//HGN44qVs2gP9fAX86UgJacEnwJPNZi1v10zAg89pDbNjOfdKe7P/yck4/Dpl1S9z469yoKfIIlkitcPtLOEgzBr1YBja2p8tHTHOdIVg4oV0HSU5d64CRZeSOCK39OTMrGj5sPwxXeg+iTGwzmLSsm3W/jeSyEeTZ2CQIKvdnwfTPPu4Sc/UZtmxvOuE/c33niZ1KY7edF7JbftEKreZ/Up8PqvIJUcvYNM9j2tfgEY122uD1CUaMKVDELFqgGnrqlRuVreOuSH8uWq2HPMKGgVbIGueuLla/j8WxWc0/MzuPA/wDP+ZFgOq5kLTygnmkjyZMH7VaMWd43mXcO7StwTyRShx39ASDr4Xtel/OjRXWxr7IJ1n1MZE3c9kn1nh16Fez8Mr9wCLylLZ8PBTpYLIznXUZb74op8HFYTbx32Q/lK1diyXb02qipG33rTzqv7O/jva1awvMozgU+quGKVyvVSsXQdXPoTOPnGCfep0WhmBu8qcX924xbOTm0gsOTD/P3friDfbuE3L+yHxZeBd7aqUpQNrTvh3mvBWw1LrlBunrpX2HTIz1l5jWCy9k2mprGaTayo8vLW4YCy3KFvUjWw9zV6MfOUv4LfXbeW96+tJhecMb+YL6xfwHXrauGkf4Ky7CNuNBrNzOZdJe5tL/4fFpGiav0/U+Cw8uFTa3h0axOH/DE49bNQ/zoceH7kTqSk5Q+fJNBjIvyBv8JVvwLfHOT9/0Rd3X5OtB1SImqxD7p0TY2PHUe6iDlKVDre5i3sbg6yZ9Nz7GU2d9x41rA5Y8aD2ST40gULqS505axPjUYzMzj+xf2tP0DnQd453MlZwUdp8p2EqURlQ7zh9DlYTCZ++9IBOPF65ZN+5MuQiA3bXc+uJygL7eA/49fwTw+0EjO54AN3ImNd/L/Uj5gd3zPI355mTY2XRFKy9Ui3mlQ99Bp/fHknS+V+KpefmfUiJY1m0rjrLrVpZjzHt7hH/WrC87azaX/gW9SY2vCe+am+w6UFDq5eU8lfNzbQlbDApT9VlY9e/ll/H4mYKib9wn9DKkXoyf+gQRZjX/NhXjvQwU1/2owsW8ZzK3/CAtGALdE9yN+eZs3sjEnVEz8Onfv5yM7PkieiFMxbN4kPQqPJkupqtWlmPMeVuCeSKe56rY5DHWHV4PTBp56lN7+K9R33EDZ7cK64asA1155cQ7w3xdM7W1TO9OXvh5d/CvufU/lfHvw8bLsfnvsh/N9FFPq38EfrNfz7e1fztYsX8/j2Zp7a0cIj4SX80PxZpNUFs88YcnzFeXZqCl1sOuSHpVfSc/FPWJg0aqRWrZ3EJ6PRZMmf/6w2zYznuFrR8uDmI3z7ge1YTIKPnFLDly9aREHhHO5e9jsST/2AS85ej/soX/jKKg+zPA4e29bE+06sgot+BEfehruugupTlR/+3G+BxYZ86rs0ySKsaz+GyST41JlzuG9TPf/52C56elMsn/NexEe+BybzsGM8bV4Rj2xpIpFMsbnsah5JbOELVXspKpw3yU9Ho8mCX6tqX3zwg1M7Ds2EOa4s93s3HKa2yMUHTqrm7jcO829/fQcpJXdvauORin+h8tx/GnSNEIL3LK/gxT3tBGMJyCtVi5vW/QvUvwHLPwBnfQVO/yIPrP4dn+z5N65aq/KzWMwmvnHJEg62h2kMRFlb6xtR2AHOXVxKMN7Lm3WdbK73c2fyIuTH/qEyRmo0Gk2OOG4UZU9LkI2H/HzklNn86L3L+epFi3hiewvf+Ps29rWG+PDJNcNee8nycnqSKZ7dpTInYnPBRf8BX9oJ770VhKA3meJne4pw16xkTrG779rzFpeybq4qL7e2dvRKRGfML8ZmNvHcrlY21weo8jkpzhscWaPRaDQT4bgR93s3HMZmNinXCvBPZ87l5DmF3LvhMHl2C5etrBj22tXVPsoK7Dy6tWnggYKKPov6b283cqgjwj+fPdB9IoTgP69ezmfOmcfyytEXHqXL0j27q5XNhwM6Qkaj0UwKx4W4xxJJ/vZWIxctK6fQrZJ1mU2Cn35gJV6XlWtPqsZlG356wWQSvGdZBc/vbiMc7+1r94d76E2mSCRT/PLZvSyv9HD+ksFx6LXFbr528eKsi0uft7iU/W1hjnTFtLhrNJpJ4biYUP3pU3voiib40MkDQ7iqfC5e/tp5uKwj+8EBLl1RwR2v1vHE9mauXlNFazDGOf/9PBUeB6fMLaK+M8r3Pn5CTsrTnbe4lO89tANAi7tmenHffVM9Ak2OmPGW+6+f389tLx7go6fW9Pm+M8mzW7LKhb52to+aQhf3v9UAwF83NhDpSZKS8Mc3DrOy2su5i3KzenR2kZu5JW4sJsGyLFw5Gs0xo7hYbZoZz4y23P+04TA/fnwXV6ycxfevWDYhq1oIwdVrKvn5M3tp8Ef405uHWTe3iLtuOJkntrewrLIgp0WlbzxzLruagziy+FWh0Rwz7rhDvX7841M5Ck0OmNHivqSigKtXV/Lja1aMqVLRcLxvTRW3PL2Xr963hfrOKP920WIsZhOXrhh+Mna8XDtC9I5GM2VocT9uGNUtI4T4PyFEqxBiW0ZboRDiKSHEXuPVZ7QLIcQvhBD7hBBbhBBrJnPwK6u9/PSDq7Cac+Ndqi50cfKcQl7d34HPZeWiE8py0q9Go9Eca7JRxTuAi49q+zrwjJRyAfCM8R7gPcACY7sR+HVuhnnsuGaNCqW85sQq7BbtMtFoNDOTUcVdSvki0HlU85XAncb+ncBVGe1/kIrXAa8QIvc+jUnk8pWzuOGMOfzTmXOneigajUYzbsbrcy+TUqZX/DQDaf9FJVCfcV6D0XbU6iAQQtyIsu6pqZk+/menzcy3L9NFLTQazcxmwhOqUkophJDjuO424DaAtWvXjvl6jUYzCTz66FSPQJMjxjsT2ZJ2txivRlIWGoHMlURVRptGo5kJuFxq08x4xivuDwLXG/vXAw9ktF9nRM2cCnRluG80Gs1051e/UptmxpNNKOS9wGvAIiFEgxDiBuC/gAuEEHuB8433AI8CB4B9wG+Bz07KqDUazeTwl7+oTTPjGdXnLqX80DCH1g9xrgQ+N9FBaTQajWZizPjcMhqNRqMZjBZ3jUajOQ7R4q7RaDTHIUK5yad4EEK0AYfGeXkx0J7D4UwmM2WsM2WcoMc6GcyUccLMGetkjXO2lLJkqAPTQtwnghBio5Ry7VSPIxtmylhnyjhBj3UymCnjhJkz1qkYp3bLaDQazXGIFneNRqM5DjkexP22qR7AGJgpY50p4wQ91slgpowTZs5Yj/k4Z7zPXaPRaDSDOR4sd41Go9EchRZ3jUajOQ6Z0eIuhLhYCLHbqNn69dGvODYIIaqFEM8JIXYIIbYLIb5otA9Ze3Y6IIQwCyHeFkI8bLyfI4R4w3i2fxZC2KbBGL1CiPuEELuEEDuFEOum6zMVQvyr8W+/TQhxrxDCMV2e6XSui5zFOP/b+PffIoT4uxDCm3HsZmOcu4UQFx2rcQ431oxjXxZCSCFEsfH+mDzTGSvuQggz8L+ouq1LgQ8JIaZLCaVe4MtSyqXAqcDnjLENV3t2OvBFYGfG+x8DP5NSzgf8wA1TMqqB/Bx4XEq5GFiJGu+0e6ZCiErgC8BaKeUywAxcy/R5pncwM+oi38HgcT4FLJNSrgD2ADcDGH9f1wInGNf8ytCIY8UdDB4rQohq4ELgcEbzsXmmUsoZuQHrgCcy3t8M3DzV4xpmrA8AFwC7gQqjrQLYPdVjM8ZShfqDPg94GBCo1XSWoZ71FI3RAxzECALIaJ92z5T+cpOFqMyrDwMXTadnCtQC20Z7jsCtwIeGOm8qxnnUsfcC9xj7A/7+gSeAdVP5TI22+1CGSB1QfCyf6Yy13Bm+Xuu0QghRC6wG3mD42rNTzS3AV4GU8b4ICEgpe4330+HZzgHagN8b7qPfCSHcTMNnKqVsBP4HZa01AV3AJqbfM81krHWRpwOfBB4z9qfdOIUQVwKNUsp3jjp0TMY6k8V92iOEyAPuB26SUnZnHpPqK3vK41CFEJcBrVLKTVM9llGwAGuAX0spVwNhjnLBTKNn6gOuRH0hzQLcDPGTfboyXZ7jSAghvolyf94z1WMZCiGEC/gG8J2pGsNMFvdpXa9VCGFFCfs9Usq/Gc3D1Z6dSk4HrhBC1AF/Qrlmfg54hRDpYi7T4dk2AA1SyjeM9/ehxH46PtPzgYNSyjYpZQL4G+o5T7dnmsmMqYsshPg4cBnwEeOLCKbfOOehvtzfMf62qoC3hBDlHKOxzmRxfxNYYEQg2FCTKQ9O8ZgANRsO3A7slFL+NOPQcLVnpwwp5c1SyiopZS3qGT4rpfwI8BxwjXHalI9VStkM1AshFhlN64EdTMNninLHnCqEcBn/F9JjnVbP9ChmRF1kIcTFKBfiFVLKSMahB4FrhRB2IcQc1GTlhqkYI4CUcquUslRKWWv8bTUAa4z/x8fmmR7LCYdJmMC4BDVjvh/45lSPJ2NcZ6B+1m4BNhvbJShf9jPAXuBpoHCqx3rUuM8BHjb256L+OPYBfwXs02B8q4CNxnP9B+Cbrs8U+B6wC9gG3AXYp8szBe5FzQUkUKJzw3DPETW5/r/G39hWVATQVI5zH8pfnf67+k3G+d80xrkbeM9UP9OjjtfRP6F6TJ6pTj+g0Wg0xyEz2S2j0Wg0mmHQ4q7RaDTHIVrcNRqN5jhEi7tGo9Ech2hx12g0muMQLe4ajUZzHKLFXaPRaI5D/n+wFaMZO2QHOQAAAABJRU5ErkJggg==\n", + "image/png": "iVBORw0KGgoAAAANSUhEUgAAAXcAAAD4CAYAAAAXUaZHAAAAOXRFWHRTb2Z0d2FyZQBNYXRwbG90bGliIHZlcnNpb24zLjUuMSwgaHR0cHM6Ly9tYXRwbG90bGliLm9yZy/YYfK9AAAACXBIWXMAAAsTAAALEwEAmpwYAABXSElEQVR4nO2dd3hc1Zn/P2d6lUa9y73ihjHG9B56C4RAGmTJstlNIW3Tk01280vZNEgPCQkkEBICJJSF0GsoxsYG9ypZ3aqjGU0v5/fHuaNiS9ZIGlmSOZ/nuc/MPffec89cW9955z3veV8hpUSj0Wg0xxamqR6ARqPRaHKPFneNRqM5BtHirtFoNMcgWtw1Go3mGESLu0aj0RyDWKZ6AADFxcVy9uzZUz0MjUaza5d6XbRoasehyYqNGzd2SilLhjs2LcR99uzZbNiwYaqHodFozjpLvT7//FSOQpMlQogDIx3TbhmNRqM5BpkWlrtGo5kmfPWrUz0CTY7Q4q7RaAY477ypHoEmR2i3jEajGWDzZrVpZjzactdoNAN86lPqVU+ozni05a7RaDTHIFrcNRqN5hhEi7tGo9HkkLca/Ww80D3Vw9DirtFoNLnkO4/v4GP3bCKdntpaGXpCVaPRDPDtb0/1CGY8/nCCtkCU9fXdrJtbNGXj0OKu0WgGOOWUqR7BjKc3kgDg4bdaplTctVtGo9EM8MoratOMm4Ah7o9taSWeTE/ZOLS4azSaAb78ZbVpxkUylSYUT7GyOh9/OMHLezumbCxZibsQwieEuF8IsVMIsUMIcbIQolAI8ZQQYo/xWmCcK4QQPxFC7BVCvC2EWD25H0Gj0WimB4FoEoBLVlSQ77Ty8OaWKRtLtpb7bcA/pJSLgZXADuCLwDNSygXAM8Y+wEXAAmO7GfhlTkes0Wg005SMS6bYY+fsRSW8ur9rysYyqrgLIfKBM4A7AKSUcSmlH7gCuMs47S7gSuP9FcAfpOI1wCeEqMjxuDUajWbakZlMzXNYKfHaCUSSUzaWbCz3OUAH8HshxCYhxG+FEG6gTErZapzTBpQZ76uAxkHXNxltGo1Gc0wTiBri7rTisVuJJFIkUlMzqZpNKKQFWA18Qkr5uhDiNgZcMABIKaUQYkwR+0KIm1FuG2pra8dyqUajmSxuvXWqRzCjyVjq+U4rXoeS11Asic9lO+pjycZybwKapJSvG/v3o8T+YMbdYry2G8ebgZpB11cbbUOQUt4upVwjpVxTUjJsCUCNRnO0WbVKbZpx0e+WcVr6xT0YnRrXzKjiLqVsAxqFEJmKuecC24GHgRuMthuAh4z3DwMfMqJm1gG9g9w3Go1mOvP002rTjIuMW2aw5Z5pO9pku0L1E8A9QggbsB/4MOqL4T4hxE3AAeBa49zHgIuBvUDYOFej0cwEvvUt9aorMo2L3kgCi0ngtJrxOqwA9E2R5Z6VuEspNwNrhjl07jDnSuBjExuWRqPRzDwCkQT5TitCCDz2ae6W0Wg0Gk129EYS5DmVxZ5xy/TFtLhrNBrNjCYQTZJniLqnf0J1anzuWtw1Go0mRwQGWe55hs89OEWWu075q9FoBvj1r6d6BDOaQCRBVYETALvFhNUspsznrsVdo9EMsGjR6OdoRiQQVROqQP+kqnbLaDSaqeeRR9SmGTNSSjWharhjALwO6/QOhdRoNO8QfvhD9XrZZVM7jhlINJEmkZL9ljtgWO46Wkaj0WhmLINTD2TwOixTNqGqxV2j0WhywODUAxm8Dqu23DUajeZokE5LHnmrhXA8t6IbGJTLPYPXoSdUNRqN5qjwyNstfOLeTTy57WBO+x1wywwV96laoaonVDUazQB//ONUj2BSSabS3Pb0HgD84XhO+x7OLZOZUJVSIoTI6f1GQ4u7RqMZoKZm9HNmMA+/1cL+zhCQ+5wvveGMW2bwhKqVVFoSTaRx2sw5vd9oaLeMRqMZ4C9/UdsxSDKV5rZn9rC0Ig+7xZTzic6A0d9gt8xU5pfR4q7RaAb45S/Vdgzyel03B7rCfPyc+Xgd1n4xzhWBSAKXzYzVPCCrGSt+KsIhtbhrNJp3BB3BGACLyr3kTUIUS28kMcTfDkxpqT0t7hqN5h1BjzGBWuCyGSGKuXbLDE09AOCxG5khtVtGo9FoJoeecAIhMvVNrZNiuQ9enQqDCnZoy12j0Wgmh95wnDyHFbNJTErOl0AkeZhbZipL7elQSI1GM8D990/1CCaNnnACn2ugBF7OQyEjCRZXeIe0TWXBDi3uGo1mgOLiqR7BpNETjuNz2YDc53yRUtLZF6PYYx/S7rar2Hbtc9doNFPLnXeq7RjEH05QcIjlnkrLnPTdF0sSS6YpctuGtFvMJlw2s46W0Wg0U8wxLO494TgF/Za7MdGZI3dJV5+KxDnUcs/cS0+oajQazSTRO8jn3u8Lz5G7pLNPxdAXew8Xd4/dQjCm3TIajUaTcxKpNMFYEp9zqOWeK3dJRtwPdcuoe01NTnct7hqNZvrRvhO2PpCz7vxGUq8Cd8bnnrHccyXuyi1TMozlPhkLprJBR8toNJrpRTwM914HoU5YdnVOusyk981Ey3j6fe65dcsUDmu5W2jtjebkPmNBi7tGoxngscem5r71/4RkFOadA89/B3rqAAFSQg7yoPdkLPdB0TKQW7eMz2UdkjQsg1owdfR97lrcNRrNAC7X1Nz3/z4LHTtgzhlQ/zLYvBAPQiICtomPqd9yP8TnnqvMkF198WEjZdS9tM9do9FMNb/4hdqONpEeKJoPzZvAUwan3aLaE+GcdJ/xuU9mtEyx53CXDKgvknA8RTKVzsm9siUrcRdC1AshtgghNgshNhhthUKIp4QQe4zXAqNdCCF+IoTYK4R4WwixejI/gEajySH33ae2o00sAAsv5A9rH+IXC38H3krVHg/lpPv+jJCGT9xuMWE1i5xOqBaNYLmXeh0AHDRSDh8txmK5ny2lXCWlXGPsfxF4Rkq5AHjG2Ae4CFhgbDcDx2bmf41GkxtSCUiEiVu9fO/FDu7fHR9wxeRM3BNYzQK3UepOCJHTzJCdfTFKhhP3aIDZzjAOYrT4Izm5V7ZMxC1zBXCX8f4u4MpB7X+QitcAnxCiYgL30Wg0xzLRAABbOiWheEqt5rS61bGcuWVUXpnBRapzFaIYTaQIRpOHu2Ue/yJ8t4ZTHjyJl+230NLTN+F7jYVsxV0CTwohNgohbjbayqSUrcb7NqDMeF8FNA66tsloG4IQ4mYhxAYhxIaOjo5xDF2j0RwTRP0AvNigrOi+WDLnlrs/nMA3TJWkXKQF6A4pl88Qt0z9y/D6L2HplSQXXkKxCHCws3vC9xoL2Yr7aVLK1SiXy8eEEGcMPiillKgvgKyRUt4upVwjpVxTUlIylks1Gs2xRExZ7tu6odRrJxxPkbLk1nIfnFcmQ4EtzUUdv1MLpiZAf+qBjLgnY/Dop8FXC1f+EsuCcwDo6vFP6D5jJStxl1I2G6/twN+AtcDBjLvFeG03Tm8GagZdXm20aTSa6c7zz6vtaGK4ZaImN9edqKQjjCGUubTcXVYVlSMlxIJ8uee/uCb0J9j24IT6HhB348vjnz+Bzt1wyY/ULxDDxdTt90/oPmNlVHEXQriFEN7Me+BdwFbgYeAG47QbgIeM9w8DHzKiZtYBvYPcNxqNRjOUaC8AC2ZVU12g3DF9acOFksNomWWiDr43G/53DvzyFBZG3yaJCWIT84V396rriz12CLbByz+CJZfDgvPVCVYnAL2B3gndZ6xks4ipDPibMRFhAf4kpfyHEOIN4D4hxE3AAeBa4/zHgIuBvUAY+HDOR63RaCaHH/xAvX7uc0fvnoZbxldQ3J8WIJi2UQE5cctIKfGHE8yiRTXMOhVCnfyl6BOcu+/blBn3H2fnXPTMu0hZllHsvgCe/C6k4nDeNwbOsSnLPRAIIKUcMqk7mYwq7lLK/cDKYdq7gHOHaZfAx3IyOo1Gc3R59FH1ehTFPdbXgx3w+ooGao6mDBdHDiz3SCJFPJWmUBgW+mW3gbuY1id30bfnR5TGgoxbbuMh3PFO3mt5Hh7/JLz1ZzjxJiiaN3COYbmLZETVWXVZh+8rx+gVqhqNZkoJB1QUSUFhwUBagIQJTJaciHt/XhmCqsHhA4xQSJykIhOw3GOqz3ZRBJvvUUJ+xueHnmOIu5MYzUcx1l3nltFoNFNKtK+boHRSmuceVCEppSYic+CW6TFCFb0yqITdrO7hdVjpk07S0eD4OzfE/V7PjdyypA8qjwfPIdF/xoSqkzjN/ghLK/PGf78xoMVdo9FMKYmQnwAuyvLseOzKZdEf654Dyz2TV8ad8oOrsL/d67DQhws5EZ+7ca3JXQgXf334cwzL3SWiR3WVqnbLaDSaAZxOtY1EbzPcfTW0bM7ZLdORAEHposTrGMizHk2C1ZUby93IK+NMBsA5IO4eu4U+nBOLljHE3eb2jXyOVUUAeUyJoyru2nLXaDQDPP74yMfC3fDHq6BzF8w/DypX5eae0V76hIs8h6U/fXswmlBRJjnxuStxt8X9kD+QCcXrsFIvHZji43fLdHV1UgR48gtHPslYbVvmSLFFW+4ajWbakIjSt+NpAndcheypU23R3MVsm+NB4mYvQghMJoHHZiEYS+ZM3A8GophNAkuse4hbJs+hLHdLok8tbBoHz729D4DzVs4b+SSL+iVU4khrt4xGo5ki/ud/1GbQvfNlYt+ehecvV2Pr3Mau025VhTRyKO62ZJCUzdO/78nkfMmRW6a1N0qZ144I9wxxy2QmVAVpVRRkjDR2h9lZrxbflx0phYrJBBYnRbYUzT1a3DUazVTwzDNqM9i3+QXsMsrvKr/BCbFfsS3vTHDk51TcHek+0vb8/n2P3TJoQnXi4n4wEKU6zwSJ0DATqsb8QmzsrpmfPrsHrzBqo9pHiYCxOvFZE7QHY8SSqTHfazxocddoNCOS7G0hLi2854MfI4RT5VHJpbhLiSsdxuQcEHevIyPunpy4ZVp7o8zzGIUyBom7y2YmajISlI1R3Lv6YjzwZjNrK8zqF4Z5lOlLq4t8i8pAebD36BTt0OKu0WhGxNTXRpepEI/DisNqyrm4h/oCWEUKi8vX3+bJ1By1upS1PQGklLT1Rql1GII6yC0jhMDi9KqdMYZD7usIkUpLZnvTo1vtADYXLqHG0BXS4q7RaKYYR6SdPmsxQgiKPXa6+uKGuPtz0n9XV6e6j6egv82bQ7dMMJYkHE9RYzf6cRUNOW7PhDCO0XJv6Fb95YkI2L2jX2B14kCJeiZ6Z7LR4q7RaAYoKlKbQV6yk5izVB3y2Onoi4EjL2eWu79bFepxegfE3WO3qFBIqxuSEUiP30d9sFf5xMusGXEfGrLo8BjuoPjYYt0busOYBDjT4SzF3YVdGpZ739ERdx3nrtFoBnjggf63vZEExbKHPq+KDS/x2Gj2R6Eid26ZXn8XoJKGZeiPlrENKtiRjYAOQ6sh7iUmw73jHCrubq+xP0bLvbE7TEW+U8XIO7Jwy1hdWMM9gLbcNRrNFNPc1o5XRLD5KgEMt0zG5x6AdHrC9wgZScPyfcX9bR67hVA8RdqaKbU3ftdMW0CJu08Y4n2I5e71qV8M8dDYvqwausPUFrqUrz5Lt4wpFcVqFnSHclOUezS0uGs0mgG+9CW1AZ1tBwBwF1cDUOSx0RWKk7bnARImsLIzQzSorFl3/tAQRYCYcKiGCUyqZtwynnRARd9Y7EOO5/vUfUPGOLJlQNyDWU6ouhGJMAUuW38is8lGu2U0Gs0Ar77a/zbQ3gSAr2wWoCz3VFoSNnnwgHLNOPKH6SR74n1KVIVjaJw7QAS7ikKfgOXeGohS6LZhifYcZrUDlPjySEgzkaCfgmGuH45IPEVHMEZNoRP2ZCnuVifEwxS6bXRrt4xGo5lKIt1K3D3Fqq5ppgB0UBrukhz43ZNhv3ozyG/tdajMkGEylvv4xf1gb5TyPIfKi+M8XNxL85z04SSeGUcWNPao8dQUOAzLPbsJVRKRo2q5a3HXaDTDkuo1Sh97ywHllgHoThurOnMxqRoLkMLcnzkRGMgMKTPVmMaftbG1N0p5vgMi3cNa7mV5dvqkk0Q4exdTQ5cS99leCcisJ1RJhCl0WbXlrtFophZLqI2ocPRbpiWG5d6ZzJ3lbooHiJk9KhWkQcYt09dfam8ClnvAEPdw12Ex7gD5Tish4RxTTvdMjHut2wjRzHJCFZmixCX6LXcpJdHE5KUi0OKu0WgGqK6G6mrSaYkz1kHIVtIvvBm3THvcmJScoLjHk2kcqT7i1qHi6B1UJBsYt1smlkzRFYobbpmeYd0yQgjiZhdiDKGQDd1h3DYzPpORBCxbtwxQ4kzjjyRIpSW9kQSLv/YP/vhqfdb3Hgt6QlWj0Qxw990AdASiFNNDwlXWfyjfacVsErTlSNz9kTheIqRsQ8UxY7n3poz7jDO/THtALRqq8Joh1jusWwYgYfHgSWT/WRq7w9QUuhAZd5E9i0llI6d7iS2JlGoNQZPhuy/Nc2R977GgLXeNRnMYjd1hyuhB5JX3t5lMgiK3jZaImvCcqLj3hBLkiRDykGiTjM+9N2XcZ5zinlnAVG3PJA073C0DIG1ebKns79EfBhk1XDljsNwL7coN0x2KD7h3Cl0jXjYRtLhrNJoBPvUp+NSnaOwOUS66sfmqhhwu9thpD6VUzHh0bMm2DqU7pCx3cUg4pdumxL0nYYj7ON0ymQVMlTbjeucIwY52L/Z0dveQUg5dwGRcPyoZcbeqzJA94QFxr9HirtFoDqNtC9x+Fjz+xdz0t3kzbN5MZ0c7DpHAU1I95HCRx0ZnKJ6TzJA94Th5IoTZNVTczSah8svEAbNt3JZ7ZgFTsdlwn4zgljE783DLCJH46JObHcEYsWSa2iLXQMqCrKJlVISRz6JWp3aH4jR2hyly2/rdULlGi7tGM1PZeCf85hxo2QQHXs5p13F/CwDW/Moh7SUeO53BWE4yQ3aH4uQRxjYo3W+G/uRhNve4Lff9nSG8DguupPElNIJbxubKwyVitPceHnIZTaSo6xz4cjkw2Noeh+WeyemecctMltUOWtw1mpmJlPDk16DqBFh8KfS157b7QCbGvWJIe7HXTmdfDJkDyz0YCOARUWy+8sOOeTIFO6zjr6O68UA3q2sLEBGVv2a4aBkAh8cHQFd392HH/vvR7Vx820v9IYvbmtVnXlzuHbDcbVmIuzGh6jUPWO797p1JQou7RjMTCXdBLMA239lQugRCHRNKjXso5tBB9cY7VHiL3DZiyTQp28TT/iYC6h6WvGHE3W5RBTtsrnGJe284we6DfayZVQC9zSBM4Ckb9lyXkW64p6dzSHt7MMr9G5qIJFJsNUR9S3OAYo9NhVfGgkrYTVnIqGG529JRXDYzHcEYLf6oFneNRjOUVNd+AH66KUEHPpBpCHUe+aJsWLgQFi7EETV+CRwi7plY96jZM2FxTwfb1JthRLe/1N44i2RvbFBW+JrZheBvAG8lWGzDnuvNV+Le6/cPaf/9P+tJGJkvNzeqY1ube1lWlY8QQk0oZ5uK2PC5YyQP29bSSyottbhrNJqh+Jt2AbAnWcpvNxmWbV/bxDu+/Xa4/XZssW5iJudATnWDYq8S95BwT1jcRUgV6sBdctgxj31QTvcsV6hKKUmlJQBv1PdgMQlW1figtxF8NSNe5zYs977AgFsmGE1w92sHuHhZBVU+J5sa/UTiKfa0B1leZUwAxwLZTabCQHqFhEoetsX4JTAtfO5CCLMQYpMQ4lFjf44Q4nUhxF4hxF+EEDaj3W7s7zWOz56ksWs071j62vaSloKzT1rDG52GRRo8mJO+02mJK+knYvUddqzIre4VxK3EbQI53a0RQ9yHsdw9/aX23Fnnlvm3P27ko3dvBGBjfQ/HVeXjtJmV5e6rHfE6YVjfocBA2t8/r28kGE3y0TPnsarGx+YGP9tbA6QlLOsX9yyThsEQcS9w24gm1HOrLZoG4g7cAuwYtP894MdSyvlAD3CT0X4T0GO0/9g4T6PR5JBk5z7aKOAjZy+lomq2auzLgbjffDOBf/84BQRJ2A+PCy8xLHe/dClX0ASSejlinaQR4C4+7Fie04o/nECOwS2zoy3AU9sP8tT2g2xu8it/eyoJgRbIH9lyzwh0OOjvb9q7awtXFTWxvDqfVTU+mv0Rnt+lXFVDLPdsxd1iV37/RIRCl4rft5qF8t1PElmJuxCiGrgE+K2xL4BzgPuNU+4CrjTeX2HsYxw/1zhfo9HkCGvgAE2UUeq1k1diLDTKhVtm9246D7RSKAKkHIeHDha5bQgB3UlDlCbgmnEluolYfGC2Hnas0uckkkgRNzmydstkapN+7q9vEU+mOXF2AQSaQaaOaLlnBDoW8vc3nd/+O74T/jpEA6yq9QHw5zcaKXLbqMg3Pnu2hTpA5eexqoLfBcavn+oCF2bT5Eljtpb7rcDngcxvsCLAL6VMGvtNQGYpWxXQCGAc7zXOH4IQ4mYhxAYhxIaOjo7xjV6jeYeSF2mix16FySQoKfTRK12kAjkQd6Db6qRQBBHuw8XdYjZR5LbTnphYfplYMkVBuoeoffjwxOoCNQHZl7ZnVYkpHE8SjqdYVeOjN6LCDU+YVaj87ZCVuKeiQeLJNFJK7PEeHDIKW+5jWWU+ZpOgIxgbmEyFsU2ogppUTYQpdClxn0x/O2Qh7kKIS4F2KeXGXN5YSnm7lHKNlHJNScnhEyoajWYE4iF8qW4iHiVYlfkO2mUBMX9rTrrvtrgoJIjZM/zfZanXTltsYuLuDycoEX7ijuHvkRF3fzK7FaoZq/19a2tZO6eQhWUe5ULyN6gTshB3j4zQ2huhoy+GVxrupjd+h9NqUnHtDHLJgLLcx1KJKlOww7Dcawud2V87DrKx3E8FLhdC1AN/RrljbgN8QojMutlqoNl43wzUABjH84GuHI5Zo3lHk+qqU28K5wBQ4XPSIfNzZrn7bTZcIoY9bwRxz7PTFDUmcccp7t2hOCX0knaXDnu82qes2p6EBVJx5TsH6OuAW1dAy+Yh53f2qeRgxV4bv7vxRP70r+uMD2NY7vlD0ygMwWQmZXbiERGaeyI0dkfII0TKbIf2bdD0hoq6YdBkajqlflGMyXJ3QSLUPyk9mWGQkIW4Sym/JKWsllLOBq4DnpVSvh94DrjGOO0G4CHj/cPGPsbxZ6WUMqej1mjewfQYYZDO0vkAVOQ7aMeHKRerVFetIlyjBNfpG37RT6nXTkNoYpkhe/piFItexAgLi/KcFrx2C53xTGZIYzVo21vgPwC7nxhyfsZyL/bY8dgt/fH4+BvAU35YYezDsHvxEqapJ0JTT5h8ESI07xK1SGnD7zh9QQl2i4nVhv99TKkHMlidkIhQZIxtysX9CHwB+IwQYi/Kp36H0X4HUGS0fwbIUUYjjUYDEGjdDUBh9SLAEHdZgC3aodISTIRbbyWxZjkAFs/hUSwApV4HB0LGj/Zxinsg0INTxLHmDy/uQgiqCpzUx40Jy17DMdBzQL22bBpyfsZyzwhnP/4DR3bJGJjchRSIEE09YRq7QuQRxlk8C1ZeB2/dywUvXc3bF9VRakQLEVC5d0bMNDkcRsz+CbMK+NaVyzhn8fCfPVeMKR2ZlPJ54Hnj/X5g7TDnRIH35GBsGo1mGBId++mVLmqqVAyD12ElYC7Amo4aC2vG4AcehmRfZnHRCOKeZ6d3gkWyoz1KHO2+ihHPqS5wsbXDmNTtqYPyZUqsAVreVF9kxuRml1G6LuPy6Ke3EarWjDoe4Sqi1NLDEz0R3ESxihS4C+DEz0PhXMSW+7A/9SUoWwjzz4Mt96vQxnnnZP+hrU7oa8dsEnxg3azsrxsneoWqRjPDsPQeoJHyASsSSLoM3/VEXTMf+ADseku9HyGLYqnXThILKYtr3OKeNPLKuAorRzynusDJxqBP7XQb8wwZy73v4ID1jLLcvXYLDqt5oIN0Cnqbjrg6tR9nAcXmPpr8Ebq7jWfo8KncNif/B/zLk+AuhddvV/2+/ReYd+5h6RmOiDGherTQ4q7RzDA84Ua6bJWYBsdIZ0QmOMFJ1aYmzBhx5SOIe4lXxXknrF5Vvm4cpINKQIdLGpahusBJa8xO2lEA3SqXDv4DA9kdB7lmuvriFHkOsdqDbZBOZuWWwVVEvuyjuSdCn9/I0eP0DRy32GDNh2HPk/DmH1T8/KrrR+93MOPMkzNetLhrNDOJVJKi5EHC7qHWqD2TNjcHq1StlhgpzMpyHYbML4aY2Ttuy90UMqxjz/DRMjAQDhn11iq3DCjLfeGFYLIo14xBZ19sGH+7EQaZn424F+JO99LaGybeZ+SYOdS9dcKHwWSGxz+v6qYuumT0fgdjxLkfLbS4azQziFRfOxZSiENC+5yFaj8ZmFisexpwmSJErXkjprLNpCAImcafPMwa6SCFacQc66B87gC9jmrllokFIdINJQtVmuPmAXHv6otTfKjlns0CpgyuIswyhVtGBmLcD/1yy6uAJZer0MxlV4F1jKkDbNoto9EcG/Q2wX03DBR1yEWX7UqwHAVDfdWFRSXEpJVQV/Nwl2VNwOzAZ+ojbhs5CsRhNZPvtKrkYeMUd0e8i4C54Ii50Kt8ynI/aK5Qz7JrrzrgmwWVq5VbxogO6godYrnHw7DnKeP8bHzu6kumQATJF8aiqcFumQynfFwVEDnhw6P3eSgZt8wEkq2NBS3uGs1ksecp2P73w8L2JkKwowkAe8HQwtWVPhcd5BP3T8zn3nXiKRSK4LB5ZQZT6rXjTzvHLe6eRBd91iPfw+ey4raZOSBLVX6Y+n+qAwWzoGq1KvPXU0cqLekOxSnORMrsfwF+uhq23AfHf2Agl/qRMOYXCgiShyHuw7mlqk6Ar7RA5aqsPucQMuNIRsd+7TiYnMqsGs07HCklIjMJ2Dsxa3ow4W4l7t5DCldX+By0Sx/VE5xQ7f7oxyn83Z+RriOH6pXm2enqckJqfOKen+ohZj9ynLcQguoCF7vixkrZ/c+rV99s5XMHeOjjJB3FnMgKir3HqbYX/heEGT78OMw6JbsBGcWzC0x9+EwRJAKRbVKwbLEaufET4f6ye5OJttw1mhzz+JZWTvr2MyQ6DXEPNOWs72SvEu+CsqHiXpnvpEP6MIfGEQo5aOFTV1+MQhHANEJemQylXgftcYey3EdZOJVOS3a1DbimookURfhJOEfPKVVd4OTtsOEiOvBPsHmUEJcuhTlnQl871n1P8BHLYxS57crl0boZFl2UvbBDv+U+2xGl0h5FOPKzK583FgZVYzoaaHHXaHLM5kY/7cEY4Ta1kjSXlrsMttElvZTkD1327rSZaTeX4Y00jc2ne+el8ORX+3e7f3gbPvpGzCuTodRrpy1uP3JO93A37HyMp3cc5IJbX+SNfQfh5R9z4MmfUUTvERcwZagqcLKl1wEWhxJF3yy1cMls5f5lv6D9hpfpqL2E1abdFLmt0LVHjadqdfbPAPpXmi7MS1DliA3vb58omdDSQfH5k4kWd40mx7T0RgGJI2iE4gVyJ+6WcDvdogCb5fA/3R7XbGwylv39UgloeBXeuEMJMRBKRjALiSN/5BBFgNI8Bz3pUVapPv8d+PP1NDWp57D++Ufg6W+w6I2vYxMpauYuGnWI1QVOeqNpUj7DTVSgXus7Q3zur2/x8+f20pq3nCIRpCLdOjC/UXn8qH0PweEDYeLqJU7WlJkmvMp3WKpOUK9NG3Lf9zBocddockyrP0IJfuzSmDjLoeXujHYQsAw/ERnxqiyRdO3JrjN/g1rkk4zAprtVHw61MMoyqlvGTuBIKQjSKdiucgmmDm4HoKderXy9Kvlt7ljwC6yrPzDqECuNiJmI2whnNER+fZ36Mnpq+0H2O5YBUOx/S4m71QXFC0ftewgmEzgLsMf9WGK9I8b4TwhvmQrLbFqf+76HQYu7RpNjWnujrPP5Aehz1+bU5+5JdhG2D5/zRRapLJF07s2qL2mEFnZJL+F//ordrX7eKDQWQw1TqGMwpV47AY4g7g2v9S+osnfvotBtY65spFfksSk5ixPPvGT0TI1ARb4S9x67ER1kWO6vG+Le0hvl4WYvQenE1bZRiXvFSrXYaKy4iiDcpaJwJsMtA1B9IjS+MTl9H4IWd40mh6TSkrZAlIsq1WKVzeZlSvxi46812k86TUG6h4RzeJeJt7iaoHSSbN+VVXexg2pO4EfJ9+AKN3Pnnb+mXPrVwRFSD2QozXMQPJLlvu1vYHGCPY+Cvn2cPK+I4x1t7ExVsqDUO7ToxRGo9KmFQm0Wwz+fsdzruzhxdgEmAS/s7WK7aSGi4VVofVvFwI8HZ6FyT0X8k2O5A1SvhWCLitufZLS4azQ5pD0YJZWW1NBGCjOP+g1fcQ787qlQFxZSpD3D52Op8DnZJyuIH8xO3KMH99ArXRSd8RFaZSHXR+7lEy5jmb9r+F8HGY5ouWdcMgvOR5YdR2W8nmqfg3k0sjtdzdUnVJNtWeVSrwOzSbDDslSJb8UKWvyqoMaFyypYM6sQKWGvfSl07FAuprH62zO4CiHSoz7PZFnuNSeq16bJt961uGs0OaTFr/zspYkWYp5q9icMkcyBpebvUBOTlvzho0wq8p3sl5WYu7Nzy9C5lzpZzsnzy/Gf+S2Wm+qodbyqjo1iubvtFtI2w/o+VNwbXoVQOxx3FRHfQhaIRha5gtiSfcxZeuKY0t2aTYIyr53NyVr4Qh3kV/NGvXLJnDSnkPOXqlj5Zs/ygYsmIu6BZkjFJs9yL1uuIn+OgmtGi7tGk0Nae5U7Ji/aiKloLi2Z2vA5sNyD7aqPQ1MPZKj0OdifrsAebs2q7qi1t446WUFZnp0l57wfzvqyEmqbJ6u8Kb4C47MdKu67HlcCtvACOpxzyRNhloZVhMhpp5yGxz62tZMVPiet/oFVna/XdeOxW1hSkdcv7v7CFeqgPQ8K546p/36chuUOk2e5W2xQseqoTKpqcddocogSIYkjcAB76XyC1hIkIicRMxFjdaqnePh6oGV5DvZLw6rv2nfkzhJRXJFW6tPllOYZQn7m56GjBDqzk4XKonwi2IeIezieZOf2zbSYK8HmpsGiolxqWp9UJ5QsyarvwVTkO/q/NEFFyqyZXYDZJJhd7Ob6tTWctny+soqr14x/8dHgXyuTZbmDcs20vgXJ2OTdAy3uGk1OaemNUGULI2IBRNE8akt89JoKchIxk/CrjI++suHF3WE10+UwQgZHC4fsqUMgabFUDVjSQsC2angzO8u3tshFQLqQET8Auw8GueynL5PoaWJnOI+DgSg7k2qsrqaXwF0yahTOcFT5nLT0RpFS0tUXY297H2vnDGST/M67V3Dx8gq4/k9wxS/G3H8/rkEZKicjzj1D9VqVWbL17cm7B1rcNZqc0uqPstpj/LQvmMP8Ug8tsnD8lnvbVvjBIujcQzrYRkC6KCnwjXh6wjeHNGL0cEgjDDLoHi4dbnaTnTWFLnqli1if+rzfeHgbPeEE8+29tMlCXtvfxZ6QnS7yETIFJYuz6vdQKvIdxJNpukJxNjf6ATihdpislb5alZZ3vAy23CfLLQNQcxKs+sCk55fR4q7R5JDW3ghLHUYln8K5zC/1cCBZQHq84r7vGehrg413Ygm30ykKsVtGjuEu8vloFyWjW+6G2yaeN07/NFBb6CKAm1jIj5SSbS0BLl1agDPRQ5elhFf3ddHYHaHZOltdUDp2lwwonzuoL863mnoxCVhePQmW9eDc8pPplvGWwZU/h7LjJu8eaHHXaHJKsz/KInObykpYMIt5JW5aZREy0Dxqgq1hySynf/s+XNE2ei0jF7cAqMx3sE9WQOdo4r6XbvLx+o7c35GoLVRumVTYT1sgSm8kwep8lRTLXTyLV/d30eQP0+Oepy4Yp7hXGguZWnojvN3kZ0GpF5dtEhLaDrHcR85nP1PQ4q7R5IhYMkVnX4xZqQNQNB8sdsMtU4Q5ERpf7vPmN5VFGWpnbmwnYduR0wJU+JzsTpYju/Yc8ctEdu9jvyynLO+QqJhLL1VbFlT5nARwYYr1sqM1AMBit8r+WF47nwNdYZp6IkQLjVQA45hMBZXOGFRah7ebelkxGVY7HD2f+1FCi7tGkyMO9qroh9Lo/n4rdVaRm3YxznDIcLcqCL3u3/stydHS5FbkO6iX5Yh4CEIdw5+UTiM797A/Vd5fD7Wfz31ObVlgs5hI2fKwJILsaFWiPsvqB2DhAuVflxIC866Ad/0/qFmbVb+HUuS2YbOYeONAD92h+OSJu8MHCBVOOZ70BdMMLe4aTY5o6Y3gJIon3KTyjQNWswmZZ+RFGavfPeOSqTkJuewaAKTnyAUuKvKdqnIRqLqjhxLthT9fjynUwXq5+HDLfYyYnD6cqSA7WwNU+Zw4wyqiZ+7c+fhcVgDKS0tUebpxCqYQgsp8B8/tVLnqV1T7JjTmETFblMU+mf72o4gWd40mR7T2RpgvWhDIIf5lV7ERkTJGyz2VKQBdsZKehdeSloJ0wZEnQCvyHTRmxL2nfujBRATuuAD2Ps2eE7/J/akzDhf3s85SW5bY3AWYSVPf2s6SCq/6jM5CTHY3JxnhijUFE48Kqch3Eo6nsJoFiyu8o18wXlxF4Jz5LhnQ4q7R5IwWf5RFJlXAOmO5A5SU15KWglRv9kUa/rG1jWee+Qd+Zy1+6eKjz6Y4I/5jfKuO7A8vz3fQRKlaONUz1HKXrW9Dxw6Sl/yEzWVXA4KyvNEzMx4JZ54S8O6uDhaX56lfJ/nql8oVq6pYUOrpT9s7ETJ+9yUVeUeMFpoweZUwQu6emYauoarR5IgWf4Tl1hYw26FwTn/73HIfHeTj7DhAtlU532zo4Uaxnxf6FvOl7z5LIpXmB+99FyfMPnJCL6vZRL7Hg58SCg5xy2x56w1WAH9pr8LvMOYHvBNzy3h9aj7BLUPKot7fDPk1AFy8vEItLsoBmYiZSfO3Z7ji5yCODZv32PgUGs00oNkfYZm1GUoWDfEvzy/10CYLifVk75bpbmukUnRTu+xUagpc3PnhtVyxqiqrayt8TtpE+RC3TDCaYPOmN4hJC7/dkqStN0qew4LTNjEruKBQTfDmEVKWe2DAcs8lGct90vztGQpmga9mcu9xlNCWu0aTI5p6IsxNN0DpeUPa5xS7eVUWUN3XlnVfzs4tABy/7hyeGEuhZ1Sse72/lCU9W/rbfvjkbk5LNOB311LXHSO4tW3Ck6kAxcXKv19kiTA7D5V4K2/4xGYTYWW1jzyHhZPnjj19wTuVUS13IYRDCLFeCPGWEGKbEOKbRvscIcTrQoi9Qoi/CCFsRrvd2N9rHJ89yZ9Bo5lypJQEetopSHUetljH67Dit5bgjGQn7lJKyvu2kcYE5SvGPJaKfCe7YsWqElI8xP6OPv7waj2rXB0Uz15GgctKZ19seHG/9lq1ZUl+gXITLcxPY+lTkTLkDZ/7ZiIsq8rn7W9cQE3h5C7ZP5bIxi0TA86RUq4EVgEXCiHWAd8DfiylnA/0ADcZ598E9BjtPzbO02imBW/Ud9MTiue8365QnNqkURB70GRqhqS7HFe6L6tUvO3BGMvlbvye+WD3jHkss4pc7EsavvmeA2w80INFJiiKt2AuXcTVq5X4lg43mfof/6G2LBFGDpYTy80DOesnwS2jGTujirtUZGqEWY1NAucA9xvtdwFXGu+vMPYxjp8rsi27otFMIrFkivf/5nU+9qc3keNJBXAEmnoiLDIZ4jbMMntzRvACraP2Vd8RZKVpH9Gy8RWdmF/q4YA04uF76tjb0cc8c4dK3lW8kOvWqtDMivxhLPdwWG3ZYldTxGfU2AZCPfO0uE8HsppQFUKYhRCbgXbgKWAf4JdSJo1TmoDMv2gV0AhgHO8FDnOUCSFuFkJsEEJs6OgYYSWdRpNDGrvDxFNpXtnXxd83Tzy/+mCaeyIsFI2krB7IP9wt4SpWk3R9nQ2j9tXduIN8EcY+68RxjWWIuHfXsa+9j3X5RjKz4gXML/Xw6w+ewAfXzT784osvVlu2WGyqVmrUDwEj1HMSfO6asZOVuEspU1LKVUA1sBYYX+7OoX3eLqVcI6VcU1Jy5CXVGk0uqOtUFmmxx863Ht1BbziRs76besIsMjUhS5eovOiHUFA+G4CulvpR+5JGCbb8BWObSM1Q6rWTtucTMXuU5d7exyqnYUAVLQDgguPKKR/Och8PjnxVfKJxvcrZbplY7LwmN4wpFFJK6QeeA04GfEKITLRNNZAxhZqBGgDjeD7QlYvBajQToa5TeRd/ev3xzI9u4d7Hn8lZ303dStwt5cOnca2oUStL+zoOjNqXp3MzIZxYysZnQwkhmFfqpc1UTqqrjobuMAtMLcpdMg4f/qiUL4P6l2DvU1AwZ/TzNUeFbKJlSoQQPuO9Ezgf2IES+WuM024AHjLeP2zsYxx/Vubawak59jm4De59H4RyZxfUdYYodNs4uTjCH63f4ZydX89Z34GuZgoIDjuZClBdVkxAuohnEete2beNevuiCSWvmlfiYV+yhFTXftISKhJNULxg3P0dkff9FT61FT70EFz9m8m5h2bMZGO5VwDPCSHeBt4AnpJSPgp8AfiMEGIvyqd+h3H+HUCR0f4Z4Iu5H7bmmMbfAHdfDbv+D1o35azbus4Qc4rd8Oy3sBFnYWIX8uD2nPTt7N6p3oyQs9xuMdNpKsYUPPKEqoyHmJWqp9M39hDIwcwv9bAnUYw50IiZFHmhOiheOKE+R8RkUgt/5p4FBbMn5x6aMTPqIiYp5dvAYdP2Usr9KP/7oe1R4D05GZ3mnUe4Wwl7yJgADPfkrOu6zhDXVXfDW39mT9WVzGp6hOTrd+K6/H8n1K+UEl/fPmUqjWC5A4TsJTiiB4c91vLwf+M+8CzW4y7FRYrYOCNlMswv9fBIehZmmeQh29cwJ/qyE/cbb5zQfTXTB51+QDO9ePH7qgTcNb9T+5HunHQbiiU5GIhxTfdvwFlA96nf4On0CVi3/RWS44x7f/BmeOa/6QknmJNuIGIrBPfIuV8S7goKkp2HhWGur+smtvFP5HdtwvXi/wDgmH3S+MZkoMR9HV9LfoRisxFbX7Zs9AtvvFEL/DGCTj+gmT6Eu2HjXbD8PbDICMcL58bnXt8VIo8QNf71cNaXmFNTwS9TZ3FxbD3segyOu3JsHaYSsO3vYLHTPO9mFpkaCfsWcqT8h5b8Soo6/bT3hijzqYnNrc29fObOZ3lZtHK3/Tpa+9IUiiBnVg9XuDp7agqcWM0W/pg8h5a57+aOCxxQdcLoF3ZmQiaPnKBMM/3Rlrtm+rD+N5AIwam3DBROCOfGcq/rDLFQGOl4K1dT4rHzlm01fmspvPXnsXfYuRtSMYgFSOx8ggWiadQaoc7iWsxC0tRQ39/2tYe2staqilWfe/F7+KP1Gr6V+iDVE8yBbjGb1PwCMKe0IDthB7jmGrVpZjxa3DXTg3gIXv8VLLwQygy/tbMwZ26Z+s4QizO51suWqnDBsnw2m1dA6+axd9j6tno126ne9mvcIoazavkRLymomAVAZ4tKxSulZM/BPq4oagZhpmLJydxx44l88cLFOKwTz1k+r9RtvE5C+KNm2qPFXTM92PwniHTzWuWHuPDWFwnFkqoqTo4s9/2dIVbbW8Ce3788fkGZh82xMgi2jr14detbYHUhV72f0uA2AFzVRxb3wjIl7t76f8Dz36O3biN9sSQL4jtUrLjNzYmzC/m3M+eN/QMOw/wSJerztbi/I9Hirpke7HiEROFCbn7ews62IDvbAqoafY4s97rOEEstTepXgbGCdH6pl60xo+pO556s+4omUnTtXc9e02z+ZdOgsnclR150ZPLVkEZwStvd8Py3sT/yMSwk1ZdD9fiKRx+JsxaXsrwqn6UV2ZYI0RxLaHHXTD2xIPLAKzweX0kkkQJgb3ufcsvkKBSyvqOPWckDQ0IVF5R62CeNPCgdu7LqR0rJu3/+MrbO7WxKzKJ0yWn0OSpIeqvAMYqIugr5aeX3+JTz/8FlP8HZs5NPWh7EkgxBTe7FfXVtAY984jTcdh038U5E/6trpp79zyPSCe7pWsw3rjyObz6yXYl7jix3fziOM9KG09E34M9HuWUaZCkpYcXcmZ24dwRjhA/uwWuPcPUlF/GeE46H3bdBLJjV9bHaM3m0fj8/WHE+vc/cysdDxsLu6vElCcs5//7vUz0CTY7Qlrtm6tnzJGHhoqdoFe9bW8vcYjf7OkLKco/3jT8O3WBnW3BQ4eqB3C/leQ6cdjud9mro2J3dUNv7OE6o/DCmylWqceEFsDy7CJM5xW6SaUlTb5xHC2/AJKRKtjVdVna+971q08x4tLhrphYpkXue4sX0ck5ZWGEkvfIYlnuBOmeC1vv2lgCLM2GQg8IVhRDML/VQRxVkY7m3vk1dcxvLTHVIkxVKjhz6OBxzS1QES11niP9LrGGPdRHMO3fYTJJTQmOj2jQzHi3umqnl4FZEsJVnkis5bb5aODO/xENjT5i4zafOmWDEzI7WACtsTZBfA0bloAyLy728FS1D9tRDIjpyJ5174NdncNnLV3KRZYP6krDYxjyWOcUqcmV/Z4gDPVF+O/8XcOUvxtzPpPHBD6pNM+PR4q6ZWvY8CcBLchUnzS0EVFy2lNAaNxbyTHCV6vbWAMdZmofN+7K0Mo9t8QqETEP3vpE72fogAL3SxWxaERUrxzWWApeVfKeVHa0BDgZiVBXnTyj7o0YzElrcNVOHlLDlfnZbFlJVMwevwwoMxGfXhY2iDxNwyyRSaeoO+qlKNg6ZTM2wpCKPvdlEzGx7EGpP5pr097iv+itwxn+OazxCCOaWuHlxtyqeMatIF3zWTA5a3DVTR+N6aN/OnZHTOXX+QC6TuSVuhIA9QcPtMQG3zL6OPpak92CWSRjG2l5c7mWfrEQiVEqB4Ti4HTp20rfgMtrDaQKLroGCWeMe05xiN+3BGAC1hVrcNZODFnfN1LHx9yQtLh5KndLvbwdwWM3UFLjY5jcidcdquUd7YdM9kE6zozXAeeY3kSYLzDvnsFO9DivlRT66LOUji/u2B0GY2FWorl9Q5h3beA5hrpHzBWBWkfsIZ2o040fHuWuy5untB1lY5qV23z0qfG+smRQHE+mBbX9jY/6FyLiHVTW+IYfnl3rY2RlRxZfHarlvfQAe/TS4CtneMotrzZuQtacgHPnDnr6kPI899ZUUD+OWSSRTRDbch7v2VLYHVc7HBRNczp+ZVPXaLRS4rBPqK+d89rNTPQJNjtCWuyYr6jtD3PzHDfz26U3wxFfgjd+Or6NoL7RthVd+BskoP+w6hbMXl2KzDP2vOL/Uw/7OENJVqL4IxkL3fvW6/nY6G3ezQDRhWnTRiKcvrczjldhcOLhV5ZIfxLMvPEde+ACPpU9m78EgbpuZigkWls5ka6wtciGmSwhkhssuU5tmxqMtd01W3P6SqsVZ3PikSnXbN3xFoZFo641iNUHRHaeDXy0C6i5YyfrWav609vDc5fNK3MSTaRK2fGxjtdy7VdZF9j3LycKwjBddOOLpSyry+HLybD5t+zum138NF6vKTFJK6jb8A4Dv7q2Gjnbml3knLMizi5WffVpOpu4yfr0sWjS149BMGG25a0alPRjl/o1NmE2CtcGnVGMwe3GPxFOc/YPnufLbfwL/AbZXXwfvf4DPW7/E7CIX6+YWHXbNonKVpyUg8rL2uf9zbyf/fvdG0t31ULkaabZxrXwCv3suFM4d8bqllXl04KOu/ALYfE9/hsj1dd3M6ttM0FFJwlNFsz8yYZcMgMtm4ZIVFZy3pGzCfeWcf/s3tWlmPFrcNaNy5z/rSaTSfOpEF2vZTspRALFeSESyuv5Ad4hIIsXNc1S8+mf3reTbe6p4uiHNdWtrMZkOt4SXVHixmU10pNxZ+9yf2n6Qx7e2kujcD9Unstl7FgCJee864nWV+Q7ynVYec12p0h1sutv43HWcZN6Jc8EZfP0yFUa5aIKTqRl+/r7VvHt1dU760miGQ4u75ogEowkefG0Hly0t5CrLK5iEpHHudcbBtqz6qO8MA3CxrwFp81C5cDW3v7gfq1lwzQnDC5zdYmZJZR6NUUfWlnuzP0IhQezpMA812PjqwTOJmj2UnPz+I14nhGBJhZeneyuh9mR4/Vc0dwfZu+NNCglimXMqlyyv4PYPnsB719ZkNRaNZqrRPnfNEXngn9t4Wn4U9/44mKxsTC+gzbKU2aD87oVzRu3jQJcq0Ozr2oSoOoGfv+9EPnHvJmoLXRR77CNed3yNj/0bbEhzDyKdBtORbZHmngiX1MSgAx5usFGycC3WG5pgmF8Gh3JcZT53v3aA5PUfw/LXD7DvuT9yotirDs46FSEE7zqufNR+NJrpgrbcNSMSS6bY9OpTeEQUseI6xJLL+J3tfewIGhOB2VruXWGqXCnM7dug5iQcVjO/+dAavnbp4StGB7Oqxkd7yqNSA0T9o96npTfCCpey8pctW8Vt7z0ecxbCnrlXLJlmZ/5pULKEBbtv50zbTqSn/Ij+eo1muqItd82IPLSphbnR7UirCXHx98HuIXDH62zyt6sTsoyYaegOcW5eI/jTYypKsarGxwvSmMCM9Kj87iMQiiXxhxPUCrWs/9PXnAe27GPIM3H2m5oCLDv9M1Q8+K+UcgAx+6rpk7HxaPDVr071CDQ5QlvummFJpyW/enEfpzvqVMItuxLZhWVe3uw0qRWfY/C5n2Q1Ys+r12Q9hllFLpJ2n9oZZVK1xa8md8vTreApB9vYwgyrC5wUuW1sbvDjn3spB9KlmEnDrFPG1M+M57zz1KaZ8Whx1wzL28291HUEWc4exCBre2GZh0gCUs6SkS33VBL2Pg3pFLFkipbeCEuSO1SNUWdB1mMQQlBSVqF2hptU7W2Gupegu46WblUJqTDWnNU8wHD3WlXjY3NjD5ub+/hp6iqkMMGcs8bc14xm82a1aWY8Wtw1w7L7YJAFohlrsm9I8eZMXpWQrWhky33PE3D31fDC92jqieCQUar7toyrlFxljfJ3xxo2Dj0gJdzzHrjrUvjJKo77x3swkcYVahp3VaNVNT72dYR4aU8nD6TPJPTxrVA8f1x9zVg+9Sm1aWY8WtyPFVIJaNkMe55WwjdB9rX3sdZsRIsMstwzi3i6RcHIlnsmR8sL3yOy6T7usn1PfUksu3rM45g/fxFPpU7AvP6XEPEPHGh+E9q3wam3wBmfp7h3K++xvISprxUKxm65A6yq9QHw1w2NLCj14CmqGlc/Gs10QIv7scCep+G7tXD7mXDP1f0FMCbC3vY+TnfWgatoSLSI12GlyuekJZU3suXetQ9cxVC2nGWvfJrjxV76Lv0NzDt7zONYUe3jx8mrscQD8NqgikWb/qCSip3+OTjrS7Ta5/BFy70I5Lgt9xXVPgAC0eRhicw0mpnGqOIuhKgRQjwnhNguhNgmhLjFaC8UQjwlhNhjvBYY7UII8RMhxF4hxNtCiNWT/SHe8az/NTh8cPUdkF8LL35/7NZ7oFWtzDSu29fRx0p2K1fKIdEia2YXsLXXCeFO9YvhULr3QckiuPYuDrhX8Gk+i2f12K12gEK3jS7vIt7yngmv/kJNrMbDsOUBlZXSkQcmE/c5rqWAgLponOKe77Qyz6hxuqom+7kBjWY6ko3lngQ+K6VcCqwDPiaEWAp8EXhGSrkAeMbYB7gIWGBsNwO/zPmoNQPEQ7D/BVh6BSy/Bk67BZregLoXx9bP89+Bhz4GjeuJJlL4u9spTzQM6yd/19JyGhIq9wt97Yf31bUXiuZB0Tz+q/iH1BefMaFkW4vL8/iluFalBrj7auRLP4R4kC/WreAL978NwP3RE2m3GatdxzGhmiEj6itrhk8PrNHMFEYVdyllq5TyTeN9ENgBVAFXAHcZp90FXGm8vwL4g1S8BviEEBW5HrhGZS0MbH8aUjFic438Kas+oEIBX/pB9h3Fw/01Qll/O/VdIa41Paf2h3GlnLmohG5hxJz3HeKaifZCqAMK5wFwoCvMrMKJFaRYXOHl2a4iklf/DvwHEC/9gLp0GQ92zeaBN5vo7IvREkzw0tzPwnHvVrnmx8kVqyo5c2FJznLIzDi+/W21aWY8Y/K5CyFmA8cDrwNlUspW41AbkElxVwU0DrqsyWg7tK+bhRAbhBAbOjo6xjpuDfDdx3fyf/f/noB0cuqfIwSjCbA64JRPKMu99a3sOtrxCMSDykrf/nda923ho5ZHCNacBVUnHHa6x26hvFqVmZOH+t0z+dCL5pNMpWnsDk84te2S8jziqTT7S8+n58Mvc2fyAl6e80nu++gpJNOSu16pJ5WWxOacB+/5/YQWHZ2xsIS7/mUtFvM7dDrqlFPUppnxZP0/WAjhAR4APiWlDAw+JqWUwJicvFLK26WUa6SUa0pKxm9pvZN5alsr77K9RVvJqXRG4PEthtAuvkS9tr6dXUeb71F+6it/BekkJ7x4E4WiD+t5Xx/xkhWLVb7v9paG/jYpJdu2vAnA1fcdZNV/P0UyLZk9wVJyiyuUFb2jNcArrYJvJG/guHPfz8rqfOaWuLnrlXoAqgqcE7qPBnjlFbVpZjxZibsQwooS9nuklMbvdw5m3C3Ga8b52gwMTp1XbbRpckhTTxhX9zaK0t0sOO1q5ha7uX9jkzqYXwPCDD31o/bz7GsbSNe9iFx5vYrpnn8eebFWXjCdhGPW4VZ7hlNWLgHgQP1A5aK/bmziyZdeIY1gweLlXLumhn8/ax7vOm5iecvnFnuwmgU724K8vLcTr93Ciqp8hBBcvrKSQDQJQJVvYhWSNMCXv6w2zYwnm2gZAdwB7JBS/mjQoYeBG4z3NwAPDWr/kBE1sw7oHeS+0eSIN7bv46OWR5AIxIJ3cfUJ1ayv76ahKwxmC/hqoKfuiH2k0pK6p3+DCcmbBUYZulM+SVB4eKLsI0e8tqwgj16RR3f7gAfu6e0HOc7Rgciv4bvvXcvXL1vKFy5cjM9lm9BntVlMzCvxsLM1wD/3drJuXlG/2+TylZX951X6tOWu0WTIxnI/FfggcI4QYrOxXQx8FzhfCLEHOM/YB3gM2A/sBX4D/Efuh/0OZ8v9XPDMBVxsXg8nfxzcxVx1fBVCwANvGtZ7wZxRLffntjZweeIxnk+v4raNMQBSs89gTeK3uKqWjTqMhKuUivBuon/6IPLOy9i4v53FtnZE0byJfsLDWFKRx+t13TR0hzltfnF/+9wSD8ur8vG5rLhsOg+eRpNh1L8GKeXLwEgzVOcOc74EPjbBcWkG8/KtMO8cqFgBgHz2WzSli/nbnG/whQveDSir9dR5xTy4qYlbzl2AqWA2bH9o5D6B+mfv4DwRoGv5R3lxQwfbWnrx2q3EkmnmZ1FOzlFQycrQiyT3tiDSUS5MzKfc0gxFp0/0Ex/G4nIvf9ukvHunDhJ3gP+6bClNPdlVhdJo3im8Q0MCZhDhbnj6v+Cft6r9QAuip46/JE5n3rKh6XOvOr6Kxu4IW1t6Vax3pLu/Huih7Grt5ezu+2j3LOG8i67GbTPzzUe286W/qUnYbMTdfe7n+Yn5Q3xh1l84mL+Sz1nuw5YMqhj3HLO4QsXVl+c5+hcaZVgzu5Arj9epAjSawWhxn8ak0xIOblM7+56DdAoOqEiG19OLOXX+0MLSpy9QFu2r+7oGVmke6prp2ge7/sGeh77LPFMrrrM/Q77bxgfWzWJ9XTf72kPccu4CTpg1+gpNMed0Wo/7V57YH+O39g9RIPrUgaLcJ9taUq4iZk6dXzyhBVGaUbj1VrVpZjzaSTlN6eqLcdb3n+e+VW+xBJQV3rqZ+L4XSeAiVrSUivyhE4ilhlX76v4u/m2hsUqzpx4qVg6c9Kf3QtceLgW6bVUUrlJunc+8ayGXrKjguMr8rKsXAZy9qJR71zfym4YKrik5iUXB1yelclGJ187n3rWQ85ZOLPJGMwqrVk31CDQ5Qov7ZJOMQyKkcrYcoZLQoWxq8BOMJWnds5ElNg/E+0jsfpqOrc+yO72QL1+6fNjrTp5XxN/ebCaRvw4rDLXc4yHo2sP/2S7i0fQ6vv+Rq1RkDaogdSZx1lg4dX4xNrOJeCpN48nfYlHyxUkRdyEEHz9nQc771RzC00+rV12wY8ajxX2SSKUl6Vgflp+tRoSMJQAn3AiX3ZbV9Vuala+8oG8vsdqV2FJ9dL56D1XJRnqWXcuyxaXDXnfy3GLufq2BLZ2S1a4i6B4UDtm+E4CH+xZz5fXvxVM68awQbruFk+YW8tKeTlYsXwHe7MvoaaYh3/qWetXiPuPR4j4JtAeinPujFzgz/hI/s7WzdfaNLMuLwMY7YcllMH/0P5ytzb2UuC0sTDayM70Gd14e81tvB2DZKRePeN26uerXwav7ulhdMHuI5f766y9xElA0dxUXLiufwCccysfOns/xtQWUevUiIo1muqAnVCeB53d3EIwm+WTpW3SKQv61+WJiF98KxQvhkU9DrG/UPrY093LF7CRuEeMfnUV8f5/KeCit7qE+9EMo8thZVObltf1dRqy7stx//txetm16lZiw85X3X5TTScl1c4v4zPkLc9afRqOZOFrcJ4F/7u1ktjvJguBrRBdeTmswyUNbuuCyn0BvA7z4v0e8vj0QpT0YY51b5Yp5JVjGs6HZpKweVc/UbD3i9SfPK2JDfQ+p/Fngb6QnGOaHT+7iFG871vKluJ32nH1WjUYzPdHinmOklPxzbyc3l2xDpOJUnf5BjqvM41cv7iNdsw4WXAA7HztiHxl/+2JTIxLBQfsc3n/yfMzvvQsu+H+jjmHd3CIiiRT1shRkije3bCEtYZ5swFR2XE4+p0ajmd5on3uO2XUwSGdfnHMKXgbfLETVCfzbma188t5NPLXjIBdUrVZl8OIhsA2fLXFLcy9CQHl0P6JgNk/cfBFeuwVM2QnzaQuKsVtMvNDuZh6wd9dW5jjysEY7oWxpDj+t5pjj17+e6hFocoS23HPMy3s68RGkrPM1VRBaCC5eVk55noO/b2qG8uWAhPYdI/axtbmXucVuLB3boew48p1WTGOIPffYLZy7pJT76lTCLtm4nquq/OpgqRZ3zRFYtEhtmhnPO1fc0ym1tH+stUZH4ZV9XVzhq0PIFCxSmRYtZhMnzS3kzYYeZMYt0jZyrvUtzb2srTCrWqTjFOPLV1ayM+Shoeg0rk89zLkuIzWvdstojsQjj6hNM+N554r7Y/8J/zsHvjsL7r0eEhNPPJVIpXltfxcXePeDxQkVq/qPra4t4GAgRgulYM+Htq3D9tEejCICLXyu5TMgTLDg/HGN5axFpXjtFj7Zfhn5IszS/b8DV9GEStBp3gH88Idq08x43nHi/uzOg/zwD/eT3vA7DhSdDksvg12PwRu/nXDfmxv9hOMpjktug+o1YBnIY766VuVqebPBr1wzbVuGXty1j+Sfrqf9ZxfyiP0r+GIt8P77oWZ8i4IcVjPvOq6czYkanrWeiUjF1K8AnZdFo3lH8I4S90QqzRfvf5vT9v2IAG4ua/4gW9d8W6XTfelHEA2M3slgogFofrN/d31dN24i5Pl3QO3JQ05dXOHFYTXxZkOPEveD2yCd7j8eev33iN3/IBYJkSxbifmmJ4YtTj0WLl+lClnsWPxxMFmOGB+v0WiOLd5R4v741jZWhl/hJLEN+/lfQ9p9/OqFfXDOV1Virtd+ObYOH/0U/OZsWP8bADYe6OHSwiaETEPtuiGnWs0mVlT7DMt9mco3YywwqusM0fTG/7FJLiTwvv+j4j8eVedMkNPmF/PJcxdw+dmnwc3Pwxmfm3CfGo1mZvCOEvc/vLyXr9jvQxYvxLnuI7xvXS2PbWnlgGMxLL4UXv2ZmmQdhZ8+s4fP3P4IctvfwVkIj32O9D9/wob6bt7l3q985cO4U1bXFrC9pZdY8cCk6q62IDf94nEWyDqqTriEs0fIGTMezCbBZ85fSE2hS/1acI6exlej0RwbHPvibkTDvNXop7L5CWbLJsTZXwazhZtOnYPFZOI3L+2Hs78MsQCsv/2I3UUTKW5/cT9LG+4hJSH24adh6ZWYnvoalyX+wYr0dihfAXbvYdeurvWRSEm2xMuVm6RtC3e9Ws/q5FuYkFSsvmRSHoFGkzV//KPaNDOeY1vcU0m4+93w+u384Z/7uMX6IKmSJbDkCkDlP3/36ir+uqGJXu9CWHSxcs3EgiN2+cS2NogF+IDtBR5NncQtT/iRV/+WppIz+B/L7ynu2QSzThn22tVGAYyNTWEoXgRtW9nc4Ocyzy5w+KByVa6fgEYzNmpq1KaZ8RxT4p5Ipfnjq/Uc6AqphmQEzDZ4/D+5ace/ME+0YD7ri2Aa+NjXra0llkzz9I6DcPpnIepX2RsHEw+rLwrggY1NfM79OI50mMTaj/GPbW08tbOLnxZ+hS1iISKdPGwyNUOxx05toYuNB3qg+gRk3Qs42jdxfHITzD0TTOZJeCoazRj4y1/UppnxHFPi/vDmFr720DbO/eEL/NdDWwlIB1x3L+vn38JCGogWLoYllw+5ZmV1PpX5Dh7f2qrCF+ecCa/8DBJRdYK/AW5bCT9ZRfDZH/HB+i9xQ+oBWHoFV118MfNK3Hzn8Z28fCDMXbP/F87/H1h4wYhjPGVeEa/u6yJx1leJOUq50/Jt8uLtKmJHo5lqfvlLtWlmPMeUuN+7voHZRS6uPbGGu19v4D//+hZSCL7cfi6fKvwFjhv/NsRqB1Xh56LlFby4u5NgNKEiSvra4C/vB38j/Ok6SMbAV4v3xW9yhultuk79L7jmTixmE1++eAl1nSGa/RGWzquFUz8JlpGzLp69uJRgLMkbHWYeWnYbKQxrfe7Ewh41Go1mMMeMuO8+GGTDgR7ef9Isvn3Vcj5/wSKe2HaQL/9tK3vb+zjjlNMgr3LYay9eXk48lebZne0w5wy49Mew/wX4ySro2AnX3knyQ49yo/M2bin5DUXnf6b/S+KcxaWcPFcVql4ze/QyeqcZZeme29nOC135fMbxTbjkh1AwK2fPQqPRaI4Zcb93fQM2s4mrT1BFLT5y+lzWzink3vUNeOwWLl05ckm542sKKMuz89iWVtWw5l/gw4+r4hqX/BDmncODm5p5vqeEq88ZOlkqhOA7717Ov581j+VV+aOOM1OW7tmd7Wxu8OOetRpO/Mj4P7hGo9EMwzEh7tFEigffbOaCZeUUutWSf7NJ8KNrV+JzWbnuxBpctpGzG5tMgouWVfD8rg5CMTVxSs2J9NzwAsnjbyCRSvPTZ/ewvCqf85YcHoc+u9jNFy5cjDnLzI3nLC5lX0eIlt4oq2p8Y/68Go1GMxrHRD73Hz21m95IguvXDg3hqi5w8fIXzsFlHT0K5ZIVFdz5Sj1PbGvj3auraQ9GOev7z1OR7+CkuUU0dkf45o3H5aQ83TmLS/nmI9sBtLhrphf33z/VI9DkiBlvuf/y+X3c/uJ+PrCutt/3PRiP3ZJVLvQ1swqoLXTxwJtNAPx1QxPheIq0hD+93sDKGh9nL8rN6tFZRW7mlrixmATLsnDlaDRHjeJitWlmPDPacv/z+ga+94+dXL6ykv++fNmErGohBO9eXcVtz+yhqSfMn99o4OS5RfzxprU8se0gy6ryclpU+ubT57KzLYgji18VGs1R48471euNN07lKDQ5YEaL+5KKPN59fBXfu2bFmCoVjcTVq6u59ek9fP7+t2nsjvCfFyzGYjZxyYqRJ2PHy3Vra3Pep0YzYbS4HzOM6pYRQvxOCNEuhNg6qK1QCPGUEGKP8VpgtAshxE+EEHuFEG8LIVZP5uBX1vj40XtXYTXnxrtUU+hi7ZxCXtnXRYHLygXHleWkX41GoznaZKOKdwIXHtL2ReAZKeUC4BljH+AiYIGx3QzMuKVu16xWoZTXnFCN3aJdJhqNZmYyqrhLKV8EDs2DewVwl/H+LuDKQe1/kIrXAJ8QIvc+jUnkspWV3HTaHD5y+typHopGo9GMm/H63MuklMaKH9qAjP+iCmgcdF6T0dbKIQghbkZZ99TWTh//s9Nm5muXjq8otUaj0UwXJjyhKqWUQgg5jutuB24HWLNmzZiv12g0k8Bjj031CDQ5YrwzkQcz7hbjtd1obwYGrySqNto0Gs1MwOVSm2bGM15xfxi4wXh/A/DQoPYPGVEz64DeQe4bjUYz3fnFL9SmmfFkEwp5L/AqsEgI0SSEuAn4LnC+EGIPcJ6xD/AYsB/YC/wG+I9JGbVGo5kc7rtPbZoZz6g+dynl9SMcOneYcyXwsYkOSqPRaDQTY8bnltFoNBrN4Whx12g0mmMQLe4ajUZzDCKUm3yKByFEB3BgnJcXA505HM5kMlPGOlPGCXqsk8FMGSfMnLFO1jhnSSlLhjswLcR9IgghNkgp10z1OLJhpox1powT9Fgng5kyTpg5Y52KcWq3jEaj0RyDaHHXaDSaY5BjQdxvn+oBjIGZMtaZMk7QY50MZso4YeaM9aiPc8b73DUajUZzOMeC5a7RaDSaQ9DirtFoNMcgM1rchRAXCiF2GTVbvzj6FUcHIUSNEOI5IcR2IcQ2IcQtRvuwtWenA0IIsxBikxDiUWN/jhDidePZ/kUIYZsGY/QJIe4XQuwUQuwQQpw8XZ+pEOLTxr/9ViHEvUIIx3R5ptO5LnIW4/y+8e//thDib0II36BjXzLGuUsIccHRGudIYx107LNCCCmEKDb2j8oznbHiLoQwAz9H1W1dClwvhJguJZSSwGellEuBdcDHjLGNVHt2OnALsGPQ/veAH0sp5wM9wE1TMqqh3Ab8Q0q5GFiJGu+0e6ZCiCrgk8AaKeUywAxcx/R5pncyM+oi38nh43wKWCalXAHsBr4EYPx9XQccZ1zzC0MjjhZ3cvhYEULUAO8CGgY1H51nKqWckRtwMvDEoP0vAV+a6nGNMNaHgPOBXUCF0VYB7JrqsRljqUb9QZ8DPAoI1Go6y3DPeorGmA/UYQQBDGqfds+UgXKThajMq48CF0ynZwrMBraO9hyBXwPXD3feVIzzkGNXAfcY74f8/QNPACdP5TM12u5HGSL1QPHRfKYz1nJn5Hqt0wohxGzgeOB1Rq49O9XcCnweSBv7RYBfSpk09qfDs50DdAC/N9xHvxVCuJmGz1RK2Qz8AGWttQK9wEam3zMdzFjrIk8H/gV43Hg/7cYphLgCaJZSvnXIoaMy1pks7tMeIYQHeAD4lJQyMPiYVF/ZUx6HKoS4FGiXUm6c6rGMggVYDfxSSnk8EOIQF8w0eqYFwBWoL6RKwM0wP9mnK9PlOR4JIcRXUO7Pe6Z6LMMhhHABXwa+PlVjmMniPq3rtQohrChhv0dK+aDRPFLt2ankVOByIUQ98GeUa+Y2wCeEyBRzmQ7PtgloklK+buzfjxL76fhMzwPqpJQdUsoE8CDqOU+3ZzqYGVMXWQhxI3Ap8H7jiwim3zjnob7c3zL+tqqBN4UQ5Rylsc5kcX8DWGBEINhQkykPT/GYADUbDtwB7JBS/mjQoZFqz04ZUsovSSmrpZSzUc/wWSnl+4HngGuM06Z8rFLKNqBRCLHIaDoX2M40fKYod8w6IYTL+L+QGeu0eqaHMCPqIgshLkS5EC+XUoYHHXoYuE4IYRdCzEFNVq6fijECSCm3SClLpZSzjb+tJmC18f/46DzToznhMAkTGBejZsz3AV+Z6vEMGtdpqJ+1bwObje1ilC/7GWAP8DRQONVjPWTcZwGPGu/nov449gJ/BezTYHyrgA3Gc/07UDBdnynwTWAnsBX4I2CfLs8UuBc1F5BAic5NIz1H1OT6z42/sS2oCKCpHOdelL8683f1q0Hnf8UY5y7goql+poccr2dgQvWoPFOdfkCj0WiOQWayW0aj0Wg0I6DFXaPRaI5BtLhrNBrNMYgWd41GozkG0eKu0Wg0xyBa3DUajeYYRIu7RqPRHIP8fwB3MyLu3B7/AAAAAElFTkSuQmCC\n", "text/plain": [ "
" ]