Thanks, Qingsong! I noticed a problem in my label data and now the negative errors disappear but I still get -nan loss. Do you think it's because the loss value is too large to display? 'coz when it stops the validation error seems have more than 20 digits.
If you would like to refer to this comment somewhere else in this project, copy and paste the following link:
I tried very small learning rate, but still nan. My data is large so I started from a relatively simple model. The training error after first epoch is 89.68%, but after that it goes to 100%. like this:
Epoch | Duration | Training error | Validation error | Test error | New best
-------+----------+------------------+------------------+------------------+----------
1 | 637.1 | 89.68% -nan |100.00% -nan | | no
2 | 638.2 |100.00% -nan |100.00% -nan | | no
3 | 643.5 |100.00% -nan |100.00% -nan | | no
4 | 648.4 |100.00% -nan |100.00% -nan | | no
If you would like to refer to this comment somewhere else in this project, copy and paste the following link:
Siince I'm doing many to one training, like described in this blog , when I changed the layer from blstm to lstm, the training become reasonable. But when I use the same setting and run it again, it goes back to nan.
Last edit: Yishan 2016-03-05
If you would like to refer to this comment somewhere else in this project, copy and paste the following link:
HI all,
I'm running my multiclass classification task with Currennt. And I keep getting negative training and validation errors and nan loss, like this:
Does anyone know what's the reason for this? Thanks!
Yishan
Last edit: Yishan 2016-03-05
what's the value of learning rate?
try a smaller value
On Sat, Mar 5, 2016 at 2:22 PM, Yishan jieralice@users.sf.net wrote:
--
Qingsong Liu
liuqs.ustc@gmail.com
Univ. of Sci.& Tech. of China
Thanks, Qingsong! I noticed a problem in my label data and now the negative errors disappear but I still get -nan loss. Do you think it's because the loss value is too large to display? 'coz when it stops the validation error seems have more than 20 digits.
I tried very small learning rate, but still nan. My data is large so I started from a relatively simple model. The training error after first epoch is 89.68%, but after that it goes to 100%. like this:
Siince I'm doing many to one training, like described in this blog , when I changed the layer from blstm to lstm, the training become reasonable. But when I use the same setting and run it again, it goes back to nan.
Last edit: Yishan 2016-03-05