-
Notifications
You must be signed in to change notification settings - Fork 2k
dynamic_decode raise Segmentation default & inference.load_data filter some lines #22
Comments
I download the tmp.zip file, and there is nothing in it? |
@mingfengwuye tmp.zip is updated. |
@hxsnow10 I have download the tmp.zip file and begin to train model untill now It works well. The final result will be next week. One more question, how could get the vocab.zh or vocab.en file. Before, I get vocab follow the wmt16_en_de.sh, but it seems different with that. |
@mingfengwuye some strange, try use my update bigger data, I make sure i meet Segmentation fault this time without anychange to code. my vocab is count from en-zh corpus myself. By the way, my system is centos7, tensorflow version 1.2.1, i run on cpu, memory is enough.
|
@hxsnow10 I train the model in the weekend, and trainning completed without any error. I did not change the source code. OS is Ubuntu, cuda 8.0 + nvidia-375, tensorflow version is 1.2.1. My script is: |
@mingfengwuye i'm sorry, can you try use dev3 and cpu once..., thanks...
|
@hxsnow10 I will try it later. |
@mingfengwuye I find in my environment and cpu, the problem lies in crossent = tf.nn.sparse_softmax_cross_entropy_with_logits( in model.py, when max_time is big like 39, it will raise Segmentation fault error. |
@hxsnow10 How could you get vocab. Could you give me some guides? Thank you very much. |
@mingfengwuye I can't catch you much.. My vocab is count and sort word by several english-chineses corpus from http://www.statmt.org/wmt17/translation-task.html#download after tokenize(using nltk and chineses tokenizer). Segmentation fault is because sprase.softmax not support big tensor, so at last i try batch_size=32 with sparse.softmax, it works. Another question: what iterator.source looks like, when i make source_reverse=False |
@hxsnow10 can you give a complete and clear solution? I face the same question too. |
@zhangpengGenedock use small batch_size。 |
Hello,
I use en-zh data like in the tmp.zip
put these files in /tmp/nmt_data
then i get a Segmentation fault result.it should be reproduced.
After some bug fix, i location the problem lieing in dynamic_decode , line 326 of model.py. But i can't go on to solve it.
Can you give some suggestions? Really Thanks.
Another small problem is inference.load_data filter some lines, which make loaded data of zh and en has different lens.
notes: tmp.zip is placed in https://github.com/hxsnow10/nmt_problem
The text was updated successfully, but these errors were encountered: