You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@uhauha2929 the text body and summarized text use different vocabulary, which might explain what you observed. Also the max_sequence_length is set on the text body, meaning it does not read any text from the text body after the max_sequence_length of words are read from the text body. One way to address the issue u mentioned is to use a single vocabulary for both text body and summarized text or read more texts from the text body. Also depending on the language of the text body (for example, the chinese requires chinese tokenizer may give a better result i think)
我用中文语料做的测试,词语和字符我都试过了,生成的句子貌似也有可读性,但是和原文出入较大,不知道是什么原因,是模型太简单了?
The text was updated successfully, but these errors were encountered: