add japanese bert pretrained model#118
add japanese bert pretrained model#118gorogoroyasu wants to merge 6 commits intonlpyang:masterfrom gorogoroyasu:japanese
Conversation
hi guys, Can you give me the ways to get new BERT pre-trained model ? Thank you. |
|
Hello, thanks for your contribution. I notice that you didn't change some functions designed to preprocess the english datasets in the data_builder file, and you use the multilingual model to substitute the old one, so I guess that you use the english datasets to train your Japanese model. Is my guess correct? Look forward to your soonest reply. Thank you! |
|
@congdoanit98 @beanandrew
You're correct. But, I used Japanese dataset. I think these were all changes. Thanks. |
|
Thanks for your reply.
I am still on the way to change the code, and I haven't do experiments on this, so I wanna know how you solve problems like this, or after your experiments, these problems won't change the experiment results? |
|
Well, You are correct again.. I conducted my experiments by using the original step5 codes, and I could generate some summaries, and it showed reasonable result. I hope this information will support your work. |
|
Thanks for your reply, your answer helps me lot! When I try to debug the code in the Step5, I notice that, because of this code, the labels like [CLS] and [SEP] are skipped, and that cause the Error "CUDAType Error" when I use the dataset it preprocessed. Thus, I changed the code as followed, and the Error no longer appear. I want to know, do you come with Errors like this when you use the original code? Or is this just my personal problem? |
|
Hmm, I couldn't find that code you mentioned on master branch. Anyways, your suggestion seems to work beautifully. I remember that I commented out the code below (rouge score calculator), because I couldn't solve errors from pyrouge library. Instead of the pyrouge library, I used this one. I fixed the code to export the summarized text, and after all results were written, I evaluated the rouge score. If you could correctly install the pyrouge library, you don't have to care I think. Additionally, the token size is very important. In Japanese token.txt, I couldn't find something like |
|
Thank you very much for your quick reply! |
in order to use
bert-base-japanese-whole-word-maskingmodel, installed transformer independently, and fixed few codes.