Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

why is very slow about the model on CPU platform? #8

Open
wshzd opened this issue Nov 28, 2019 · 7 comments
Open

why is very slow about the model on CPU platform? #8

wshzd opened this issue Nov 28, 2019 · 7 comments

Comments

@wshzd
Copy link

wshzd commented Nov 28, 2019

l run the model in windows10 with CPU, but it will spend 4 hours every epoch, that is, 100 epoches need 400 hour in order to run the whole model. it claims it is faster than biLSTM+CRF, actually,it is not.
ok, l run the BERT+biLSTM+CRF on same envirment(windows10 with CPU), it only costs 10 hours, however, it's accuracy is 0.92
Please can you tell me that is why?

@chiyuzhang94
Copy link

Hi, I am also trying to use the tool. But I am wondering where the dataset is. I saw you run this model. What dataset did you use? Could you please share with me?

@wshzd
Copy link
Author

wshzd commented Dec 5, 2019

the format of the dataset is normally two columns,one is char, another is corresponding label
as follows:
人 o
民 B-pro

@chiyuzhang94
Copy link

@wshzd Thanks for your reply. Are the two elements split by white space or anything else?

@wshzd
Copy link
Author

wshzd commented Dec 6, 2019

anything

@wudaoyunqi
Copy link

I run this model with the WSJ dataset on GPU platform,it seems that each epoch will take 50 minutes:(

@Nealcly
Copy link
Owner

Nealcly commented Jan 21, 2020

I run this model with the WSJ dataset on GPU platform,it seems that each epoch will take 50 minutes:(

Could you kindly share your log with me via email?

@wudaoyunqi
Copy link

I run this model with the WSJ dataset on GPU platform,it seems that each epoch will take 50 minutes:(

Could you kindly share your log with me via email?

Sorry for the late reply, I went through the log and found a parameter setting error. Then I ran the model on a high-performance GPU server, and it was fast. Thanks for your kindly reply.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants