-
Notifications
You must be signed in to change notification settings - Fork 37
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
train problem #41
Comments
What is your GPU specifications? |
@vdobrovolskii. I have similar issue during train and evaluate. |
Please, share your GPU specs and also the output of |
@vdobrovolskii
|
@vdobrovolskii I am not even able to evaluate here in this machine. Loading Bert model is really slow. |
You won't be able to train the model on your machine without modifying the code... But for evaluation it should be more than enough, can you show me the exact sequence of steps you're taking? (Commands and outputs) |
@vdobrovolskii These are the steps I followed starting from data processing. |
@vdobrovolskii |
I would recommend you to use print statements to see where actually the code is stuck. |
@shantanu778 I am not exactly sure what is happening there. The problem is I no longer work where I did when this paper was written, so I don't have access to the server where the original environment was hosted. So I am afraid I can't tell you the exact versions of the packages that I had back then. However, I'm inviting everyone who's got it working to share their own |
On your screenshot I can see that the word-level evaluation is going OK, but something is wrong with predicting the spans. Would you mind taking a look at the data and confirming that the data preparation went well and everything looks normal? I would pay extra attention to the |
Sorry to bother! After I execute "python run.py train bert" ,I've been stack here for a long time(just like the picture) .It growed to 15% and nothing changed. Is that normal?
PS: I've changed batch_size from 512 to 64,unless my GPU can not run...
The text was updated successfully, but these errors were encountered: