Skip to content

How to train with a dataset that doesn't fit into memory #12551

Discussion options

You must be logged in to vote

@rmitsch I think I've figured out the issue.

The actual issue was with my test dataset. Spacy needs the test dataset to fit entirely in the memory during the eval. And my dev.jsonl file was still pretty large! That's why I was facing that issue

I'm going to reduce that file.

Thanks for your help!

Replies: 4 comments 13 replies

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
13 replies
@rmitsch
Comment options

@delucca
Comment options

@delucca
Comment options

@delucca
Comment options

Answer selected by delucca
@rmitsch
Comment options

@delucca
Comment options

@rmitsch
Comment options

@delucca
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
training Training and updating models gpu Using spaCy on GPU
2 participants