You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Import fails at the tokenization step when the document contains multiple languages. This is due merging different spaCy docs with different vocab objects.
Installation
pip install goldenverba
pip install from source
Docker installation
If you installed via pip, please specify the version:
Weaviate Deployment
Local Deployment
Docker Deployment
Cloud Deployment
Configuration
Reader: Any
Chunker: /
Embedder: /
Retriever: /
Generator: /
Steps to Reproduce
Upload longer documents over 500.000 tokens with different languages
Additional context
This will be fixed with v2.2
The text was updated successfully, but these errors were encountered:
Description
Import fails at the tokenization step when the document contains multiple languages. This is due merging different spaCy docs with different
vocab
objects.Installation
If you installed via pip, please specify the version:
Weaviate Deployment
Configuration
Reader: Any
Chunker: /
Embedder: /
Retriever: /
Generator: /
Steps to Reproduce
Upload longer documents over 500.000 tokens with different languages
Additional context
This will be fixed with
v2.2
The text was updated successfully, but these errors were encountered: