Skip to content

Commit

Permalink
docs: Resolve typos, missing import (#3443)
Browse files Browse the repository at this point in the history
Closes #3429

Hello!

# Description

Resolves typos and a missing import as described by #3429

**Type of change**

- [x] Documentation update

**How Has This Been Tested**

I ran the snippet that I modified - it worked.

**Checklist**

- [x] I added relevant documentation
- [ ] follows the style guidelines of this project
- [x] I did a self-review of my code
- [ ] I made corresponding changes to the documentation
- [ ] My changes generate no new warnings
- [ ] I have added tests that prove my fix is effective or that my
feature works
- [ ] I filled out [the contributor form](https://tally.so/r/n9XrxK)
(see text above)
- [ ] I have added relevant notes to the CHANGELOG.md file (See
https://keepachangelog.com/)

---

- Tom Aarsen
  • Loading branch information
tomaarsen committed Jul 26, 2023
1 parent 40dc948 commit 4d04835
Showing 1 changed file with 4 additions and 3 deletions.
7 changes: 4 additions & 3 deletions docs/_source/guides/train_a_model.md
Original file line number Diff line number Diff line change
Expand Up @@ -89,6 +89,7 @@ Options:

```python
import argilla as rg
from argilla.training import ArgillaTrainer
from datasets import load_dataset

dataset_rg = rg.DatasetForTokenClassification.from_datasets(
Expand Down Expand Up @@ -126,18 +127,18 @@ It is possible to directly include train-test splits to the `prepare_for_trainin
*TextClassification*

For text classification tasks, it flattens the inputs into separate columns of the returned dataset and converts the annotations of your records into integers and writes them in a label column:
By passing the `framework` variable as `setfit`, `transformers`, `spark-nlp` or `spacy`. This task requires a `DatastForTextClassification`.
By passing the `framework` variable as `setfit`, `transformers`, `spark-nlp` or `spacy`. This task requires a `DatasetForTextClassification`.


*TokenClassification*

For token classification tasks, it converts the annotations of a record into integers representing BIO tags and writes them in a `ner_tags` column:
By passing the `framework` variable as `transformers`, `spark-nlp` or `spacy`. This task requires a `DatastForTokenClassification`.
By passing the `framework` variable as `transformers`, `spark-nlp` or `spacy`. This task requires a `DatasetForTokenClassification`.

*Text2Text*

For text generation tasks like `summarization` and translation tasks, it converts the annotations of a record `text` and `target` columns.
By passing the `framework` variable as `transformers` and `spark-nlp`. This task requires a `DatastForText2Text`.
By passing the `framework` variable as `transformers` and `spark-nlp`. This task requires a `DatasetForText2Text`.

*Feedback*
For feedback-oriented datasets, we currently rely on a fully customizable workflow, which means automation is limited and yet to be thought out.
Expand Down

0 comments on commit 4d04835

Please sign in to comment.