Running models on seperate devices #485
-
Hi, I was wondering if it it possible to run 2 models on seperate devices. For example if I want to run all-MiniLM-L6-v2 on CPU with torch engine Is that possible to do with the CLI? If so what would the CLI command look like infinity_emb v2 ... |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 2 replies
-
@Sens612 There is a section in the Readme on CLI usage. e.g. infinity_emb v2 \
--model-id all-MiniLM-L6-v2 --engine optimum --device cpu --batch-size 8 \
--model-id cross-encoder/ms-marco-MiniLM-L-6-v2 --engine torch --device cuda --batch-size 16 |
Beta Was this translation helpful? Give feedback.
-
Aha, I had no clue you could do it that way, I thought majority of the arguments were forced on both models. Looking in the README I don't think this is shown clearly with an example. Either way thank you, and great work on the repo. |
Beta Was this translation helpful? Give feedback.
@Sens612 There is a section in the Readme on CLI usage.
e.g.