You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Exporting text-generation models (e.g., LlamaForSequenceClassification) was disabled in #1308
Given https://arxiv.org/abs/2310.01208, these models can outperform typical encoder models for sequence classification (I can confirm this on my own datasets)
What would it take to support this in transformers? CC @fxmarty since it was mentioned specifically in the PR above
Motivation
I would like to export fine-tuned sequence classification models that use a decoder-only model as their base architecture to ONNX
Your contribution
Happy to submit a PR to transformers with guidance on what would be needed
The text was updated successfully, but these errors were encountered:
Ah, I see what's happening -- ONNX doesn't support int64 inputs to argmax which is how these models are computing the sequence lengths for pooling. Will open a PR over in transformers and leave this open until we can enable export in this repo
Feature request
Exporting text-generation models (e.g., LlamaForSequenceClassification) was disabled in #1308
Given https://arxiv.org/abs/2310.01208, these models can outperform typical encoder models for sequence classification (I can confirm this on my own datasets)
What would it take to support this in transformers? CC @fxmarty since it was mentioned specifically in the PR above
Motivation
I would like to export fine-tuned sequence classification models that use a decoder-only model as their base architecture to ONNX
Your contribution
Happy to submit a PR to
transformers
with guidance on what would be neededThe text was updated successfully, but these errors were encountered: