Skip to content
This repository has been archived by the owner on Nov 22, 2022. It is now read-only.

RoBERTa-QA JIT #1088

Closed
wants to merge 1 commit into from
Closed

Conversation

debowin
Copy link

@debowin debowin commented Oct 30, 2019

Summary:
This diff torchscriptifies RoBERTa-QA to make it exportable in the same manner as BERT-QA.
It also converts the has_answer_logits tensor to float before doing an argmax in squad_output_layer in order to use the FP16FairseqOptimizer. This otherwise breaks as argmax_cuda isn't supported for "Half" datatype.

Differential Revision: D18218091

@facebook-github-bot facebook-github-bot added the CLA Signed Do not delete this pull request or issue due to inactivity. label Oct 30, 2019
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D18218091

debowin pushed a commit to debowin/pytext that referenced this pull request Nov 5, 2019
Summary:
Pull Request resolved: facebookresearch#1088

This diff torchscriptifies RoBERTa-QA to make it exportable in the same manner as BERT-QA.
It also converts the has_answer_logits tensor to float before doing an argmax in squad_output_layer in order to use the FP16FairseqOptimizer. This otherwise breaks as argmax_cuda isn't supported for "Half" datatype.

Differential Revision: D18218091

fbshipit-source-id: a53ea099da52e13784f327d533348524dfd6e5d5
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D18218091

debowin pushed a commit to debowin/pytext that referenced this pull request Nov 5, 2019
Summary:
Pull Request resolved: facebookresearch#1088

This diff torchscriptifies RoBERTa-QA to make it exportable in the same manner as BERT-QA.
It also converts the has_answer_logits tensor to float before doing an argmax in squad_output_layer in order to use the FP16FairseqOptimizer. This otherwise breaks as argmax_cuda isn't supported for "Half" datatype.

Differential Revision: D18218091

fbshipit-source-id: 96a112260a20ae836fbfad88994b6a934be2062c
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D18218091

debowin pushed a commit to debowin/pytext that referenced this pull request Nov 8, 2019
Summary:
Pull Request resolved: facebookresearch#1088

This diff torchscriptifies RoBERTa-QA to make it exportable in the same manner as BERT-QA.
It also converts the has_answer_logits tensor to float before doing an argmax in squad_output_layer in order to use the FP16FairseqOptimizer. This otherwise breaks as argmax_cuda isn't supported for "Half" datatype.

Differential Revision: D18218091

fbshipit-source-id: d917a82663be0b42f337b84b048a18346ad39b41
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D18218091

Summary:
Pull Request resolved: facebookresearch#1088

This diff torchscriptifies RoBERTa-QA to make it exportable in the same manner as BERT-QA.
It also converts the has_answer_logits tensor to float before doing an argmax in squad_output_layer in order to use the FP16FairseqOptimizer. This otherwise breaks as argmax_cuda isn't supported for "Half" datatype.

Reviewed By: hikushalhere

Differential Revision: D18218091

fbshipit-source-id: 7c315b17fedfe82915d396179bef28367dce1d56
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D18218091

@facebook-github-bot
Copy link
Contributor

This pull request has been merged in 8afd065.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
CLA Signed Do not delete this pull request or issue due to inactivity. Merged
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants