Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

TAPAS unable to use weak supervision labels to finetune #141

Open
shabbie opened this issue Sep 30, 2021 · 1 comment
Open

TAPAS unable to use weak supervision labels to finetune #141

shabbie opened this issue Sep 30, 2021 · 1 comment

Comments

@shabbie
Copy link

shabbie commented Sep 30, 2021

I am trying to Fine-tune the pretrained TAPAS WTQ model on a custom dataset. I have used both Hugging face Pytorch code. My dataset has majority of samples with arithmetic operations, so they rely on scalar answer as supervision. The details of the problem faced with both codes are described below:

  • As we do not have the answer coordinates available, the coordinates are predicted by the utility by computing the cost matrix in the utility provided.
  • However, the utility returns 'None' as the result when it does not find a matching candidate from the table (which is the case whenever the answer is the result of an aggregation operation over cell values)
  • Now, if we pass 'None' as the answer coordinate to the TAPAS Tokenizer, we don't get the labels tensor as the result of that.
  • While using that tokenization output and passing it to the TAPAS model, it does not compute loss rather just returns the predicted answer coordinates and predicted aggregation operator (this is the case while we do inference)

I have tried passing the labels tensor to be all zeros as try but that makes the model learn to not select any column

@SyrineKrichene
Copy link
Collaborator

SyrineKrichene commented Nov 23, 2021 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants