Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ValueError: Too many rows #177

Open
Gayatri-95 opened this issue Jun 7, 2023 · 0 comments
Open

ValueError: Too many rows #177

Gayatri-95 opened this issue Jun 7, 2023 · 0 comments

Comments

@Gayatri-95
Copy link

Hi,

I am trying to fine tune TAPAS WTQ model on dummy data of 990 rows and 18 columns ('Nobel Laureates, 1901-Present' dataset: https://www.kaggle.com/datasets/nobelfoundation/nobel-laureates). I am running the notebook on Kaggle with max RAM of 30 GB. However, I am encountering issues while encoding using TapasTokenizer.
When the parameter truncation=False, the error is - "ValueError: Too many rows" and when the parameter truncation=True, it is - "ValueError: Couldn't find all answers".

The encoding code looks like below:
encoding = tokenizer(table=table, queries=item.question, answer_coordinates=item.answer_coordinates, answer_text=item.answer_text, truncation=True, padding="max_length", return_tensors="pt")
encoding.keys()

Can anyone let me know what is the maximum size of the data i.e., maximum number of rows and columns which can be handled by the model?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant