Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

the size of pre-train model 's input data #127

Open
willow-yll opened this issue Mar 23, 2022 · 1 comment
Open

the size of pre-train model 's input data #127

willow-yll opened this issue Mar 23, 2022 · 1 comment

Comments

@willow-yll
Copy link

willow-yll commented Mar 23, 2022

Hi,
I want to use pretrain-model and add some layers to train a downstream task,
what kind of data should I input?
Should I process the amino acid sequence into one-hot encoding?
Or if there is a file of tape I can use to deal with the amino acid sequence?
I don't know the data size to input....

my model like this:

pretrained = pretrained = UniRepModel.from_pretrained('babbler-1900', force_download=False)

class UniRep_bilstm(nn.Module):

    def __init__(self,emb_dim,hidden_dim,num_layers,output_dim,max_len):

        super(UniRep_bilstm,self).__init__()
        self.unirep = pretrained
        self.bilstm = bilstm(emb_dim,hidden_dim,num_layers,output_dim,max_len)

    def forward(self,input_ids):
        unirep_outputs = self.unirep(input_ids)
        outputs = self.bilstm(unirep_outputs)
        return outputs
@willow-yll
Copy link
Author

Could you please tell me the size and type of input data , and the way to deal with amino acid sequence? thanks :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant