You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
在微调baichuan2-7B-Base模型的时候,发现,输入的token长度不能超过512,其官方给出的最大长度为4096。
微调遵循官方教程,使用lora的方式,在微调过程中使用官方数据集,在数据中添加超长数据时出现
Token indices sequence length is longer than the specified maximum sequence length for this model (1759 > 512). Running this sequence through the model will result in indexing errors
请问,是否需要修改相关配置文件或是相关代码?还是7B的模型本身不支持超长序列的输入?
The text was updated successfully, but these errors were encountered:
同样的问题
Traceback (most recent call last):
File "/home/ma-user/work/mindformers/research/baichuan2/run_baichuan2_pipeline.py", line 145, in
outputs = pipeline_task(inputs,
File "/home/ma-user/anaconda3/envs/MindSpore/lib/python3.9/site-packages/mindformers/pipeline/base_pipeline.py", line 121, in call
outputs = self.run_multi(inputs, batch_size, preprocess_params, forward_params, postprocess_params)
File "/home/ma-user/anaconda3/envs/MindSpore/lib/python3.9/site-packages/mindformers/pipeline/text_generation_pipeline.py", line 208, in run_multi
outputs.extend(self.run_single(item, preprocess_params,
File "/home/ma-user/anaconda3/envs/MindSpore/lib/python3.9/site-packages/mindformers/pipeline/base_pipeline.py", line 170, in run_single
model_outputs = self.forward(model_inputs, **forward_params)
File "/home/ma-user/anaconda3/envs/MindSpore/lib/python3.9/site-packages/mindformers/pipeline/text_generation_pipeline.py", line 224, in forward
output_ids = self.network.generate(input_ids, **forward_params)
File "/home/ma-user/anaconda3/envs/MindSpore/lib/python3.9/site-packages/mindformers/generation/text_generator.py", line 1120, in generate
output_ids = self._greedy_search(
File "/home/ma-user/anaconda3/envs/MindSpore/lib/python3.9/site-packages/mindformers/generation/text_generator.py", line 315, in _greedy_search
raise ValueError(
ValueError: the input_ids length 757 exceeds the max length config 512.check your inputs and set max_length larger than your inputs length.
在微调baichuan2-7B-Base模型的时候,发现,输入的token长度不能超过512,其官方给出的最大长度为4096。
微调遵循官方教程,使用lora的方式,在微调过程中使用官方数据集,在数据中添加超长数据时出现
Token indices sequence length is longer than the specified maximum sequence length for this model (1759 > 512). Running this sequence through the model will result in indexing errors
请问,是否需要修改相关配置文件或是相关代码?还是7B的模型本身不支持超长序列的输入?
The text was updated successfully, but these errors were encountered: