描述bug
在模型预处理过程中捕获到异常:在获取对象的字符串表示时超过了最大递归深度。
重现步骤
重现问题的步骤:
从github上采样:
results = model.train(dataset=df)
print(results)
请提供代码、yaml配置文件和一个数据样本,以便完全重现问题。无法重现的问题将被忽略。
导入yaml
import yaml
config_str = """
model_type: llm
base_model: /mnt/d/AI/Models/llama-7b-hf
quantization:
bits: 4
adapter:
type: lora
prompt:
template: |
### Instruction:
{instruction}
Input:
{input}
Response:
input_features:
* name: prompt
type: text
preprocessing:
max_sequence_length: 256
output_features:
* name: output
type: text
preprocessing:
max_sequence_length: 256
trainer:
type: finetune
learning_rate: 0.0001
batch_size: 1
gradient_accumulation_steps: 16
epochs: 3
learning_rate_scheduler:
warmup_fraction: 0.01
preprocessing:
sample_ratio: 0.1
"""
config = yaml.safe_load(config_str)
1条答案
按热度按时间y0u0uwnf1#
@Bdl-1989 Odd error. Are you able to use the tokenizer for
/mnt/d/AI/Models/llama-7b-hf
on your outside of Ludwig?