LLaMa 3.1: rope_scaling 错误
原因是因为一些基本库和model不匹配了:
Q:rope_scaling
must be a dictionary with with two fields, name
and factor
, got {'factor': 8.0, 'low_freq_factor': 1.0, 'high_freq_factor': 4.0, 'original_max_position_embeddings': 8192, 'rope_type': 'llama3'}
pip install --upgrade transformers
Q: ImportError: cannot import name 'top_k_top_p_filtering' from 'transformers'
pip install --upgrade trl
Q: ImportError: Using the Trainer
with PyTorch
requires accelerate>=0.26.0
:
pip install -U accelerate