diff --git a/scripts/nlp_language_modeling/convert_hf_llama_to_nemo.py b/scripts/nlp_language_modeling/convert_hf_llama_to_nemo.py index d6007aa771c0..58abcd20442d 100644 --- a/scripts/nlp_language_modeling/convert_hf_llama_to_nemo.py +++ b/scripts/nlp_language_modeling/convert_hf_llama_to_nemo.py @@ -98,6 +98,9 @@ def load_config(llama_config): nemo_config = OmegaConf.load( os.path.join(os.path.dirname(__file__), '../../examples/nlp/language_modeling/conf/megatron_llama_config.yaml') ).model + + if llama_config.get('rope_theta', None): + nemo_config['rotary_base'] = llama_config['rope_theta'] nemo_config.encoder_seq_length = llama_config['max_position_embeddings'] nemo_config.num_layers = int(llama_config['num_hidden_layers']) nemo_config.hidden_size = llama_config['hidden_size']