@@ -880,20 +880,21 @@ def parse_args() -> argparse.Namespace:
880
880
881
881
hparams = Model .load_hparams (dir_model )
882
882
883
- model_class = Model .from_model_architecture (hparams ["architectures" ][0 ])
884
- model_instance = model_class (dir_model , ftype_map [args .outtype ], fname_out , args .bigendian )
883
+ with torch .inference_mode ():
884
+ model_class = Model .from_model_architecture (hparams ["architectures" ][0 ])
885
+ model_instance = model_class (dir_model , ftype_map [args .outtype ], fname_out , args .bigendian )
885
886
886
- print ("Set model parameters" )
887
- model_instance .set_gguf_parameters ()
887
+ print ("Set model parameters" )
888
+ model_instance .set_gguf_parameters ()
888
889
889
- print ("Set model tokenizer" )
890
- model_instance .set_vocab ()
890
+ print ("Set model tokenizer" )
891
+ model_instance .set_vocab ()
891
892
892
- if args .vocab_only :
893
- print (f"Exporting model vocab to '{ fname_out } '" )
894
- model_instance .write_vocab ()
895
- else :
896
- print (f"Exporting model to '{ fname_out } '" )
897
- model_instance .write ()
893
+ if args .vocab_only :
894
+ print (f"Exporting model vocab to '{ fname_out } '" )
895
+ model_instance .write_vocab ()
896
+ else :
897
+ print (f"Exporting model to '{ fname_out } '" )
898
+ model_instance .write ()
898
899
899
- print (f"Model successfully exported to '{ fname_out } '" )
900
+ print (f"Model successfully exported to '{ fname_out } '" )
0 commit comments