@@ -50,7 +50,7 @@ class GptParams:
5050 # If chat ended prematurely, append this to the conversation to fix it.
5151 # Set to "\nUser:" etc.
5252 # This is an alternative to input_prefix which always adds it, so it potentially duplicates "User:""
53- fix_prefix : str = " "
53+ fix_prefix : str = ""
5454 output_postfix : str = ""
5555 input_echo : bool = True ,
5656
@@ -75,7 +75,7 @@ def gpt_params_parse(argv = None, params: Optional[GptParams] = None):
7575 parser .add_argument ("--top_p" , type = float , default = 0.95 , help = "top-p samplin" ,dest = "top_p" )
7676 parser .add_argument ("--top_k" , type = int , default = 40 , help = "top-k sampling" ,dest = "top_k" )
7777 parser .add_argument ("--temp" , type = float , default = 0.80 , help = "temperature" ,dest = "temp" )
78- parser .add_argument ("--n_predict" , type = int , default = 128 , help = "number of model parts " ,dest = "n_predict" )
78+ parser .add_argument ("--n_predict" , type = int , default = 128 , help = "number of tokens to predict (-1 = infinity) " ,dest = "n_predict" )
7979 parser .add_argument ("--repeat_last_n" , type = int , default = 64 , help = "last n tokens to consider for penalize " ,dest = "repeat_last_n" )
8080 parser .add_argument ("--repeat_penalty" , type = float , default = 1.10 , help = "penalize repeat sequence of tokens" ,dest = "repeat_penalty" )
8181 parser .add_argument ("-b" , "--batch_size" , type = int , default = 8 , help = "batch size for prompt processing" ,dest = "n_batch" )
0 commit comments