@@ -1394,7 +1394,9 @@ void gpt_params_print_usage(int /*argc*/, char ** argv, const gpt_params & param
1394
1394
options.push_back ({ " *" , " --keep N" , " number of tokens to keep from the initial prompt (default: %d, -1 = all)" , params.n_keep });
1395
1395
options.push_back ({ " *" , " --chunks N" , " max number of chunks to process (default: %d, -1 = all)" , params.n_chunks });
1396
1396
options.push_back ({ " *" , " -fa, --flash-attn" , " enable Flash Attention (default: %s)" , params.flash_attn ? " enabled" : " disabled" });
1397
- options.push_back ({ " *" , " -p, --prompt PROMPT" , " prompt to start generation with (default: '%s')" , params.prompt .c_str () });
1397
+ options.push_back ({ " *" , " -p, --prompt PROMPT" , " prompt to start generation with\n "
1398
+ " in conversation mode, this will be used as system prompt\n "
1399
+ " (default: '%s')" , params.prompt .c_str () });
1398
1400
options.push_back ({ " *" , " -f, --file FNAME" , " a file containing the prompt (default: none)" });
1399
1401
options.push_back ({ " *" , " --in-file FNAME" , " an input file (repeat to specify multiple files)" });
1400
1402
options.push_back ({ " *" , " -bf, --binary-file FNAME" , " binary file containing the prompt (default: none)" });
@@ -1409,7 +1411,9 @@ void gpt_params_print_usage(int /*argc*/, char ** argv, const gpt_params & param
1409
1411
" halt generation at PROMPT, return control in interactive mode\n "
1410
1412
" can be specified more than once for multiple prompts" });
1411
1413
options.push_back ({ " main" , " -sp, --special" , " special tokens output enabled (default: %s)" , params.special ? " true" : " false" });
1412
- options.push_back ({ " main" , " -cnv, --conversation" , " run in conversation mode (does not print special tokens and suffix/prefix, use default chat template) (default: %s)" , params.conversation ? " true" : " false" });
1414
+ options.push_back ({ " main" , " -cnv, --conversation" , " run in conversation mode, does not print special tokens and suffix/prefix\n "
1415
+ " if suffix/prefix are not specified, default chat template will be used\n "
1416
+ " (default: %s)" , params.conversation ? " true" : " false" });
1413
1417
options.push_back ({ " main infill" , " -i, --interactive" , " run in interactive mode (default: %s)" , params.interactive ? " true" : " false" });
1414
1418
options.push_back ({ " main infill" , " -if, --interactive-first" , " run in interactive mode and wait for input right away (default: %s)" , params.interactive_first ? " true" : " false" });
1415
1419
options.push_back ({ " main infill" , " -mli, --multiline-input" , " allows you to write or paste multiple lines without ending each in '\\ '" });
@@ -1453,6 +1457,7 @@ void gpt_params_print_usage(int /*argc*/, char ** argv, const gpt_params & param
1453
1457
options.push_back ({ " main" , " --cfg-scale N" , " strength of guidance (default: %.1f, 1.0 = disable)" , (double )sparams.cfg_scale });
1454
1458
options.push_back ({ " main" , " --chat-template JINJA_TEMPLATE" ,
1455
1459
" set custom jinja chat template (default: template taken from model's metadata)\n "
1460
+ " if suffix/prefix are specified, template will be disabled\n "
1456
1461
" only commonly used templates are accepted:\n "
1457
1462
" https://github.com/ggerganov/llama.cpp/wiki/Templates-supported-by-llama_chat_apply_template" });
1458
1463
options.push_back ({ " grammar" });
0 commit comments