You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
options:
-h, --help show this help message and exit
-i, --interactive run in interactive mode
--interactive-first run in interactive mode and wait for input right away
-ins, --instruct run in instruction mode (use with Alpaca models)
-r PROMPT, --reverse-prompt PROMPT
run in interactive mode and poll user input upon seeing PROMPT (can be
specified more than once for multiple prompts).
--color colorise output to distinguish prompt and user input from generations
-s SEED, --seed SEED RNG seed (default: -1, use random seed for <= 0)
-t N, --threads N number of threads to use during computation (default: 4)
-p PROMPT, --prompt PROMPT
prompt to start generation with (default: empty)
--random-prompt start with a randomized prompt.
--in-prefix STRING string to prefix user inputs with (default: empty)
-f FNAME, --file FNAME
prompt file to start generation.
-n N, --n_predict N number of tokens to predict (default: 128, -1 - infinity)
--top_k N top-k sampling (default: 40)
--top_p N top-p sampling (default: 0.9)
--repeat_last_n N last n tokens to consider for penalize (default: 64)
--repeat_penalty N penalize repeat sequence of tokens (default: 1.1)
-c N, --ctx_size N size of the prompt context (default: 512)
--ignore-eos ignore end of stream token and continue generating
--memory_f32 use f32 instead of f16 for memory key+value
--temp N temperature (default: 0.8)
--n_parts N number of model parts (default: -1 = determine from dimensions)
-b N, --batch_size N batch size for prompt processing (default: 8)
--perplexity compute perplexity over the prompt
--keep number of tokens to keep from the initial prompt
--mlock force system to keep model in RAM rather than swapping or compressing
--mtest compute maximum memory usage
--verbose-prompt print prompt before generation
-m FNAME, --model FNAME
model path (default: ./models/7B/ggml-model-q4_0.bin)
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
-
Hi,
Can you add more explanations and examples ?
options:
-h, --help show this help message and exit
-i, --interactive run in interactive mode
--interactive-first run in interactive mode and wait for input right away
-ins, --instruct run in instruction mode (use with Alpaca models)
-r PROMPT, --reverse-prompt PROMPT
run in interactive mode and poll user input upon seeing PROMPT (can be
specified more than once for multiple prompts).
--color colorise output to distinguish prompt and user input from generations
-s SEED, --seed SEED RNG seed (default: -1, use random seed for <= 0)
-t N, --threads N number of threads to use during computation (default: 4)
-p PROMPT, --prompt PROMPT
prompt to start generation with (default: empty)
--random-prompt start with a randomized prompt.
--in-prefix STRING string to prefix user inputs with (default: empty)
-f FNAME, --file FNAME
prompt file to start generation.
-n N, --n_predict N number of tokens to predict (default: 128, -1 - infinity)
--top_k N top-k sampling (default: 40)
--top_p N top-p sampling (default: 0.9)
--repeat_last_n N last n tokens to consider for penalize (default: 64)
--repeat_penalty N penalize repeat sequence of tokens (default: 1.1)
-c N, --ctx_size N size of the prompt context (default: 512)
--ignore-eos ignore end of stream token and continue generating
--memory_f32 use f32 instead of f16 for memory key+value
--temp N temperature (default: 0.8)
--n_parts N number of model parts (default: -1 = determine from dimensions)
-b N, --batch_size N batch size for prompt processing (default: 8)
--perplexity compute perplexity over the prompt
--keep number of tokens to keep from the initial prompt
--mlock force system to keep model in RAM rather than swapping or compressing
--mtest compute maximum memory usage
--verbose-prompt print prompt before generation
-m FNAME, --model FNAME
model path (default: ./models/7B/ggml-model-q4_0.bin)
thank you
Beta Was this translation helpful? Give feedback.
All reactions