You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi there! I'm currently experimenting with llama.cpp to summarize large text files. My approach involves splitting a large text file into smaller chunks and adding chat format markers at the beginning and end of each chunk.
My current setup demands reloading the model for each promote, which is not efficient. I'm seeking a solution to automate parsing from a text file without the need for reloading the model each time. I haven't come across any command-line options that handle a single shot answer for multiple promote within a single file or allow passing multiple files to automate the parsing process.
If there is a way through the system command prompt or via llama.cpp, I would appreciate guidance.
@echooffREM Check if directory argument is providedif"%~1"=="" (
echo Usage: %~nx0 directory
exit /b
)
REM Path to the programsetPROGRAM="main.exe"REM Desired parameters for the programsetPARAMETERS=-c 2048 --temp 0.0 --top_p 0.0 --top_k 1.0 -n -1 -m Phi-3-mini-128k-instruct.Q8_0.gguf
REM Directory containing the text filessetFILES_DIR=%~1REM Navigate to the specified directorycd /d FILES_DIR
REM Initialize response.txttypenul> response.txt
REM Process each part filefor%%iin (*_part*.txt) do (
echo Processing %%i...
REM Call the program with the specified parameters and input file%PROGRAM%%PARAMETERS% -f "%%i"|findstr /R /C:"<|assistant|>">> response.txt
)
echo All files processed.
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
-
Hi there! I'm currently experimenting with llama.cpp to summarize large text files. My approach involves splitting a large text file into smaller chunks and adding chat format markers at the beginning and end of each chunk.
My current setup demands reloading the model for each promote, which is not efficient. I'm seeking a solution to automate parsing from a text file without the need for reloading the model each time. I haven't come across any command-line options that handle a single shot answer for multiple promote within a single file or allow passing multiple files to automate the parsing process.
If there is a way through the system command prompt or via llama.cpp, I would appreciate guidance.
Beta Was this translation helpful? Give feedback.
All reactions