mirror of
https://github.com/nomic-ai/gpt4all.git
synced 2025-06-21 13:10:35 +00:00
Clean up the docs a bit.
This commit is contained in:
parent
832b5d1a96
commit
ac7ecd2cef
@ -72,26 +72,26 @@ void llmodel_llama_destroy(llmodel_model llama);
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Load a model from a file.
|
* Load a model from a file.
|
||||||
* @param model A pointer to the LLMODEL_C instance.
|
* @param model A pointer to the llmodel_model instance.
|
||||||
* @param modelPath A string representing the path to the model file.
|
* @param model_path A string representing the path to the model file.
|
||||||
* @return true if the model was loaded successfully, false otherwise.
|
* @return true if the model was loaded successfully, false otherwise.
|
||||||
*/
|
*/
|
||||||
bool llmodel_loadModel(llmodel_model model, const char *model_path);
|
bool llmodel_loadModel(llmodel_model model, const char *model_path);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Check if a model is loaded.
|
* Check if a model is loaded.
|
||||||
* @param model A pointer to the LLMODEL_C instance.
|
* @param model A pointer to the llmodel_model instance.
|
||||||
* @return true if the model is loaded, false otherwise.
|
* @return true if the model is loaded, false otherwise.
|
||||||
*/
|
*/
|
||||||
bool llmodel_isModelLoaded(llmodel_model model);
|
bool llmodel_isModelLoaded(llmodel_model model);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Generate a response using the model.
|
* Generate a response using the model.
|
||||||
* @param model A pointer to the LLMODEL_C instance.
|
* @param model A pointer to the llmodel_model instance.
|
||||||
* @param prompt A string representing the input prompt.
|
* @param prompt A string representing the input prompt.
|
||||||
* @param response A callback function for handling the generated response.
|
* @param response A callback function for handling the generated response.
|
||||||
* @param recalculate A callback function for handling recalculation requests.
|
* @param recalculate A callback function for handling recalculation requests.
|
||||||
* @param ctx A pointer to the PromptContext_C structure.
|
* @param ctx A pointer to the llmodel_prompt_context structure.
|
||||||
*/
|
*/
|
||||||
void llmodel_prompt(llmodel_model model, const char *prompt,
|
void llmodel_prompt(llmodel_model model, const char *prompt,
|
||||||
llmodel_response_callback response,
|
llmodel_response_callback response,
|
||||||
@ -100,14 +100,14 @@ void llmodel_prompt(llmodel_model model, const char *prompt,
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Set the number of threads to be used by the model.
|
* Set the number of threads to be used by the model.
|
||||||
* @param model A pointer to the LLMODEL_C instance.
|
* @param model A pointer to the llmodel_model instance.
|
||||||
* @param n_threads The number of threads to be used.
|
* @param n_threads The number of threads to be used.
|
||||||
*/
|
*/
|
||||||
void llmodel_setThreadCount(llmodel_model model, int32_t n_threads);
|
void llmodel_setThreadCount(llmodel_model model, int32_t n_threads);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Get the number of threads currently being used by the model.
|
* Get the number of threads currently being used by the model.
|
||||||
* @param model A pointer to the LLMODEL_C instance.
|
* @param model A pointer to the llmodel_model instance.
|
||||||
* @return The number of threads currently being used.
|
* @return The number of threads currently being used.
|
||||||
*/
|
*/
|
||||||
int32_t llmodel_threadCount(llmodel_model model);
|
int32_t llmodel_threadCount(llmodel_model model);
|
||||||
|
Loading…
Reference in New Issue
Block a user