Configuration
The application may be configured by editing the settings in Config.toml. The main settings are:
-
embeddings: the model to create embeddings.
- dense_model: the model to create dense embeddings. For example, OpenAI embedding are set by "OpenAI/text-embedding-3-small"
- sparse_model: the model for sparse embeddings. The sparse embeddings supported by Qdrant are used, i.e. "Qdrant/bm25"
-
major: the main language model used, for example, for chatting.
- model: the provider and the model name, e.g. "OpenAI/gpt-4.1-mini"
- temperature: temperature (defaults to 0.1)
- max_retries: number of connection attempts when using model (default to 2)
-
major.provider_params: other parameters set for the model (model-dependent)
-
minor: the accessory language model used, for example, for creating questions and for creating summaries.
- model: the provider and the model name, e.g. "OpenAI/gpt-4.1-nano"
- temperature: temperature (defaults to 0.1)
- max_retries: number of connection attempts when using model (default to 2)
-
major.provider_params: other parameters set for the model (model-dependent)
-
aux: the auxiliary language model used, for example to classify text. This model should have low latency.
- model: the provier and the model name, e.g. "Mistral/mistral-small-latest"
- temperature: temperature (defaults to 0.7)
- max_retries: number of connection attempts when using model (default to 2)
-
aux.provider_params: other parameters set for the model (model-dependent)