Skip to content

AI Settings

All the settings that control the behavior of Atuin AI are specified in an [ai] section in your config.toml. See the configuration documentation for more detailed information about Atuin's configuration system.

enabled

Default: false

Whether or not the AI feature are enabled. When set to false, the question mark keybinding will output a message with instructions to run atuin setup to enable the feature.

endpoint

Default: null

The address of the Atuin AI endpoint. Used for AI features like command generation. Most users will not need this setting; it is only necessary for custom AI endpoints.

api_token

Default: null

The API token for the Atuin AI endpoint. Used for AI features like command generation. Most users will not need this setting; it is only necessary for custom AI endpoints.

Capabilities

Settings that control what capabilities are sent to the LLM. These are specified under [ai.capabilities].

Default: true

Whether or not to include the "history search" capability in the context sent to the LLM. This allows the AI to request to search your Atuin history for relevant commands when generating suggestions or answering questions.

Example config

[ai.capabilities]
enable_history_search = false

Opening context

Settings that control what context is sent in the opening AI request. These are specified under [ai.opening].

send_cwd

Default: false

Whether or not to include your current working directory in the context sent to the LLM. By default, only your OS and current shell are sent.

Example config

[ai.opening]
send_cwd = true

send_last_command

Default: false

Whether or not to send your previous command as context in the initial request, allowing the AI to provide more relevant suggestions.

Example config

[ai.opening]
send_last_command = true