AI Settings
All the settings that control the behavior of Atuin AI are specified in an [ai] section in your config.toml. See the configuration documentation for more detailed information about Atuin's configuration system.
enabled
Default: false
Whether or not the AI feature are enabled. When set to false, the question mark keybinding will output a message with instructions to run atuin setup to enable the feature.
endpoint
Default: null
The address of the Atuin AI endpoint. Used for AI features like command generation. Most users will not need this setting; it is only necessary for custom AI endpoints.
api_token
Default: null
The API token for the Atuin AI endpoint. Used for AI features like command generation. Most users will not need this setting; it is only necessary for custom AI endpoints.
Capabilities
Settings that control what capabilities are sent to the LLM. These are specified under [ai.capabilities].
enable_history_search
Default: true
Whether or not to include the "history search" capability in the context sent to the LLM. This allows the AI to request to search your Atuin history for relevant commands when generating suggestions or answering questions.
Example config
Opening context
Settings that control what context is sent in the opening AI request. These are specified under [ai.opening].
send_cwd
Default: false
Whether or not to include your current working directory in the context sent to the LLM. By default, only your OS and current shell are sent.
Example config
send_last_command
Default: false
Whether or not to send your previous command as context in the initial request, allowing the AI to provide more relevant suggestions.
Example config