ollama settings

This commit is contained in:
m3tam3re 2024-06-04 12:47:41 +02:00
parent d581be532d
commit 4d9313407b

View File

@ -829,7 +829,7 @@ require('lazy').setup({
{ {
'David-Kunz/gen.nvim', 'David-Kunz/gen.nvim',
opts = { opts = {
model = 'dolphin-mistral', -- The default model to use. model = 'dolphin-llama3', -- The default model to use.
host = 'localhost', -- The host running the Ollama service. host = 'localhost', -- The host running the Ollama service.
port = '11434', -- The port on which the Ollama service is listening. port = '11434', -- The port on which the Ollama service is listening.
display_mode = 'split', -- The display mode. Can be "float" or "split". display_mode = 'split', -- The display mode. Can be "float" or "split".
@ -863,6 +863,7 @@ require('lazy').setup({
extract = '```$filetype\n(.-)```', extract = '```$filetype\n(.-)```',
model = 'codellama:latest', model = 'codellama:latest',
} }
require('gen').select_model()
end, end,
vim.keymap.set('n', '<leader>lo', vim.cmd.Gen, { desc = '[O]llama AI' }), vim.keymap.set('n', '<leader>lo', vim.cmd.Gen, { desc = '[O]llama AI' }),
}, },