nvim changes
This commit is contained in:
parent
951c29c4c8
commit
5f84e0dcff
@ -94,7 +94,9 @@ vim.g.maplocalleader = ' '
|
||||
vim.g.have_nerd_font = true
|
||||
|
||||
-- [[ Setting options ]]
|
||||
-- l
|
||||
-- See `:help vim.opt`
|
||||
--
|
||||
-- NOTE: You can change these options as you wish!
|
||||
-- For more options, you can see `:help option-list`
|
||||
|
||||
@ -829,13 +831,18 @@ require('lazy').setup({
|
||||
{
|
||||
'David-Kunz/gen.nvim',
|
||||
opts = {
|
||||
model = 'dolphin-llama3', -- The default model to use.
|
||||
model = 'dolphin-llama3:latest', -- The default model to use.
|
||||
quit_map = 'q', -- set keymap to close the response window
|
||||
retry_map = '<c-r>', -- set keymap to re-send the current prompt
|
||||
accept_map = '<c-cr>', -- set keymap to replace the previous selection with the last result
|
||||
host = 'localhost', -- The host running the Ollama service.
|
||||
port = '11434', -- The port on which the Ollama service is listening.
|
||||
display_mode = 'float', -- The display mode. Can be "float" or "split".
|
||||
show_prompt = true, -- Shows the Prompt submitted to Ollama.
|
||||
show_model = true, -- Displays which model you are using at the beginning of your chat session.
|
||||
display_mode = 'float', -- The display mode. Can be "float" or "split" or "horizontal-split".
|
||||
show_prompt = false, -- Shows the prompt submitted to Ollama.
|
||||
show_model = false, -- Displays which model you are using at the beginning of your chat session.
|
||||
no_auto_close = false, -- Never closes the window automatically.
|
||||
file = false, -- Write the payload to a temporary file to keep the command short.
|
||||
hidden = false, -- Hide the generation window (if true, will implicitly set `prompt.replace = true`), requires Neovim >= 0.10
|
||||
init = function(options)
|
||||
pcall(io.popen, 'ollama serve > /dev/null 2>&1 &')
|
||||
end,
|
||||
|
Loading…
x
Reference in New Issue
Block a user