You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When trying to use deepseek coder (via ollama) and its tokenizer and tokens for fim, the result seems completely irrelevant (or, maybe, cut off). However, when using the prompt I would expect to go to the model directly in the ollama, everything works fine:
vs
Here is my config for llm.nvim:
require("llm").setup({
model="deepseek-coder:1.3b-base",
enable_suggestions_on_startup=true,
accept_keymap="<C-M-j>",
dismiss_keymap="<C-M-k>",
tokens_to_clear= {
"<|endoftext|>",
},
fim= {
enabled=true,
prefix="<|fim▁begin|>",
middle="<|fim▁hole|>",
suffix="<|fim▁end|>"
},
backend="ollama",
debounce_ms=0,
url="http://localhost:11434/api/generate",
context_window=240,
-- cf https://github.com/ollama/ollama/blob/main/docs/api.md#parametersrequest_body= {
-- Modelfile options for the model you useoptions= {
num_predict=4,
temperature=0.2,
top_p=0.95,
},
},
lsp= {
bin_path=vim.api.nvim_call_function("stdpath", { "data" }) .."/mason/bin/llm-ls",
},
tokenizer= {
repository="deepseek-ai/deepseek-vl-1.3b-base", -- not working for some reason
},
})
I believe it is a problem with how llm-ls handles it, but if I am wrong, I will open an issue on the llm.nvim github
The text was updated successfully, but these errors were encountered:
When trying to use deepseek coder (via ollama) and its tokenizer and tokens for fim, the result seems completely irrelevant (or, maybe, cut off). However, when using the prompt I would expect to go to the model directly in the ollama, everything works fine:
vs
Here is my config for llm.nvim:
I believe it is a problem with how llm-ls handles it, but if I am wrong, I will open an issue on the llm.nvim github
The text was updated successfully, but these errors were encountered: