Replies: 3 comments 2 replies
-
I am having this exact problem. i have tried many different local models, and they all do this. It might be this needs a huge model to work. like 70b or something. I've tried 13b - 20+ ones as well but same thing. |
Beta Was this translation helpful? Give feedback.
-
If you use local models hosted by lets say ollama you create a model file in ollama and set num_ctx to the amount the model supports. Qwen27b supports 131k for instance. The new modified model us created through the ollama create statement so now make that new model name the default model or call it direct in fabric with the --model switch |
Beta Was this translation helpful? Give feedback.
-
This is to piggyback on what @timrohrbaugh said. Thank you for the input @timrohrbaugh but I couldn't quite get my head around what you were talking about, it did help lead me in the right direction for what to even look for as a solution to the problem. Here is an article another user posted in the issues section that walks through creating a custom model: |
Beta Was this translation helpful? Give feedback.
-
I have installed fabric as per the readme file
And I use ollama with the model llama2
I have set llama2 as default model
The output just provides a summary of the transcript and doesn't follow the pattern extract_wisdom
Can somebody help me with this issue?
Beta Was this translation helpful? Give feedback.
All reactions