Replies: 10 comments 14 replies
-
I have noticed the same thing. I have a much smaller setup but running Fabric in WSL on same host as Ollama so local only. I also tried creating a custom pattern but still get similar results to what is listed by the OP. I is there a source that could explain how to format a pattern so that local llama3 would output in a structured format? Thanks! |
Beta Was this translation helpful? Give feedback.
-
I had the same problem on llama3 locally. what's weird is that when I give the same prompt to Groq llama3-8b-8192, I get the correct result |
Beta Was this translation helpful? Give feedback.
-
have you tried with llama3 8b running locally with LM Studio ? |
Beta Was this translation helpful? Give feedback.
-
what hardware does everyone have? is this related to gpu limitations by any chance? I am having the same issue. |
Beta Was this translation helpful? Give feedback.
-
I have a 7+ yo GTX 1070 running under Windows 10 WSL that works without any problems.
…On Jun 29, 2024 at 6:34 PM -0400, Asentient ***@***.***>, wrote:
what hardware does everyone have? is this related to gpu limitations by any chance? I am having the same issue.
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you are subscribed to this thread.Message ID: ***@***.***>
|
Beta Was this translation helpful? Give feedback.
-
Llama3 (and any other LLM for that matter) is really a part of a family of models fine-tuned on different datasets. Different fine tunings will result in models that follow certain instructions to different degrees. You might have more luck with the "instruct" variant of llama3: https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct. Of course, how you word and format the prompt and examples matters, but we're trying to abstract that away here. |
Beta Was this translation helpful? Give feedback.
-
I am experiencing the exact same issue. My setup is: |
Beta Was this translation helpful? Give feedback.
-
Hey everyone, check out the potential solution here: https://medium.com/@celobusana/solving-fabric-and-local-ollama-context-issues-a-step-by-step-guide-1d67e443e27e |
Beta Was this translation helpful? Give feedback.
-
I had the exact same problem when I first started using fabric on the command line. The format of a pattern, like extract_wisdom, came out wonderful when using my OpenAI API key, and using their models. When using local models pulled from Ollama, I was only getting a very short summary. I moved on, playing around with the fabric framework, and started stitching (piping) patterns together. Randomly, I tried stitching extract_wisdom together twice, just to see what it would output. Voila! Llama3 output a very nice response, with proper format for that pattern! ex: See if that gives you the proper response and format for that specific pattern (extract_wisdom). This allows for the proper response without changing any parameters. You do, however, have to pipe to the command a second time. I have no idea why this is; just a random discovery. Maybe the prompt isn't being input as a system prompt locally, in Ollama (or whatever you're using for implementing your local models). Or maybe it is another issue entirely. Someone smarter than me will figure it out. This is an amazing framework, so I'm sure it will be addressed in its due time. |
Beta Was this translation helpful? Give feedback.
-
After reading all of the suggestions and trying to resolve it by creating a custom model with num_ctx PARAMETERS like that example: Parameters License I was able to narrow it down to the ollama configuration, because it is working properly with LM Studio using any of the open source models. For people with the same issue I may suggest trying LM Studio until they find another solution for themselves. |
Beta Was this translation helpful? Give feedback.
-
Hi all,
I am playing with the fabric extract wisdom and while the gpt-4 creates a really nice output with all the ideas, insights, quotes, habits, facts, references, takeaway, recommendations, etc. the local LLM produces something like 5-15 sentences without formatting or anything similar to the gpt-4 outputs.
Is there some settings for ollama that I can change or this is LLM limitation and nothing can be changed for it to produce similar results locally?
GPU: nvidia 4090
RAM: 32gb
CPU: intel 11900k
OS: Arch linux
If there are options I can change in the ollama settings or some parameter I can add to the input for it to generate longer text I would be happy to understand :)
The input:
yt --transcript https://www.youtube.com/watch?v=UbDyjIIGaxQ | fabric -m llama3:latest -sp extract_wisdom
The output:
"The video is discussing the concept of fabric, an AI tool that helps users extract surprising, insightful, and interesting information from text content. The speaker, who is also the creator of fabric, shares his thoughts on how to use fabric effectively, including defining what you're trying to do with the tool and using it to distill down overwhelming amounts of information.
The video also touches on the idea of AI augmenting human capabilities, rather than replacing them. The speaker believes that AI can help humans become better by taking current capabilities and increasing them at a faster rate than before.
Additionally, the video discusses how the speaker uses fabric in his daily life, including recording conversations with friends and family, transcribing them with Whisper AI, and then using fabric to extract relevant information from those conversations. He also mentions that he has started using fabric to process Bible study discussions, which has helped him to better retain important details.
The video concludes by highlighting the Obsidian Save feature in fabric, which allows users to save notes directly to their Obsidian note-taking app. The speaker demonstrates how to set up this feature and uses it to save a note from his GC analyzer to Obsidian.
Overall, the video is about exploring the potential of AI tools like fabric to improve human capabilities and make life easier."
Expected outcome should be similar to what gpt-4 outputs if that is possible at all.
Thank you all!
Beta Was this translation helpful? Give feedback.
All reactions