About 26,200,000 results
Open links in new tab
  1. Question about privacy on local models running on LM Studio

    Nov 5, 2023 · Question about privacy on local models running on LM Studio Question | Help It appears that running the local models on personal computers is fully private and they cannot connect to …

  2. LM Studio, Which model to use with rtx 3060 ? : r/LocalLLaMA - Reddit

    Nov 3, 2023 · Hi everyone, Pardon my ignorance I'm new around here. Since yesterday, I was looking for a GPT4 alternative so I downloaded LM Studio with - speechless llama2 hermes orca platypus …

  3. Correct way to setup character cards in LM Studio? : r/LocalLLaMA

    Oct 11, 2023 · Character cards are just pre-prompts. So use the pre-prompt/system-prompt setting and put your character info in there. LM studio doesn't have support for directly importing the cards/files …

  4. LLM Web-UI recommendations : r/LocalLLaMA - Reddit

    Extensions with LM studio are nonexistent as it’s so new and lacks the capabilities. Lollms-webui might be another option. Or plug one of the others that accepts chatgpt and use LM Studios local server …

  5. Re-use already downloaded models? : r/LMStudio - Reddit

    Jan 4, 2024 · trueIn the course of testing many AI tools I have downloaded already lots of models and saved them to a dedicated location on my computer. I would like to re-use them instead of re …

  6. What is considered the best uncensored LLM right now?

    WizardLM is really old by now. Have you tried any of the Mistral finetunes? Don't discount it just because of the low parameter count. I was also running WizardLM-33b-4bit for the longest time, but Mistral …

  7. LM Studio Alternative that Supports Custom GPU Offloading : r

    Apr 7, 2024 · 14 votes, 29 comments. Are there any open-source UI alternatives to LM Studio that allows to set how many layers to offload to GPU. I have tried the…

  8. Why do people say LM Studio isn't open-sourced? - Reddit

    LM Studio is a really good application developed by passionate individuals which shows in the quality. There is nothing inherently wrong with it or using closed source. Use it because it is good and show …

  9. Failed to load model Running LMStudio ? : r/LocalLLaMA - Reddit

    Dec 3, 2023 · Personally for me helped to update Visual Studio. I.e. exactly what Arkonias told below Your C++ redists are out of date and need updating.

  10. Why ollama faster than LMStudio? : r/LocalLLaMA - Reddit

    Apr 11, 2024 · There's definitely something wrong with LM Studio. I've tested it against Ollama using OpenWebUI using the same models. It's dogshit slow compared to Ollama. It's closed source, so …