• micheal65536@lemmy.micheal65536.duckdns.org
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    1 year ago

    WizardLM 13B (I didn’t notice any significant improvement with the 30B version), tends to be a bit confined to a standard output format at the expense of accuracy (e.g. it will always try to give both sides of an argument even if there isn’t another side or the question isn’t an argument at all) but is good for simple questions

    LLaMa 2 13B (not the chat tuned version), this one takes some practice with prompting as it doesn’t really understand conversation and won’t know what it’s supposed to do unless you make it clear from contextual clues, but it feels refreshing to use as the model is (as far as is practical) unbiased/uncensored so you don’t get all the annoying lectures and stuff

  • librecat@lemmy.basedcount.com
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Anything based on llama2 tbh. It’s fast enough and logical enough to handle the kinds of programming related tasks I want to use a llm for (writing boilerplate code, generating placeholder data, simple refactoring). With the release of the vicuna and codellama models things are getting even better.

  • DrakeRichards@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    I do image generation for RPGs, so AZovya’s RPG v3 model is easily my favorite. It does a wide range of styles very well and understands a lot of RPG-specific tokens. I’m really hoping they update it for SDXL, because all of the models I’ve seen so far are disappointing compared to what’s available with SD 1.5.

    I don’t have an answer for LLMs, but I’m curious what others will reply with. Aren’t there only like… 3 or 4 models in common use for LLMs? I’m used to having hundreds to pick from with Stable Diffusion; I don’t think I understand how LLM models are different.

    • micheal65536@lemmy.micheal65536.duckdns.org
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      There are only a few popular LLM models. A few more if you count variations such as “uncensored” etc. Most of the others tend to not perform well or don’t have much difference from the more popular ones.

      I would think that the difference is likely for two reasons:

      • LLMs require more effort in curating the dataset for training. Whereas a Stable Diffusion model can be trained by grabbing a bunch of pictures of a particular subject or style and throwing them in a directory, an LLM requires careful gathering and reformatting of text. If you want an LLM to write dialog for a particular character, for example, you would need to try to find or write a lot of existing dialog for that character, which is generally harder than just searching for images on the internet.

      • LLMs are already more versatile. For example, most of the popular LLMs will already write dialog for a particular character (or at least attempt to) just by being given a description of the character and possibly a short snippet of sample dialog. Fine-tuning doesn’t give any significant performance improvement in that regard. If you want the LLM to write in a specific style, such as Old English, it is usually sufficient to just instruct it to do so and perhaps prime the conversation with a sentence or two written in that style.