this post was submitted on 30 Jul 2024
1200 points (98.0% liked)

linuxmemes

21222 readers
41 users here now

Hint: :q!


Sister communities:


Community rules (click to expand)

1. Follow the site-wide rules

2. Be civil
  • Understand the difference between a joke and an insult.
  • Do not harrass or attack members of the community for any reason.
  • Leave remarks of "peasantry" to the PCMR community. If you dislike an OS/service/application, attack the thing you dislike, not the individuals who use it. Some people may not have a choice.
  • Bigotry will not be tolerated.
  • These rules are somewhat loosened when the subject is a public figure. Still, do not attack their person or incite harrassment.
  • 3. Post Linux-related content
  • Including Unix and BSD.
  • Non-Linux content is acceptable as long as it makes a reference to Linux. For example, the poorly made mockery of sudo in Windows.
  • No porn. Even if you watch it on a Linux machine.
  • 4. No recent reposts
  • Everybody uses Arch btw, can't quit Vim, and wants to interject for a moment. You can stop now.

  • Please report posts and comments that break these rules!

    founded 1 year ago
    MODERATORS
    1200
    submitted 3 months ago* (last edited 3 months ago) by Smokeydope@lemmy.world to c/linuxmemes@lemmy.world
     

    List of icons/services suggested:

    • Calibre
    • Jitsi
    • Kiwix
    • Monero (Node)
    • Nextcloud
    • Pihole
    • Ollama (Should at least be able to run tiny-llama 1.1B)
    • Open Media Vault
    • Syncthing
    • VLC Media Player Media Server
    you are viewing a single comment's thread
    view the rest of the comments
    [–] Smokeydope@lemmy.world 1 points 3 months ago* (last edited 3 months ago) (2 children)

    Thank you thats useful to know. In your opinion what context size is the sweet spot for llama 3.1 8B and similar models?

    [–] brucethemoose@lemmy.world 1 points 3 months ago (1 children)

    4 core i7, 16gb RAM and no GPU yet

    Honestly as small as you can manage.

    Again, you will get much better speeds out of "extreme" MoE models like deepseek chat lite: https://huggingface.co/YorkieOH10/DeepSeek-V2-Lite-Chat-Q4_K_M-GGUF/tree/main

    Another thing I'd recommend is running kobold.cpp instead of ollama if you want to get into the nitty gritty of llms. Its more customizable and (ultimately) faster on more hardware.

    [–] Smokeydope@lemmy.world 1 points 3 months ago* (last edited 3 months ago) (1 children)

    Thats good info for low spec laptops. Thanks for the software recommendation. Need to do some more research on the model you suggested. I think you confused me for the other guy though. Im currently working with a six core ryzen 2600 CPU and a RX 580 GPU. edit- no worries we are good it was still great info for the thinkpad users!

    [–] brucethemoose@lemmy.world 1 points 3 months ago

    8GB or 4GB?

    Yeah you should get kobold.cpp's rocm fork working if you can manage it, otherwise use their vulkan build.

    llama 8b at shorter context is probably good for your machine, as it can fit on the 8GB GPU at shorter context, or at least be partially offloaded if its a 4GB one.

    I wouldn't recommend deepseek for your machine. It's a better fit for older CPUs, as it's not as smart as llama 8B, and its bigger than llama 8B, but it just runs super fast because its an MoE.

    [–] brucethemoose@lemmy.world 1 points 3 months ago* (last edited 3 months ago)

    Oh I got you mixed up with the other commenter, apologies.

    I'm not sure when llama 8b starts to degrade at long context, but I wanna say its well before 128K, and where other "long context" models start to look much more attractive depending on the task. Right now I am testing Amazon's mistral finetune, and it seems to be much better than Nemo or llama 3.1 out there.