fishynoob

joined 1 week ago
[โ€“] fishynoob@infosec.pub 1 points 1 week ago

That means it's likely a problem with DNS.

[โ€“] fishynoob@infosec.pub 3 points 1 week ago (8 children)

Send the link to the discussion and the screenshot of your comment

[โ€“] fishynoob@infosec.pub 1 points 1 week ago

Yeah I'm not going to run them on CPUs, that's not going to be very good. I'll buy the GPUs when I can.

[โ€“] fishynoob@infosec.pub 1 points 1 week ago (2 children)

Yes, just thought if you could check that the correct ports are opened. I.e. is port 443 open for NGINX on Unraid? Is NGINX forwarding traffic to the correct port to your backend? Is the backend configured to allow traffic on a certain domain/all domains if it is handling HTTPS?

[โ€“] fishynoob@infosec.pub 1 points 1 week ago (1 children)

Has anybody here actually worked for a company that uses Qubes OS as their corporate OS? I think Qubes is awesome and with some work can definitely be used for corporate work but most people don't know, don't care and are scared of security

[โ€“] fishynoob@infosec.pub 2 points 1 week ago (4 children)

Don't know why people care about overwriting their comments. It's not like Reddit and Discord lose your content anyway. Having my content sold by Reddit is just as bad as OpenAI or Discord, so I just don't say private stuff on those platforms

[โ€“] fishynoob@infosec.pub 1 points 1 week ago

Thank you, that makes sense. Yes, I will look to create templates using AI that I like. Thanks again for the help

[โ€“] fishynoob@infosec.pub 2 points 1 week ago* (last edited 1 week ago)

Thanks for the edit. You have a very intriguing idea; a second LLM in the background with a summary of the conversation + static context might make performance a lot better. I don't know if anyone has implemented it/knows how one can DIY it with Kobold/Ollama. I think it is an amazing idea for code assistants too if you're doing a long coding session.

[โ€“] fishynoob@infosec.pub 4 points 1 week ago

Better be AGPL or she's never getting cloned on my PC, that's for sure!

[โ€“] fishynoob@infosec.pub 2 points 1 week ago

I see. Thanks for the note. I think beyond 48GB of VRAM diminishing returns set in very quickly so I'll likely stick to that limit. I wouldn't want to use models hosted in the cloud so that's out of the question.

[โ€“] fishynoob@infosec.pub 1 points 1 week ago

Absolutely. TheBloke's fine-tuned models with their guardrails removed are the only conversational models I will run. I get enraged looking at AI telling me to curb my speech.

I do use Python but I haven't touched AI yet so it's going to be a learning-curve if I go down that route. I am hoping to get finetuned models OOTB for this kind of stuff but I know it's a hard ask.

I was going to buy 2-3 used GPUs/new budget GPUs like the B580 but with the tariffs the prices of these are INFLATED beyond what I can afford to pay for them. Once something changes (financially speaking) I'll probably throw enough VRAM at it to at least get the 8B models (probably not FP16 but maybe quantised to 4K/8K) running smoothly.

Thanks for the reminder. I have wanted to use character AI for so long but couldn't bear to give away my thought patterns to them (look at my hypocrisy: I'm giving it all away anyway when everyone is free to scrape Lemmy). I guess I'm an idiot.

[โ€“] fishynoob@infosec.pub 1 points 1 week ago* (last edited 1 week ago) (2 children)

I was going to buy the ARC B580s when they come back down in price, but with the tariffs I don't think I'll ever see them at MSRP. Even the used market is very expensive. I'll probably hold off on buying GPUs for a few more months till I can afford the higher prices/something changes. Thanks for the Lexi V2 suggestion

view more: โ€น prev next โ€บ