this post was submitted on 11 Jun 2023
1 points (100.0% liked)
Technology
37727 readers
619 users here now
A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.
Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.
Subcommunities on Beehaw:
This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I've been using LLMs a lot. I use gpt 4 to help edit articles, answer nagging questions I can't be bothered to answer, and other random things, such as cooking advice.
It's fair to say, I believe, that all general purpose LLMs like this are plagiarizing all of the time. Much in the way my friend Patrick doesn't give me sources for all of his opinions, Gpt 4 doesn't tell me where it got its info on baked corn. The disadvantage of this, is that I can't trust it any more than I can trust Patrick. When it's important, I ALWAYS double check. The advantage is I don't have to take the time to compare, contrast, and discover sources. It's a trade off.
From my perspective, The theoretical advantage of bing or Google's implementation is ONLY that they provide you with sources. I actually use Bing's implementation of gpt when I want a quick, real world reference to an answer.
Google will be making a big mistake by sidelining it's sources when open source LLMs are already overtaking Google's bard's ai in quality. Why get questionable advice from Google, when I can get slightly less questionable advice from gpt, my phone assistant, or actual, inline citations from bing?