this post was submitted on 19 Nov 2024
1019 points (97.7% liked)
People Twitter
5383 readers
943 users here now
People tweeting stuff. We allow tweets from anyone.
RULES:
- Mark NSFW content.
- No doxxing people.
- Must be a tweet or similar
- No bullying or international politcs
- Be excellent to each other.
- Provide an archived link to the tweet (or similar) being shown if it's a major figure or a politician.
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
It depends upon what you use ChatGPT for and if you know how to use it productively. For example if I ask ChatGPT coding questions it is often very helpful. If I ask it history questions it constantly makes things up. You also again need to know how to use it, like people who claim ChatGPT is not helpful for coding you ask them how they use it and they basically just ask ChatGPT to do their whole project for them and when it fails they claim it is useless. But that's not the productive way to use it, the productive way to use it is like a replacement for StackOverflow or to provide you examples of how to use some library, or things like that, not doing your whole project for you. Of course, people often use it incorrectly so it's probably not a good idea to allow its use in the workplace, but for individual use it can be very helpful.
For coding it heavily depends on the language. For example, it's quite decent at writing C#, but whenever I try to ask it any question about rust, it's either flat out wrong or doesn't even fucking compile.
Also found it most useful when I know exactly what I want, just don't know the syntax. Like when I was writing C# code generation for the first time. Also unsurprisingly sucks at working with libraries.
I used it today to find out how to do something on my Juniper that would have taken 45 minutes of sifting bullshit documentation. One question and I figured it out in 2 minutes.
This is similar to gabe Newell's idea of piracy. This is a convenience issue. And GPT solves some of it.
And thank god it doesn't get them all the way there, because if it were able to completely do everything accurately with the level of ambiguous prompts the layperson gives it, anyone technical would essentially be out of a job.
And honestly, the world would be better off not making people complacent just being end users of everything, and instead have to have a modicum of understanding what they are doing.
I used to think its just neophobia having all these kids using smart phones and touch screens for everything at increasingly earlier ages, but its like they only know how to use/consume things, never an inkling of trying to tinker with things and understand how to repurpose the mechanisms , figure out how things work (tbf everything now is super integrated, much harder to repair).
It just doesn't bode well to me when it seems like the future labor force is so disconnected from the underlying systems they use.