this post was submitted on 22 Jun 2025
104 points (73.4% liked)

Programming Humor

3193 readers
1 users here now

Related Communities !programmerhumor@lemmy.ml !programmer_humor@programming.dev !programmerhumor@kbin.social !programming_horror@programming.dev

Other Programming Communities !programming@beehaw.org !programming@programming.dev !programming@lemmy.ml !programming@kbin.social !learn_programming@programming.dev !functional_programming@programming.dev !embedded_prog@lemmy.ml

founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] spankmonkey@lemmy.world 2 points 6 days ago (2 children)

MY biggest disappointment with how AI is being implemented is the inability to incorporate context specific execution if small programs to emulate things like calculators and chess programs. Like why does it doe the hard mode approach to literally everything? When asked to do math why doesn't it execute something that emulates a calculator?

[–] otacon239@lemmy.world 3 points 6 days ago

I’ve been waiting for them to make this improvement since they were first introduced. Any day now…

[–] Ephera@lemmy.ml 1 points 5 days ago

That's definitely being done. It's referred to as "tool calling" or "function calling": https://python.langchain.com/docs/how_to/tool_calling/

This isn't as potent as one might think, because:

  1. each tool needs to be hooked up and described extensively.
  2. the naive approach where the LLM generates heaps of text when calling these tools, for example to describe the entire state of the chessboard as JSON or CSV, is unreliable, because text generation is unreliable.
  3. smarter approaches, like having an external program keeping track of the chessboard state and sending it to a chess engine, so that the LLM only has to forward the move that the user described, don't really make sense to incorporate into a general-purpose language model. You can find chess chatbots on the internet, though.

But all-in-all, it is a path forward where the LLMs could just do the semantics and then call a different tool for each thinky job, serving at least as a user interface.
The hope is for it to also serve as glue between these tools, automatically calling the right tools and passing their output into other tools. I believe, the next step in this direction is "agentic AI", but I haven't yet managed to cut through the buzzword soup to figure out what that actually means.