this post was submitted on 10 Jul 2025
9 points (84.6% liked)

World News

48359 readers
2183 users here now

A community for discussing events around the World

Rules:

Similarly, if you see posts along these lines, do not engage. Report them, block them, and live a happier life than they do. We see too many slapfights that boil down to "Mom! He's bugging me!" and "I'm not touching you!" Going forward, slapfights will result in removed comments and temp bans to cool off.

We ask that the users report any comment or post that violate the rules, to use critical thinking when reading, posting or commenting. Users that post off-topic spam, advocate violence, have multiple comments or posts removed, weaponize reports or violate the code of conduct will be banned.

All posts and comments will be reviewed on a case-by-case basis. This means that some content that violates the rules may be allowed, while other content that does not violate the rules may be removed. The moderators retain the right to remove any content and ban users.


Lemmy World Partners

News !news@lemmy.world

Politics !politics@lemmy.world

World Politics !globalpolitics@lemmy.world


Recommendations

For Firefox users, there is media bias / propaganda / fact check plugin.

https://addons.mozilla.org/en-US/firefox/addon/media-bias-fact-check/

founded 2 years ago
MODERATORS
 

The U.S. government recently announced plans to incorporate artificial intelligence into troop command systems, a move that promises enhanced efficiency and responsiveness. However, news of OpenAI tightening its safety protocols raises a critical question: how much trust can we place in technologies still needing refinement and protection against potential threats?

Despite advancements in safety, even cutting-edge AI systems remain vulnerable to manipulation or errors. In military operations, such failures could be catastrophic. This begs the question: is it justified to entrust such sensitive functions to algorithms, no matter how advanced?

Integrating AI into the military isn’t just a step toward the future — it’s a decision demanding rigorous risk assessment. Given past failures with innovations in other fields, can we rely on AI when national security is at stake?

you are viewing a single comment's thread
view the rest of the comments
[–] Jumuta@sh.itjust.works 1 points 3 days ago

do two wrongs make a right?