Friday, April 04, 2025

Philosophizing AGI

 kw: simulated intelligence, ai, artificial intelligence, companies, philosophy, artificial ethics


In the recent issue of Wired, in an article about the company Anthropic, where the founders plan to develop AGI (artificial general intelligence), this photo and caption appear. The caption reads, "Amanda Askell / A trained philosopher who helps manage Claude's personality". Claude is the AI agent that the Anthropic folks are trying to develop into an AGI that is benevolent and ethical.

The first thought I had was, "Trained philosopher? Huh! What does an untrained philosopher look like?" My inner philosopher immediately replied, "Like a human being."

My second thought: "Who decides what is ethical?" In a hyper-divided America, struggling to stay afloat sociologically in a chaotic world, we find this spectrum (not at all autistic…):

  • Radical (these days, Woke Leftists) - The bleeding-edge elites define ethics, to which you'd better kowtow, or else.
  • Liberal - Liberty, the most freedom for the greatest number, favoring plenty of government care and oversight.
  • Moderate - "Leave us alone."
  • Conservative - Don't change what works; keep government out of most affairs.
  • Reactionary - Whatever I say is good, is good. Contradict me at your own peril.

This doesn't even touch on religions, which have their own ethical standards, based on whatever god or scripture they believe. BUT! One thing is for darn sure: I don't want any trace of "what is ethical" to be decided within government.

No comments: