Let’s not forget the old, well-read post: Dreams of AI Design. In that essay, Eliezer correctly points out errors in imputing meaning to nonsense by using suggestive names to describe the nonsense. This error is very relevant to understanding the problems facing LessWrong’s intellectual contributions.

A major source of simple-mindedness in AI programs is the use of mnemonics like “understand” or “goal” to refer to programs and data structures. This practice has been inherited from more traditional programming applications, in which it is liberating and enlightening to be able to refer to program structures by their purposes. Indeed, part of the thrust of the structured programming movement is to program entirely in terms of purposes at one level before implementing them by the most convenient of the (presumably many) alternative lower-level constructs.

... If a researcher tries to write an “understanding” program, it isn’t because he has thought of a better way of implementing this well-understood task, but because he thinks he can come closer to writing the first implementation. If he calls the main loop of his program “understand”, he is (until proven innocent) merely begging the question. He may mislead a lot of people, most prominently himself, and enrage a lot of others.

What he should do instead is refer to this main loop as G0034, and see if he can convince himself or anyone else that G0034 implements some part of understanding. Or he could give it a name that reveals its intrinsic properties, like NODE-NET-INTERSECTION-FINDER, it being the substance of his theory that finding intersections in networks of nodes constitutes understanding...

When you say (goal . . ), you can just feel the enormous power at your fingertips. It is, of course, an illusion.1

Of course, Conniver has some glaring wishful primitives, too. Calling “multiple data bases” contexts was dumb. It implies that, say, sentence understanding in context is really easy in this system...

Consider the following terms and phrases:

  • llms are trained to predict/simulate”
  • “Attention mechanism” (in self-attention)
  • “AIs are incentivized to” (when talking about the reward or loss functions, thus implicitly reversing the true causality; reward optimizes the AI, but AI probably won’t optimize the reward)
  • “Reward” (implied to be an attractive quantity to the decision-maker)
    • “Advantage function” and “value function”
    • “The purpose of RL is to train an agent to maximize expected reward over time” (perhaps implying an expectation and inner consciousness on the part of the so-called “agent”)
  • “Agents” (implying volition in our trained artifact... generally ‘cuz we used a technique belonging to the class of algorithms which humans call “reinforcement learning”)
  • “Power-seeking” (AI “agents”)
  • “Shoggoth”
  • “Optimization pressure”
  • “Utility”
    • As opposed to (thinking of it) as “internal unit of decision-making incentivization, which is a function of internal representations of expected future events; minted after the resolution of expected future on-policy inefficiencies relative to the computational artifact’s current decision-making influences”
  • “Discount rate” (in deep RL, implying that an external future-learning-signal multiplier will ingrain itself into the AI’s potential inner plan-grading-function which is conveniently assumed to be additive over timesteps, and also there’s just one such function and also it’s Markovian)
  • “Inner goal /mesa objective /optimization daemon (yes, that was a real name)”
  • “Outer optimizer” (perhaps implying some amount of intentionality; a sense that ‘more’ optimization is ‘better’, even at the expense of generalization of the trained network)
  • “Optimal” (as opposed to equilibrated-under-policy-updates)
  • “Objectives” (when conflating a “loss function as objective” and “something which strongly controls how the AI makes choices”)
  • “Training” (in ML)
    • Yup!
  • “Learning” (in ML)
  • “Simplicity prior”
    • Consider the abundance of amateur theorizing about whether “schemers” will be “simpler” than “saints”, or whether they will be supplanted by “sycophants.” Sometimes conducted in ignorance of actual inductive bias research, which is actually a real subfield of ML.

Lest this all seem merely amusing, meditate on the fate of those who have tampered with words before. The behaviorists ruined words like “behavior”, “response”, and, especially, “learning.” They now play happily in a dream world, internally consistent but lost to science. And think about this: if “mechanical translation” had been called “word-by-word text manipulation”, the people doing it might still be getting government money.

Some of these terms are useful. Some of the academic imports are necessary for successful communication. Some of the terms have real benefits.

That doesn’t stop them from distorting your thinking. At least in your private thoughts, you can do better. You can replace “optimal” with “artifact equilibrated under policy update operations” or “subjectively maximal expected utility relative to [entity X]’s imputed beliefs.” One nice thing about brains is these long phrases can compress into single concepts which you can instantly understand.

For years, bad habits have held back clear thinking in the AI alignment community. Let’s do better.

Black and white trout

Find out when I post more content: newsletter & RSSRSS icon

Thoughts? Email me at alex@turntrout.com

  1. How many times has someone expressed “I’m worried about ‘goal-directed optimizers’, but I’m not sure what exactly they are, so I’m going to work on deconfusion”? There’s something weird about this sentiment, don’t you think?