I teach a course called “Human and Machine Intelligence Alignment.” The basic idea is that there are productive analogies to be made in looking at
I teach a course called “Human and Machine Intelligence Alignment.” The basic idea is that there are productive analogies to be made in looking at
I’ve been working with Claude to craft some tutorials on interpretability. One channel we are working on is an analogical infrastructure that permits the exploration
When we talk with one another we are simultaneously aware that we are in the world together and that we have to work very hard
Suppose you have a transformer that handles tokens of dimension Dmodel.D_{model}. And suppose it has ten tokens in the input. The residual stream that’s updated
I just re-read Max Weber’s essay “Science as a Vocation” in the context of thinking about human and machine intelligence alignment. In particular, I’m thinking
I suspect that this is a naive question already thought through by people much smarter than I, but here goes. We’ve heard some concerns expressed