How Google engineer Blake Lemoine became convinced an AI was sentient
Current AIs aren’t sentient. We don’t have much reason to think that they have an internal monologue, the kind of sense perception humans have, or an awareness that they’re a being in the world. But they’re getting very good at faking sentience, and that’s scary enough.
Over the weekend, the Washington Post’s Nitasha Tiku published a profile of Blake Lemoine, a software engineer assigned to work on the Language Model for Dialogue Applications (LaMDA) project at Google.
LaMDA is a chatbot AI, and an example of what…