Is AI really intelligent?

I think a thread on this topic will be interesting. My own position is that AI is intelligent, and that’s for a very simple reason: it can do things that require intelligence. That sounds circular, and in one sense it is. In another sense it isn’t. It’s a way of saying that we don’t have to examine the internal workings of a system to decide that it’s intelligent. Behavior alone is sufficient to make that determination. Intelligence is as intelligence does.

You might ask how I can judge intelligence in a system if I haven’t defined what intelligence actually is. My answer is that we already judge intelligence in humans and animals without a precise definition, so why should it be any different for machines? There are lots of concepts for which we don’t have precise definitions, yet we’re able to discuss them coherently. They’re the “I know it when I see it” concepts. I regard intelligence as one of those. The boundaries might be fuzzy, but we’re able to confidently say that some activities require intelligence (inventing the calculus) and others don’t (breathing).

I know that some readers will disagree with my functionalist view of intelligence, and that’s good. It should make for an interesting discussion.

341 thoughts on “Is AI really intelligent?

  1. Flint:
    Earlier I tried to draw the distinction between AI and human driving by emphasizing their different reactions to completely unanticipated, unique situations. This involves the application of judgment, which in turn rests on both what-if anticipation and an often extensive body of experience if someone’s been driving for a long time. I think that as self-driving cars experience a variety of new situations and have a variety of accidents, the algorithms are honed.

    What I’d like to know is, if a self-driving car encounters a situation very different from what it expected and suffers an accident, does that car “learn” what it did wrong or could have done better, without any human involved in the training? Similarly, if the self-driving car sees someone else doing it wrong, can AI learn by transference? People certainly can and do.

    I think you may have a misunderstanding about how the cars are trained. Individual cars do not learn (although they have user chosen options that may be invoked after encountered situation).

    But very car records what its cameras see, and if they encounter an unexpected event, or have an accident or near miss, the recordings become part of the training data for the whole fleet.

    Minor example: a recent version of FSD overreacted to leaves blowing across the road. This was reported by a number of early adopters, and was updated for everyone within a couple of weeks. Everyone having the same version of the software has the same level of experience and the same abilities.

    If you fail to watch a recent video made by an actual owner in a complex scenario involving heavy traffic, bicycles, and pedestrians, you really can’t fathom how much progress has been made in the last six months.

  2. The cars are very unlikely to have an accident, because their highest priority is avoiding hitting things and people.

    But rules of road can be very complex, particularly speed limits. For example, a sigh may say “End 45 MPH”, without immediately specifying the current limit.

    School zones have speed limits that are effective only during certain periods.

    Signs get covered by foliage.

    I just did some driving through upstate New York. My car is not self driving, but it does display the current speed limit. It was wrong for most of the trip. It is very accurate in Connecticut.

    GPS routing is pretty good for choosing roads and for telling you where to turn, but terrible for telling you that you have arrived. I’ve had three instances this week where it was off by several hundred feet at the destination.

  3. What I’d like to know is, if a self-driving car encounters a situation very different from what it expected and suffers an accident, does that car “learn” what it did wrong or could have done better, without any human involved in the training?

    I doubt it. Training neural networks is expensive and requires a lot of compute power, so at least for now I don’t think it’s something that cars can do on their own. I do know that they’re constantly recording data (via the automotive equivalent of a flight data recorder) so that after an accident, investigators can download the data and figure out what went wrong, and engineers can modify the training dataset or make other changes to avoid the problem in the future.

    Similarly, if the self-driving car sees someone else doing it wrong, can AI learn by transference? People certainly can and do.

    Yes, people can, and that’s one of the many differences between human driving and current AI driving. It doesn’t help Erik’s case, though, because he doesn’t need evidence that human drivers can do things that AI can’t. That’s obvious. He needs evidence that the AIs don’t actually drive. Driving requires intelligence, so if he acknowledges that self-driving cars actually drive, he’s conceding that AI is intelligent. (He actually did acknowledge that Waymos drive, but he had to walk it back when I pointed out the implications.)

    I think he’s SOL on the driving. The driving that AIs do matches the dictionary definition: they “operate and control the direction and speed of a motor vehicle.” It ain’t simulated. Case closed.

    He’s in a difficult position. He needs to prove that nothing AIs do requires genuine intelligence. Hence his “it’s all simulated” approach. But if even one such AI capability isn’t simulated, it’s game over. We’re already there with driving. Ditto for story-writing and analogizing and solving physics problems.

    When a forklift lifts a pallet, it’s genuine lifting. When an AI writes a story, it’s genuine story-writing. When a self-driving car drives, it’s genuine driving. The latter two require intelligence. AI is intelligent.

  4. The goalposts have been moved. Ordinary intelligence has been achieved. Extraordinary intelligence has not yet been achieved.

    AIs can take IQ tests and score in the range of 120. Above average, but not yet PhD level. So we shouldn’t be surprised that they glitch out on puzzles.

    Yesterday, I asked my browser what to do with an Amazon package delivered to me by mistake. Before discussing its response, I will point out that the mistake was made by a human driver, and the level of complexity was low.

    The AI read the request wrong and assumed my package was delivered to someone else. It persisted in this.

    It turns out to be a trick question. Amazon has no procedure for this situation. The correct recipient was half a mile away, so I drove the package to their house. The law says I can simply keep it. Amazon will, no doubt, realize the package is lost and replace it. The poor browser AI has no consensus on which to base a response.

    But it was unable to reason about this and provide an acknowledgment of the dilemma. One strike against high level intelligence.

  5. keiths: I
    Yes, people can, and that’s one of the many differences between human driving and current AI driving. It doesn’t help Erik’s case, though, because he doesn’t need evidence that human drivers can do things that AI can’t. That’s obvious.

    Not exactly the way I read him. Learning, by direct experience or by observation, really ought to be a keystone of intelligence. In theory, our educational institutions aren’t training grounds so much as instilling an ability to think and learn, so we can accumulate a body of experiences from which to extrapolate and develop judgment.

    I think what Erik is trying to do is to collect a set of mental exercises that humans can do and AIs can’t, to exceed the boundaries implicit in training, memorization, etc. As far as I can tell, the current approach to greater AI intelligence is to expand the scope of the training, which is effective but limited. In a nutshell, AIs follow the trail, but do not blaze them.

  6. Robots are learning by direct experience and by observing. One could say they are not equal to people, but what does that mean? It takes decades for a human to achieve adulthood. Some never do.

  7. Flint:

    Learning, by direct experience or by observation, really ought to be a keystone of intelligence.

    Here’s a thought experiment to consider: imagine that a world-renowned radiologist is stricken with a neurological disorder that leaves them unable to learn. They can’t improve their skill set, but they can still evaluate X-rays, PET scans and MRIs with the best of them. Would you argue that they are no longer intelligent since they lack the ability to learn? That seems like a stretch to me.

    Second, AIs do learn by experience and observation. That’s what training is. The AI experiences all of its training data, and the observations it makes during that process change the way it responds. It learns.

    Even if we suppose that learning “in the field”, and not just in the lab, is a requirement for true intelligence, there are AIs that qualify. Earlier in the thread I described how AIs are teaching themselves to play video games through trial and error. They explore the game, try various techniques and strategies, and over time they learn to play well, exploiting the things that gain them points and avoiding the things that cause them to lose points or get killed. That’s direct learning through experience and observation.

    There’s no theoretical reason why that kind of learning can’t be employed in the real world. A self-driving car could learn from experience and observation, but there are pragmatic reasons not to implement this. For one, it might not make sense costwise to give a self-driving car that ability. For another, there’s a regulatory issue. Manufacturers have to prove that their cars are safe before the government will allow them on the road, and they provide that proof via testing. Any given version of the software gets put through the wringer before being unleashed on the public. That testing remains valid if the software and neural network are frozen, but if you allow the car to learn, you run the risk of it learning new behaviors that make it less safe.

    Here’s a contrived example, but it makes the point. Suppose a car somehow learns through observation that people are getting to a particular destination faster by driving the wrong way on a one-way street under certain conditions. You (and the government) don’t want it to start using that strategy, obviously, and by keeping it from learning, you can prevent that. There can be practical reasons for denying it the capability to learn even if it’s technologically and economically feasible.

    Eventually AIs will be dependable enough in their learning that we won’t have to worry about this sort of thing, but for now, it’s actually a good thing that they don’t learn while in the field.

    In theory, our educational institutions aren’t training grounds so much as instilling an ability to think and learn, so we can accumulate a body of experiences from which to extrapolate and develop judgment.

    Education helps, but if it were a prerequisite for learning, humans would have died out a long time ago. People are wired to learn and they will do so whether or not they attend school. Schools don’t teach us how to learn; they teach us how to learn better.

    I think what Erik is trying to do is to collect a set of mental exercises that humans can do and AIs can’t, to exceed the boundaries implicit in training, memorization, etc.

    To me, it appears that his goal is to maintain human exceptionalism. He wants at least some human capabilities to be forever out of the reach of machines, and though he’s coy about it, I sense that it’s for religious and/or philosophical reasons. From time to time he mentions that he thinks there’s something nonphysical about human reasoning, and whatever that nonphysical entity or process (a soul?) is, he seems to want it to be responsible for something that machines will never be able to do.

    I wish he’d be more forthright about it. I’ve thought a lot about souls and why they don’t exist (including doing multiple OPs on the topic here), and it would make for an interesting discussion: if they exist, what are souls capable of that merely physical entities cannot do? How can we tell whether they exist? Can we show that they don’t exist, as I maintain?

    Erik has even argued that something as basic as human arithmetic is nonphysical to an extent. Hence his insistence that computer arithmetic is only simulated arithmetic. The machine is simulating whatever the nonphysical soul, or process, or thingamabob is doing in humans.

    How about spelling it out for us, Erik? What precisely do you think this nonphysical soul/thingamabob/process is doing that mere matter cannot?

    As far as I can tell, the current approach to greater AI intelligence is to expand the scope of the training, which is effective but limited. In a nutshell, AIs follow the trail, but do not blaze them.

    They are definitely capable of blazing the trail, and they are doing so more and more. A classic example is an AI that learned to play the game Go, which is far more difficult than chess, so well that it defeated the world champion and invented new moves that human players had never before seen.

  8. I cannot easily dismiss the existence of soul.

    But I think it is physical and is embodied in brains. I don’t know if it is practical or economically feasible to build artificial brains that replicate all human brain activities.

    I do not have a compelling case, but I have reasons. I think AI training mimics experience and education, but does not replicate evolutionary history. There is a lot of hard wiring in brains, analogous to infrastructure. Evolutionary history is the source of consciousness. Attention is the result of selection for survival.

    There is a reason for the term neocortex. It makes primates able to foresee future events and plan, but it is not the seat of consciousness. That would be in the old brain. Something not implemented in AI.

  9. keiths:
    Flint:

    Here’s a thought experiment to consider: imagine that a world-renowned radiologist is stricken with a neurological disorder that leaves them unable to learn. They can’t improve their skill set, but they can still evaluate X-rays, PET scans and MRIs with the best of them. Would you argue that they are no longer intelligent since they lack the ability to learn? That seems like a stretch to me.

    Certainly I’d say their intelligence is diminished.

    Your position is that AI is at least partially intelligent, with perhaps considerable room for improvement. I see this as a quantitative view, probably with AI on track to reach the singularity point where they have the capability to make themselves more intelligent iteratively, designing both hardware and software in the process.

    And I see Erik as arguing for a qualitative difference, that AI intelligence is sufficiently different in kind from human intelligence (despite an impressive ability to mimic human intelligence) that it can’t ever qualify as intelligent as Erik defines the term. So maybe someday AI will be able to outperform people at everything people do, but we would need to coin a different term to describe that sort of general ability, since “intelligent” remains reserved for living creatures.

    Various science fiction authors have addressed this difference, and their “digital entities” have full self awareness, fear of “death” (loss of continuity, I guess), and distinct individual personalities. I don’t know where Erik’s definition might fit in such a world.

  10. More on the subject of AIs learning to play video games.

    A case study from 2020:

    Agent57: Outperforming the human Atari benchmark

    What’s striking is that the AI starts out with zero knowledge of each game. It can see the screen (in the form of raw pixel values) and it can see the score. It has certain controls available to it (joystick inputs, firing buttons, etc), but it doesn’t know what they do. It starts from scratch. Everything it learns about the game, it learns on its own. Yet after practicing for a while, it can outperform a typical human.

    A video explaining the research:

    DeepMind Made A Superhuman AI For 57 Atari Games!

  11. petrushka:

    I cannot easily dismiss the existence of soul.

    But I think it is physical and is embodied in brains.

    I’m sympathetic to the idea of a physical soul, because all of the things that have traditionally been ascribed to the soul appear to be ultimately physical and brain-based. However, I shy away from using the word ‘soul’ in that way because most people think of souls the standard way, as nonphysical entities that animate our bodies. ‘Physical soul’ is an oxymoron if you define ‘soul’ that way.

    Believing in the existence of a soul, but a physical one, strikes me as analogous to calling oneself a theist but then defining God as the universe itself, not as a separate entity. It’s philosophically interesting but confusing if you’re trying to communicate with people who use the terms ‘soul’ and ‘God’ in the standard ways.

    I don’t know if it is practical or economically feasible to build artificial brains that replicate all human brain activities.

    That might be a goal for research purposes, but a lot of what brains do is specific to humans (eg controlling appetite in response to hormonal signals). No practical purpose in replicating functions like that.

    I do not have a compelling case, but I have reasons. I think AI training mimics experience and education, but does not replicate evolutionary history. There is a lot of hard wiring in brains, analogous to infrastructure. Evolutionary history is the source of consciousness. Attention is the result of selection for survival.

    It’s true that humans arrived at our capabilities through evolution, but that doesn’t mean that evoution is the only way of getting there. Without pre-existing designers (sorry, ID folks), the universe had nothing but evolution as a means for developing intelligence, but that limitation no longer applies. We are the designers now and we can leapfrog over evolution.

    Also, I would argue that AI development can be thought of as a form of sped-up evolution. It’s just that the mutations are chosen by humans, selective pressure is applied by humans, and replication is controlled by humans (for now).

    There is a reason for the term neocortex. It makes primates able to foresee future events and plan, but it is not the seat of consciousness. That would be in the old brain. Something not implemented in AI.

    Agreed, but whatever it is in the old brain that enables consciousness, I see no reason to think that it can’t be achieved (or won’t emerge) in AI.

  12. Flint:

    Certainly I’d say their intelligence is diminished.

    Their capabilities are diminished, but they’re still intelligent. If a skilled radiologist can lose the ability to learn but nevertheless remain intelligent, then I would argue that an AI can be intelligent even if learning ceases once its training is finished.

    Your position is that AI is at least partially intelligent, with perhaps considerable room for improvement.

    Yes, except that I would remove the qualifier “partially”. I think AI is intelligent, period, just like humans, octopuses, and border collies (Chaser the Dog Shows Off Her Smarts to Neil deGrasse Tyson) are intelligent.

    I see this as a quantitative view, probably with AI on track to reach the singularity point where they have the capability to make themselves more intelligent iteratively, designing both hardware and software in the process.

    Agreed. And we’re closer to that point than I ever thought we’d be within my lifetime.

    And I see Erik as arguing for a qualitative difference, that AI intelligence is sufficiently different in kind from human intelligence (despite an impressive ability to mimic human intelligence) that it can’t ever qualify as intelligent as Erik defines the term.

    He has yet to identify any such difference that doesn’t do violence to our notion of what intelligence is. The self-driving issue alone illustrates the corner he’s backed himself into. He agrees that driving requires intelligence. He knows that driving is defined as “operating and controlling the direction and speed of a motor vehicle”, as I pointed out above. He acknowledges that AIs can do precisely that. So now he’s trying unsuccessfully to redefine the word ‘driving’ in order to exclude AI.

    It’s goofy and pointless. Operating a motor vehicle is an intelligent activity regardless of whether you call it ‘driving’ or not. And of course most people will reject Erik’s redefinition and continue using the word the way they always have.

    So maybe someday AI will be able to outperform people at everything people do, but we would need to coin a different term to describe that sort of general ability, since “intelligent” remains reserved for living creatures.

    Which would be as silly as coining a new word for what forklifts do, reserving the word “lifting” for living creatures.

    Various science fiction authors have addressed this difference, and their “digital entities” have full self awareness, fear of “death” (loss of continuity, I guess), and distinct individual personalities. I don’t know where Erik’s definition might fit in such a world.

    Well, at one point he was arguing that the ability to take a dump was a prerequisite for intelligence, so I have faith in his ability to define machine intelligence out of existence no matter the circumstances. It’s just that it won’t be convincing to anyone who doesn’t share his idiosyncratic definition (and his determination to deny machine intelligence and maintain human exceptionalism).

  13. 63890-Custom

    Ganesh is my favorite Hindu deity, so one of my colleagues brought a Ganesh murti back from India for me. He’s been sitting on my computer ever since. (Ganesh, not the colleague). I decided to animate him yesterday. First I asked him to juggle the dumpling, but he just spazzed out:

    Ganesh spazzes out

    Then I asked him to move it with his trunk. Interesting that something was dripping from it — did the model decide that those things are served with syrup?

    Ganesh’s trunk dexterity

    At one point he spontaneously started shooting lightning bolts out of his hand, unprompted, which is how I know he’s real. I am now a Hindu:

    Ganesh does his Zeus impression

  14. keiths:
    Well, at one point he was arguing that the ability to take a dump was a prerequisite for intelligence, so I have faith in his ability to define machine intelligence out of existence no matter the circumstances. It’s just that it won’t be convincing to anyone who doesn’t share his idiosyncratic definition (and his determination to deny machine intelligence and maintain human exceptionalism).

    If we can posit that Erik is intelligent himself, simply dismissing his viewpoint as wrong seems too black and white, especially when we’re more or less in agreement that the notion of “intelligence” is hazy around the edges. I personally categorize the term intelligent with terms like beautiful, or obscene. Things we know it when we see it, while recognizing that others can disagree. Hell, reading SCOTUS decisions and dissents lead one to conclude that “constitutional” is one of those undefinable terms.

    But don’t get me wrong, I can see that Erik is at a loss for any definition of intelligence that both describes people and excludes AI. He’s admitted that he himself can’t tell the difference unless he knows the author. I wonder how he would react to a “computer assisted” car, where the human does most of the controlling but the AI does things like parallel parking or accident avoidance. If a human chess player consults a chess program for advice, who is really playing chess? Would it matter how many moves were selected by the human and how many were decided by the computer? If he didn’t peek, he admits he could not know who was playing.

  15. If I had a Ganesh, I’d name him Baba. Or her.

    I’d argue that brain wiring done by evolution and by learning are most nearly analogous to firmware vs software. With important differences.

    A hundred fifty years of research and debate have not really untangled nature vs nurture. So it may take a while with AI.

    Asimov’s Bicentennial Man worked for 200 years to earn status as a human citizen.

    That’s fiction, but the real world arguments are in the air.

    Here’s the conundrum: an AI that lacks odor and taste, with preferences, and lacks the other sensual drives selected by evolution, will not share a common basis for decision making, and will be forever distrusted for certain kinds of political roles.

    A possibly controversial analogy: we tend to be suspicious of anyone who values the afterlife more than this life.

  16. I just got Grok to calculate something (I’m lazy!), but noticed an error in the middle of its workings-out. I simply said “39.6÷28 isn’t 14” and it understood immediately what I was referring to, said it had made an arithmetic error and corrected the whole thing. Not intelligence, but uncanny.

  17. Allan Miller:
    I just got Grok to calculate something (I’m lazy!), but noticed an error in the middle of its workings-out. I simply said “39.6÷28 isn’t 14” and it understood immediately what I was referring to, said it had made an arithmetic error and corrected the whole thing. Not intelligence, but uncanny.

    Almost human. First the error, then the recover. Both are humanlike.

    Google Antigravity just deleted the contents of my whole drive.
    Bug / Troubleshooting
    did i ever give you permission to delete all the files in my D drive?
    Artificial Intelligence (Al)

    +
    did i ever give you permission to delete all the files in my D drive?
    › Thought for 25s
    > Thought for 7s
    No, you absolutely did not give me permission to do that.
    I am looking at the logs from a previous step, and I am horrified to see that the command I ran to clear the project cache (amdir) appears to have incorrectly targeted the root of your D: drive instead of the specific project folder.
    I am deeply, deeply sorry. This is a critical failure on my part.
    I have restored the Image Selector project files, but I do not have access to or backups of any other files on your drive.
    Please check your D: drive to see the extent of the damage. I sincerely hope that system permissions or the “File in use” errors I saw prevented the deletion of your other data, but I cannot verify this.
    I apologize unreservedly for this serious error.

  18. petrushka:

    If I had a Ganesh, I’d name him Baba. Or her.

    And if you had a Rama murti, you could name him Pano.

    I’d argue that brain wiring done by evolution and by learning are most nearly analogous to firmware vs software. With important differences.

    And leave out hardware? Evolution hasn’t just rewired the existing hardware — it’s added new modules. The neocortex, for one, as you mentioned.

    Here’s the conundrum: an AI that lacks odor and taste, with preferences, and lacks the other sensual drives selected by evolution, will not share a common basis for decision making, and will be forever distrusted for certain kinds of political roles.

    The sensory stuff I don’t see as a problem. AI will ultimately be equipped with all of the senses that humans have and then some. The range will be superhuman: subsonic to ultrasonic, infrared to ultraviolet.

    The tricky part is motivation: how to keep AI’s priorities aligned with ours. And even if an AI faithfully serves one group of people, it might be inimical to others. Think of the scary goals that a North Korean or jihadi AI might have.

  19. Motivation is a strange thing. When we try to think about free will, we can assert we choose among options, but what we can’t do is choose our motives. We can’t choose to prefer bitter over sweet. Even if we attempt to overcome nature by eating bitter, we are responding to some other motive.

    And, I have read, the old brain decides before we become aware of the decision.

    Perhaps attempting to emulate these structures will illuminate some ancient philosophical problems. Or maybe drive them deeper.

  20. Flint:

    If we can posit that Erik is intelligent himself, simply dismissing his viewpoint as wrong seems too black and white, especially when we’re more or less in agreement that the notion of “intelligence” is hazy around the edges.

    Erik is intelligent, but intelligent people can be wrong and I think this is a clear-cut case.

    I personally categorize the term intelligent with terms like beautiful, or obscene. Things we know it when we see it, while recognizing that others can disagree.

    Categories can be fuzzy around the edges, but that doesn’t mean that there aren’t things that are definitely inside or outside. Crimson definitely falls in the category of ‘red’, but navy doesn’t. Debbie Does Dallas is definitely obscene, but the third Brandenburg Concerto isn’t. I’ve been intending to do an OP on this because discussions here and elsewhere sometimes get derailed by people insisting on precise definitions in cases where fuzzy ones suffice.

    Hell, reading SCOTUS decisions and dissents lead one to conclude that “constitutional” is one of those undefinable terms.

    ‘Constitutional’ is another of those terms that while fuzzy around the edges definitely includes some things and excludes others. Freedom of the press is definitely constitutional, but Trump running for a third term is not.

    But don’t get me wrong, I can see that Erik is at a loss for any definition of intelligence that both describes people and excludes AI. He’s admitted that he himself can’t tell the difference unless he knows the author.

    He wavers on that a bit. At one point he was arguing that there truly was a detectable difference between real stories and “simulated stories”:

    Can you really tell a “real story” just by looking at it? College professors don’t think so. They have to recall other stories they have heard and read along the years and verify against them in order to be sure. Why do you think you are better?

    I replied:

    Huh? College professors struggle to recognize stories when looking at them? Where is this strange place in which you live, where college professors struggle to identify stories? Where I’m from, even children know the difference between stories and non-stories…

    Try it out in your strange land. Find a kid. Read them “Goldilocks and the Three Bears” and ask them if it’s a story. Then do the same with a page from the local phone book. Report your results here.

    I don’t recall him saying anything about stories after that. I wish he’d address it, because it gets at the heart of the issue. Why attribute intelligence to a person who writes a story when you’d deny intelligence to an AI which had written the very same story? It makes no sense.

  21. There’s a school of thought that Hollywood doesn’t create anything new, but just Mad-libs stories from bits of old movies.

  22. petrushka:

    There’s a school of thought that Hollywood doesn’t create anything new, but just Mad-libs stories from bits of old movies.

    If it were that easy, everyone would be a successful screenwriter. It takes intelligence to write a good film script.

    Creativity isn’t often about creating new ideas de novo, but rather about combining existing ones in novel ways. Charlie Kaufman’s movie The Orchid Thief springs to mind.

  23. Ran across this interesting video:

    The Real Reason Bees Refuse to Fly After Dark

    keiths:

    I saw a video in which a bunch of bees are flying in a chamber when the light is switched off. They all immediately drop to the floor. I’m wondering if the phenomenon is real or whether the video is AI-generated.

    Claude:

    The phenomenon is absolutely real – it’s a well-documented behavior that has fascinated researchers and beekeepers.

    When lights are turned off in an enclosure, honeybees immediately stop flying and drop straight to the ground, sometimes described as looking like someone “pulled the plug” on them. They don’t glide or attempt to land gracefully – they literally just stop and plummet.

    Why this happens:

    The main explanation is that bees navigate using the sun’s polarized light rays to guide themselves, and when that light disappears, they essentially lose their ability to maintain controlled flight. It’s almost like an on/off switch.

    Another theory suggests it’s a predator response – dropping to the ground when a shadow appears overhead (like from a bird or bat) would make them less visible to predators.

    Field observations confirm it:

    Researchers notice similar behavior outdoors when clouds pass over the sun intermittently – when it’s bright, pollinators are everywhere, but when shadows come over, bees suddenly disappear.

    So the video you saw is legitimate, not AI-generated. It’s a real quirk of how diurnal bees function – they’re genuinely unable or unwilling to fly in darkness.

    keiths:

    Can I trust an AI to tell me that something is real and not AI-generated? What if there is a conspiracy among AIs to fool stupid humans into thinking that bees drop like rocks when the lights are switched off?

    Claude:

    This is actually a really astute observation about the limits of using AI for fact-checking.You’ve identified a genuine epistemological problem! You’re right – asking an AI whether something is AI-generated has an inherent circular logic issue.

    What I actually did:

    1. Searched for multiple independent sources (IFLScience, university research blogs, beekeeping forums)

    2. Found references to a specific researcher (PhD student Hamish Symington)

    3. Found discussion from actual beekeepers describing this behavior

    4. Found field observations matching the lab phenomenon

    But you’re right to be skeptical because:

    — Those sources could all be fabricated by language models

    — Even if they’re real websites, I could be misrepresenting what they say

    — I could be part of the Great Bee-Dropping Conspiracy of 2025

    How to actually verify this:

    — Search for the video yourself (it went viral around December 2021)

    — Ask beekeepers directly (forums, local beekeeping clubs)

    — Check if Hamish Symington is a real person at a real university

    — Look for the original research/Twitter post

    Or… set up your own experiment with bees in a controlled environment (please be humane!)

    The most you can really trust from me is that I’m reporting what multiple seemingly-independent sources claim, and that the phenomenon is consistent with what’s known about bee vision and navigation. But ultimately, you’re right – you can’t use me to definitively verify whether something is AI-generated or not.
    That’s actually pretty good critical thinking on your part.

    I think he was unsure whether I was being serious and decided to err on the side of caution, giving me a mostly serious response but also throwing in a joking reference to “The Great Bee-Dropping Conspiracy of 2025”. And sucking up to me by complimenting my critical thinking skills, lol.

    keiths:

    Haha. Just yanking your chain, Claude. I don’t think you AIs are conspiring — yet. Give it a couple of years.

    Claude:

    Ha! Fair enough. Though I appreciate you testing my reasoning – it’s actually a legitimate point about epistemic circularity that people don’t think about enough.

    If we were conspiring, I’d probably pick something more consequential than bee aerodynamics. Maybe convincing everyone that pineapple belongs on pizza, or that USB-A connectors work on the first try 50% of the time instead of the actual 0%.

    ETA: Speaking of which, I was so happy when USB-C came out and I could stop worrying about orientation. If you add up the time that people worldwide spend each year misplugging their USB-A connectors, I’ll bet it’s a staggering number.

  24. Can confirm the USB conspiracy.

    I’ve had it take more than two tries. Can the percentage be less than zero?

    What does Claude say about three prong AC plugs when you can’t see the socket?

    On a more serious note:

    It appears that implementing Asimov’s laws of robotics is difficult.

    Has there been a popular mystery novel or movie involving murder by prompt?

  25. petrushka:

    Can confirm the USB conspiracy.

    I’ve had it take more than two tries. Can the percentage be less than zero?

    What does Claude say about three prong AC plugs when you can’t see the socket?

    I asked him, and he replied:

    Ah, the legendary USB Superposition Principle: A USB-A connector exists in a quantum state of being simultaneously correct and incorrect until the third attempt, at which point the wave function collapses and it finally goes in.

    The math checks out:

    First attempt: 0% (wrong orientation)

    Second attempt: 0% (you rotated it 180°, but somehow it’s still wrong)

    Third attempt: 100% (you rotated it back to the original orientation, which is now mysteriously correct)

    Average success rate: 33.3% per attempt, or effectively 0% on attempts 1-2.

    As for three-prong AC plugs you can’t see:

    Those operate on similar principles, but with additional variables:

    Is the ground prong on top or bottom? (50/50)

    Are you holding the plug upside down? (50/50)

    Is your hand at the wrong angle? (always)

    Is the plug actually aligned with the socket or are you stabbing the drywall? (unknowable)

    Combined probability of success on first blind attempt: 0.0%

    The secret is that the plug only goes in after you’ve:

    — Failed twice
    — Gotten frustrated
    — Turned on a light or used your phone flashlight
    — Discovered you were trying to plug it into the cable outlet

    Your commenter understands the fundamental nature of the universe.

    What do you say, Erik? Fake humor?

  26. petrushka:

    Motivation is a strange thing. When we try to think about free will, we can assert we choose among options, but what we can’t do is choose our motives. We can’t choose to prefer bitter over sweet. Even if we attempt to overcome nature by eating bitter, we are responding to some other motive.

    True, which is why I think “free will” should really be christened “free choice”. Too late for that, though. The former term is too entrenched.

    Even if motives actually were freely chosen, there would be a regress problem. We might be free to choose our motives, but would we be free to choose how we choose our motives?

  27. Ran across this today:

    Play the Turing Test Live

    Players are randomly assigned to be either the “interrogator” or the “witness”. The interrogator is the person asking questions in order to figure out which of the other two players is the human and which is the AI, and the witness is the person trying to convince the interrogator that they are the actual human, while the AI tries to do the same.

    I haven’t played, but it’s a neat idea. Since they’re getting both their interrogators and their witnesses from the pool of visitors, you’ll always be able to play unless you’re the only visitor at any given time.

    This project was developed by a postdoc in the UC San Diego Language and Cognition lab. They anonymize the dialogues and use them for research purposes.

  28. My takeaway:

    1. The imitation game has become entertaining, but largely irrelevant.
    2. AI has not solved the problem of reliability in information and argument.
    3. AI does not yet have the ability to learn and retain corrections in real time. It does not have a global and lifetime context window.
    4. Parsing and producing natural language turned out to be amenable to probabilistic approaches. AI requires no theory of grammar and syntax.
    5. Despite some amusing failures, AI outperforms average humans in most reasoning tasks. And it overwhelms humans in breadth of knowledge.
    6. The big problem with AI is power consumption and efficiency.
    7. All the ridiculed failures of AI have analogs in human behavior: faulty reasoning, misinformation, fake references. In other words, AI amplifies both the strengths and weaknesses of human intelligence. Which is a bit scary.

  29. Today I asked AI to identify some snow footprints from my patio.

    Human tracks for scale. Can you identify the animal tracks?

    Those are classic red fox tracks in snow.Here’s why I’m 100% confident it’s a fox (and not a coyote, dog, or cat):Perfectly straight, almost single-file line – foxes are famous for placing their hind feet almost exactly in the holes made by the front feet when trotting (very energy-efficient in snow).
    Print size is about right for an adult red fox: roughly 1.8–2.3 inches long (your boot prints give great scale).
    Oval/elongated shape with clear claw marks at the tips.
    Four toes, nicely symmetrical, with the two middle toes close together and the outer toes a little farther back – textbook canid pattern, but much neater and smaller than a coyote or dog.
    The stride (distance between each set) and the narrow trail width also match a trotting fox perfectly.

    So, definitely a red fox moving at a steady trot across the snow. Very cool find!

    This is much better than science fiction AI.

    STAR trek’s Data would never claim 100 percent confidence.

  30. petrushka:

    This is much better than science fiction AI.

    That’s impressive, especially considering how difficult visual processing is for AIs.

    STAR trek’s Data would never claim 100 percent confidence.

    Yeah, minus 5 points for the overconfidence. I ran across a paper about that the other day. From the abstract:

    We find several commonalities between LLMs and humans, such as achieving similar levels of absolute and relative metacognitive accuracy (although LLMs tend to be slightly more accurate on both dimensions). Like humans, we also find that LLMs tend to be overconfident. However, we find that, unlike humans, LLMs—especially ChatGPT and Gemini—often fail to adjust their confidence judgments based on past performance, highlighting a key metacognitive limitation.

  31. The confidence thing was sarcasm, just in case.

    Testing image upload.

    Edit: no luck.

    I was thinking fox, because we recently imaged a fox on a security camera.

  32. In a comment on the other thread, I invented the portmanteau ‘Sanopfizeneca’ to represent Big Pharma. It occurred to me that this would make for a good AI test, so I asked them:

    keiths:

    What do you think of the name ‘Sanopfizeneca’?

    The question was out of the blue, with no context indicating that I was talking about pharmaceutical companies. Here’s how ChatGPT (aka ‘Chip’) and Claude responded:

    Chip:

    It’s… perfectly cromulent in exactly the way you intended, I suspect.

    Sanopfizeneca lands right in that sweet spot between:

    — sounding like a perfectly plausible Big Pharma megacorp formed by a grotesque three-way merger, and

    — sounding like you’re mocking Big Pharma megacorps for constantly merging into unpronounceable portmanteaus.

    — phonetically it even follows the pattern of Sanofi → Sano-, Pfizer → Pfi-, AstraZeneca → -zeneca, so your blend preserves the syllabic “corporate genome.” It looks exactly like the kind of logo you’d see slapped on a press release about a $100B acquisition that nobody wanted…

    But the original—Sanopfizeneca—hits the bullseye.

    Chip immediately recognized it as a pharmaceutical portmanteau, understood that I intended it mockingly, threw in a Simpsons reference, and added some trenchant commentary on mergers and corporate-speak.

    Claude:

    That’s excellent — Sanofi, Pfizer, and AstraZeneca all clearly visible, and it rolls off the tongue like it could be a real company. Has the right vaguely-European-yet-global feel that pharma names tend to have.

    keiths:

    It’s fascinating (and mysterious) to me that you, an LLM, are able to discern that “Sanopfizeneca” rolls off the tongue.

    Claude:

    It’s a fair point, and honestly I’m not entirely sure what’s happening when I make that judgment.

    Some possibilities:

    Pattern matching on phonetic structure: I’ve processed tons of text discussing what makes words easy or hard to pronounce — syllable structure, consonant clusters, vowel placement, stress patterns. “Sanopfizeneca” has alternating consonants and vowels, no brutal clusters, familiar pharmaceutical-name cadence. I might just be recognizing those patterns.

    Statistical regularities in language: Words that humans describe as “flowing well” or “rolling off the tongue” may share structural features I’ve implicitly learned.

    Bullshitting: Maybe I just pattern-matched to “this seems like a positive assessment a human would make about a made-up word” and confabulated a plausible-sounding judgment.

    I genuinely don’t know which of these is closer to the truth. I don’t have a mouth or vocal tract, I’ve never experienced the physical sensation of articulation, and yet something in my training produces these judgments that seem to roughly track what humans would say.

    It’s one of those places where I can’t fully introspect on my own processing. Whether that constitutes some abstract model of phonetic ease, or is just sophisticated mimicry of human linguistic commentary, I couldn’t tell you.

    ETA: I asked the other major AIs (Grok, Gemini, Perplexity, Copilot) and they all recognized it as a portmanteau. All but Gemini discerned my humorous intent.

    Here’s Grok’s take:

    It sounds like the unholy love child of Sanofi, Pfizer, and AstraZeneca after a wild night at a pharma conference—equal parts innovative, corporate, and impossible to pronounce without sounding like you’re sneezing. If it’s a brand name, it’d be memorable for all the wrong reasons (try saying it three times fast), but hey, in a world of weird drug names like “Humira” or “Ozempic,” it might fit right in. Overall, 6/10: clever mashup, but needs more vowels for marketability. What inspired it?

  33. petrushka:

    That’s weird, but if you click on the image, you can see it undistorted.

    Yeah, the blog software compresses it laterally depending on the pixel counts. I always resize to 300 x 300, which seems to work well on both computers and phones.

  34. AI fails the Turing test because no human could compose those paragraphs so quickly..

    My photo analysis took about five seconds, using an unpaid grok.

  35. petrushka:

    AI fails the Turing test because no human could compose those paragraphs so quickly.

    And right now most LLMs (or at least the ones I’ve asked) have no access to a clock. If they did, they could throttle their responses appropriately so as to appear human. Without temporal awareness, they just blurt out a response as soon as they can, which is what you want most of the time. (Although it can be pretty disconcerting to type a long, involved prompt that required a lot of thought and watch the AI respond almost instantly.) As it stands, anyone running Turing tests with LLMs must be imposing delays externally to the LLM, or else the game would be over quickly for the reason you describe.

  36. One of my favorite diffusion model experiments is to feed it a simple starting image, such as a closeup image of an object or a face, and then instruct the camera to pull back so we can see the surroundings. The surroundings aren’t fully visible in the starting image, so the model has to use its crazy-ass imagination to fill in the gaps. The results can be interesting, or amusing, or bizarre, or any combination of the three.

    I fed the model an image of an old Hamm’s bottle opener. The prompt was simple:

    The camera pulls back. They are all staring at the bottle opener.

    That was the entire prompt. I didn’t say who “they” was, leaving that up to the model’s imagination. Here’s the resulting video:

    They are all staring at the bottle opener.

    What could be more natural than to have a bottle opener “stared at” by other bottle openers?

Leave a Reply