I think a thread on this topic will be interesting. My own position is that AI is intelligent, and that’s for a very simple reason: it can do things that require intelligence. That sounds circular, and in one sense it is. In another sense it isn’t. It’s a way of saying that we don’t have to examine the internal workings of a system to decide that it’s intelligent. Behavior alone is sufficient to make that determination. Intelligence is as intelligence does.
You might ask how I can judge intelligence in a system if I haven’t defined what intelligence actually is. My answer is that we already judge intelligence in humans and animals without a precise definition, so why should it be any different for machines? There are lots of concepts for which we don’t have precise definitions, yet we’re able to discuss them coherently. They’re the “I know it when I see it” concepts. I regard intelligence as one of those. The boundaries might be fuzzy, but we’re able to confidently say that some activities require intelligence (inventing the calculus) and others don’t (breathing).
I know that some readers will disagree with my functionalist view of intelligence, and that’s good. It should make for an interesting discussion.
Recorded instances are rare.
But several things need to be said: the car can see in all directions and can quickly evaluate possible escape routes. The specific options are not programmed. The scenarios are trained, and no one can predict the actual action taken.
Another point: the cars are constantly evaluating distant objects, and in actual cases, avoid getting into desperate scenarios. There are dozens of videos of situations that could be tragic, but are avoided so smoothly that humans may not even realize the problem.
Then there are scenarios where no effective action is possible. I took a defensive driving course some years ago, and we were told to avoid head on collisions at all cost, even if it meant steering into a solid object.
Simple crash avoidance systems have been around for a while. Statistically, they are much better than humans. AI is better, and it is improving quickly.
One other thing: Tesla has been updating software frequently this year. They are able to take incidents from beta testers and distribute updates in a week or two.
I’m aware of one recent head on collision between a Tesla truck and A BMW driving on the wrong side at high speed. Only ten percent of Tesla owners have FSD, and not everyone has it activated all the time.
I noticed something interesting. If you look at the initial, fully randomized noise at the beginning of the sequence above, there happens to be a dark patch, which I’ve circled here:

Her eye ends up developing in that spot. You can tell it’s the same spot by noting the distinctive yellow squiggle that’s above it in both of these images:


That’s interesting, because knowing how diffusion models work (which I’ll explain in a future OP), I can see how it would be tempted to put a dark feature in a spot that was already dark in the original random noisy image.
Is that what’s going on here? I don’t know, but perhaps I’ll do some experiments to see if I can doctor some original pure noise images in order to coax the model into putting features at predetermined locations.
Beyond Weasel?
I read that the designers of the 286 were concerned that a malicious program able to get control in real mode could rewrite the interrupt and other tables, or abuse call gates, or otherwise trash (or even become) the OS. Which is why the BIOS had to go through all that exercise to test memory and support VDisk via the 8042.
I understand that you have seen protected mode OSes switch to real mode, but I’m pretty sure that that wasn’t possible on the 286. If the OS running on the 286 could “transition to real mode” the BIOS could have done it also.
Obviously there is no such difference.
There was an instance in Ukraine war when dormant AI-driven drones were transported (by unsuspecting Russians) close to several targets, then at a given moment the drones broke out of their packaging and started flying around. The moment was pre-programmed – the same moment for the entire fleet of drones. The target areas were pre-determined coordinates, and upon arrival the drones had to identify specific military airplane-like, tank-like and other such objects to detonate themselves on. This is as close as drones became to “selecting targets themselves” – frankly not at all. And if it were any other way, it would be a scandalous war crime.
It is astonishing how little both of you know on this topic. Seems like you have been through intensive unlearning courses and excelled at that.
petrushka: literally everything you say about self-driving is catastrophically wrong. You swallow market hype uncritically and you are not allowing real-life user feedback correct you.
I’m reminded of the joke of the lady watching the parade and noticing that “everyone in the whole parade is out of step except my son – and that includes the drummers!” I guess nobody but Erik can see the obvious – even those who have long professional careers in the discipline!
So we’re back to Dawkins:
In the face of this position, even Dawkins was helpless.
Flint,
If you got something to refute then why don’t you? Because you got nothing, that’s why.
keiths and petrushka have abandoned their expertise, if they ever had any in the first place. They don’t know what simulation is, they don’t know what software is, and, as it turns out, they also don’t know what hardware is. They don’t know how any of these things work, either in broad principle, technically, or legally. Do you? Can you post a fact for a change? For now, I’m the only one who cited actual facts in this thread instead of blather.
Erik:
Identifying potential targets and then picking one to go after is selecting a target.
It is astonishing to me that you can’t grasp petrushka’s simple point. Here are two scenarios that illustrate the difference:
Scenario #1:
You’re a soldier in combat. You see a squadron of enemy tanks approaching. You select a tank, point your Javelin at it, and fire. The Javelin hits the tank you selected.
Scenario #2:
You’re a soldier in combat. You launch an autonomous drone that has instructions to fly to a predetermined point and loiter. While it is loitering, a squadron of enemy tanks enters its field of view. It selects a tank, flies to it, and detonates.
In scenario #1, the soldier selected the target. In scenario #2, it was the drone that selected the target. The soldier didn’t select a target, because he didn’t know what the available targets were or would be. He was depending on the drone to select a target, which it did.
Here’s an analogy. The commander of a squadron of A-10s gets a radio call. Some ground forces are pinned down near Kandahar. The commander sends a pilot to that location to provide close air support. The pilot flies to that location, selects a target on the ground, and fires at it.
In that scenario, who selected the target? Was it the squadron commander, or the pilot? It was the pilot, obviously. The squadron commander’s role was to give instructions to the A-10 pilot. The pilot’s role was to fly to the combat zone and select and destroy targets.
The squadron commander is analogous to the soldier who launched the drone, and the pilot and his aircraft are analogous to the drone. It was the latter who selected the targets.
Erik:
You crack me up, Erik.
Are you ever going to answer my question?
You wrote:
I asked:
“No evidence, no matter how overwhelming, no matter how all-embracing, no matter how devastatingly convincing, can ever make any difference.” At first, I thought Dawkins was exaggerating. You have proved him right.
Flint:
But they made writes to the MSW privileged in protected mode. Since writes to the MSW are privileged, there is no danger of rogue programs trashing the IDT or other critical structures. The OS has full control of what code does and doesn’t get to run in real mode, and it can limit that access to trusted code such as legacy device drivers.
It’s analogous to CPL 0. The OS has control of who gets to run at that privilege level, so there’s no danger of user programs mucking with sensitive structures like the IDT, the page tables, or the hardware itself.
My point is that if you make MSW writes privileged, which is what the 286 architects did, then there is no reason to block the processor from entering real mode when PE is cleared. That’s why the 386 and beyond permit it.
Right. The 286 couldn’t do it, and that was a major architectural flaw. I’m just questioning the reason for that architectural mistake, since the architects were aware that writing to the MSW was a privileged operation. My best guess is that they didn’t think that re-entering real mode would ever be necessary, not anticipating that real mode would be needed to handle legacy device drivers and certain BIOS calls from legacy programs.
petrushka:
[A note for anyone who is unfamiliar with ‘Weasel’. Weasel was a toy program written by Richard Dawkins to demonstrate the basic evolutionary principle of random variation and natural selection. It engendered some lively discussion between us and proponents of Intelligent Design, both here and at William Dembski’s Uncommon Descent blog.]
Interesting question. There are some parallels and some disanalogies. Let me think out loud.
Targets:
Mutations:
Selection:
None of the three — Weasel, evolution, or diffusion models — do what became known in the discussion as “explicit latching”. That is, they don’t lock changes into place in order to prevent further mutations from “undoing” the beneficial ones. That was a hot topic in the Weasel discussion, because the IDers were erroneously convinced that Weasel cheated by latching, which is something that doesn’t happen in nature.
The latching business was pretty funny, so I went back and googled parts of the discussion. I think my favorite bit was when the inimitable kairosfocus, having been shown that Weasel didn’t latch, insisted that it was “implicit quasi-latching”.
Good times.
ETA: kairosfocus is still at it, using the same turgid prose we found so funny:
When you look at your example of evolving an image, consider that self driving computers do not have to be explicitly programmed. The actual process of training requires one of the largest supercomputers in existence, and the largest dataset in existence.
Over the course of twelve years, the set of scenarios has been refined. It started with billions of miles of actual driving and has been refined. I’ve read that the current training data is synthetic, not because the situations are too complex, but because actual human drivers are too sloppy.
Driving is a bit like Douglas Adams definition of flying: aim for the earth and miss.
The critical part of driving is to aim for the destination, and avoid crashing.
I think you’re pretty much correct here. Seems clear to me that the 286 architects were clueless about the POST, but I think it went well beyond. My reading (long ago, but written by one of the 286 team) was that they figured the 286 would come out of reset and the OS would take control immediately. No POST, no DOS, no legacy programs, no device drivers not written for this hypothetical OS.
This isn’t a stupid or far-fetched picture – it’s pretty much what linux does. A dedicated linux PC has only a tiny ROM that knows little more than how to load sector 0 from the disk. The sector 0 code then loads a few more sectors in real mode, enough code to build the required code, date, and interrupt descriptor tables, hooking interrupt entries to protected mode drivers, after which it’s all protected. So there are no backward compatibility issues, no legacy drivers, no stupid software tricks like we discussed earlier. My linux experience is limited to boot ROMs, so I may have the rest of this wrong…
petrushka:
I was thinking recently about an alternate universe in which we somehow didn’t know that human cognition was based on biological neural networks. Would we have stumbled upon the neural network architecture as a way of building AI, or did we absolutely need the hint from nature? Are there other ways of implementing robust machine learning that don’t depend on neural networks or something mathematically equivalent, like transformers? Where would we be today if Minsky and Papert hadn’t proven the limitations of perceptron networks, thus putting the brakes on the field, or if someone had invented back propagation (the algorithm that allows deep networks to learn) sooner?
I’ve lost track of what I’ve posted here, but I’ve been thinking about variations in human intelligence for fifty years.
In evolution, taking a path can preclude alternative paths. Humans are unlikely to develop wings.
I’ve wondered if humans taking certain paths in early learning are precluded from becoming proficient in some tasks. And vice versa. There is the somewhat disturbing possibility that biological evolution could predispose individuals to certain paths.
I’m not a big believer in “g”. I think g is academic proficiency, and our world favors that. But I’m thinking it’s possible to be born with greater or lesser propensity toward a set of skills, and life experience amplifies initial conditions.
Not unlike your image evolver amplifies variations in noise.
petrushka:
I once read that learning to read co-opts neural circuitry that is used for other purposes in illiterate people. Presumably that means that illiterate people are better at certain things because that portion of their neural circuitry hasn’t been hijacked.
That’s a specific example, but it isn’t hard to believe that it’s a more general phenomenon, so that certain learning paths preclude others or make them more difficult by virtue of the fact that they’re recruiting neural circuitry that could otherwise be employed differently.
Isn’t that a given? Most of what we do is either selected for specifically, or is a byproduct of something else that was selected for.
Except that performance is correlated on a wide range of seemingly unrelated skils, some of which aren’t academic. Think Raven’s progressive matrices, for instance.
Definitely. Twin studies demonstrate both the heritability of g and of specific skills, like mathematical ability.
You seem to be rediscovering the old saw that as the twig is bent, the tree is inclined. Or, give me the child until age 6 and I will give you the man. I think there’s no question that the entire nature of education and training is a tacit recognition that neural circuitry can be purposed more or less permanently.
(As a footnote, I notice that all the guitar gods began playing early in childhood, none of them past puberty. I tried to learn guitar in my 60s, and with the help of a childhood playing musical instruments, I made it all the way to intermediate level before arthritis made further progress impossible, but to be honest I never would have become much better despite 4 hours of practice a day.)
In an earlier comment I explained why AI video generators can’t generate frames sequentially but instead have to do them all at once, which limits the length of the clips. That presents problems when trying to stitch clips together, and the problem I described in that comment is that motion won’t necessarily match between a clip and its successor, so that the “seam” is visible because the motion changes noticeably. I talked about some workarounds in that comment.
Thinking about it some more, there are other difficulties. Suppose you’re doing a scene where someone enters their boss’s closed-door office, talks to the boss, and then leaves. You see them outside in the cubicle area first, then you see them in the office. Suppose your first clip ends there. Now you have to generate a second clip showing them leaving the office and returning to the cubicle area. How do you guarantee that the cubicle area looks the same on the way out as it did on the way in? When it’s generating the second clip, the AI has no idea what that area is supposed to look like, other than what is specified in the prompt, and there are many different ways of rendering that area to match the verbal description in the prompt. Odds are that the cubicle area will look quite different in the second clip.
Another problem is that if you’re generating the second clip using the final frame of the first clip as your starting point, there’s an underdetermination problem. The information in a 2D image isn’t usually sufficient to imply the correct 3D situation, so while the second clip will match the first clip right at the seam, it might look quite different as the frame sequence proceeds. I’ll explain with an example.
In a 2D photo or image of a person’s face, you see them from one particular angle. Your view reveals a lot about the 3D structure of their face, but it doesn’t reveal everything. Some information is missing. For instance, if they’re facing the camera directly, you’re seeing their nose “edge on” and you can’t really tell what it’s going to look like when seen from the side. So when an AI is creating a video of them, it has to fill in the gaps and decide on a 3D structure that matches the 2D image.
My experiment was to take this image* produced by Midjourney…








…and generate a bunch of video clips from it on my home PC, using a prompt that would cause her to turn her head and show her profile to the virtual camera. I ended up with a whole range of profiles. Here is a sampling:
They look like different people, but I’ll note that the videos all looked very smooth. In each case, the diffusion model picked a 3D structure that dovetailed perfectly with the 2D starting image. All quite natural-looking. One 2D image, many compatible 3D structures.
The diffusion models are very good at maintaining a consistent 3D structure within a clip, but when you’re stitching two clips together by using the final frame of the first clip as the starting frame of the second clip, that 3D structure isn’t carried over. The model has to fill in the gaps again, and it might do so in a way that’s very different from the way it filled them the first time. You could start out with one of the women above in your first clip and end up with a different woman in your second clip.
* It was one of the four images that Midjourney generated when I asked it simply to “Generate an image”, with no other instructions. An exercise in free association.
petrushka:
The consensus in the field is that g really does exist and is independent of academic proficiency. Here’s a relevant paper:
Spearman’s g Found in 31 Non-Western Nations: Strong Evidence That g Is a Universal Phenomenon
keiths:
Flint:
There’s a general decline in plasticity from childhood to adulthood, which is why adults find it harder to learn new things, but I’m talking about something other than that decline. I’m referring to the fact that neural circuitry that has been repurposed is no longer available to carry out its original function. If skill A and skill B are competing for neural resources, proficiency at one may come at the expense of proficiency at the other.
Sorry to hear about your arthritis. Growing old sucks, doesn’t it?
I found that photo online during the “Trump Always Chickens Out” period. It’s one of my all-time favorite AI images, so now that I’m doing video generation on my home PC, I couldn’t resist using it as a starting image. The results:
Trump in action
It not just about repurposing.
Everything we learn about how the world works makes it more difficult to learn alternative schemas. Woody Allen made fun of this in “Sleeper” by depicting a future in which all the dietary recommendations are turned upside down.
The serious point is, science requires reappraisals of everything, and individuals find this hard to do.
“One death at a time…”
petrushka:
I know, which is why, in the comment you just quoted, I made the point about plasticity. Even if repurposing didn’t happen, the inevitable reduction in plasticity would make it harder to unlearn things and learn new things in their place.
An excerpt from a comment in another thread:
I’m reposting it here because the script exercise was interesting. It took ChatGPT only four tries to get the script right, and one of those was my fault for not specifying that I wanted the script to count ‘boarders’ along with ‘boarder’. The entire exercise took less than ten minutes.
Before the advent of sophisticated AI, I would never have bothered to write such a script. Learning the library calls alone would have taken more than ten minutes, and the whole project would have taken much longer. It wouldn’t have been worth it. With AI, it only cost me ten minutes of my time, and that made all the difference.
As a curious person, I find lots of questions popping into my head every day. Before AI, I would triage those, bothering to investigate only the ones that were particularly interesting and worth the investment in time. Now I can get answers easily, sometimes within a minute, from AI.
Example: I was on US 101 the other day behind a semi that had just merged into traffic from an on-ramp. It was evidently fully loaded and taking forever to get up to speed, and I wondered “how underpowered would a typical sedan have to be in order to accelerate as slowly as a fully loaded semi?”
Here is ChatGPT’s answer:
I checked with other AIs and they got the same answer. Curiosity sated in less than five minutes.
Some experiments with diffusion models. I found some collage faces online that I thought they might have trouble with, and used them as the starting images for video clips.
I wondered if the model would treat this as a jumble versus recognizing that it was a person and consolidating the pieces into a normal face. I kept the prompt neutral — it was literally the single word ‘nothing’ — because I didn’t want to give any hints. I tried it five times, and the model always consolidated the pieces into a normal-looking face. Here’s an example video:
Fragmented woman gets reassembled
Next image:
I wondered if the model would recognize this as a face, but do so without morphing it into something more humanlike. It succeeded. I could make ‘her’ sing, and she would blink her eyes realistically. I even asked her to cry, and she did so in a style that perfectly matched the style of the collage. That’s impressive:
Ms CollageFace cries
Next image:
This one is so clearly a person that I thought the model would irresistibly normalize her face. It usually did, but not always. Sometimes it maintained all four eyes and both mouths. Once I saw it consolidate the mouths but maintain all four eyes. Here’s an example where all of the eyes and mouths are maintained:
Ms FourEyesAndTwoMouths smokes a cigarette
ETA: These videos were all generated using Wan 2.2, the latest model from Tongyi Lab, a subsidiary of Alibaba.
I’m not disagreeing, but I don’t like the term plasticity. It implies a hardening of something and a slowing down of learning.
I don’t think that describes what is happening. I prefer to think of learning as a kind of evolution, and taking a path precludes taking the alternate paths. You may converge on a similar solution, but maybe not as quickly or efficiently.
Case in point. Children in multilingual homes use a common brain area to learn multiple languages. They do not need to translate.
People who learn multiple languages later in life use multiple physical locations in the brain, and are not as adept at translation.
I think that in addition to any biological advantage individuals may have that contribute to intelligence, early learning bends the twig. Twin studies are mostly flawed, but they document the possibility of at least a full SD of environmental influence. The difference between having to work hard in college, and getting a PhD.
Video of the outlet singing
The story behind this one: I have a friend who was in a session with her new therapist, who didn’t know her well yet. My friend casually remarked that the wall outlets resembled faces — it’s pareidolia, and they look like faces to me, too. Her therapist asked, very seriously, “Do they speak to you?”
We still laugh about that, so I had to make a video for her of an outlet that does speak and even sings. It’s interesting, because it demonstrates that the diffusion model itself also experiences pareidolia of a sort. It recognizes the “face”. It knows where the “eyes” and “mouth” are, and it isn’t thrown off by the absence of a nose. It put a red “tongue” inside of the mouth, and the eyes are quite expressive as the outlet sings. There’s some very abstract representation happening in that neural network.
ETA: I should add that the prompt was simply “It starts singing.”
The image above is very flat and 2D, so it’s not hard to animate once the model decides where the “mouth” and “eyes” are. I wanted to challenge it a bit more by giving it a photograph of a real outlet:
Similar to the above, the prompt was simply “they start singing”. The most common response was that the model recognized the rigidity of the plastic and didn’t deform the slots or the ground hole. It did recognize that the outlets weren’t attached to the plastic plate, and so it exploited that by moving them in and out. In this particular video, the model made the mouths “sing” by putting a flashing light within them. Pretty clever: move what physics allows you to move, and if you can’t move or deform something, animate it by making it flash:
Rigid outlets “sing” by flashing
In some attempts, the model decided to actually deform the slots and ground holes in order to make the outlets sing:
Expressive eye and mouth movements
In this video, someone hurts the upper outlet and makes it wince, and the lower outlet winces in sympathy (I have no idea where that came from. It wasn’t in the prompt):
Wincing outlets
ETA: No outlets were harmed in the making of this video.
keiths,
‘Wincing outlets’ asks for a login to Google Drive
Allan:
Thanks. Should be fixed now.
Purely subjective ink blot impression: the schematic drawing suggests to me, a bit of surprise mixed with worry. Perhaps witnessing a minor accident.
The photo suggests disapproval of something.
I suspect most people will read something into the face.
I wonder if AI has an opinion.
petrushka:
That’s my impression too. I have a related anecdote I’ll share later if I can find a certain photo I took.
I asked Claude:
Claude: