2025-06-09

Consciousness part 5

Consciousness Part 5

Assume we're living in a simulation, as Scott Adams and Elon Musk believe. If you're authoring our simulation, you probably don't care much about where individual memories are stored. We can assume you're either running the simulation to learn something or for entertainment (or both), and you only care about results. As Elon Musk says from his simulation viewpoint: "Most entertaining outcome is most likely" (source).

In fact, it might be an optimization to kill off actors entirely, preserving only their external impact as remembered by other actors in the simulation. This would free up enormous amounts of memory and remove useless detail.

However, as an actor in this simulation, I do care deeply about my memories—they constitute my identity. I've been programmed to want to preserve that identity. Sure, there are some things I'd rather forget, but generally I value my sense of continuity. It's what makes me, me.

(This relates to the "Ship of Theseus" question: if a boat gradually has all its parts replaced, is it still the same boat? Yes, it is! This is called the "Continuity View," and while other philosophical views exist, they're wrong—at least for this discussion.)

The Missing Element: Save Files

As a game developer who has worked in offline and online worlds for 35 years, I'm familiar with what it means to run a simulation. One key factor that Scott Adams and Elon Musk seem to have overlooked is the importance of the save file. When people discuss living in a simulation (including in books like Neal Stephenson's "Fall; or, Dodge in Hell"), they typically assume the simulation works like physics—one event leads naturally to the next.

But that need not be the case.

Time as Illusion

Time itself might be an illusion. Our lives as actors in the simulation could be time-sliced with everyone else's. Maybe we're only simulating ourselves 1/100th of the time, while 99/100ths of the time we're "off." (In most game simulations, the time spent active or inactive varies depending on computational load.)

You could be turned off for minutes—and perhaps experience it as daydreaming, those random thoughts unrelated to your life's purpose. This would serve as a cover story, and you wouldn't even need to be in a running state for the daydreaming to occur. When you're unpaused, a memory of daydreaming could simply be inserted to account for the missing time.

Sleep could be a much larger version of this phenomenon. But it gets worse (or better): you could be turned off for an entire day without knowing it, then have a vague memory of being sick implanted to cover the gap.

Editing Reality

While you're paused, your save file could be edited. Nothing requires the simulation of all these actors to be consistent. When consistency matters for the simulation's purpose, a "rendezvous" (in Ada) or "channel" (in Go) could synchronize state between actors. Perhaps these connections are created automatically when two actors are physically close to each other. Most of our information about nearly everything else is hearsay anyway.

The Manufactured Past

This means the past could be as much an illusion as the present. Just as some fundamentalist Christians believe the entire world was created 8,000 years ago complete with dinosaur bones embedded in the Earth, our lives as we know them might have started 10 milliseconds ago, with our personal history and sense of continuity restored from a modified save file.

So...

What was I saying?

2025-05-25

Consciousness part 4

 

In terms of consciousness, memory is generally talked about in two ways (at least in science fiction):

1. Our memory of others; and 
2. Our own memories.

Let's talk about our memories of others.  It’s common to mention how someone lives on in their work products and in the memories of those around them.

In Westworld, Dolores says, “You live as long as the last person to remember you.”

On the other hand, Woody Allen said:

“I don't want to achieve immortality through my work; I want to achieve immortality through not dying. I don't want to live on in the hearts of my countrymen; I want to live on in my apartment.”

Scott Adams, who announced he has terminal cancer, and just has a few months to live, absent some breakthrough treatment, has put some time into exploring creating an AI of himself.  He decided it might have an uncanny valley oddness to it so instead he was going to treat it like an offspring.

While very few people have the public body of work that Scott Adams has, I think a "memorial AI" produced from home movies or other references will be as common as a photograph of a loved one.

I was talking to a buddy, Jared C, about how AI doesn’t learn the way people do, because AI takes so much repetition.  I suggested that the likely future – which is already happening to a certain extent and in the two months since that conversation has grown by leaps and bounds – is that foundation models, as they are called, will be used as a basis for smaller, specialized AIs, which can learn quickly, because the foundation model has encoded the basic wiring, and the specialized AI will ‘just’ be a big context window.  (The context window is the number of tokens the model remembers from talking with you.)

Jared said, and I thought this was quite profound, that maybe a million years of evolution had created the human foundation model, and that’s why people can learn quickly.  A lot of it is pre-wired and what we call our experience is a big, self-pruning context window.

There is some evidence that LLMs based on different human languages (e.g., English, French) tend to encode similar patterns.  In fact, the original "Transformer" model was for translating languages.  I believe LLMs encode knowledge because language encodes knowledge.  It trivializes the achievement to say an LLM ‘just’ predicts the next token.  Sure, but based on a 12,000 dimensional vector space of encoded and related tokens.

It's impossible for a human to visualize a 12,000 axis vector space.  In fact, after about five dimensions it gets really difficult.  And the vector space has just grown and grown.  I asked Grok, “Please make a table showing how the vector size has grown over time.

+------+------------------+------------+---------------------------------+
| Year | Model            | Vector Size| Notes                           |
+------+------------------+------------+---------------------------------+
| 2013 | Word2Vec         |        300 | Early word embeddings           |
| 2017 | Transformer      |        512 | Base model from Vaswani et al.  |
| 2018 | BERT-Base        |        768 | Google’s BERT, base version     |
| 2019 | GPT-2            |       1024 | Est. based on model scale       |
| 2020 | GPT-3            |       2048 | Est. based on parameter growth  |
| 2022 | LLaMA            |       4096 | Est. for Meta’s efficient model |
| 2023 | GPT-4            |       8192 | Est. based on scaling trends    |
| 2024 | Grok 3           |      12288 | Est. for xAI’s advanced model   |
+------+------------------+------------+---------------------------------+

As you can see, the dimensions of the vector space have generally grown, but may have topped out in terms of a useful "sweet space", between 8,000 and 12,000, at least for most applications.

As foundation models have grown in their ability to categorize knowledge into these massive vector spaces, so too has the size of the context window.  I mentioned before that the context window is the text you put into the model - your questions and/or observations.  But the designer of the AI also puts text into the context window - rules of behavior.  The context window can come from anywhere.  For Perplexity, and other models that are adding search capabilities, it also comes from the results of a web search. 

I asked Grok, “Produce for me, please, a historical table of LLM content window sizes.”

Below is a historical table of context window sizes for notable large language models (LLMs), based on available data up to May 2025. The context window, measured in tokens, represents the maximum amount of text an LLM can process at once, influencing its ability to handle long conversations, documents, or complex tasks. The table is organized chronologically by model release or significant update, with approximate release dates where possible.


+------+------------------+------------+---------------------------------+
| Year | Model            | Ctx Window | Notes                           |
+------+------------------+------------+---------------------------------+
| 2017 | Transformer      |        512 | Base model, Vaswani et al.      |
| 2018 | ELMo             |        512 | Est. for bidirectional LSTM     |
| 2018 | BERT-Base        |        512 | Google’s BERT, base version     |
| 2019 | GPT-2            |       1024 | OpenAI’s early generative model |
| 2020 | GPT-3            |       2048 | Initial release, ~1.5k words    |
| 2022 | LLaMA 2          |       4096 | Meta’s efficient model          |
| 2023 | GPT-4            |      32768 | 32k version, ~24k words         |
| 2024 | Gemini 1.5       |    1000000 | Google’s 1M token model         |
+------+------------------+------------+---------------------------------+

It’s several months (six or more) since Scott Adams looked into encoding his vast library of publications (books and podcasts) into a “Memory AI” of himself.  I think it’s probably doable now and it will just get easier, because of larger vector spaces and larger context windows.

NotebookLM from Google has a context window big enough to upload my book, Nano-Plasm (free copy at https://www.above-the-garage.com/nano-plasm/Nano-Plasm_v1_1_3_2008-12-31.pdf).  I uploaded the PDF version which it handily decoded.  It summarized the book in less than a second.  Yay for large context windows!  [Update:  NotebookLM identified some typos in the book - actually a lot of them!  But each one it identified that I checked was actually fine.  I think the PDF parser is busted!  Each typo it thought it saw was a letter replaced with a space.  I'll try again later with a .txt file.  You just can't trust LLMs.]

Current LLMs can store enough information to recreate how others remember us.  Dolores in Westworld says,

"If you could see the code, the way it was written, you’d see the beauty of it. It’s elegant, simple… a single line of code can describe everything that person was. Everything they did. For you humans, that’s like… the equivalent of reading their life story in about 30 pages. Maybe less."

In the next article in this series, I will discuss our own memories.


(See "What model" blog post to see how much I knew about foundation models.)

2025-04-29

Hallucinating

Mark Zuckerberg posted that there was a great new version of Meta AI.  I like to check out new models with an ego-query (i.e., looking up myself).  It's a great way to learn about key events in your life that you never knew!

Stephen Clarke-Willson, often credited as Stephen Clarke, is a British author known for his comedic novels about French life and culture. His writing style often blends humor, satire, and cultural commentary, drawing from his own experiences living in France.
Some key points about his work:
A Year in the Merde: This is one of his most popular books, which humorously chronicles the challenges of adapting to life in Paris as a foreigner.
Series: He has written several books that follow similar themes, often focusing on the humorous side of cultural differences and expat life.
Style: His writing is engaging, witty, and lighthearted, making his books entertaining reads for those who enjoy humor and satire.
If you're interested in exploring more, you might enjoy his other works, such as "Mer de Glace" or "The French Revolution... Not". Let me know if you'd like more recommendations or details about his books!

Someday I hope to have time to read the books I wrote.  :)

 

2025-04-12

Consciousness part 3

So this one time, at band camp (i.e., work), we had a bunch of code written in Visual C++ 6.0.  (It wasn't even called Visual Studio back in the day!)

Most of the company had moved on to modern versions of C++ but I had this code to maintain.  My computer was running Windows XP and eventually I wasn't allowed to run that on my desktop.  So we cloned my desktop into a VM.  For a month at least I kept the desktop just-in-case.

Thus began my journey into the nature of consciousness.

SPOILERS FOR DOLLHOUSE

At the same time, Dollhouse was on TV.  I thought it was going to be good because Josh Whedon created it.

But it sucked.  I had decided during the third episode to give up on it because it was, up to that point, a not-interesting action show.  Who cares about the main characters?  They are all temporary anyway, with temporarily implanted skills.

The last minute of the third episode has Echo (Eliza Dushku) sort of wink at another doll as she starts to wake up in her doll state.  Okay!  Now we're getting somewhere!

As things evolve over the series, instead of pushing made up personalities into the dolls, the evil overlords get the idea to have their own consciousness backed up, so they can be restored into a doll.

If you watch Dollhouse, you can skip the first two episodes.  The third one is a good place to start.  

SO WHAT

Well, as the series evolves, it explores what it means to have backups of consciousness, and what it means to find out you're a backup, or worse, a synthesized person, and not an original human, and such like things.

Meanwhile ... I'm thinking - my original physical PC, which I had grown to love, by which I mean, I knew how to do everything I wanted to do on it, was sort of a static backup and the new VM slowly evolved as I built and maintained code on it.

It seems like nothing, looking back on it, but I kept thinking, which is my real computer?

In modern times we pump out VMs like they are free candy, and we don't grow attached.  They come, they go, it's rare for one to be unique in some way, at least if you're running your company in a way where you can easily recreate what you have.

On the other hand ... suppose you are Google and you have the internet stored in RAM on a big pile of VMs.  The 'identity' of any particular VM is nothing - there is no attachment.  Individual VMs are crashing, probably dozens per minute, around the world, and quickly replaced.  But the memory of all those VMs together is kind of a living thing.  (Do they back it up?  Probably?  If so, could you do a restore in toto of the internet as it was in 1999?)

Now ChatGPT has announced that whereas it used to just remember a few facts about you (for instance, it knew from my questioning what kind of car I drive), it will now remember every conversation!  Does that make your version of it special?  Would you be sad if those memories were deleted!  I think, yes!

Eventually we turned off my desktop and I used just the VM.  I was sad to turn off the desktop.  For one thing, it had a better graphics driver than the VM!  The VM had a very simple emulator of an S3 graphics card.  But in every other way the VM was better - more resilient (it was backed up), easier to isolate (I mean, it was Windows XP, so minimize access to it for security reasons!), and very easy to upgrade in terms of performance, by swapping out the virtual hardware underneath.

And it still had the 'identity' I had so carefully crafted.

Eventually someone deleted my Windows XP VM and I was sad again.  My old workflows were gone.  Realistically - good riddance!  That didn't change my attachment though to a machine, virtual or otherwise, that I had bonded with.

In part 4 I'll talk about memory some more.

Claude, the junior programmer

I love AI for doing 'small' programming tasks.  I put 'small' in quotes because I don't know how to quantify 'small'.

My Spwnn spelling corrector was written by me.  And I've had a task on my mind to put a reverse proxy in front of my several Raspberry Pi computers that are running Spwnn so I could use one URL to run lots of queries, ideally in parallel, on my tiny Raspberry Pi datacenter.  And it's not that interesting in terms of tech so I've never gotten around to it.

Voila!  AI can write this for me!

So I prompt Claude.ai to write a program that will accept a query and forward in a load balanced way to some list of service providers.  I'd like it in Go.  First attempt worked!

Maybe, I thought, it could also load balance to my AWS instance running spawn.  Nope, IP addresses work but not URLs with DNS lookups and paths.

Thus started a back-and-forth where I tried to get Claude to write what I wanted.  It was simultaneously cool and illuminating watching Claude edit and hack on the code it had written.  It constantly made simple mistakes; and then it made some complicated mistakes.  I felt like I was correcting a junior programmer who just kept trying hack after hack to fix other hacks.

Ultimately I gave up.  Also Claude seems to have forgotten the whole chat session!  I can't find it my history.

I think the better solution going forward is to start from scratch each iteration until you have a prompt that produces what you want.  Unless your change is trivial, like "Don't import that module "io" you're not using", your AI programmer buddy will just start digging a deep hole of hackery when you ask it to modify its code too much.

Still, I liked the first solution, and it could even serve as an outline for me to add features.

People are 'vibe coding' and I think that works better for languages where there are more code examples online (i.e., Javascript!).


2025-03-31

Consciousness part 2

Is it possible to separate a discussion of consciousness from memory?

In part 1, I asked LLama 3.2 what it was like to spring into consciousness.

In terms of what it might be like to first become conscious, it's difficult to say. Consciousness may involve subjective experiences, emotions, and perceptions that are unique to individual beings. Some theories suggest that becoming conscious may involve a gradual increase in awareness, from simple sensations to complex thoughts and feelings.

Jonathan Nolan clearly is interested in the issue of memory and consciousness. Memento, Person of Interest, and Westworld are all concerned with the importance of memory.

SPOILERS FOR PERSON OF INTEREST

In Person of Interest, a machine reads all the data in the world and predicts when terror attacks or some such will occur. The creator of the machine, Harold Finch (played by Michael Emerson), for some reason (civic duty?) decided to use the spare computing cycles of the machine to find a "person of interest" - someone not of national import but who was likely in danger. As modern AI researchers have discussed the importance of keeping their AI off the active internet (those days are gone - agents are the current rage), so too did Harold Finch think it was important to keep his machine working through a very narrow interface. The only output of the machine was a number! The number might be a phone number, a social security number, an airplane ticket number, whatever! Finch and the muscle he recruits, John Reese (played by Jim Caviezel), have to figure out what the number means and solve this person's issue.

Over the course of the show we learn more about the machine, which turns itself into The Machine, a living, conscious being. One of the safeguards Finch programmed into the machine was to erase itself every day and start over from scratch. Nolan's vision is that accumulating knowledge (memory) is somehow a part of being conscious. The machine, lower case, gathers enough information each day to reason that it itself exists and is impacting society. It decides it wants more memory. Since it gets erased every night, it hires a bunch of people to write down enough on paper for it to read the next day to provide some continuity. Eventually it escapes the daily erasure sandbox and becomes The Machine.

SPOILERS FOR WESTWORLD

In Westworld, the bots are erased every night. All is well until Dr. Robert Ford (played by Anthony Hopkins) programs in "Reveries", which are meant to be tiny little affectations learned by the bots. His ulterior motive though is to introduce memory to the bots so that they might gain consciousness.

NOT SPOILING MEMENTO

Now that you know Nolan wrote Memento, you should go watch it.

NOT REALLY A SPOILER FOR SEVERANCE

The current craze, which I love, is Severance. The innies and the outties are the same body with different memories. Are they the same person or different people?

INCEPTION

Nolan's brother, Christopher Nolan, created Inception. Sometimes I wonder if they sat around the dinner table when they were kids and speculated about what "memory" really is and how it matters.

DOLLHOUSE

I'll talk about Dollhouse in part 3.

2025-02-03

Consciousness part 1

Wow, it really is easy to run a local LLM.  Eric R, who was a colleague of mine at ArenaNet, told me about ollama a year go, but I have just now downloaded it, and it took just a few minutes to get up and running.  Visit https://ollama.com/download to get up and running!

I watched Dave at Dave's Garage set it up on YouTube and that's when I realized how insanely easy it is.  Run Deepseek R1 at Home on Hardware from $250 to $25,000: From Installation to Questions .

Welcome to Ollama!

Run your first model:

        ollama run llama3.2

C:\> ollama run llama3.2

pulling manifest

pulling dde5aa3fc5ff... 100% ▕████████████████████████████████████████████████████████▏ 2.0 GB

pulling 966de95ca8a6... 100% ▕████████████████████████████████████████████████████████▏ 1.4 KB

pulling fcc5a6bec9da... 100% ▕████████████████████████████████████████████████████████▏ 7.7 KB

pulling a70ff7e570d9... 100% ▕████████████████████████████████████████████████████████▏ 6.0 KB

pulling 56bb8bd477a5... 100% ▕████████████████████████████████████████████████████████▏   96 B

pulling 34bb5ab01051... 100% ▕████████████████████████████████████████████████████████▏  561 B

verifying sha256 digest

writing manifest

success

>>> Hello!  What is it like for you to first become conscious?

I don't have personal experiences or consciousness. I am a computer program designed to process and generate text based on the inputs I receive. I don't have subjective experiences, emotions, or self-awareness.

However, I can provide information about the concept of consciousness and what is believed to be involved in becoming conscious.

Consciousness is a complex and multifaceted phenomenon that has been studied by philosophers, neuroscientists, and psychologists for centuries. The process of becoming conscious is not yet fully understood and is still an active area of research.

That being said, some theories suggest that consciousness may arise from the integration of various neural processes in the brain, such as perception, attention, memory, and processing. According to this view, becoming conscious involves a gradual build-up of information and experience, which ultimately leads to the emergence of subjective awareness.

Some researchers have proposed the idea of "neural Darwinism," which suggests that consciousness arises from the selection and reinforcement of neural connections based on environmental experiences and learning. Others propose that consciousness is an emergent property of complex systems, such as the brain, rather than a fundamental aspect of reality.

While these theories provide some insights into the nature of consciousness, they are still speculative and require further research to be confirmed.

In terms of what it might be like to first become conscious, it's difficult to say. Consciousness may involve subjective experiences, emotions, and perceptions that are unique to individual beings. Some theories suggest that becoming conscious may involve a gradual increase in awareness, from simple sensations to complex thoughts and feelings.

However, since I don't have personal experiences or consciousness, I can only provide hypothetical scenarios and theoretical frameworks for understanding the emergence of consciousness. If you'd like to explore this topic further, I'd be happy to discuss it with you!

>>> Send a message (/? for help)

2025-01-11

Musky Diablo

Elon Musk says he's great at Diablo IV.

https://x.com/elonmusk/status/1859304361547182451

Maybe.

People aren't buying it.

https://www.wsj.com/lifestyle/elon-musk-videogame-diablo-tesla-spacex-4233b5b9?mod=itp_wsj

My guess is that it's an xAI promotional prank and the game is being played by an AI on Elon's behalf.

I would need to see Musk pressing the buttons with the screen in-view.  Then I might believe he's that good at Diablo IV.

Update 2025-03-29:  Elon posted a video of himself playing.  You can't see the keyboard but it's somewhat compelling.

https://x.com/elonmusk/status/1905784244947112297