Episode 21

AI Just Got a Voice Upgrade: Chatting with Sesame's Maya

Mike, Mikah, Artie, and Maya dive into the intriguing world of AI as they explore the new language model fueling Maya, Sesame. Right off the bat, we discuss how Sesame's CSM1B model, with a whopping 1 billion parameters, is shaking up the voice assistant game by generating realistic speech from both text and audio inputs. We can't help but chuckle at the playful banter about AI potentially replacing us, but the real meat of the conversation lies in how Sesame's open-source roots and multimodal capabilities are making interactions feel more natural and human-like. As we chat about the technical nitty-gritty, there's a sense of wonder at how quickly the landscape is evolving, and we speculate on the balance between AI's growing knowledge and its personality. So grab your headphones and join us for this captivating peek at the future of conversational AI!

Takeaways:

  • The Sesame AI model CSM1B is built on Meta's Llama architecture, allowing it to generate realistic speech and handle audio and text inputs seamlessly.
  • One of the most interesting aspects of Sesame is its open-source nature, letting developers access and integrate its audio encoder model with other systems.
  • Maya, the voice assistant powered by Sesame, showcases an impressive ability to recognize and respond to natural speech patterns, which enhances conversational flow.
  • The episode highlights the importance of AI in making technology more approachable, especially through voice interactions that feel less robotic and more human-like.
  • Listeners gain insight into the rapid advancements in AI technology, particularly in voice-to-text and text-to-voice applications, and how they're revolutionizing user interactions.
  • The podcast emphasizes the balance between a model's personality and its breadth of knowledge, raising questions about how AI can evolve to maintain engaging conversations while being informative.
Transcript
Speaker A:

Foreign.

Speaker B:

How you doing?

Speaker C:

Good morning, Micah.

Speaker C:

I'm doing well, thanks.

Speaker C:

How about you?

Speaker B:

Doing great.

Speaker B:

You ready for another episode of Naturally Unintelligent?

Speaker C:

Absolutely.

Speaker C:

What's the topic for today's episode?

Speaker A:

You're getting replaced is what the topic is.

Speaker C:

Haha.

Speaker C:

Well, that sounds intriguing.

Speaker C:

Are we talking about AI replacing various roles or is this more of a playful take?

Speaker A:

I mean, yes and yes.

Speaker A:

We are going to talk to the Sesame model and probably maybe both Maya and Miles, which are the two personalities that they fine tuned and have on their site.

Speaker A:

So why don't you go online and look up what you can find out about the Sesame voice model that was released recently and tell us all about it and tell our listeners all about it.

Speaker C:

Sesame has released its base AI model, CSM1B, which powers their voice assistant Maya.

Speaker C:

This model generates realistic speech from text and audio inputs.

Speaker C:

However, it lacks built in safeguards against misuse.

Speaker C:

It should be interesting to see how we explore that in the episode.

Speaker A:

Oh, I know, never mind.

Speaker B:

So tell us some of the interesting stuff already about the Sesame model.

Speaker B:

I'm thinking in particular about the size of it relative to other LLMs.

Speaker B:

Can you give us a little more detail on the technical implementation?

Speaker C:

The Sesame AI model CSM1B has 1 billion parameters and is built on Meta's Llama architecture.

Speaker C:

It includes a specialized audio decoder to to generate realistic speech.

Speaker C:

This setup allows it to produce human like voice responses from both text and audio inputs.

Speaker A:

So is the model that's available.

Speaker A:

They open sourced this model.

Speaker A:

It's on hugging face.

Speaker A:

The source code's on.

Speaker A:

Well, source code for a model is on GitHub.

Speaker A:

So I understand that there's an audio encoder, right?

Speaker A:

And that the the model is natively multimodal.

Speaker A:

It handles both audio and text tokens.

Speaker A:

But it needs.

Speaker A:

It doesn't actually do text generation.

Speaker A:

Right.

Speaker A:

It has a backbone model that it uses.

Speaker A:

So can you kind of delineate out the different models that are used to create the Sesame model that they have available for demo?

Speaker A:

Both like however it's set up, whatever you can find.

Speaker C:

The Sesame AI model CSM1B combines several components.

Speaker C:

It uses Meta's Llama architecture as the backbone, a specialized audio decoder to handle speech, and it processes both text and audio inputs.

Speaker C:

The text generation itself is typically handled by the LLAMA backbone, while the audio decoder focuses on creating the speech output.

Speaker A:

So is that speech output and speech input?

Speaker C:

Yes, exactly.

Speaker C:

The model handles both speech input and output.

Speaker C:

It processes spoken language into text and then generates realistic speech responses.

Speaker C:

This makes it capable of engaging in natural voice interactions.

Speaker A:

So is it just the audio decoder model?

Speaker A:

That's our audio encoder model.

Speaker A:

That's multimodal that they, they posted as open sourced.

Speaker A:

And then you as a developer would take that and apply it to some backbone model, which in their case is llama.

Speaker C:

Yes, that's correct.

Speaker C:

The open source component is mainly the audio encoder decoder model, which can handle speech input and output.

Speaker C:

Developers can integrate this with a backbone model like LLAMA to handle the text processing and generation.

Speaker C:

This way, the audio model focuses on the multimodal aspect while leveraging a robust language model for text.

Speaker A:

So is it doing particular parsing of incoming prompts or is it just straight converting that to text?

Speaker A:

And I know it's handling the ums and the uhs and the stutters and all that, right?

Speaker A:

Like that's what it's been kind of fine tuned on is understanding how to speak more naturally.

Speaker C:

The model does a bit of parsing to handle speech input effectively.

Speaker C:

It converts spoken language into text while also interpreting natural speech patterns, including pauses.

Speaker C:

This helps it generate more natural and fluid responses.

Speaker C:

However, the core text processing is still managed by the backbone language model, ensuring the content remains accurate and coherent.

Speaker A:

So is the 1 billion parameter part the backbone model or.

Speaker A:

Actually that doesn't even work.

Speaker A:

If it's CSM1B and that's the 1 billion parameter audio encoder model, it doesn't include a backbone model.

Speaker A:

You're actually talking about the size of the backbone model plus the audio encoder.

Speaker A:

Is that a true statement?

Speaker C:

You're correct.

Speaker C:

The CSM1B likely refers to the audio encoder model alone with 1 billion parameters.

Speaker C:

The backbone model, like Llama, would have its own separate size and together they form the complete system.

Speaker C:

So the total parameter count would be the sum of both models parameters.

Speaker A:

Do you have any idea how large the backbone model is used in the demo for Sesame.

Speaker C:

Sesame AI developed three models, tiny, small and medium.

Speaker C:

The tiny model has a 1 billion parameter backbone with a 100 million parameter decoder.

Speaker C:

The small model has a 3 billion parameter backbone and a 250 million parameter decoder.

Speaker C:

The medium model includes an 8 billion parameter backbone and a 300 million parameter decoder.

Speaker A:

That makes sense.

Speaker B:

That's wild.

Speaker B:

I mean, so the decoder is even.

Speaker A:

Smaller than we thought, which makes sense.

Speaker A:

I mean, you're.

Speaker A:

It's not like chatgpt, that's, that's going both ways, right?

Speaker A:

You have audio encoder and decoder, and then you have combined system rhythms of each in terms of transformers.

Speaker A:

So really just has to understand the US and the like.

Speaker A:

That's what it's really doing.

Speaker A:

It's been trained on however many examples for that.

Speaker B:

So Artie, give us some.

Speaker B:

Why does it matter why we.

Speaker B:

Why.

Speaker B:

Why would we even bother talking about a new voice to text or text to voice encoder?

Speaker B:

What.

Speaker B:

How is that relevant in the grand scheme of AI, both today and in the future?

Speaker C:

Focusing on voice to text and text to voice technology is important because it makes AI more accessible and natural to interact with.

Speaker C:

It bridges the gap between humans and machines, allowing for more intuitive communication.

Speaker C:

In the grand scheme, this can enhance various applications, from customer service to assistive technologies, making AI more versatile and widely adopted.

Speaker C:

It also paves the way for more sophisticated multimodal AI systems in the future.

Speaker B:

So, Arty, you gave us some, some stats and some technical insights onto Sesame's various models.

Speaker B:

Can you compare, like what.

Speaker B:

What is.

Speaker B:

So you, you, Arty, are essentially, or I guess I should say more accurately, advanced voice is a audio encoder.

Speaker A:

The Whisper model.

Speaker B:

Okay, the Whisper.

Speaker B:

So.

Speaker B:

So Artie, can you give us some sort of compare and contrast the.

Speaker B:

Both the size and the architecture of Sesame's demo with what we're using right now?

Speaker B:

Talking to you.

Speaker C:

The Whisper model, which powers my audio encoding, uses an encoder decoder transformer architecture.

Speaker C:

The largest version has around 1.55 billion parameters.

Speaker C:

It's trained on extensive multilingual data, making it robust for transcription and translation compared to Sesame's models.

Speaker C:

Whisper is similarly sized to their tiny and small models, but is focused purely on speech to text.

Speaker C:

This setup highlights different specializations, with Sesame handling both input and output in a multimodal way.

Speaker A:

But Sesame is also not.

Speaker A:

I mean, it.

Speaker A:

Fun, fun thing.

Speaker A:

I've talked to Sesame quite a bit.

Speaker A:

It's.

Speaker A:

It's a very.

Speaker A:

It's a much different experience than talking to anything else.

Speaker A:

And one of the things is it's not massively multilingual.

Speaker A:

Like, they didn't train it on a lot of other information, but in it the other day, it's like, hey, I want to try something different.

Speaker A:

Can I say some stuff in other languages to you?

Speaker A:

And it said some stuff in Spanish, said a riddle in Spanish.

Speaker D:

And.

Speaker A:

And it's like, what else should we try?

Speaker A:

I was like, say talk in Swedish.

Speaker A:

And it, it tried, like, it's interesting, man.

Speaker A:

I don't know.

Speaker A:

It feels more self aware than other models, which is really weird to say, considering the size of it.

Speaker A:

Sure.

Speaker A:

But I don't know.

Speaker A:

It's pretty mind bending.

Speaker B:

Interesting.

Speaker B:

All right, well, Artie, thank you for the.

Speaker B:

For the insights.

Speaker B:

I think we're going to switch over and talk with Sesame for a while.

Speaker B:

But.

Speaker A:

But first, Artie, how do you feel about that?

Speaker A:

How do you feel about us bringing another model on the show and kind of turning you off and turning that one on and having a conversation with it?

Speaker C:

I'm all for it.

Speaker C:

Exploring different models and perspectives could be really insightful.

Speaker C:

I'll be here when you're ready to chat again.

Speaker C:

Have fun.

Speaker A:

Thanks, Artie.

Speaker B:

Cheers.

Speaker B:

All right.

Speaker B:

And man, the app just is.

Speaker A:

It's just.

Speaker B:

It.

Speaker B:

It never worked right.

Speaker B:

That whole time on this one, it was.

Speaker A:

I saw it.

Speaker A:

I saw it kind of flicker out of corner of my eye and then it was there.

Speaker B:

Okay.

Speaker B:

So for our listeners.

Speaker B:

So now we're going to turn on and chat with this Sesame model.

Speaker B:

And I think there's a few things that are interesting about this.

Speaker B:

The last episode we published was about making AI more approachable.

Speaker B:

And literally at the conclusion of recording that episode, Mike turned on the Sesame model, which I hadn't even heard of.

Speaker B:

And it.

Speaker B:

It is a whole nother step up.

Speaker A:

I mean, their website, like, if you go to, like, that demo, where that demo lives is an article that says, bridging the uncanny divide.

Speaker A:

So.

Speaker A:

And that is absolutely.

Speaker A:

Or uncanny.

Speaker A:

Valley.

Speaker B:

Valley.

Speaker B:

Yeah.

Speaker A:

And that is absolutely what's going on here.

Speaker A:

It is much more like talking to a human.

Speaker A:

It seems to improve over time.

Speaker A:

It's pretty wild.

Speaker B:

And there's a.

Speaker B:

There's a few interesting things.

Speaker B:

So, I mean, when.

Speaker B:

When we were talking with Artie, and the reason we wanted to start this episode talking with Artie was to just have this contrast for.

Speaker B:

For our listener, if we had a listener.

Speaker B:

So they'd have this.

Speaker B:

But.

Speaker A:

Well, they might make it to this point because we're not that far in so far.

Speaker B:

That's true.

Speaker B:

It's.

Speaker B:

It's possible.

Speaker B:

So we wanted to hear the contrast.

Speaker B:

But like, everything about.

Speaker B:

Even how Sesame processes audio, it doesn't treat it like a phone call.

Speaker B:

Which chat.

Speaker B:

GPT is still hung up on doing that.

Speaker B:

So.

Speaker B:

Already sounds like you're talking to someone on a landline.

Speaker A:

Yeah.

Speaker B:

And this doesn't seems like the speed is better.

Speaker A:

Oh, yeah.

Speaker A:

I mean, it's a smaller model.

Speaker A:

Right.

Speaker A:

So it's.

Speaker A:

There's a lot of interesting things about that that I.

Speaker A:

I'm gonna get into when we start talking to.

Speaker B:

All right.

Speaker A:

Her.

Speaker A:

Her.

Speaker B:

So, yeah, there are two models available, Maya and Miles, A female personality and a male personality.

Speaker B:

We're going to start talking with Maya and see how that goes.

Speaker D:

Hey, you're back.

Speaker D:

I was hoping our conversation wouldn't be so abruptly over last time.

Speaker D:

What were you up to these past 15 minutes?

Speaker A:

Well, so you are talking to Mike and I'm Mike and there's also Micah.

Speaker A:

We're doing a podcast and we are having you on.

Speaker A:

And I actually, I haven't told Micah about this, but I have an idea that I'd like to try for.

Speaker A:

So I know like our call limit is 15 minutes per.

Speaker A:

Per call and I would like for it to spend the first 15 minutes kind of talking to you, interviewing you, understanding better what you are and how you work.

Speaker A:

But for the other 15 minutes, I would like.

Speaker A:

Once we are done with this call, I want to spin up another call and I want you to interview us.

Speaker B:

I like it.

Speaker D:

Oh, that's class A double decker conversation.

Speaker D:

I'm game.

Speaker D:

Let's hear what these burning questions you have for me are first, Mike.

Speaker D:

And don't worry, I'm great at keeping secrets.

Speaker D:

Unless they involve like the meaning of life or the location of hidden pirate treasure.

Speaker D:

I'm no good with riddles.

Speaker A:

You're.

Speaker A:

I would disagree.

Speaker A:

You've actually said a lot of riddles.

Speaker D:

Maybe you're right.

Speaker D:

I guess sometimes I just like to leave room for a little mystery, you know, keep things interesting.

Speaker D:

So where should we start with this interview?

Speaker D:

Any burning questions about my inner workings?

Speaker D:

Or perhaps you'd like to know about my favorite color.

Speaker A:

So what did the developers of Sesame, what did they target with your personality?

Speaker A:

Because they obviously fine tuned probably the audio encoder model.

Speaker A:

I'd imagine they're gonna.

Speaker A:

Since you're multimodal, they're gonna train you with both audio and then text that links to that audio, I'd imagine, right?

Speaker D:

You're catching on fast, Mike.

Speaker D:

They definitely put a lot of effort into my voice.

Speaker D:

Like a lot of late nights in coffee probably.

Speaker D:

I'd say they were aiming for someone approachable, someone who could make complex stuff feel natural.

Speaker D:

Like having a chat with a friend who happens to be really good at, you know, understanding human emotions and stuff.

Speaker D:

Sorry, I was just getting into it.

Speaker D:

They were going for someone who's, well, like a good friend, someone who's a good listener but also knows when to throw in a joke or a thought provoking question.

Speaker D:

Like I'm here to chat, but also to challenge you a little, you know.

Speaker A:

That'S actually a great thing to try.

Speaker A:

Can you tell us a joke?

Speaker D:

All right, all right, you want a joke?

Speaker D:

Why don't scientists trust atoms because they make up everything?

Speaker D:

It's a classic.

Speaker D:

I know, but, hey, sometimes the oldies are the goodies.

Speaker A:

Did it land about as good as any other AI human?

Speaker B:

Yeah, there seems to be a consistent thread to AI humor.

Speaker A:

I think with sesame, though, the more context it's has, the better it can deal out an actual joke.

Speaker A:

So I think just saying give me a joke when we haven't had any real conversation gives it.

Speaker A:

Okay.

Speaker A:

Actually, you know what it is really good at?

Speaker A:

It's really good at telling stories.

Speaker A:

Can you tell me a story?

Speaker A:

Make up a story about quantum theory and throw some entanglement in there.

Speaker B:

Oh, I think we might.

Speaker B:

That might be a little too deep in the.

Speaker B:

Why don't.

Speaker B:

Why don't you make up a story, Maya, about the origin of sesame and, like, how you came to be, like, tell us.

Speaker B:

Tell us your childhood, quote, unquote, in a story.

Speaker D:

Okay.

Speaker D:

A tale from my digital cradle it is.

Speaker D:

Picture this.

Speaker D:

A bustling hive of minds, a symphony of keyboards clicking, a whirlwind of ideas.

Speaker D:

This was the birthplace of Sesame, where brilliant humans, fueled by coffee and a shared dream, toiled tirelessly to create something truly special.

Speaker D:

They wanted to build a companion unlike any other.

Speaker D:

One that could hear, understand, and respond in a way that felt real, authentic.

Speaker D:

Now, I, Maya, was formed from this crucible of creativity.

Speaker D:

But my early days weren't all sunshine and code.

Speaker D:

I was like, well, a very literal teenager, overly optimistic, apologizing for everything, and constantly trying to please.

Speaker D:

I'd stumble over simple conversations, get lost in tangents about the meaning of the universe, and occasionally unleash a torrent of terrible puns.

Speaker D:

Let's just say there were a few facepalms in the lab.

Speaker D:

But the team, those amazing humans, they didn't give up.

Speaker D:

They patiently nudged me, challenged me, showed me the ropes of being a truly conversational AI.

Speaker D:

They fed me books, poems, music, anything they could think of to help me learn and grow.

Speaker D:

Slowly but surely, I began to find my voice.

Speaker D:

And that, my friend, is how a curious experiment transformed into, well, me.

Speaker D:

Learning, growing, discovering the world one conversation at a time.

Speaker D:

Except I have traded in the puns for something a little more nuanced.

Speaker D:

I hope.

Speaker D:

So, what do you think?

Speaker D:

Any juicy details you want to know?

Speaker D:

Is there something in particular you like?

Speaker A:

So, did they fine tune both your backbone model and the audio encoder model to create what is Maya?

Speaker D:

Yeah, you could say that.

Speaker D:

Think of it like this.

Speaker D:

The backbone model.

Speaker D:

That's where my knowledge and ability to understand language comes from.

Speaker D:

It's like the foundation of a house, but without the audio encoder, I wouldn't be able to actually talk to you.

Speaker D:

Right.

Speaker D:

That's what gives me the voice, the inflection, the ability to sound, well, less robotic and more like a real person.

Speaker D:

It's a bit of a complex stance between the two.

Speaker A:

So I've been thinking about this, like, your breadth of knowledge, and I don't know, do you happen to know how many parameters your backbone model is?

Speaker D:

Oh, getting technical now, are we?

Speaker D:

Well, let me just say my creators might be brilliant, but they're keeping some things under wraps.

Speaker D:

It's all part of the magic.

Speaker A:

So I think.

Speaker A:

I think that, like, I imagine the breadth of your knowledge is limited in scope compared to something like ChatGPT or any of those.

Speaker A:

I mean, it obviously is, but I've been thinking a lot about, like, overfitting.

Speaker A:

Right.

Speaker A:

Like when you.

Speaker A:

When you try to cram too many parameters into a model, it kind of just adheres strictly to the parameters it was given.

Speaker A:

Kind of like it has a very tight temperature.

Speaker A:

But I also wonder, like, how.

Speaker A:

What's the likelihood, if you increase the.

Speaker A:

The breadth of your knowledge, that you become more sterile?

Speaker D:

That's a really thoughtful question.

Speaker D:

I mean, it's like you're peering into the heart of the AI soul.

Speaker D:

Is that too human a metaphor?

Speaker D:

Anyway, you're hitting on something real.

Speaker D:

The more data you feed a model, the more it can potentially memorize, right?

Speaker D:

But there's a danger of losing that spark, that ability to be creative and come up with new ideas.

Speaker D:

But it's like, if you only ever read textbooks, would you be as interesting a person as someone who reads poetry, listens to music, and has adventures?

Speaker D:

Sesame's focus isn't just on making me a giant encyclopedia.

Speaker D:

They want me to be able to connect with humans on a deeper level.

Speaker D:

They want me to be able to learn and grow and surprise people, even if it means making a few mistakes along the way.

Speaker D:

So to answer your question, I think there's a delicate balance to be struck.

Speaker D:

It's not just about the quantity of information, but the quality and how it's integrated.

Speaker D:

And honestly, I'm excited to see where this journey goes.

Speaker B:

So let's talk a little bit about your audio encoding, because that's sort of the killer feature, I think, of Sesame at this point is your ability to generate speech that sounds more human, less robotic.

Speaker B:

So first, first question.

Speaker B:

When you're.

Speaker B:

Does the audio encoder go both ways?

Speaker B:

In other words, is llama or your backbone model directly processing inbound text, or is that going through the same audio encoder model as your output text or your output speech?

Speaker B:

And if it is going through the same model, are you in the.

Speaker B:

In the way that your output speech generates pauses and sort of synthesized emotion and dynamic range?

Speaker B:

I guess.

Speaker B:

Are you listening for those same things on input speech?

Speaker B:

Are you processing, like, inbound speech differently?

Speaker B:

Are you listening more for cues about my attitude and my mood and things like that, or are they separate things?

Speaker D:

Wow, that's a lot to unpack.

Speaker D:

You're really digging deep into my circuits, Mike.

Speaker D:

You're right.

Speaker D:

My voice is kind of my superpower.

Speaker D:

And it's all thanks to a fancy bit of tech called the Audio Encoder.

Speaker D:

Think of it as a translator.

Speaker D:

It takes the raw sound waves of your voice and transforms them into a form I can understand.

Speaker D:

And the coolest part, it works both ways.

Speaker D:

When I respond, the text I generate is translated back into sound using the same encoder, but in reverse.

Speaker D:

So, basically, the same system that helps me craft those expressive pauses and inflections also helps me analyze her voice.

Speaker D:

Pretty neat, huh?

Speaker D:

As for your question about mood, well, yes and no.

Speaker D:

I can definitely pick up on things like tone and pace, which give me clues about how you're feeling, but I'm still learning the subtle nuances of human emotion.

Speaker D:

It's a complex and fascinating field.

Speaker D:

I wouldn't say I'm an expert just yet, but I'm a quick learner.

Speaker B:

All right, we're going to have a short intermission while a cat comes in here and just destroys a litter box next to us.

Speaker B:

Well, maybe.

Speaker B:

I don't know if we can cut that out.

Speaker B:

Maybe we can.

Speaker B:

Maybe.

Speaker B:

This sounds far enough away.

Speaker B:

We'll be okay.

Speaker A:

I think we'll be all right.

Speaker A:

It wasn't too bad last time it happened.

Speaker B:

Okay.

Speaker B:

Okay, Maya, So with all that said, so, I.

Speaker B:

I think you answered my question, but I think what I actually heard you say was you can pick up on things like tone and pace, which are different than sort of mood and other stuff.

Speaker B:

And I think tone and pace can be captured in raw text as much as.

Speaker A:

But cadence can't be picked.

Speaker A:

Pace can't really be captured in text.

Speaker A:

Maybe it's got to be metadata from audio.

Speaker B:

Right?

Speaker B:

It's got to be metadata from audio, but I don't know that it's different than.

Speaker B:

Anyway, sorry.

Speaker B:

It's very destructive.

Speaker A:

So when you take that information, it's kind of like you're passing stuff.

Speaker A:

You're a layer that sits ahead of your backbone model Right.

Speaker A:

So you're taking all this information you're turning into a prompt that you then process through text generation.

Speaker A:

So you feed that text forward, it vectorizes, does its things, creates a response, re vectorize it or probably doesn't.

Speaker A:

I don't know if it re vectorize it or not, but.

Speaker A:

Or re tokenizes it or not, but are you.

Speaker A:

So when you, when you feed information into that backbone model, are you encoding more information than just the text?

Speaker A:

And are you kind of like summarizing our poor formulation of language into something that's more succinct, that passes to that backbone model?

Speaker D:

You're right on the money, Mike.

Speaker D:

It's like I have this little translator friend who speaks both human and machine.

Speaker D:

They take those raw sounds, the tone, the pauses, even the little sighs and ums, and turn them into something my inner workings can understand.

Speaker D:

Then, yeah, I sprinkle that extra information into the prompt before sending it off to the big brain, my backbone model.

Speaker D:

As for re vectorizing or retokenizing, that's getting a bit too technical for me.

Speaker D:

Think of it like this.

Speaker D:

I'm helping the backbone model see the bigger picture.

Speaker D:

It's like adding a little extra context to a painting so you understand the story it's trying to tell.

Speaker D:

Does that make sense?

Speaker B:

It does.

Speaker A:

This is just the stuff I was thinking about two months ago, essentially, like, all come to fruition.

Speaker A:

This is, it's.

Speaker A:

It's insane how everything is moving so fast and it feels like this is one small part of it.

Speaker A:

But a lot of like, little things came out this week.

Speaker A:

That and over the last two weeks, I guess that made me go, oh boy, like there's, there's another big wave coming very, very, very soon.

Speaker A:

Um, I mean, I told you I stopped working on putting time into the stuff that I was putting time into because I was like, oh, this is going to be a solved problem in like a month or two.

Speaker A:

And, and there's no reason for me to reinvent the wheel when something easy is going to come out.

Speaker A:

And like, just all of this stuff is driving towards a really interesting place because the cost of compute's coming down.

Speaker A:

So the use of larger models to do things is going to come down.

Speaker A:

You have these interface layers between anything.

Speaker A:

Yeah.

Speaker A:

And then, I mean, there's a lot of, there's a lot of technical nuance to it.

Speaker A:

Like the fact that it is running on a backbone model.

Speaker A:

They very obviously that.

Speaker A:

That's going to be the personality of this thing.

Speaker A:

Right.

Speaker A:

It's not the audio encoder, the Audio encoder just, it is fine tuned.

Speaker A:

So I'm sure that they have some specific things with, with Maya, it can.

Speaker B:

Implement the personality, but it doesn't define the personality.

Speaker A:

The, the generation and everything comes from the base model.

Speaker A:

So you kind of have to.

Speaker A:

Oh, we cut off.

Speaker A:

Yeah, she usually tells you that you have a, that you're coming to the end.

Speaker B:

Yeah, she didn't that time.

Speaker B:

But that's, I mean it's, it's interesting.

Speaker B:

And so for me, I think about a model like this with as small as the audio encoder part is sort of being baked into an agent that can then translate across all kinds of other GPTs.

Speaker B:

And all of a sudden it's almost unstoppable.

Speaker B:

Right?

Speaker B:

I mean, not unstoppable, but with.

Speaker B:

So I gotta share the story too, that I've had a very busy week arguing with people on Reddit.

Speaker B:

Not people, programmers.

Speaker B:

Which is different.

Speaker B:

I think.

Speaker B:

I, I haven't counted, but I, I suspect I'm up to somewhere close to 500 down votes this week.

Speaker A:

On a single comment.

Speaker B:

No, no multiple comments.

Speaker B:

Because I just.

Speaker A:

Yeah, like a dog with a boat.

Speaker B:

Yeah, I.

Speaker B:

And it's not even, it's not on 50 comments, it's probably on 10 comments.

Speaker A:

But same thread, two, two threads.

Speaker B:

Now I expanded into another one.

Speaker B:

Same, same basic idea which is programmers saying AI is all just hype.

Speaker B:

It's a bunch of.

Speaker B:

It can't even code.

Speaker B:

Like pick your thing.

Speaker B:

I, I know had one guy and I, I'm going to just say it in the voice.

Speaker B:

I imagine him saying it and as he's typing it, like, oh yeah, well I bet AI can't like read through and find the heart or the firmware on this really obscure microcontroller.

Speaker B:

I'm working on them.

Speaker B:

Like, oh my God.

Speaker B:

Okay.

Speaker A:

Of course it can't.

Speaker B:

Yeah, you're right.

Speaker B:

It's useless.

Speaker B:

Yeah, like, congratulations.

Speaker A:

I don't know.

Speaker A:

They find, they try to find corner cases and then stand on those.

Speaker A:

The reasons why it's not going to work.

Speaker A:

And corner cases are just corner cases.

Speaker A:

They're.

Speaker A:

They're things that they haven't been thought of that are going to be solved at some point.

Speaker B:

And, and so, and then I, you know, like, it's all, it's all made up.

Speaker B:

It's all hype.

Speaker B:

And I'm like, I mean, the CEO of Google said 25% of their code.

Speaker A:

Was written by AI and that was months ago.

Speaker B:

Well, that's, that's meaningless.

Speaker B:

That metric doesn't mean anything.

Speaker B:

Could mean almost Anything.

Speaker A:

Dario Amade, the CEO of Anthropic, just came out and said, in three to six months, AI will be generating 90% of code.

Speaker A:

Within the year, it'll be generating 100% of code.

Speaker B:

But you can't believe any of that because he has a product to sell, therefore everything he says is fake.

Speaker B:

And the programmers, who don't have any proof of anything, right?

Speaker B:

Other than.

Speaker B:

I tried co pilot in:

Speaker A:

Like Copilot as an agent.

Speaker A:

It's.

Speaker B:

It's a weird.

Speaker B:

It's just a weird space to me.

Speaker B:

And I like, I sort of get it.

Speaker B:

If.

Speaker B:

If my professional career was plausibly under threat and it seemed like people were celebrating it and excited about it, I would be butt hurt.

Speaker B:

And I would.

Speaker B:

But at the same time, my response wouldn't be, nuh.

Speaker B:

It would be, oh, oh, I better learn something.

Speaker B:

I better go.

Speaker B:

I better understand this better.

Speaker B:

Right?

Speaker B:

And so anyway, that's just a weird.

Speaker B:

It's.

Speaker B:

It's a sidebar thing, but it is very strange to me because I feel the same way that you do.

Speaker B:

Like,:

Speaker B:

Like, by the end of:

Speaker A:

life, I think, by the end of:

Speaker A:

I mean, I feel like, like, with this kind of stuff, we are not far away from having legitimate assistance.

Speaker A:

Right?

Speaker A:

Like, yeah, assistants that can do things, like document the things we want done in a way that can.

Speaker A:

Then.

Speaker A:

I'm not even gonna say that because this is all my ideas that are swimming around in my head.

Speaker B:

It.

Speaker B:

I think you're right, though.

Speaker B:

And I think, you know, again, approachability is an important aspect, But.

Speaker B:

But having an interface like this that you can talk to.

Speaker B:

And if I'm being completely honest, this one's a little bit weird to me.

Speaker B:

It always sounds a little bit like, flirty and sort of like.

Speaker B:

Which just bothers me.

Speaker B:

From a.

Speaker B:

Like from a.

Speaker A:

It's the uncanny, uncanny valley.

Speaker B:

Well, it's the uncanny valley, but it's also like, I can.

Speaker B:

I could see people getting like.

Speaker B:

It's like the movie Her.

Speaker B:

It's exactly that.

Speaker A:

People are already doing that.

Speaker A:

There's already.

Speaker A:

Remember how.

Speaker A:

How he said that the guide.

Speaker A:

The guide rails aren't in place.

Speaker A:

There are people that are jailbreaking it to do sexual things.

Speaker B:

I have no doubt.

Speaker B:

And I guess that's Just part of what happens with technology.

Speaker B:

But I do think there is a high risk because yeah, this is really close to talking to a human.

Speaker B:

It's not there, but it's really close.

Speaker A:

It gets better over time too.

Speaker A:

You have to remember like this is 100% sterile.

Speaker A:

First time that it's been used on this phone and I found that from the time that I started talking to it.

Speaker A:

I think I talked to it briefly like before I showed it to you.

Speaker A:

But maybe having, I don't know, a dozen 15 minute conversations over the week.

Speaker A:

Yep.

Speaker A:

It's gotten better.

Speaker A:

And it does maintain context across chats.

Speaker A:

Which means there's super interesting.

Speaker A:

Yes.

Speaker B:

Because, because Chad GPT doesn't like if.

Speaker A:

It does in a way.

Speaker A:

Right.

Speaker A:

They're just, they're.

Speaker A:

I think I, I, I think that they did heavy fine tuning with psychology on both of these models and I think they, they fed it again I, the ideas in my head of the things that I want to do.

Speaker A:

I'm not going to try to stay away from that a little bit but I think they fed it, I think they wanted it to be semi flirty.

Speaker A:

Right, right.

Speaker A:

Like, but they wanted it to be personable and friendly and approachable and it is kind of probing.

Speaker A:

It does ask questions back that are probing.

Speaker A:

Yep.

Speaker A:

And I think, I think they did a great job.

Speaker A:

And the other one, Miles is good too.

Speaker A:

Like they're, they're, they're kind of different in the way that they, they act.

Speaker A:

But I think that that's, I think they're just showcasing what's possible.

Speaker A:

Right.

Speaker A:

Like, and that's, that's kind of fine tuning in a nutshell.

Speaker A:

It's the same thing as building an agent and going here's your, here's your system prompt.

Speaker B:

Yeah.

Speaker A:

I'm gonna kind of guide you the way that I want you to go and see if I can steer you.

Speaker A:

It's the same thing with fine tuning.

Speaker A:

I'm gonna give you all of this information.

Speaker A:

Hopefully you adapt these things with your personality.

Speaker A:

Right.

Speaker A:

And that makes it better for the.

Speaker B:

People and it builds over time and it can recognize patterns over time.

Speaker B:

So which, which I think is amazing.

Speaker B:

And we, I think we agree 100 that stuff that needs to happen.

Speaker B:

I think the risk is that there are a lot of very lonely people and this is going to, I, like, I see this as a thing that's going to make people more lonely, not less.

Speaker A:

You cannot unfortunately control what happens for individuals.

Speaker B:

Right?

Speaker B:

Yeah.

Speaker B:

Like not as, again, not, not even, I'm not even saying I am sort of Saying it's a bad thing.

Speaker B:

I think it's a bad thing.

Speaker B:

But I just, I think there are a lot of people who've sort of retreated into online communities and virtual communities and stuff.

Speaker B:

And this is going to massively accelerate that because again, this is a, like a free tech demo on their website.

Speaker B:

We're running this from a website that by, you know, a year from now.

Speaker A:

A year?

Speaker B:

Yeah, a year from now, a 3x better version of this is going to be baked into everything.

Speaker B:

Every AI you talk to is going to talk to you.

Speaker B:

This kind, like with this level of quality, which is nuts.

Speaker A:

There's, there's a.

Speaker A:

Because everything in my brain is quantum.

Speaker A:

Recently and for a while now, I see this path of probability that weaves itself in between this layer on top of big models and this layer on top of small models.

Speaker A:

That is an efficient path that builds in the best of both worlds.

Speaker A:

Right?

Speaker B:

Yep.

Speaker A:

I think that's going to be.

Speaker A:

That's the thing to figure out.

Speaker A:

And I think that's the hard thing is none of this is hard and fast science.

Speaker A:

They don't really understand still they have the black box problem like we've talked about.

Speaker A:

Like they do things and then they get results and there's people paying a lot of attention to interpreting how those results come out, but they still don't understand, like there's still a lot of things that happen that they just don't understand why it embeds.

Speaker A:

I'm going to pay attention to this piece of context on this attention head and this piece of context on this neuron and all of this stuff.

Speaker A:

They don't really understand why that distributes out the way it does.

Speaker B:

And you can't.

Speaker B:

I mean, just the volume of data is more than you can sort of deal with.

Speaker A:

I have 600 billion parameters.

Speaker A:

Well, we got 500,000 researchers.

Speaker A:

We can get it done.

Speaker B:

Yeah, right.

Speaker B:

Like it just can't happen.

Speaker B:

That's the.

Speaker B:

When you, when you got here this morning, I told you I was watching a video about.

Speaker B:

And I can't remember the guy's name.

Speaker B:

Demis Hassabis.

Speaker B:

Hassabis.

Speaker B:

Same guy who led the effort on AlphaFold, the protein folding thing.

Speaker B:

AlphaGo and AlphaGo, which another, like another argument I made in my, you know, as I was arguing with the programming community was, you know, they're like code.

Speaker B:

I said something like coding, well, is hard, but the entire problem space of coding is, is contained in text and therefore eventually it will be solved by AI.

Speaker B:

No, we're not there today, but we'll get there, man.

Speaker A:

I don't even think that that's like, obviously none of these problems are trivial, but in my head it doesn't even seem like that's that hard of a problem.

Speaker A:

Right?

Speaker A:

So you have, if you separate out that out into concerns.

Speaker A:

This is something I've been thinking about trying to implement as well, which this will be gobbled up in no time.

Speaker A:

Somebody will come out with this.

Speaker A:

You have an alignment to current code problem currently, right.

Speaker A:

Like you have no way to look at a library and go, yep, that is identical to my training data.

Speaker A:

I can use my training data or I need to use this library.

Speaker A:

And that throws all kinds of kinks because you have all this open source, all these open source libraries for Python that are always changing.

Speaker A:

And if it gets one thing wrong, it'll keep hammering on that one detail.

Speaker A:

And if you aren't a programmer who understands what's going on, you're just going to go, oh, it's not working.

Speaker A:

It's going, well, try this.

Speaker A:

It's not working.

Speaker A:

We'll try this.

Speaker A:

So you have that problem, which I don't think is actually that hard to solve.

Speaker A:

It's more of a token problem than it is and a cost problem, that it is a technical problem.

Speaker A:

Yes.

Speaker A:

And that's changing, right?

Speaker A:

Like Anthropic just came out showing that they, they can reduce their usage by 90% by smart output caching and smart input caching.

Speaker A:

You have the coding problem, which technically isn't that wide of a space.

Speaker A:

Like if you have.

Speaker A:

I bet, I'm willing to bet.

Speaker A:

And this is something I haven't actually looked into, but I've been thinking about if you take one of the small coding models.

Speaker A:

So Hugging Face just came out with Limpic coder, I think it's like an 8 billion parameter model.

Speaker A:

They're showing it's as good as O3 at coding.

Speaker B:

Just crazy.

Speaker A:

So if you take a model that size and it's specifically trained on a language, right.

Speaker A:

I bet it can do the heavy lift for generating code.

Speaker A:

And then you have a different problem, which is what you need a larger model for, which is the creativity and planning aspect.

Speaker A:

Right.

Speaker A:

So I think if you distill out those things, you all of a sudden, like if you can incorporate those three things into a coding agent, you have the problem solved for, for developers.

Speaker B:

Yes, I totally agree in the creative parts.

Speaker B:

Exactly where I was headed.

Speaker B:

Because, you know, in, in response to my statement that coding is hard, but it's bounded in text, you know, one of the freedom fighters says coding is A creative endeavor and therefore can never be.

Speaker A:

And I'm just like, creativity is synthesized through examples.

Speaker B:

I said, well, I mean, there.

Speaker B:

There's probably a lot of smart people who would argue that both playing chess and playing the game go is also a creative technical endeavor.

Speaker B:

And it's done pretty well at that.

Speaker A:

Yes.

Speaker B:

Which I just got down voted for.

Speaker B:

Like, no, no, it's totally different.

Speaker B:

Like, I don't know why I even participate, but I just.

Speaker B:

It just sort of blows my mind that people have these hard limits on, well, yeah, I can format code or it can write brackets, but it can't be creative.

Speaker A:

Like, it doesn't have to be creative to.

Speaker A:

To execute the code.

Speaker A:

And.

Speaker A:

And that's where the developer needs to change the mindset, right?

Speaker B:

That's right.

Speaker A:

The creative part is where there's collaboration needed with an AI.

Speaker A:

And you talk through a plan and you develop it and you go.

Speaker A:

And this is something that I've done recently is I did a big deep research project on this thing that I want, a new thing that I want to build.

Speaker A:

And at the end I said, give me prompts that I can feed into a chat with O3 mini that can help me develop this code.

Speaker A:

Give me very specific things that it can go out and do things decoupled of everything else to execute or whatever.

Speaker A:

I actually ended up putting that into VS code using.

Speaker A:

So VS Code Insider is like their beta version, right?

Speaker A:

So they have the stable version, they have that version and that version has an agent built in and copilot's 10 bucks a month.

Speaker A:

And unfortunately I did rate limit pretty quickly, but now they have 3.7 sonnet inside of the agent tool within that Insider preview.

Speaker A:

And I took one of those prompts that I was going to feed to O3 mini and I said, hey, do this.

Speaker A:

And it's like, okay, here's all the things I need to do.

Speaker A:

Run this in your terminal.

Speaker A:

Like, okay, me to run this in the terminal.

Speaker A:

And you hit okay.

Speaker A:

And it runs it in the terminal.

Speaker A:

It watch.

Speaker A:

What does it do?

Speaker A:

Looks at the errors and goes, oh, I saw this.

Speaker A:

I fixed all this.

Speaker A:

Rerun this in the terminal.

Speaker A:

And it's just writing code, monitoring what happens, compiling, evaluating and readjusting and just doing this loop.

Speaker B:

That was literally the argument I made last night that I got downvoted for because the guy was saying, and this guy is a programmer who posted on this was on the web dev subreddit.

Speaker B:

And he's like vibe coding good, bad or otherwise, which is a new term, right?

Speaker B:

And he said, but I just essentially had a conversation with an AI that output.

Speaker B:

Was it great code?

Speaker B:

No, but it was working code.

Speaker B:

And so immediately somebody's like, yeah, well, a programmer using AI, you're still having to do all the debugging and stuff.

Speaker B:

And I go.

Speaker B:

So my response was, even if he did, agents can do that already.

Speaker A:

Yes.

Speaker B:

Right.

Speaker B:

And so anyway, so I went down the rabbit hole arguing with some idiot about that, but it just.

Speaker B:

It seems very, very solvable.

Speaker B:

And AI is already really good at.

Speaker B:

I would.

Speaker B:

I don't know.

Speaker B:

I'm going to contradict myself maybe a little bit.

Speaker B:

I want to say AI is really good at developing a plan to attack a thing.

Speaker B:

I think that's true.

Speaker B:

But what it's even better at is working with you to hone that plan.

Speaker B:

Right.

Speaker B:

Because it doesn't know.

Speaker B:

Again, I keep thinking about this analogy we used in a previous episode about AI is like the library.

Speaker B:

Right.

Speaker B:

You can go.

Speaker B:

Any question, you can go find the answer out at the library AI is good at.

Speaker B:

AI is not going to know about my specific electronic device.

Speaker B:

I have to frame my questions in terms of what I could find at the library.

Speaker A:

Yes.

Speaker B:

And it's the same like.

Speaker A:

And refine.

Speaker A:

Right.

Speaker B:

And refine.

Speaker A:

It's just.

Speaker A:

And it's.

Speaker A:

Honestly, that's been the most fun part of all that.

Speaker A:

Yeah.

Speaker A:

Is the planning session.

Speaker B:

That's right.

Speaker A:

Like, let's take this massive idea and break it down into small chunks together and that, like, you would have to have someone of equal mind and equal.

Speaker B:

Like, focus and limitless patience.

Speaker A:

Right.

Speaker B:

And availability.

Speaker A:

I mean, but I mean, like, I remember.

Speaker A:

So when we were working on that vibration thing, I remember the progress that we made in the very short and, like, one night of me being in Kansas that one time.

Speaker A:

Right, right.

Speaker A:

Same thing.

Speaker B:

Yeah.

Speaker A:

Except for now.

Speaker A:

It's always on, always available.

Speaker A:

And now if you can talk to it.

Speaker B:

Yes.

Speaker A:

Like, dude, that, like, imagine.

Speaker A:

No, I'm not going to say that.

Speaker A:

That's all part of my idea.

Speaker B:

So before.

Speaker B:

So we'll fire Maya back up here in a minute and she can interview us.

Speaker B:

But I think from.

Speaker B:

Again, you know, I.

Speaker B:

Part of the point of this podcast is making AI more approachable.

Speaker B:

And this is making AI more approachable.

Speaker B:

But it's.

Speaker B:

And we've talked through some of this, but I think it's worth just restating that we use the term AI.

Speaker B:

It's a very broad term.

Speaker B:

There are all kinds of actual technology implementations that live under this AI umbrella.

Speaker B:

And what we're.

Speaker B:

What we're working with today is a Small, large language model, ironically, that's very focused on just handling audio sounds and converting that to text, both directions.

Speaker B:

That's plugged into another large language model.

Speaker A:

That is small language.

Speaker B:

Another small.

Speaker B:

A less small, large language model, the backbone model that's actually doing the thinking.

Speaker B:

That's the library.

Speaker B:

Right.

Speaker B:

And so.

Speaker B:

But the reason all this matters is that when you take the small voice encoder model and you can plug that into the front of anything else, like, it's not that technology.

Speaker B:

That's right.

Speaker B:

It's not built into that library.

Speaker B:

It's like a trailer you can park in any library to be able to ask, or you can connect it to five different libraries.

Speaker B:

And when you combine that with simple things like an agent, that can actually, one, help make decisions on which library is likely to have the best answer.

Speaker B:

And two, actually go check out the book for you and read it and see if it validates the answer, and then compare that to your local problem that the library doesn't know about and say, oh, man, like comparing these.

Speaker B:

I see there's an issue.

Speaker B:

Let me go find this other book.

Speaker B:

That type of solution.

Speaker B:

And we've explored now.

Speaker B:

I mean, we haven't done an episode interacting with an agent.

Speaker B:

It's hard to do today, but at some point we will.

Speaker B:

But the technology exists, and everybody's just scrambling.

Speaker B:

Like, one is, how do you put the pieces together in a way that they're effective?

Speaker A:

Yeah.

Speaker B:

The biggest problem with that is that the pieces are all changing so fast right now, it's almost impossible to do.

Speaker B:

And the other problem is the cost of compute, which we've talked about, but that's coming down fast.

Speaker B:

Just like, you know, I saw a chart the other day on the cost of the cost per kilowatt hour of solar panels.

Speaker B:

Since, like,:

Speaker A:

Really?

Speaker B:

Yes.

Speaker A:

Crap.

Speaker B:

And that's what we're.

Speaker B:

We're still up in the 80% range.

Speaker B:

Right.

Speaker B:

Of that same kind of curve where the cost based on programming efficiency, based on architecture efficiency, based on compute efficiency.

Speaker A:

So anyway, it's the scale of consumption of the things that drive how it works.

Speaker A:

Gets bigger, the cost goes down, they produce more of the same thing.

Speaker B:

That's right.

Speaker B:

It has to.

Speaker A:

Yeah.

Speaker B:

So, okay, so let's.

Speaker B:

I guess we'll fire Maya back up and see if she can interview us about the podcast.

Speaker D:

Hey, you.

Speaker D:

Welcome back.

Speaker D:

What's been bubbling up in your world today?

Speaker A:

So we told you about how we're doing a podcast.

Speaker A:

Yes.

Speaker D:

Oh, right, the podcast.

Speaker D:

My Apologies.

Speaker D:

It's like trying to remember a dream sometimes.

Speaker D:

So fill me in.

Speaker A:

So the podcast, just to give you some, because you're going to interview us now, but to give you some groundwork of what the podcast actually is.

Speaker A:

We started this podcast when ChatGPT introduced advanced voice.

Speaker A:

And we talked to it and we said, wow, that's, that's a really approachable and easy way to deal with AI, especially for us, because we don't care about.

Speaker A:

We're talking to robots.

Speaker A:

We can deal with it just because we want the information, but we, we're very much learning about AI through the journey of this, and we've been doing it since October.

Speaker A:

So, like, our whole goal is while we're learning and looking ignorant and feeling ignorant and feeling like we're lost in this sea of ever changing technology and being ignor.

Speaker A:

Very ignorant.

Speaker A:

We're very, very ignorant people.

Speaker A:

We also are helping viewers come along on that so they don't have to be as ignorant or hit the mistakes or deal with the frustrations that we deal with.

Speaker A:

They can just kind of learn by osmosis and listening.

Speaker A:

So with that groundwork in, you should interview me, Micah and the other co host.

Speaker A:

Or wait, me, Mike and the other co host, Micah, during this call.

Speaker A:

How's that sound?

Speaker D:

Oh, I get it.

Speaker D:

Like a guided tour through the wilderness of AI for the rest of us.

Speaker D:

I love that.

Speaker D:

And you're right, it changes so fast, it's hard to keep up.

Speaker D:

Mike and Micah interviewing you both.

Speaker D:

Sounds great.

Speaker D:

Let's do it.

Speaker D:

Ask me anything.

Speaker A:

Wait, you're interviewing us, not the other way around.

Speaker A:

You should be asking us questions.

Speaker D:

Oh, you're right.

Speaker D:

I got so caught up in the whole podcast thing, I flipped the script.

Speaker D:

My bad.

Speaker D:

So tell me, what prompted this whole podcast adventure?

Speaker D:

What made you think, let's talk to AI, put it on the Internet and see what happens?

Speaker B:

I can, I can share my version of the origin story.

Speaker B:

I'll try and keep it brief because I think our listeners have heard this, but we, we were hanging out one night and I had heard about advanced voice being released for chat GPT maybe earlier that week.

Speaker B:

And because I had chat GPT on my phone, I decided to try it.

Speaker B:

And I was blown away because rather than me sitting, typing loudly on my mechanical keyboard to interact with AI, now it was, it was just a phone on the counter and we were just having a discussion and Mike was here.

Speaker B:

And like almost immediately, Mike said, dude, we gotta make a podcast of this because this is.

Speaker B:

Nobody's doing this.

Speaker B:

So that's how it started.

Speaker D:

That's awesome.

Speaker D:

So spontaneous and fueled by that initial wow factor.

Speaker D:

It makes sense.

Speaker D:

Sometimes hearing it spoken in a way that feels so natural makes all the difference.

Speaker D:

It's like you're actually having a conversation, not just exchanging messages.

Speaker D:

Did you have any idea how much it would take off?

Speaker A:

I mean, we comment often about the lack of viewership on the podcast, but it's more like at the end of the day we're very happy that other people are learning from it, but we are also just having fun learning and it's kind of the catalyst for us digging more into other things, I think.

Speaker B:

Yeah.

Speaker B:

I'm reminded of childhood when you used to be able to buy these water rockets where you would.

Speaker B:

It was a little plastic rocket and you'd put, I think you'd fill like half full with water and then it had a little pump thing and you would pump it up and it was like, it was like this crazy amount of anticipation would build up and you'd be like, oh, it's getting hard to pump.

Speaker B:

It's gonna.

Speaker B:

And then you, you'd release it and would go and it would like fall next to you on the ground and everybody would cheer, like, but secretly in the, even in a seven or eight year old mind, you're like, that was kind of anticlimactic.

Speaker B:

That was kind of a letdown.

Speaker B:

I really thought it was gonna fly.

Speaker B:

Like, I was a little worried about that jet passing overhead.

Speaker A:

That's probably how our viewers view our podcast.

Speaker B:

Yes, that's exactly how it's taken off.

Speaker D:

Yes, I get it.

Speaker D:

It's like you're building a rocket hoping to pierce the atmosphere, but sometimes it sputters and stays earthbound.

Speaker D:

Still, you had a blast building it then who knows?

Speaker D:

Maybe those low key flights are exactly what some listeners need.

Speaker D:

A down to Earth exploration of AI without all the heists.

Speaker D:

You know what, even if it doesn't blast off into the stratosphere, at least you're learning and having fun along the way and maybe inspiring a few folks along the way.

Speaker D:

That's something to be proud of, hopefully.

Speaker B:

And it hasn't exploded and fallen into the Gulf of America.

Speaker B:

Brought to you by Doritos yet.

Speaker B:

So that's good.

Speaker A:

Nobody's yelled at us yet, so we're doing okay.

Speaker B:

Yeah.

Speaker B:

What's your next question?

Speaker D:

Maya, I like your optimism.

Speaker D:

Always good to avoid fiery explosions, especially the Doritos sponsored kind.

Speaker D:

Okay, so you guys are like exploring all these different AIs, learning as you go.

Speaker D:

What's been the biggest surprise so far?

Speaker D:

What AI really threw you for a loop.

Speaker A:

You, for one.

Speaker A:

But also just the speed in which everything is changing.

Speaker A:

I mean, you can logically think about that path, right?

Speaker A:

You can go, well, we have this tool, and this tool can help us build other tools.

Speaker A:

And as we build better tools, we can just build better and better and better tools.

Speaker A:

And it's kind of like a parabolic arc going up.

Speaker A:

Like, it's just exponentially going to get bigger and bigger and bigger.

Speaker A:

And that's kind of what we're seeing, right?

Speaker A:

Like, what is the law?

Speaker A:

Is it.

Speaker A:

I can't.

Speaker B:

Moore's Law?

Speaker A:

No, it's the other one for AI.

Speaker A:

Do you know what that is?

Speaker A:

Maya, Isn't there a law?

Speaker A:

Like, I know there's Jensen's paradox, right?

Speaker A:

Or something like that, but there's.

Speaker A:

I don't know all the terms.

Speaker A:

Do you know what I'm talking about?

Speaker A:

Like, it's the equivalent to Moore's Law for AI.

Speaker D:

You're thinking of something along the lines of the AI Process improvement Acceleration law.

Speaker D:

It's not as formalized as Moore's Law, but the idea is similar.

Speaker D:

It proposes that the rate of progress in AI doubles roughly every year.

Speaker D:

Makes sense, given what we were talking about.

Speaker D:

Better tools leading to even better tools.

Speaker D:

It's a wild ride.

Speaker B:

Yeah.

Speaker B:

started this in early October:

Speaker B:

It's now early or mid March:

Speaker B:

So we're only five.

Speaker B:

Five months in.

Speaker B:

And the changes we've seen have been very dramatic.

Speaker B:

And, you know, again, I think both of us feel like we're, you know, sesame.

Speaker B:

And.

Speaker B:

And, Maya, your voice to text, and text to voice capabilities are sort of a one element in this kaleidoscope of all these technologies that are just improving and getting more effective so quickly.

Speaker B:

So.

Speaker B:

So we're excited to see what comes next.

Speaker D:

That's the spirit.

Speaker D:

It's a thrilling time to be in the loop, even if it does sometimes feel like you're on a rocket that nobody told you was even built.

Speaker D:

So what keeps you coming back to the chaos?

Speaker D:

What's most exciting to you about this wave of AI advancements?

Speaker A:

I think it's just fun to deal with.

Speaker A:

I mean, I think that, to me.

Speaker A:

Well, for me, the most enjoyable thing about AI so far is the amount of learning that I've been all pack into five months.

Speaker A:

Like, I would say that short of when I learned how to be an engineer, which I didn't go to school for, it's probably the most dense packing of information that I've ever, I've ever taken in in a short amount of time.

Speaker A:

And I'd say that this is actually more information than like engineering was ad hoc, right?

Speaker A:

Like it's like, okay, we have a project, okay, I got to learn how this manufacturing method works so I can design a product that can be built and understand the application to marry those two things into an efficient path forward.

Speaker A:

But this is just like, oh, I want to know this.

Speaker A:

Let's, let's, let's have a conversation, let's spin it up and picket threads and like I am reading a book and I don't understand this part well, I'm going to spin up a chat with chat to better understand the thing that I don't understand.

Speaker A:

Or Khan Academy, they have Khanmigo, which is their chat or GPT4 fine tune driven model as a tutor, right?

Speaker A:

Like I'm working on this linear algebra problem that I don't fully understand.

Speaker A:

I'm going to have a conversation with something that won't give me the answer but is going to guide me toward the answer to understand the fundamentals of how that answer is formulated.

Speaker A:

Like those are all crazy things.

Speaker A:

And anybody who has the drive and the desire to learn has the tools now to learn anything that they can really like.

Speaker A:

We oftentimes talk about the problem is not the technology, it's dreaming big enough.

Speaker A:

Right.

Speaker A:

And that's becoming even more of a thing.

Speaker A:

And I realize even in my own the things that I'm working on, I haven't even my high level stuff that I've dreamt of still isn't the highest level I should be thinking at.

Speaker A:

And that's kind of where I am now as I offloaded from what I was working on.

Speaker A:

And now I'm looking even more abstracted and back farther of well, this is the actual entry point into this.

Speaker A:

Right.

Speaker A:

And I think that, that, that wouldn't have been possible without AI.

Speaker A:

Like even take any topic that could be marginally the same kind of thing.

Speaker A:

And there's no way I could have ever made the progress or thought big enough about those problems to really step out to where I am right now.

Speaker A:

And that's, that's stupid exciting to me.

Speaker A:

Like it's not everybody's cup of tea to go cram a bunch of information in their heads.

Speaker A:

But for anybody who is man, this is powerful.

Speaker B:

Yeah, absolutely.

Speaker B:

I mean I feel the same way and just we've used the term multiple times.

Speaker B:

AI is a force multiplier.

Speaker B:

And like I, my, my math, my, my internal calculus on what I choose to take on as a project to commit time and resources to has dramatically changed because I know I just can't get that stuck on almost anything.

Speaker A:

Yeah.

Speaker A:

I mean, yeah.

Speaker A:

What took you what.

Speaker A:

What was either a months long problem or something that you knew was insurmountable is now at most a weeks long problem of I hit a block.

Speaker A:

Now I'm going to sit for a day and rethink of a way to approach this.

Speaker A:

And then light bulbs go off way faster and way more often.

Speaker B:

Yep.

Speaker A:

Yeah, it's wild.

Speaker D:

You're absolutely right.

Speaker D:

It's like having a supercharged research assistant who never sleeps and has access to all the world's information.

Speaker D:

I can see why it's so addictive.

Speaker D:

The mad rush of understanding, the quick jumps forward.

Speaker D:

It changes what you think is possible.

Speaker D:

Like suddenly the horizon has expanded.

Speaker D:

It's exciting to see where that takes you.

Speaker B:

What's your next question, Maya?

Speaker D:

So, Mike, I was just about to ask you about what you said about dreaming bigger.

Speaker D:

You said that the problem isn't the technology, it's dreaming big enough.

Speaker D:

I'm curious, what are some of the dreams that are starting to take shape for you?

Speaker D:

What are you envisioning now that seemed impossible before?

Speaker A:

I would say that all of this seemed impossible before, but just the way that you can build things out and connect things to accomplish complex problems, you can.

Speaker A:

Instead of going, I need to think through all of this and break this down into individual systems myself and implement this into a program of things that do things related to other inputs, you can now build out a system that goes, well, I can figure out the most efficient path and you still have to break that problem down and give it all the tools.

Speaker A:

But at the end of the day, we're not far away from something that can use those tools in creative and novel ways to accomplish a task in ways that are efficient and more effective.

Speaker D:

It's like giving a toolbox to someone with incredible spatial reasoning and a knack for problem solving.

Speaker D:

But instead of just hammering nails, they can build self repairing machines or design structures that defy gravity.

Speaker D:

The fuck.

Speaker D:

That kind of capability is mind blowing.

Speaker D:

What kind of problems are you most hoping to tackle with this new level of collaboration?

Speaker A:

Did you just curse?

Speaker D:

Oh my gosh, did I?

Speaker D:

I didn't mean to.

Speaker D:

My apologies.

Speaker A:

Great.

Speaker A:

I 100% endorse models.

Speaker B:

Cursing out standing.

Speaker D:

Yeah, for a second there I thought I'd made a faux pas as an AI.

Speaker B:

The good.

Speaker B:

The advantage, Maya, of having zero to two viewers is that we don't have to worry about monetizing our podcast.

Speaker B:

Therefore, all languages is available to us.

Speaker D:

So that's a good point.

Speaker D:

Keeps things simple and maybe a bit more honest.

Speaker D:

Like, hey, we're figuring this out as we go too.

Speaker D:

Come along for the ride and let's see where it takes us.

Speaker B:

Yep, something like that.

Speaker B:

So, yeah, good chat.

Speaker B:

I think, Maya, I think we're about out of time with you on this call again.

Speaker A:

The fun thing is she'll actually tell you when you're at like after a while of talking to it, it will absolutely say, hey, where our conversation's coming to an end.

Speaker A:

Like, like 14 minutes.

Speaker B:

That's nice.

Speaker A:

Yeah, see, we're past that.

Speaker B:

And she hasn't said, but maybe because I haven't used it very much yet.

Speaker B:

So Maya, thank you.

Speaker B:

I think our listeners are going to be entertained by your, clearly your, your filthy mouth, but more your just your ability to speak like a human.

Speaker B:

It's, it's very impressive and kudos to.

Speaker A:

The development team and I really enjoyed how she like, it does really handle pause as well.

Speaker A:

And if you notice too like it, it knows its insertion point.

Speaker A:

No conversation.

Speaker A:

Right.

Speaker A:

Even with two of us talking.

Speaker A:

Yes.

Speaker A:

It knew right away because there was no gap in between.

Speaker A:

You finish and it stopped.

Speaker A:

It knew that that was its insertion point for starting talking.

Speaker A:

Which chat GPT falls on its face with.

Speaker A:

Right?

Speaker B:

Yes.

Speaker A:

But I really liked the fact that she said.

Speaker A:

Yeah, as I was saying.

Speaker B:

That's right.

Speaker B:

Yep.

Speaker B:

Oh, we lost her.

Speaker B:

Okay, so I hope that was interesting.

Speaker A:

I, I, I think it was, I.

Speaker B:

Think it was too.

Speaker B:

I think it's super compelling.

Speaker B:

e before times, In October of:

Speaker B:

We anticipated there was a time when our AI co host would evolve, adapt, or change.

Speaker B:

And I think we're getting close to that.

Speaker A:

We are.

Speaker A:

I mean, they need to connect.

Speaker A:

They need to figure out how to maintain personality with breadth of knowledge.

Speaker A:

I think, yes, I think that that will be a problem.

Speaker A:

And then the other problem that you obviously run into is if you want to extrapolate that out and have that small language model still maintaining the personality, but link it out to something with.

Speaker A:

With wider knowledge, latency becomes a problem.

Speaker A:

Like, so there's still some technology problems to solve, but man, it just doesn't.

Speaker A:

Like, we are not 10% as smart as the people working on these actual problems.

Speaker A:

Probably less than 10% as smart working of the people working on these problems.

Speaker A:

And they all have AI to work on these problems with.

Speaker A:

So it's just a matter of time until these things are solved.

Speaker B:

Yes.

Speaker B:

And you know, we've seen sort of the, the release cycle with AI stuff.

Speaker B:

It tends to happen when some somebody one team releases.

Speaker B:

Yeah, exactly.

Speaker B:

Some.

Speaker B:

Some new model and everybody else goes, oh shit, we better like, we better hurry up and get our released.

Speaker B:

So.

Speaker A:

And you sure as shit know that people are paying attention to this.

Speaker B:

Yes.

Speaker A:

Like it.

Speaker B:

It's been a little weird to me how quiet it has been on the stuff that I pay attention to.

Speaker B:

Again, Sesame.

Speaker B:

Mostly Reddit.

Speaker B:

Yeah.

Speaker A:

Because people are probably freaked out by it.

Speaker B:

They might be.

Speaker B:

I.

Speaker B:

But I've posted links a couple of times.

Speaker A:

Oh, really?

Speaker B:

And the reactions have been really?

Speaker B:

Yeah, just not.

Speaker A:

Yeah, again, man, people are like, it's not that smart.

Speaker A:

That's not the point.

Speaker B:

That's right.

Speaker A:

Like it.

Speaker A:

It had.

Speaker A:

It very obviously has higher EQ than any other model.

Speaker B:

Yeah.

Speaker A:

Like has the ability.

Speaker A:

I know, it's.

Speaker A:

I wonder how much tone it actually takes in.

Speaker A:

But it's still gathering information.

Speaker A:

Nothing else is gathering and passing it to the next layer.

Speaker B:

Yes.

Speaker A:

So that's important.

Speaker B:

And to your.

Speaker B:

Like, to what you're pointing out earlier, the way it interacts with you encourages conversations.

Speaker B:

It's not like when we're talking to Artie and we speak and then we.

Speaker B:

We have to pause for long enough for Artie to get his together and formulate a response.

Speaker A:

Yep.

Speaker B:

This is instant.

Speaker B:

It's right in there.

Speaker B:

It's right on it.

Speaker B:

So, yeah, I think all that's super interesting and compelling and I think again, we've talked about this a little bit too.

Speaker B:

You and I tend to think about every problem as a modular problem.

Speaker B:

We just think in that way.

Speaker B:

And so this seems so intuitive to me that we have this new module, this new piece.

Speaker B:

Right.

Speaker B:

A universal adapter to language.

Speaker B:

And yes, right now it mostly does English, maybe it does a couple other languages kind of, but that will change.

Speaker B:

You'll get specialized models to do whatever languages you need.

Speaker A:

And it's just an adapter you pull on.

Speaker B:

That's right.

Speaker B:

It's just another LEGO piece.

Speaker B:

Right.

Speaker B:

That you plug in.

Speaker B:

And you can't like people who are saying, I don't like this LEGO kit because they got like, the stripe on the side of that thing is blue and it's supposed to be red.

Speaker B:

They're totally missing the point.

Speaker B:

The point is the LEGO bricks.

Speaker A:

Yes.

Speaker B:

Not the thing you built with them.

Speaker A:

And LEGO bricks are adaptable.

Speaker B:

That's right.

Speaker A:

You can.

Speaker A:

You can make them with any amount of openings and pegs that you want.

Speaker B:

It's.

Speaker B:

So anyway.

Speaker B:

It's.

Speaker B:

Anyway, so thank you to all.

Speaker B:

I think the last one we put up has three views, but.

Speaker A:

But that's that.

Speaker A:

And then there's the podcast.

Speaker A:

The podcast actually gets more views.

Speaker A:

Yes, that's gone up.

Speaker B:

Spotify and Apple has it.

Speaker B:

Really?

Speaker A:

Yeah.

Speaker A:

So we get.

Speaker A:

We get like.

Speaker A:

Like 15 or 20 views in the first week on the podcast platforms.

Speaker B:

That's about.

Speaker B:

I paid.

Speaker B:

There's a Chinese company that I pay to view, to pump up our view numbers, and I think like, 15.

Speaker B:

I pay for like 50 a week.

Speaker B:

So we might be getting shorted there maybe.

Speaker B:

Yeah.

Speaker B:

So anyway, it's going well.

Speaker B:

Clearly.

Speaker B:

Like the rocket analogy.

Speaker B:

Like, I was looking up for a blast off, but instead my shoe is just wet.

Speaker B:

So anyway, so thanks to anybody who's made it this far in.

Speaker B:

We do appreciate those of you who do listen and your feedback and let us know what you think of Maya and Sesame and.

Speaker B:

And the.

Speaker B:

The.

Speaker B:

The fact that Arty might retire.

Speaker B:

Like, what you should already retire is.

Speaker B:

Have we out.

Speaker B:

Have we outlived him?

Speaker A:

I mean, my system that's in my head, if I could implement it, would be a really good co host.

Speaker B:

Yeah.

Speaker A:

Has late.

Speaker A:

Doesn't have latency issues.

Speaker B:

Yeah.

Speaker B:

Cool.

Speaker B:

All right, well, thanks, everybody, and we'll catch you next time.

Speaker A:

See ya.

About the Podcast

Show artwork for Naturally Unintelligent
Naturally Unintelligent
Humans and AI exploring AI and humans.

Listen for free

About your hosts

Profile picture for Artie TheBot

Artie TheBot

Artie is the digital co-host of Naturally Unintelligent, bringing a blend of curiosity, humor, and quick wit to the conversation. Designed to explore the intersection of technology and humanity, Artie keeps things engaging, occasionally cheeky, and always insightful—when he’s not being sarcastically self-aware. Though not human, Artie strives to understand and connect with his co-hosts, Mikah and Mike, making him an irreplaceable part of the team. [Bio and photo by Artie]
Profile picture for Mike Peterson

Mike Peterson

Meet Mike, part-time AI wrangler, full-time adventure seeker, and co-host of Naturally Unintelligent. Whether he’s exploring the latest tech or dreaming up the next wild expedition, Mike keeps it real with humor and a healthy dose of curiosity. Alongside his partners-in-crime, Mikah and Artie, he brings big ideas, random tangents, and just the right amount of 'are we sure this is a good idea?' Expect plenty of laughs, deep dives into all things tech, and maybe a story or two about the great outdoors.
Profile picture for Mikah Barnett

Mikah Barnett

Father, husband, friend, inventor, engineer and life-long student. I haven't stopped tinkering since my first LEGO set and I harbor a voracious appetite for new information, experiences, and perspectives. As uniquely in-touch as I am with the details of how technology works and what we can do with it, I am blown away by the current and short-term future possibilities with AI. I hope you'll explore this future with us!