Podchaser Logo
Home
The Correct Amount of Rocks

The Correct Amount of Rocks

Released Friday, 31st May 2024
Good episode? Give it some love!
The Correct Amount of Rocks

The Correct Amount of Rocks

The Correct Amount of Rocks

The Correct Amount of Rocks

Friday, 31st May 2024
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:00

There is a way to get John

0:02

Syracuse to leave his house. For those

0:04

of you who remember in November, leave

0:06

us November, I traveled up, I

0:09

made a multi-hundred mile train

0:11

trip to go see Marco and try the Vision

0:13

Pro prior to release at a lab. And I

0:15

said to John in advance, hey, I know you're

0:18

not going to the lab, but wouldn't it be

0:20

neat if you came down to New York and

0:22

visited with us. I think that would be a lot

0:24

of fun. And John told me to go outside and

0:27

play hide and go screw myself. Those were his words

0:29

exactly. In those words precisely. And

0:32

so I thought there was no possible way

0:34

to get John to leave his house, even

0:36

for his beloved friends of years and co-workers

0:38

for years and years and years. But

0:41

it turns out John, there is a way to get you out

0:43

of the house and I am happy to report the

0:45

ATP reunion for the first time since 2019 is

0:47

a go. John,

0:50

what's going on? We are

0:52

going to WWDC. All

0:54

of us? At Apple Park. All of us. All

0:56

13. Sorry, all three of

0:58

us. Indeed.

1:00

I am extremely extremely extremely excited.

1:03

I cannot overstate how excited I

1:05

am, especially to see the two of

1:07

you fellas, but also to be able to go to Apple

1:09

Park. The only place I've ever been on Apple Park is

1:11

a visitor center. You know, the public

1:13

visitor center that you don't need like any special privileges

1:15

to get to. I'm excited that

1:17

we are press for the purposes of this

1:19

event, which I'm really really excited about. I

1:22

swear I have never done a press

1:25

WWDC. I don't think so. And

1:27

I meant to look through my badges to see if I

1:29

could find a press badge. Maybe I have and it's

1:31

just been so darn long and I don't recall knowing

1:33

me. That's probably true. But one way or another, it's

1:35

happening again and I am so

1:38

excited. So I am glad that you are

1:40

willing to make the trek instead

1:42

of just a couple of convenient hours down

1:44

the interstate and or train from Boston to

1:46

New York. You're just going to cross the

1:48

country instead and I appreciate the effort. Yeah,

1:51

I mean I have done press WWDCs. We've

1:54

all attended many WWDCs, but I've

1:56

never been to one at Apple Park and neither has

1:58

Casey. So that's what I'm most excited for. is

2:00

to be officially allowed onto

2:03

Apple's campus. And I

2:05

think they don't let you take real cameras, which is kind of

2:07

a bummer, so I'll just take lots of pictures of my phone,

2:09

but whatever. I'm excited to do it. It'll

2:11

be a fun experience. Yeah, looking forward,

2:13

not looking forward to taking the plane flight, but you

2:15

know, you do what you have to. All

2:19

right, let's do a little follow-up. I

2:22

wanted to briefly call attention to there's

2:24

some new Vision Pro content. What,

2:27

really? Yes, there is. There's new Vision Pro

2:29

content. Who'd have thunk it? So

2:31

there's three things I wanted to call everyone's attention to,

2:34

two of which are from Apple, and one of which

2:36

is not. First of all, there

2:38

is a new sizzle reel. I

2:40

don't know if that's the bestest way of describing it, but

2:42

basically it's like a three and a half minute video that

2:45

gets you interested in the Vision Pro.

2:47

And previously, I hadn't watched it in a couple

2:50

of months at least, but previously my recollection is,

2:53

or the way I remember it was, it showed like a

2:55

little bit of the kids

2:58

playing soccer, rhinoceroses, it showed

3:00

the tightrope walker, and

3:02

I think it showed a very brief

3:04

bit of sports if memory serves. I

3:07

don't entirely remember, but one

3:09

way or another, there's a new one, and

3:11

I think it's really well done. It's still,

3:13

it's a little jumpy for my taste, just

3:15

a touch. It's not like that MLS thing

3:17

from a month or so back where it

3:19

was way too jumpy. It's just a touch

3:21

jumpy, but the soundtrack is excellent, and

3:24

I really, really like it. Again, three and a

3:26

half minutes. You can find it in the Apple

3:28

TV app, and since it's new,

3:30

it is kinda front and

3:32

center. I know I was complaining and

3:34

moaning about the information architecture last week,

3:36

but in this case, it's pretty good, so you should

3:38

check that out. Additionally, there's a

3:41

new, I believe it's the Adventure

3:43

series. This is the one

3:45

that had the tightrope walker on it.

3:47

There's a new episode of that all

3:50

about parkour, and it's three, I believe

3:52

Brits based on accents, although

3:54

who knows? Anyways, three

3:56

Brits doing all various

3:58

and sundry stunts. across paris and

4:01

it's really really well done it's like

4:03

twelve to fifteen minutes long someone that

4:06

much would and uh... and

4:08

i really enjoyed it it's not you know

4:10

earth shattering but it's really good and i

4:12

gotta tell you i know spoilers for the

4:14

twelve minute video but there's a point at

4:16

which the three gentlemen are trying to jump

4:18

from one rooftop to another and they position

4:21

the camera such that if you want you

4:23

can look down and see how tall it

4:25

is and i gotta tell you three

4:28

d and immersive it

4:30

looks area fell but what

4:33

pretty pretty cool and i think both of these

4:35

i mean that's it's all together these two things

4:37

are worth you know literally fifteen minutes a year

4:39

time i really do think it's pretty cool and

4:42

then finally uh... what if has finally launched

4:44

i had planned to do the whole darn

4:46

thing earlier today and report in on it

4:48

unfortunately they seem to release on pacific time

4:50

so it wasn't available until my afternoon and

4:53

i've only had the time to do the

4:55

first like handy ish minutes

4:58

but it was very very cool uh...

5:00

the premise here no spoilers is

5:02

you know you're in the marvel cinematic universe

5:04

and what if you know

5:07

things weren't the way they were in

5:09

the movies what if things went a

5:11

little awry and different and it's your

5:13

responsibility as part of the story to

5:15

try to fix all this and so

5:17

the portion that i've done so far

5:19

again first ten minutes or so is

5:21

they teach you how to cast some spells

5:24

lately is the interact uh... with

5:26

you so it begins as fully immersive you

5:28

can't see any of your own environment and then

5:30

it converts to uh... augmented such

5:32

that there are a couple of characters in your

5:34

space and i don't know if it was an

5:36

extremely happy accident or if it was deliberate but

5:38

i was doing this in my office and i

5:41

had spun my chair around for my desk and

5:43

monitors were behind me in my office i don't

5:45

know twelve feet by twelve feet from a concept

5:47

or something like four meters by four meters not

5:49

not terribly big not going to give me the

5:51

site and i don't matter not big at the

5:53

point and other public characters wanted designed to be

5:55

floating in the air and sure enough he was

5:57

floating in the air and his feet were kind of inside my guest

6:00

bed, which is behind my desk, but you

6:02

know, that's fine. It's he's magic. So whatever.

6:04

But then the other character was designed to

6:06

be on the ground and sure enough, he

6:08

was on the, you know, guest room floor.

6:11

Like he was standing on the floor. There's

6:13

a little shadow below him and it looked

6:15

just spot on. And then they had you

6:17

do some, you go back into an

6:19

immersive environment where it's, you know, all you can

6:21

see is what they're showing you. And

6:23

they have you do some spells where you're, you're

6:26

using hand tracking in order to do stuff. And

6:28

there were three or four that I learned and one

6:30

of them didn't really work that well, but the

6:32

rest were spot on. And again, I've only done

6:35

the first few minutes, but it's really slick and

6:37

it's free. So definitely, definitely check all three of

6:39

these out. Again, that's the new sizzle reel. I

6:41

think they just call it like immersive or like

6:44

immersive demo or something like that. Uh,

6:46

and then the parkour, the adventure

6:49

series, the second episode about parkour and what if by

6:51

Marvel and I've put a link to the, what if

6:53

thing in the show notes, assuming I can dig one

6:56

up. I should be able to, um, I

6:58

don't think I can link to the other ones in the show

7:00

notes, but if I find a way, I will do so. All

7:03

right. The low storage 13 inch

7:05

iPad pros have 12 gig Ram

7:07

or 12 gigs worth of Ram

7:09

chips in them, but they don't

7:12

use 12 gigs of Ram. John, what's going on

7:14

here? This is discovered by, uh, the,

7:17

I fix it folks with teardown, uh, their

7:20

lead teardown technician, uh, Sharon McCarthy, uh, found

7:22

that readers apparently, meaning like people who saw

7:24

the YouTube video, I don't know how they

7:26

spotted this, but they apparently spotted two, six

7:28

gigabyte Ram modules on the 256 gig and

7:31

512 gig 13 inch iPad pro. I looked at

7:33

the teardown video. I looked at the teardown on

7:35

the website. I'm like, how did they spot that

7:38

these were 12 gig modules? I can barely make

7:40

out the part numbers. They're so blurry from compression,

7:42

but whatever. Um, anyway,

7:44

uh, Sharon says our chip

7:46

ID confirms this with high certainty

7:48

to six gigabyte LPDDR 5X modules

7:50

produced by micron for a total

7:52

of 12 gigabytes of Ram. To

7:55

remind you, uh, Apple says this machine has eight gigs of

7:57

Ram because remember, if you want to get 16, you got

7:59

to get the higher. storage, they say

8:01

it has eight gigs, but there's apparently 12

8:04

gigs on the chips there. So why

8:06

does Apple utilize only eight gigs of RAM? I

8:09

think it says Apple has never done this before as far

8:11

as we know. Someone asked if it's possible these RAM chips

8:13

are defective and some of the RAM is disabled for some

8:15

reason. The folks at iFixit says

8:18

that, my understanding is that if this were the

8:20

case, they would receive a different part number and

8:22

be labeled as four gig, but I don't think

8:24

that's how LPDDR5 is manufactured anyway. There's very little

8:27

doubt that there's 12 gigs of RAM. Another

8:29

possible question, what if those are the

8:32

only LPDDR5X modules they can get their hands on, but

8:34

as soon as they're able to put in four gig

8:36

ones, they will. So just to avoid problems, this

8:38

initial batch is artificially limited. And

8:40

the iFixit answer to that was that these modules have been in

8:42

production since 2020 according to the spec sheets.

8:45

So it seems unlikely based on how long they've been

8:47

around. Super weird

8:49

because like,

8:51

why would they do that? Like, you know, if you

8:53

follow the thread, the Twitter thread about this, you can

8:56

see people throwing out a bunch of theories and then

8:58

basically getting shot down by iFixit. Why

9:01

would they use bigger RAM chips? Could they

9:03

not get smaller RAM chips? And

9:05

if they couldn't get smaller RAM chips, why wouldn't they

9:07

just free up the whole 12 gigs? The

9:10

other theories are like, oh, they're reserving a certain amount of

9:12

RAM for LLM stuff. Why would they

9:14

only do it on this model and not all the

9:17

models? Because I think this is the only one that

9:19

has more RAM and the other ones have the amount

9:21

that Apple says based on the teardowns. And

9:23

this one has this particular variant of this

9:25

particular model has 12 gigs instead of eight.

9:28

Super weird. But maybe this is

9:31

the side door for Apple finally putting more RAM

9:33

in its products. They'll just install it

9:35

and not enable it for you. But it's there.

9:37

You just can't use it. How is your M4

9:40

iPad Pro treating your eyes?

9:43

Because apparently it's not all roses and

9:45

pansies. I mean, it's good for me.

9:48

But Ben writes in to

9:50

say, I'm upgrading to an M4 iPad Pro from the 2018

9:52

iPad Pro almost immediately.

9:55

I noticed that my eyes seemed unable to

9:57

properly focus on the display resulting in eyes.

9:59

strain, fatigue, blurry vision, and even headaches. I

10:01

couldn't use the display for very long before

10:04

the symptoms reappeared, so I went down a

10:06

rabbit hole researching. This is kind of

10:08

like Marco with his in-pro thing, but even

10:10

worse. It seems like I'm not the only

10:12

one experiencing this, though. I have yet to

10:14

determine the exact issue. It might be PWM,

10:16

which stands for pulse-width modulation. And by the

10:18

way, when I follow these links to look

10:21

at the research that he was doing, whatever, everyone

10:23

just says, oh, it might be PWM. I think you

10:25

have a PWM. Yeah, it's probably PWM. People will come

10:27

into a forum or a Reddit or whatever and say,

10:29

hey, I just got a new M4 iPad Pro, and

10:31

the screen hurts my eyes. What do you think the

10:34

problem is? And people would say, yeah, it's probably PWM.

10:37

And I was like, are you going to explain what

10:39

PWM is? I mean, I guessed it was pulse-width modulation

10:41

just by knowing the term or whatever. But when you're

10:43

helping somebody, don't just say, yeah, you probably have PWM,

10:45

because they don't know what PWM is. Never mind that

10:47

the term pulse-width modulation doesn't make much sense. I bet

10:49

if you hear this now, you're like, well, I know

10:51

what pulse-width modulation is, but what does that have to

10:53

do with screens and why it would be hurting their

10:55

eyes? So we'll get to that in a second. So

10:57

anyway, Ben says it might be

10:59

PWM, though I've never known this to be a

11:01

problem. And I have been using LG OLED TV,

11:03

as well as OLED versions of the iPhone Pro

11:05

for years without any issue, or maybe tandem OLED

11:07

is misaligned. There is such a thing. I

11:10

ended up going to the Apple Store and compared my device

11:12

with others. Mine appeared to be slightly different, as if the

11:14

HDR was turned on all the time. Overall,

11:17

the display was always just too much, as best

11:19

described as basically yelling at me all the time.

11:21

Since I was still within the 14-day return period, they

11:24

switched my device to replacement, which seems to be much

11:26

better now at all and not perfect. My question is

11:28

for John, are you noticing any eye fatigue with the

11:30

new Pro, especially compared to the 2018 version? If so,

11:32

do you expect this could be improved through software updates?

11:35

So here's the research on

11:37

what PWM is talking about. According

11:39

to these links that Ben provided that we'll put

11:41

in the show notes, the

11:43

way this OLED and some

11:45

other OLEDs handles brightness,

11:49

there's two ways OLEDs can handle brightness.

11:51

One is they can send less voltage

11:53

to the pixels and they're not

11:55

as bright. So if you dim

11:57

the brightness of the screen or whatever, how does it do

11:59

that? do that. One way you can do it is

12:01

you can just send less voltage to the screen and it

12:03

gets dimmer, right? But if you

12:06

use, if you only use adjusting the electricity

12:08

going to the pixels to the control brightness,

12:10

apparently when you get to low brightness levels,

12:13

you lose a lot of the color saturation too and it

12:15

looks kind of like dingy and gross. So they

12:18

tend to not want to do that

12:20

for the lower brightnesses. The

12:22

other way you can control brightness on an OLED

12:25

is you can have the screen be at

12:27

maximum brightness briefly and then

12:29

go off and then max brightness and off and

12:31

the longer, you know,

12:33

the on period is, the brighter it

12:35

is. So if they

12:37

show like a little graph and this is the pulse width modulation

12:39

of saying, you know, if you just pulse the screen, pulse, pulse,

12:42

pulse, pulse, pulse, the faster the pulses go, the brighter the screen.

12:44

If you go pulse, pulse, pulse,

12:47

it is dimmer because the light is on less

12:49

of the time. You don't notice

12:51

this because it pull Apple screen this

12:53

time. Oh, it pulses at 480 times a second, which is a pretty

12:55

high refresh

12:57

rate if you remember from the CRT days like,

12:59

oh, my screen looks flickering at 60 Hertz. I

13:02

can see the flicker but 85 I can't see

13:04

it anymore. At 120 I definitely can't see it.

13:06

At 480 I can tell

13:08

you I cannot see OLED flickering.

13:11

But the complaint about the M4

13:14

iPad Pro tandem OLED is that it

13:16

uses pulsing to

13:19

control its brightness through its entire brightness range.

13:21

Apparently even at maximum brightness, it's still pulsing

13:23

as opposed to some other OLEDs which will

13:25

use pulsing down at low brightnesses. But once

13:28

they get the high brightnesses, they will do

13:30

that by keeping it out all the time

13:32

but it's sending less voltage. I

13:34

don't know if this is just the way tandem OLEDs work. Is

13:36

this the way Apple is choosing to make it work? But some

13:38

people report that this

13:41

bothers them. I don't personally see

13:43

how it could because I am not aware of

13:45

anyone who would notice flickering at 480 Hertz. We're

13:47

not even talking about motion here. We're

13:50

just saying like put on just a full

13:52

field, you know, red slide.

13:55

I can't see it flickering at 480 Hertz.

13:57

If you had asked me whether this is

13:59

flickering to control brightness, I would have said

14:01

no, because I can't see it. But maybe

14:03

people with very young or better eyes than

14:05

mine can see it. I

14:07

don't know. Or maybe it's anyway, follow

14:09

the links, decide for yourself. I,

14:12

and by the way, I would say, do iPhones

14:14

use this or do they use voltage regulation? I

14:17

honestly couldn't tell you because it just looks like

14:19

a screen to me and I can't see whether

14:21

it's dimming or pulsing at 480 hertz. Some

14:24

of the OLEDs that are out in the market

14:26

pulsed even higher rates than that. So yeah,

14:29

is this an issue? Is this something people are

14:31

imagining? Are some people just uniquely sensitive to it?

14:34

All I can tell you is that my old

14:36

man eyes don't see this and aren't

14:38

bothered by it. But we'll see. We'll

14:40

see if Apple has an update. There was, we

14:42

didn't put this in the notes for last episode,

14:44

but there is some kind of actual software error

14:47

with displaying HDR video where some of the highlights

14:49

are getting blown out that Apple, uncharacteristically,

14:51

Apple immediately acknowledged inside a software fix

14:53

was coming for. So

14:55

it is possible that maybe whatever issue people are complaining

14:57

about here will also be wrapped up in a fix.

14:59

We'll see. Michael Thompson

15:02

writes in with regard to trillions

15:04

of operations per second measurements or

15:06

TOPS measurements. Michael writes, I

15:08

found this article on the Qualcomm website that

15:10

suggests that the TOPS measurement they use for

15:12

their NPU performance is based on 8-bit integers.

15:14

In the paragraph headed precision, they state, quote,

15:17

current interest industry standard for measuring AI

15:19

interference, inference, excuse me, and TOPS is

15:22

an, is that int 8

15:24

precision. The context here being

15:26

whether or not the new surface line in

15:28

the, what is it, the copilot plus PC

15:31

or whatever it is, line of PCs,

15:33

are they or are they not

15:36

actually faster for neural related things

15:38

than Apple stuff? And

15:40

so I don't recall what was, does

15:42

this mean they are faster than I presume? Well,

15:45

I mean, so this, this matches what we

15:47

saw. The idea that, that on this website,

15:49

there's Qualcomm is saying the industry standard is

15:51

8-bit int 8 precision, right? And what we've

15:53

seen is Apple using int 8 and

15:56

so is Qualcomm and so is Microsoft and everybody

15:58

who's, who's current talking about the current. line of

16:00

products, they're all using 16-bit

16:18

stuff. And so their numbers were half as big.

16:20

But with their new stuff, Apple is using numbers

16:22

that are twice as big, and they're using Intate

16:24

and so is Qualcomm. So the answer is, everyone's

16:27

using Intate now. Is that because Intate

16:30

is more representative of the actual jobs

16:32

we're asking our NPUs to do? Maybe,

16:34

but whatever. The industry has

16:36

decided, when measuring tops, we're

16:38

going to use Intate precision. That

16:40

may become less relevant

16:42

if it turns out that the things we ask

16:45

our NPUs to do involve 16-bit or

16:47

32-bit values, and it doesn't really matter

16:49

how fast they can do stuff on Intate things, but

16:51

I would trust that Intate is actually a relevant measure

16:53

right now. So the answer is, the

16:56

Copile Plus species and the Snapdragon X

16:59

Elite thing has 40 tops. The

17:01

M4 has 38. Those

17:03

are both Intate measures. That means they're

17:05

essentially comparable. Eric

17:07

Jacobson writes in with regard

17:10

to iCloud Drive and Node

17:12

modules. So if you

17:14

recall, this was with John Sun, who

17:16

basically nuked his MacBook Air by trying

17:20

to sync the Node modules folder through iCloud

17:22

Drive. Eric writes, I haven't used it since

17:24

I don't use iCloud Drive, but there's a

17:26

project that will add a No Sync directive

17:28

to every Node module on a file system.

17:31

I imagine it might need to be rerun whenever a

17:33

new project is kicked off, and we'll put a link

17:35

in the show notes to No Sync-iCloud. Yeah, and

17:38

I tried to look at the code to remind myself

17:40

how. I think you make a directory with a .nosync

17:42

extension that has the same name as the other one.

17:44

That's the way you signal to iCloud Drive not to

17:46

sync the directory or something like that. This

17:48

is the Node module itself, so you can look at the code.

17:51

But unfortunately, the documentation is all in, what are we

17:53

going to say here? I'm going to say Chinese? Yeah, something

17:55

like that. Yeah, so the documentation is in Chinese, and I

17:57

can't read it. But the source code is not in Chinese.

18:00

in Chinese and I still couldn't quite make any sales of

18:02

it. But yeah, I

18:04

think it's just making .nosync directories in the right places.

18:06

And it's a node module that you can use, and

18:08

it will, I think, be just included in your project

18:10

and make sure everything knows sync. So that's useful

18:12

and helpful if you want to dare to walk

18:15

that tightrope of trying to use node modules with

18:17

iCloud Drive. Indeed. And then

18:19

Eric continues, I do, however, use Time Machine.

18:21

It can attest that the ASIMOV utility works

18:23

perfectly for excluding node modules and other dependency

18:25

directories. Also, it is a background service, so

18:27

it doesn't need to be reinitialized. And we

18:29

will, again, put a link to the show

18:31

notes, to ASIMOV. And

18:34

also, John, I guess you wanted to call attention to

18:36

the list of excluded directories, which I put in the

18:38

show notes as well. Yeah, it shows what

18:40

kind of things, like when it says dependencies, what

18:42

does that mean? Obviously, it means node modules, but

18:44

it has a whole list of all the different

18:47

things it excludes, things from Gradle, Bower, PyPy,

18:49

NPM, Parcel, Cargo, Maven, I

18:52

think CocoaPods might be, yeah, CocoaPods is in there.

18:55

Marco, have you heard of any of those things

18:57

other than CocoaPods? Precisely zero.

19:00

Flutter. Anyway, it's sad

19:02

that I think I've heard of all of

19:04

these. But I installed

19:06

this, and I ran it. It installs a little

19:09

launch daemon thing or whatever. And

19:11

it essentially does the, I forget

19:13

about the

19:15

time machine. I think it's an extended attribute or

19:17

something, but there's a way to exclude things from

19:19

time machine. Maybe it just calls TMUtil. Yeah. Anyway,

19:23

excluding all these directories from your time machine

19:25

backups can make your time machine backups go faster.

19:27

What it's basically saying is you

19:29

don't need to back up the dependencies

19:31

of your code. If you're writing something in node, and

19:33

you use 17 node modules, you don't need to back

19:35

those up. You get them from the internet anyway. You

19:38

got them through NPM or YARM or whatever. They're on

19:40

the internet. Do not back them up. It's not your

19:42

code. It's a dependency. You didn't write that code. It's

19:44

just pulling it in. And there are tons of files.

19:46

So if you can exclude those directories in time machine,

19:48

it will make your time machine backups go faster. But

19:51

who remembers, oh, what am I going to do? Go

19:53

to Options in Time Machine and drag the little thing

19:55

in there, set the extended attribute. I don't remember how

19:57

to do this. This just runs in the background. all

20:00

the time, looks for directories that fit this

20:02

signature, and excludes them from time machines. I

20:04

did that, I probably saved, I

20:06

don't know, thousands, many

20:08

thousands of files are no longer on my time machine

20:11

backups, because I ran this. I hope

20:13

it doesn't have bugs and isn't excluding a whole bunch

20:15

of important files from my time machine backups, but you

20:17

know, I've got multiple backup strategies. So

20:19

for now, I'm trying the experiment of running this

20:21

as I'm off-demon in the background to see if

20:23

it helps with my time machine backups. And I'm

20:25

still not running iCloud Drive, of course. And

20:29

then finally, this is actually, I should have

20:31

moved this up by the other Vision Pro

20:33

followup, didn't think about it, but anyway, Jonathan

20:35

Gobranson writes with regard to audio routing during

20:38

Vision Pro guest mode. So if you recall,

20:40

I was doing demonstrations for my mom and

20:42

dad, and I noticed that when mom was

20:45

on Mount Hood or whatever it's called, and

20:47

I had her go fully immersive, that the

20:49

crickets and whatnot were being routed through my

20:51

iPad Pro, which was doing mirroring at the

20:53

time. And so it doesn't really make for

20:55

a very good effect if the audio is

20:58

going through there. So Jonathan writes, you can

21:00

choose during setup of each guest user session,

21:02

whether to route audio to the Vision Pro

21:04

or the iPad or whatever the case may

21:06

be, if you choose to mirror content. And

21:08

we'll put a link to the Knowledge Base

21:11

article. So what you do is you look

21:13

up, and

21:15

you get the little green down

21:17

chevron near the top of your view.

21:20

Then you go into Control Center, and

21:22

you go back into the mirror my

21:24

view button. And then in there, there's

21:26

a audio routing section that

21:29

you can choose to push everything back onto the Vision Pro.

21:31

Not entirely sure why this isn't the default, to be

21:33

honest, because pretty much every time I've always wanted this,

21:36

but here we are. At least now I know that

21:38

there is a way around it. So good deal. We

21:42

are sponsored this episode by Fast Mail, not

21:44

only a great email host, but my chosen

21:46

email host since 2007. Obviously

21:49

long before they were sponsored, long before I was a

21:51

podcaster, I've been a Fast Mail customer, and I've been

21:53

very happy because hosting email, it's

21:57

constantly a moving target. You don't wanna do it yourself.

21:59

You don't wanna... to also use one of

22:01

the big services necessarily because then your name

22:03

is tied to them. I

22:06

like having my own control. I like being

22:08

able to have standardized hosting with a domain

22:10

that I own. Fastmail is a great place

22:12

to do this. They even now have domain

22:14

sales. You can add or buy domains right

22:16

through them and they set up all the

22:18

records for you and everything so it works

22:20

immediately and painlessly, but it's just a great

22:22

email host. Fastmail has all the features you

22:24

want under an email host. Of course, obviously,

22:27

it's iMap hosting and they have all the basic

22:29

default stuff that works with every mail client.

22:31

They also have all the modern productivity features

22:33

you might want, things like scheduled sends, snooze,

22:35

folders, labels, search, etc. They have all the

22:37

great plans, all the great features. I love

22:39

Fastmail. They now also have multi-user plans. If

22:41

you want to have your whole household have

22:43

a plan, they have a duo plan for

22:45

secure private email for you and a partner

22:47

at a reduced price and for just a

22:49

little bit more. They have a family plan

22:51

for up to six people. You can upgrade

22:53

your entire family. Of course,

22:55

in those family and duo plans, you have

22:58

things like shared domains if you want. You

23:00

can have shared or private calendars and address

23:02

books. You can have a family calendar all

23:04

shared, hosted by Fastmail, all based on standards.

23:06

It is wonderful. I strongly

23:09

recommend Fastmail and they are super private

23:11

too. For over 20 years, Fastmail has

23:13

been a leader in email privacy. They

23:15

believe in working for you as people

23:17

to be cared for, not data and

23:19

products to be exploited. Fastmail

23:21

has better span filters than everyone

23:23

else and absolutely no ads. It

23:25

is super easy. It gets to serve with

23:27

Fastmail today. You can download your data from

23:29

wherever you are now. You can import it into

23:32

your new Fastmail inbox. You don't need to leave

23:34

anything behind when you switch. It is great.

23:36

Try Fastmail free for 30 days and get

23:38

10% off your first year

23:40

at fastmail.com/ATP. Once again, free for

23:42

30 days, 10% off your

23:45

first year, fastmail.com/ATP. Thank you so much

23:47

to Fastmail for being an awesome email

23:49

host for all these years and for

23:51

sponsoring our show. you

24:00

too soon, but nevertheless, we

24:02

should talk about some last-second predictions,

24:05

and I guess this is most

24:07

predominantly from Mark Gurman in today's

24:09

episode. So I'm going to read

24:11

a whole bunch of stuff. One of you guys, feel free

24:13

to pipe in and interrupt at your convenience. But

24:15

here we go. Mark Gurman writes, Apple

24:18

is preparing to spend a good portion

24:20

of its worldwide developers conference laying out

24:22

its AI-related features at the heart of

24:24

the new strategy

24:26

is Project Grey Matter, a set of AI

24:29

tools that the company will integrate into

24:31

core apps like Safari, Photos, and Notes.

24:33

The push also includes operating system features

24:35

such as enhanced notifications. Much

24:38

of the processing for less computing-intensive AI features

24:40

will run entirely on the device, but

24:43

if a feature requires more horsepower, the work will be pushed

24:45

to the cloud. There are several new capabilities in

24:48

the works for this year, including ones that transcribe

24:50

voice memos, retouch photos with AI, and make searches

24:52

faster and more reliable on the spotlight feature. Faster

24:55

would be great, particularly on my iPad, please, and

24:57

thank you. They also will improve Safari web search

24:59

and automatically suggest replies to emails and text messages.

25:03

The Siri personal assistant will get an upgrade as well, with

25:06

more natural sounding interactions based on Apple's

25:08

own large language models. There's also a more

25:10

advanced Siri coming to Apple Watch for on-the-go tasks.

25:12

Developer tools, including Xcode, are getting AI enhancements

25:14

too. So let's stop here for a second

25:16

and look at this list of features, because we're

25:18

always like, how will Apple add AI sauce

25:20

to all this stuff? There

25:24

was a big story, German had it, and I think we might

25:26

have mentioned in the show, like, oh, they're going to fix Siri.

25:29

We were speculating months ago or weeks ago, whatever.

25:31

Is this the year that they're going to fix

25:33

Siri with AI, or are they just going to

25:35

add it to a bunch of other stuff? German's

25:37

rumor was like, no, they're doing a Siri thing,

25:39

so we can expect to see that. But here's

25:41

some specifics. Specifics seem not weird. I don't know.

25:45

Sometimes these rumors aren't comprehensive. Very often, Apple emphasizes

25:47

one or two particular things, whereas we just get

25:49

a laundry list and we don't know which one

25:51

they're really going to concentrate on, and it's going

25:53

to be impressive. But let's look at some of

25:55

these in turn and see how exciting they are.

25:58

Transcribing voice memos. Apple's been

26:00

doing transcriptions, for example, on voicemail

26:03

for a long time now. Having

26:05

transcription be better, that's

26:08

good. Probably not going to really radically

26:10

change people's licenses and existing features, and it's making a

26:12

little bit better. Retouch photos with

26:15

AI could mean a lot of things. That could be

26:17

a headlining feature. That

26:19

short description doesn't tell us. Is this going to just be

26:21

like, oh, they're better at background detection for tearing people off?

26:24

Or is it going to be like, oh, they're going

26:26

to really emphasize that they're doing the feature that Google

26:28

and many others have where if there's something in the

26:30

background or photo you don't want, you can remove it,

26:33

remove a person from the background who shouldn't be there,

26:35

remove a sign or a tree or something

26:37

like that. That

26:40

could be a big headlining feature. Google has certainly had

26:42

a whole ad campaign about theirs, and obviously it's

26:45

every graphics application from Photoshop to

26:47

Lightroom, Pixelmator and everything has

26:49

features like this, and they've been touting them, and I think they're

26:51

crowd pleasing. That's part of

26:53

the problem. So far, like, all right, voicemail

26:55

transcription and fancy AI photo cleanup and removal,

26:58

that's table stakes today. Neither

27:01

of those – I'm glad Apple's

27:03

getting there. They should, but neither

27:05

of those I think is likely to make

27:07

a big splash simply because other

27:10

people have been doing that for a decent amount

27:12

of time, and the rest of the industry is

27:15

there. What I want to

27:17

see from Apple

27:19

is what we get excited about from Apple. I

27:21

want them to blow us away with stuff that we

27:24

haven't seen from everyone else. I want to see features

27:26

that are not just playing catch up with the rest

27:28

of the industry. I want to see Apple leading the

27:30

way, not following, and these are just following so far.

27:33

Yeah, I mean, as with all these things,

27:35

like, we know that this has been done

27:37

elsewhere and even on Apple's own

27:39

platforms in various applications, but when people get

27:41

one of these new phones or they update

27:44

the OS or they see an ad

27:46

on TV that Apple touts this feature of erasing a

27:48

person, chances are good that

27:50

they've never seen that before if they're not a tech

27:52

nerd, right? And so as far as they're concerned, wow,

27:54

this is amazing. Especially if they don't

27:57

know it's even possible, like, they're just saying, oh, on my phone, I

27:59

could just tap a – person and remove them. And we all

28:01

know that that's been around for a long time and they're

28:03

playing catch-up. But A, they do have to play catch-up. We

28:05

don't want them to not do this. And B, it

28:07

can be just as impressive. It can be very impressive

28:09

to people who haven't seen it before. So I just

28:11

hope they do a good job of this because there

28:13

is a lot of competition, again, even on Apple's own

28:15

platforms in, you know, applications

28:18

that you can get for the Mac and for the

28:20

iPad and for the iPhone that already do the same

28:22

job. I hope Apple at least

28:25

matches them. Historically, Apple has kind of

28:27

been, they want

28:29

to make it simple. So there's not a lot of adjustability.

28:31

And if it does a bad job, like the Apple way

28:33

is like, well, just hop on the person and remove them.

28:35

And you'll tap on a person and it will make a

28:37

terrible mess of it. And you'll be like, is

28:40

there something I can do to try that again and do better?

28:42

And the Apple way is no. Nope. You tap

28:44

them and it does a good job. Good. And if it

28:46

doesn't, oh well. Whereas like in Lightroom,

28:48

there's a million knobs you can adjust and you can take

28:50

a second attempt and you can mask differently. You do all

28:52

sorts of things that are fancy your application. So there's

28:54

that. I just

28:57

really I really hope they do catch up and I hope

28:59

they do a good job of it. The

29:01

next time about improving Safari web search.

29:05

How? I get. All

29:08

right. Well, I mean, I can say I mean, for

29:10

whatever it's worth. Obviously, we don't

29:13

know what this means yet. We might not even learn

29:15

it in two weeks. But Safari

29:17

had like whatever Apple is doing,

29:19

whatever their back end logic is

29:21

for autocomplete suggestions

29:24

in Safari is

29:26

basically a search engine like they like I'm

29:28

I can't even imagine how many like millions

29:31

and millions of web searches Apple

29:33

avoids even making on the iPhone

29:35

by people just tapping that first

29:37

autocomplete review result that comes up.

29:40

So maybe it's just related

29:42

to whatever the back end is of that and

29:44

that could be one of those like invisible things

29:47

that everybody will take for granted and nobody will

29:49

even notice but we will appreciate it getting better.

29:51

You know, like things like autocorrect. Well,

29:53

notice if they screw it up like they've changed autocomplete many

29:56

many times over the years and when they screw it up

29:58

we all notice and we don't see it. I really

30:00

hope it doesn't start suggesting ridiculous

30:02

things. I think that might all be

30:04

local, too. So maybe that is a place for LLMs to

30:06

try to do better predictions. But anytime

30:08

they're messing with an existing feature, you hope it's

30:11

going to be a big improvement. But there's a

30:13

possibility, especially when you're throwing LLMs in the mix,

30:15

it's a possibility. The current strategy

30:17

is look at your past history, look

30:19

at the number one search results. They

30:22

could even be hitting the Google back end for some

30:24

of those things. It's a very straightforward, non-complicated algorithm. But

30:28

it's deterministic, right? And if they switch

30:30

from that to like, let's just chuck it over the wall to the LM

30:32

and see what it says. I'm concerned

30:34

that there might be some wackiness there, but

30:36

we'll see. More reliable

30:39

spotlight feature similarly. When you use spotlight on

30:41

the phone, I don't know if

30:43

it's using spotlight, spotlight. It's the same technology

30:45

as the run to the Mac. But it's

30:47

like, oh, I'm searching for stuff on my

30:49

phone. And it includes contacts and all your

30:51

applications and files and things you've

30:53

done recently. And yeah, you could

30:55

probably throw LLMs into the mix there to handle.

30:58

It's basically like, so you don't have to type

31:00

things exactly. And it's better about synonyms. And you can

31:02

type in vague things. Type

31:04

in some expression about what you want to do in settings

31:06

and have the LLMs figure out the setting you want to

31:08

go to. Potential for good, also

31:11

potential for messing up spotlight on the

31:13

phone. And then suggest

31:15

replies to emails and text messages. This starts to

31:17

get into the area where we thought, will Apple

31:19

go there? We haven't gotten

31:21

to chat bots yet. But obviously,

31:24

everyone else is doing this so it's another catch-up feature. But

31:27

the idea that Apple would, within the mail application or while

31:29

sending text messages, pop up a little clippy and say, it

31:31

looks like you're trying to reply to your friend. Do you

31:33

want me to write it for you? Which

31:35

is what, I mean, just look at the Gmail interface, for

31:37

example. Every time I'm in Gmail, it's like, do you want

31:40

me to summarize this email? Do you want me to just

31:42

write the reply for you? Like, Google's been pushing on this

31:44

for years. It used to just be like, here's

31:46

a canned phrase that is probably a reasonable reply to

31:48

this email. And now it's like, you know what? I

31:51

can just write the whole email for you. Like, you

31:53

don't have to click from canned phrases. You don't have

31:55

to, you know, the autocomplete that happens in like Google

31:57

Docs and everything. It's just like, let me just write the whole

31:59

thing for you. And that, we

32:02

should find a link to Nevin Mergens blog post. Nevin

32:04

Mergens said he got an email from a friend that

32:06

was written by an AI and he kind of

32:09

flipped out about it, right? That seems

32:11

like a less, that

32:14

is not a conservative move to suggest to

32:17

people that the phone will write their email

32:19

or their text message for it. And summarizing I can see

32:22

it's like the phone is helping me if I just want

32:24

to summary it, I don't want to read it all, let

32:26

the phone summarize it. It's like asking the phone to help

32:28

you deal with your stuff. But

32:30

having generative AI write something

32:32

that you will send as if you wrote

32:35

it yourself is a step

32:38

farther than Apple has gone. So we'll see how they

32:40

present that. Again, obviously everyone else is doing this, it's

32:42

not new. I'm not saying Apple is going to be

32:44

the first one to do this and they're going to

32:46

get in big trouble or something. It just doesn't, it's

32:49

not a very safe thing to do. So I'm wondering

32:51

how daring Apple will be in like

32:53

a little presentation. They'll just be like, and it

32:55

can even suggest a reply for you and some

32:57

happy person who's happy to get a text will

33:00

tap a thing on their phone screen and they'll

33:02

be so delighted to see the three word reply

33:04

come and they'll hit the send button and they

33:06

won't talk about it anymore. But

33:09

I don't know, it wigs me out a little bit too. I

33:12

mean keep in mind like we're probably not

33:14

that far away from, I mean people are probably

33:16

doing it now. Where like you're

33:19

using your AI to compose an email

33:21

that will be then read by the

33:23

recipient's AI and we will all just

33:25

be sending even more garbage crap to

33:27

each other and having it be processed

33:29

by even more garbage AI junk

33:32

on the other end and we will finally reveal email

33:34

to be the useless thing that has been most of

33:36

this time. It's not useless and I feel like

33:38

in a business setting this is most useful because there's

33:40

a lot of boilerplate and ICDs in business that would

33:42

help. But like we already have auto, like I mentioned

33:45

auto complete, like Gmail forever has had a thing

33:47

where you start writing something and it guesses what

33:49

you're probably going to say for these next three words and

33:51

you can auto complete that. But

33:54

I feel like that kind of piecemeal auto complete,

33:56

even if it is L-empowered, that piecemeal auto complete

33:58

at least forces you to read it. everything,

34:00

whereas having a button that says, compose

34:03

a reply for me, invites the possibility that you will

34:05

not read what it generated, because that will take too

34:07

much of your time, and you'll just hit send. And

34:09

now we're sending messages that we're not even looking at

34:11

to give a thumbs up or a thumbs down on,

34:14

because it just takes too much time. And that's just,

34:16

I think it's a waste of people's time, because

34:18

if you don't read what it wrote, maybe it didn't write what you

34:20

wanted it to write, and now you have an even more confusing email

34:22

thread. And I mean,

34:24

again, we'll figure this out as a culture,

34:26

as a technology advances, but I'm just saying

34:28

character for Apple, for Apple's thus far very

34:31

conservative approach to AI, suggesting

34:34

replies to emails and text messages seems like

34:37

a big move for this particular

34:39

company. Yeah, agreed. Continuing

34:43

from Mark Gurman, one standout feature will

34:46

bring generative AI to emoji. The company is

34:48

developing software that can create custom emoji on

34:50

the fly, based on what users are texting.

34:52

That means you'll suddenly have an all new

34:54

emoji for any occasion, beyond the

34:56

catalog of options that Apple currently offers on the

34:58

iPhone and other devices. Another

35:01

fun improvement on related AI will be the revamped iPhone

35:03

home screen that will let users change the color of

35:05

app icons and put them wherever they want. For instance,

35:07

you can make all your social icons blue or finance

35:09

related ones green, and they won't need

35:12

to be placed in the standard grid that has existed since day

35:14

one in 2007. All right, so that's

35:16

a lot of stuff here. The generative AI

35:18

emojis, obviously they're not emojis, because emojis are

35:20

like Unicode things. There's a defined set of

35:22

them. You can't just make up new ones. I mean, you can't. If

35:25

you're the Unicode Consortium, but Apple can't. So what it's

35:27

actually saying is kind of like on Slack and Discord

35:30

and all these other things, where you can

35:32

sort of generate custom stickers, or I think they call

35:34

them custom emojis or custom reactions or whatever. The point

35:36

is, this is going to be a new graphic that

35:38

they'll generate on the fly for you that will be

35:40

sent as an image, because they can't just send it

35:42

as a Unicode code point, because they're

35:45

not dynamically adding things to Unicode. That's not

35:47

how this works. So they're kind of generating stickers

35:49

for you. And I guess they're

35:51

going to like, you know, they

35:53

have a bunch of like Mr. Potato Head style building parts,

35:55

they feed into the LLM, and then it

35:57

can do like sentiment analysis to figure out what kind

35:59

of of emoji we'd want to say like there's

36:02

a lot of emoji especially for things like face

36:04

reactions and stuff like that I

36:06

guess there's ones that aren't there like you ever try

36:08

to someone someone emoji and you're shocked that there's not

36:10

like a watermelon emoji or something I can't think of

36:12

one that doesn't exist sorry but this frequently I go

36:14

into emoji and I'm like oh surely there's an emoji

36:17

for this and like there's no ice cream cone like

36:19

again I don't know if that's real or not like

36:21

I'm shocked by what's not there so that's nice that

36:23

it will generate to it but it'll have to send

36:25

it as an image because it can't send it in

36:27

the other way and I do wonder

36:29

this is another place where like I feel

36:31

like Apple is taking a risk even with

36:33

a very very limited image generator that's trained

36:35

on their own source you know trained on all

36:37

their own sources for emojis that Apple makes

36:39

and can generate new ones

36:42

based on that and some other information there's

36:44

the potential to generate a little image that

36:46

maybe is embarrassing to Apple as a company

36:48

you know I'm saying so and there's

36:51

a potential to generate images that

36:53

just look ugly or not up

36:55

to Apple standards you know what I mean like this

36:58

is that's a weird thing to be touting

37:00

and the coloring icons that's you

37:02

know another I'm not saying keeps it from wanting to say a bridge

37:05

too far it's not a bridge too far but it is a bridge

37:08

like you're gonna take developers

37:10

icons and draw over them they've

37:12

done this before they used to add the shine

37:14

to your icon if you didn't opt out you

37:16

remember that like the gloss yeah yeah oh yeah

37:18

like we're gonna change the color of your icon

37:20

like I'm sure these companies that are like so

37:23

proud of their branding and change their icons or

37:25

whatever like that now the user will be

37:27

able to say you know what I want Instagram to be green and

37:30

the OS will be like sure I'll take that Instagram icon and

37:32

I'll make it green we can do that that's

37:35

weird I mean you can make little color-coded things on

37:37

your home screen I suppose but yeah

37:39

I can't imagine that's the whole story to that one

37:41

I Mean First of all, like a

37:43

lot of companies I think that have a problem with

37:45

just like trademark issues with like what if you if

37:48

you color my icon green First of all, it's not

37:50

my brand identity. Second of all, it might look like

37:52

my competitors or it might infringe someone else's trademark or

37:54

something like I could see. there's I could see like

37:56

big companies having a big problem with that but well

37:59

if users are new. Doing in a matter whether

38:01

the pressing for this a Mac Os back

38:03

in a day, someone else will look up

38:05

on version of Mac Os I'm cursed or

38:07

have been the operating system for max. I

38:09

do this when you added a label to

38:11

something akin to label the Mccoist entering classic

38:13

back on sugar levels in one version of

38:15

Mac O S or System Six or whatever

38:17

was called back in it which of something

38:19

of sigma to be Mrs and seven when

38:21

you apply to label a collie the icon.

38:24

By. To review fried green green label I would

38:27

like green tint the I can buy basically

38:29

making the ball pixels greener, the bike pixels

38:31

rad. I think it did not look good.

38:33

It was not a good. learned that that

38:35

didn't last very long time. Certainly Mac Os

38:37

Ten era. that has never been the case.

38:39

but. After the user

38:41

deciding to color I can like and that

38:43

was not on my list of things to

38:45

think of they were be doing our noses

38:47

like unrelated day. I've alleges I was eighteen

38:49

change the next week we'll probably have many

38:51

more produces area receive features but what an

38:53

odd what an odd thing to do but

38:55

us has is this stuff people clamoring for

38:57

this. I want to change the color of

38:59

my icons. I mean since I can't they

39:01

added that while ago where you can pick

39:03

from different like on the Bob Lutz but

39:05

even that that member how conservatively the added

39:07

that like it has to pop have a

39:09

dialogue and people spaces. Are you know what's

39:12

happening so apps can masquerade as the rap?

39:14

I literally just implemented that underpin the rewrite

39:16

a certificate. Not,

39:19

but I'm but I mean the I'm that

39:21

There's clearly there is a very large market.

39:23

there was I going to before I? there's

39:25

a very large my god. Therefore, these homes

39:27

ring customization apps that use all sorts of

39:29

tricks and hacks and frankly kind of bad

39:32

hacks to get people to be able to

39:34

customize icons for their apps and make like

39:36

that their aesthetic home screens that they want

39:38

that A huge demand, Huge market. I think

39:40

Apple. Wants to try to address some of the

39:42

man without having all these you tax. If this

39:45

is indeed what they're doing here, the problem is

39:47

unless you can have. Arbitrary.

39:49

Icons for apps like to cook like specify your

39:51

own image as the user as he came with

39:53

circuits right like if you it if you can

39:55

do that then all these apps a solid. These

39:58

could see if they saw it seems. The

40:00

people are just drawing it themselves like their

40:02

use an app that have liked cool theme.

40:04

Packs of here's here's like a boat, a

40:06

pipe layer, icons for all the popular apps

40:08

that you know you have like Instagram plus.

40:10

Also here's a bunch of like random ones

40:12

you can apply yourself to whatever else you

40:14

need them to be in. have like ten

40:16

generic ones. That's. A huge

40:18

market and so if we're just talking about like.

40:21

You. Can tinted seven different colors. that's not going

40:23

to get rid of us market at all and

40:25

you know some people use about that are be

40:28

great but I can imagine that being worth the

40:30

hassle in the troubled that they might get him

40:32

with other companies so. I

40:34

have a feeling this is like or embryos. Sometimes a

40:36

rumor comes out. And. Birds College.

40:38

Ah, and then you know, whenever it's actually

40:41

announced we see, oh, there's important detail x,

40:43

Y and Z. Hear that explain it better

40:45

and that now makes more sense. I think

40:47

we're missing those details right now on this

40:49

feature. Add to their sub the something else

40:51

here to explain this that we're not getting

40:53

it. I'm in a might just be like

40:56

you mention those icon packs are part of limitations of

40:58

customize your on scream using shock of through place that

41:00

by com for thought zipper cumbersome as the doing anything

41:02

else can we get behind the existing app and make

41:04

the shortcut to be out in assists to the half

41:06

on second if you get one of those icon packs

41:08

maybe it of the have an icon for the app

41:10

that you want to change. But. Of

41:12

the O S says okay, well you're making a little

41:15

region of green icons and you can use items from

41:17

the steam pack you got. Oh but there's no green

41:19

icon for this app. will just take the exit. The

41:21

actual icon for the happened to degree outfits and. So.

41:24

Maybe it's like us, you know I catch all

41:26

for if there's if busy bargain draw their own

41:28

custom icons. and if we can't

41:30

find one in a prebuilt icon pack now you

41:33

can just the least make the bygones mats i'm

41:35

i don't quite understand the idea that to conquer

41:37

everything but there's that and have the other part

41:39

of this is being able to the says not

41:41

but i'm the standard good reasons we just have

41:44

to be able to leave specks you know to

41:46

make space or of clear space or icons anymore

41:48

i'm assuming that's what they're talking about here is

41:50

like now you can arrange things on the home

41:53

screen and leave gaps if you want to i

41:55

still think doing anything on the home screen on

41:57

the i phone as credibly painful talk to listen

42:00

for years and years, I think it will continue to be

42:02

painful. I really wish there was a nice

42:04

interface where you could mess around with this

42:06

stuff and do it without destroying anything and

42:08

either commit or roll back all your changes

42:10

as opposed to the current system, which is

42:12

just pure chaos. And it's like the hardest

42:15

game on the entire operating system to try

42:17

to rearrange icons on your phone without screwing

42:19

everything up and just throwing your hands up

42:21

and saying forget it. I'm hiding

42:23

everything into the app library. Yeah,

42:25

I wonder. I feel like I

42:28

heard this from Jason maybe. I

42:30

heard it somewhere. Maybe it was Mike on upgrade. I forget

42:32

where I heard it, but somebody pointed out, well, maybe there

42:34

will be layers and that

42:37

makes me think of how it works. That was

42:39

upgrade, yeah. In SF symbols, you

42:41

can have different layers and you can color them in

42:43

different ways and whatnot. I think

42:46

that makes some amounts of sense, but

42:48

that's not going to fix the Instagrams of the world.

42:50

So I'm not sure. Or maybe it will be opt-in for

42:53

developers, but again, I don't see it. Oh, Vision OS has

42:55

this, right? Don't they have layers in their icons? Yes.

42:58

They actually require it with Vision OS. Yeah, but

43:00

if they made that a new, like, hey, for

43:02

iOS 18, if you provide a layered

43:05

icon, our new

43:07

coloration API will respect that, like in case you're

43:09

something like SF symbols where we can colorize parts

43:11

of it. We'll see. It

43:13

just seems to me like a weird place to be

43:15

spending resources. I think the arrangement part is a smart

43:18

feature. They should have that. People have wanted that for

43:20

a long time. People have been doing it with spacer

43:22

icons. It's annoying. The place where they

43:24

really should have spent resources is how hard it is to

43:26

rearrange the home screen. We'll see if

43:28

they do that. Yeah, that's the only way I

43:30

really wish I could still interact with my phone

43:32

via iTunes, which I know is now Finder, but

43:35

that was so much better, was doing home screen

43:37

rearrangement. And even that was terrible, but it was

43:39

better. Yeah, it was terrible, but it was way

43:41

better. All right, finishing up with

43:43

Mark German, a big part of the effort

43:45

is creating smart rehabs. The technology will be

43:48

able to provide users summaries of their missed

43:50

notifications and individual text messages, as

43:52

well as web pages, news articles, documents, notes, and

43:54

other forms of media. There's also no

43:56

Apple Design chatbot, at least not yet. That

43:58

means the company won't be... competing in the

44:00

highest profile area of AI. And the version

44:03

that Apple has been developing itself is simply

44:05

not up to snuff. The company's

44:07

held talks with both Google and open AI about

44:09

integrating their chat bots into iOS 18. Apple

44:12

ultimately sealed the deal with open AI and

44:14

their partnership will be a component of the

44:16

WWDC announcement. So a couple of quick thoughts

44:18

here. First of all, I think

44:20

we all are in a group chat that oftentimes

44:23

will just pop off at a time that you're not

44:25

paying attention. And so it is not uncommon, particularly I

44:27

find that I'm in a group chat with a couple

44:29

of guys that are constantly

44:32

looking at cars that they may or may not

44:34

ever buy. One of them in

44:36

particular is obsessed with finding a affordable 9-11

44:38

and those are mutually

44:41

exclusive terms. You cannot find a 9-11 that is

44:43

affordable. That's not a pile of garbage. And

44:45

so anyways, so I will often come back to my

44:47

phone or computer, what have you, with literally 40 or

44:50

50 messages, most of which I don't particularly care about

44:52

because 9-11 is not really my cup of tea. And

44:54

so if there was a way to summarize

44:56

that, I'm all in, sign me up. And

44:59

then additionally, I mean, here it is that Gurman

45:01

is saying open AI is the winner. And like we discussed,

45:04

I think this past week, that's

45:06

going to be a little dicey. So I wonder how

45:08

they spin this. Well, in typical Gurman fashion, there's

45:11

also no Apple design chat bottle. It's not yet Apple

45:13

designed. Do they mean Apple's in-house LLM that they've been

45:15

working on? Is that what he's saying? There won't be

45:17

any of it because I bet whatever, like they say

45:20

they did a deal with open AI, but he doesn't

45:22

say, okay, well, what does that deal, how is that

45:24

deal going to manifest? Is it going to manifest by

45:26

iOS 18 having a chat bot that is powered by

45:28

open AI? And will that not be

45:31

Apple designed or will it be open AI designed?

45:33

Will it be open AI branded? These are all,

45:35

I guess questions that he doesn't have the answers

45:37

to yet. But yeah, the news is they did

45:39

a deal with open AI. There's also, you

45:41

know, news this week that like Apple is still talking

45:43

to Google, obviously, you know, anyway,

45:46

so we'll see what, what,

45:48

uh, if the words open AI appear at WWC,

45:50

if they're going to announce this as a partnership,

45:52

they're just going to use them on the backend.

45:54

Like I said last week, maybe they've been building

45:57

features using the existing open AI API. And now

45:59

they just think deals and how they can

46:01

officially do it in a released OS instead of

46:03

just doing in dev builds but I'm just not

46:05

quite sure how this is going to manifest because

46:07

this paragraph begins making it seem like oh there's

46:09

gonna be no chatbot but they have licensed the

46:11

open AI chatbot. Well is there gonna be

46:13

a chatbot or not? Is there gonna be part of

46:15

iOS where you can start typing to

46:17

an LLM and getting answers in the chat

46:20

GPT style or will there be an open

46:22

AI app or will it be integrated into

46:24

Siri? Questions we do not have

46:26

answers for yet but that is the the capper

46:28

on all this stuff. A bunch of strange features

46:30

which I mean they all seem plausible to me

46:32

but this is not the list I would have come

46:35

up with about how to add AI sauce to iOS

46:37

18 and at the very bottom is you know

46:40

open AI something something something

46:42

open AI. Yep so

46:45

yeah we'll see what happens as I think John said a

46:47

moment ago typically you know in the couple of days leading

46:49

up to WDDC we oftentimes or

46:51

maybe the week before we hear some

46:53

last-minute leaks and so on I was

46:55

gonna say announcements but they're not announcements so we'll

46:57

see what happens but I am very

47:00

much looking forward to sitting next to you

47:02

fine gentleman and drinking this all in while

47:04

at Apple Park hopefully not getting a sunburn.

47:06

Well there was did you see there was

47:09

one kind of last-minute thing that came out

47:11

today also about Siri being used to script

47:13

apps but not yet and only

47:15

Apple apps? Yeah

47:18

it was basically the gist of this rumor I think it

47:20

was also Gurman was just in

47:22

a separate report that Apple

47:24

will also preview Siri

47:26

being able to like take more complex actions

47:29

within apps kind of what we've been talking

47:31

about you know we've been talking we've been

47:33

speculating about how wouldn't it be great if

47:36

you know apps could expose you know

47:39

kind of stock actions similar to the intent system

47:41

and you kind of describe what they are in

47:43

some kind of language and then have an LLM

47:45

based you know Siri interface be able to analyze

47:47

what the app is telling the system here is

47:50

how you can do it and here's the actions

47:52

to call to do it and

47:54

be able to offer to users the ability

47:56

to perform actions like that via voice commands

47:59

that are not necessarily phrasing exactly the right way,

48:01

that we're not set up in advance as shortcuts and

48:03

things like that. The rumor

48:05

that came out today from German is basically exactly

48:08

that is happening, but

48:10

not yet. That basically that

48:12

it is, that exactly that kind

48:14

of thing will be previewed. It won't

48:16

actually ship until a point release, probably

48:18

next, they said next year, so 2025, so

48:21

presumably in the first half of next year, you know, iOS 18.2 or 3 or

48:23

4 or 5 or whatever, and

48:27

that it would only work in Apple's apps to start. So

48:30

that tells me, not an API

48:32

yet. Or an API

48:34

that is released immediately, you know, released

48:36

at WWDC, but obviously nobody's going to

48:39

be able to implement it until

48:41

at the earliest of the fall, and it may not

48:43

be an instant adoption.

48:46

That's interesting. You know, I wonder if

48:49

maybe we will see, you know, the current

48:51

API to do all this is called AppIntents.

48:54

This is, they launched just a couple

48:56

of years ago, and it's like the

48:58

new Swift-based method of Intents, which I've

49:00

mentioned before, is way better

49:02

for developers and way simpler and way

49:04

easier to implement than the old system

49:06

of these weird Xcode files that had

49:09

custom configuration that would generate files, and

49:11

the generation process would always break and

49:13

cause weird bugs, and then you'd have

49:15

this extension that processed it and would

49:17

have to communicate to your app via

49:19

weird signaling mechanisms for extensions to apps. That

49:21

was a whole mess before, and they've redone

49:24

that system two or three times. I

49:26

wonder if, you know, sometimes when Apple is

49:29

kind of going in a direction before a

49:31

future update for something, sometimes

49:33

the APIs will launch for

49:36

it that don't come out and

49:38

explicitly say this is what this is for, but

49:40

kind of lay the groundwork for it. So

49:43

for instance, AutoLayout coming out, this is the

49:45

famous example, AutoLayout coming out before the iPhone

49:47

6 came out with its first, like, really

49:49

new screen size. You know, not counting the

49:51

iPhone 5, of course. That was just adding

49:53

a table row. Anyway,

49:55

so that's – I wonder if what

49:57

we're going to see this summer is maybe – maybe

50:00

they will add something to the AppIntense

50:02

API to give us more

50:05

ways to describe to the

50:07

system what these actions

50:09

do and maybe hide them from the

50:11

user but make them available

50:13

to this indexing system that presumably the new

50:15

Siri would have. Maybe we'll

50:18

see something like that. So I think

50:20

this, if for some reason they

50:22

demo this and don't tell us how to use

50:24

an API for our apps in the future, maybe

50:27

we'll be able to see hints in like

50:29

what's new in AppIntense this year. So

50:32

keep an eye on that. It sounds

50:34

more like they're considering their

50:36

own apps first. They're not quite sure what that API

50:38

should look like so they can experiment with their own

50:40

apps because if they change their minds, they'll just rewrite

50:43

their own apps to fit with it or whatever and

50:45

once they get it worked out, because it's very often,

50:47

especially in the old days of Mac OS X, they

50:49

would very often roll out entire new frameworks. There would

50:51

be private frameworks that only Apple's apps could use and

50:54

they would work out the kinks. Is this the right

50:56

API? Because they could change the API all they want because

50:59

no one else should be using this private framework and

51:01

once they finally got the framework where

51:03

it looking like they wanted it to,

51:06

then the next year they roll out essentially the same framework

51:08

under a new name but now it's a public framework. It

51:10

kind of sounds like what they're doing here, that they're not

51:12

at the point, like especially if they're kind of like rushed

51:14

to catch up, they're not at the point where they're ready

51:16

to even perhaps even to tell developers

51:19

in the size class type of way, do this thing and

51:22

then next year you'll be folded into the system because it

51:24

seems like they don't even know what that would be yet.

51:26

Like how to indicate, like they're working that out themselves but

51:28

they do want to get something out there and you can

51:30

demo it if they do it on their own apps. They

51:33

say coming in spring or whatever, not

51:35

an 18.0, whatever released down the road, we're

51:39

gonna have this and it's only gonna be our apps to

51:41

start basically because they haven't even figured out what they want

51:43

to ask developers to do and

51:45

related to that, maybe think of Windows

51:47

and the recall feature or whatever they

51:49

have. Isn't there something on iOS that's like activity,

51:52

something rather where you essentially make

51:54

a call let the OS know

51:56

what your app is doing. Yeah,

51:58

the user activity. framework.

52:01

Yeah, anyway, there's a similar, we had a very similar name

52:03

on Windows. It's some kind of like activity something something. I

52:05

think I have the exact same name. I

52:07

saw the story of reason behind like user

52:09

activity. That's our API. Yeah, and it's like,

52:11

it's basically that's how they did the thing

52:13

where like you can see like a PowerPoint

52:15

slide and recall and click on it and

52:18

go to that slide in PowerPoint. It's because

52:20

like during the, during the

52:22

screen recording like PowerPoint itself called the

52:24

user activity API to let it know

52:26

that it's viewing slide number 123 and

52:28

like recall recorded that because it's an

52:30

OS API and you're just telling the OS what's going

52:32

on and then when you click on the thing and

52:34

the recall thing, it looks back at that timestamp and

52:36

see during that time PowerPoint had this activity user activity

52:39

thing and it jumps back like same type of deal.

52:41

There's cooperation between APIs of

52:43

the app has to call that just says here's what's going on. Here's

52:47

my deal. And then another part of the system, you

52:49

know, cooperates with that sees the

52:52

stuff that those APIs called and connects the dots.

52:54

And that's exactly how we would think. I mean,

52:56

the thing you described for user intents, I

52:58

think that's probably better than when Apple is going to roll

53:00

out. I have dim hopes that they're going to do what

53:02

you describe what you described. I think is what they should

53:04

do, but it sounds much harder than what they

53:06

might do this, especially since they're

53:09

not even rolling out publicly. It seems like it might be a little

53:11

bit simpler than that and like not

53:13

involved the existing intent system at all,

53:15

which would kind of be a shame. But you know,

53:17

we'll see like that when they demo the feature, we'll

53:19

see the utility of it because they won't tell you

53:22

how it's implemented. But I do hope there is a

53:24

developer story for this. There's not, you know, people are

53:26

going to have a lot of questions of like, hey,

53:28

you demo this feature in the keynote, but it's only

53:30

for Apple things. Is that coming for third party soon

53:32

and they're going to not comment on

53:34

future products? Yeah, I

53:36

hope this is a direction

53:38

they're going. I was very I was

53:40

simultaneously very happy to see this rumor come today,

53:43

but also a little bit disappointed that we probably

53:45

won't be able to really use it until next

53:47

year or something like next year, because

53:50

this is an example of something that only

53:53

the platform owners can do and

53:56

it takes advantage of Apple strengths and

53:59

you know and kind of doesn't really rely on their weaknesses so

54:01

much, and gives them a

54:03

way to lead again. I

54:06

think so many of Apple's AI features are

54:09

going to be rightfully seen as playing catch-up,

54:11

as I was saying earlier. They

54:14

need to show us that they're not, especially

54:16

in this age of government, looking

54:19

into what they're doing and seeing does this look like

54:22

a monopoly in ways that

54:24

hurt consumers or hold things back. Apple

54:27

needs to show that they're leading and that they're not

54:29

just being complacent in the next big thing. On lots

54:31

of levels, I think they need to show governments that,

54:33

they need to show probably the stock market that, and

54:36

they definitely need to show their customers that. They

54:38

don't need us all to be looking over

54:40

in Google Land saying, hmm, they're

54:43

doing a lot of good things with AI when

54:45

they're on those Google models these days. Maybe I

54:47

should switch to Android for this coming fall or

54:49

whatever. They don't need anybody thinking

54:51

that way, so they need to show everything's

54:54

fine over here in Apple Land in this

54:56

new age of AI. Please stay here, it's

54:58

nice and comfortable, don't switch. We

55:01

have features too, you can be happy with our version.

55:03

And I

55:06

think there's enough reporting

55:08

now around this narrative that it's

55:10

probably true. They were probably caught

55:12

off guard with the rise of

55:14

new L and base techniques. They

55:16

did probably decide very

55:18

late, possibly like a year ago, to

55:21

hey, we should probably start investing

55:23

heavily in this. We kind of missed the boat on

55:25

this. It does seem like

55:27

there might have been some kind of, Bill

55:29

Gates, we missed the internet kind of realization there. But

55:33

we'll see how quickly they can actually

55:35

take action here. I think on

55:37

some of the big stuff that Google and

55:40

Microsoft are demonstrating, I think Apple's gonna be

55:42

behind for a while because that stuff that

55:44

takes years to develop, it takes priorities and

55:46

skills and connections that Apple doesn't have. Not

55:49

to say they can't have it, but they currently choose not

55:52

to. So we'll

55:54

see what happens. It took

55:56

a lot of problems and

55:58

rot to get where we

56:00

are today with Siri. Have

56:03

they fixed those problems? Have they cleaned

56:05

up that rot? Have they

56:07

even realized and admitted to themselves that they

56:09

needed to do that? We don't

56:11

know yet. Or they're just pouring AI sauce

56:13

over the rot. Right, I mean that is

56:15

very possibly what we're gonna see. I

56:17

have a feeling, I think what we're going to see, I

56:20

think it's going to disappoint a lot

56:22

of people for not being enough AI

56:24

sauce, but I

56:27

don't really care that much if we don't get all

56:29

of our AI wish list items in

56:31

two weeks. What I want to

56:33

see is have they turned the ship

56:35

around? Have they actually realized current

56:38

Siri is garbage? Have they actually

56:40

started moving in a

56:42

better direction with that? Or are

56:45

they kind of half asking this and

56:47

being complacent and thinking, what do you mean Siri

56:49

is the best voice system? We are the most

56:51

private, et cetera. If they start going

56:54

in that direction, kind of defending where they

56:56

already are and suggesting that we don't need

56:58

anything better than this, I will be

57:00

concerned. But what I'm

57:02

looking for is not for them to solve every single

57:04

problem in two weeks. That's unrealistic. I want to just

57:06

see are they going in the right

57:08

direction? And so if we get a preview of this

57:11

cool app AI interaction

57:13

based system that only

57:15

works in Apple's apps and only comes next spring or whatever,

57:18

if we get a preview of that and it's really

57:20

cool and it shows we will be able to do

57:22

that in our app, maybe in iOS 19 next year,

57:24

that would be great. I

57:26

will be happy with that. I'll be a

57:28

little upset that it didn't come sooner, but

57:30

hey, as long as good stuff is coming

57:32

soon, I can be patient and wait,

57:34

just like Apple is often a very patient company when it

57:36

comes to this stuff. I just want

57:38

to make sure that what we see are

57:40

signs that they're going in the

57:43

right direction, not just sitting back and hoping

57:46

that we're all going to start talking about

57:48

Division Pro again and stop looking about AI

57:50

features. When it comes

57:52

to these sort of inside Apple things,

57:54

especially when they cite specific executives, the

57:56

reliability is never great, right? Because this

57:59

is a game. telephone, people have grudges,

58:01

who knows what actually goes on. But

58:03

the two characterizations

58:06

of Apple's

58:08

decision to go all

58:11

in on AI this year exemplify

58:14

kind of the worst case scenario as far as

58:17

I'm concerned. I don't know if they're true, they

58:19

could be total BS. But here are the two

58:21

things. One, and we think we've mentioned this snarkily

58:23

on for shows that there's various rumors and supposed

58:27

tales from the inside saying, recently,

58:29

as a year ago or a year or a

58:32

half ago, important Apple executives who

58:34

are often named saw a

58:36

chat GPT and that made them realize

58:38

Siri sucks. And that's depressing to say,

58:40

how did you not know? That's what

58:43

it took for you to realize Siri sucks. I

58:47

hope that I really hope to God that's

58:49

not what it took. We hope

58:51

that's not true. We hope that it's just like

58:53

someone sort of extrapolating or exaggerating

58:57

a story they heard or whatever, we

58:59

would hope that Apple has known Siri as a

59:01

problem for a long time. And chat GPT was

59:03

just really like the straw that broke with Campbell's

59:05

back. But it's an unflattering story about Apple, which

59:07

is again, maybe widespread. And the

59:09

second unflattering story related to this is that

59:11

I think Federighi was named here.

59:13

It was like a mandate a year ago

59:16

that every team under Federighi has

59:18

to add some AI feature this

59:21

year. And that is an unflattering

59:23

story, but that's exactly what big stupid corporations do.

59:25

They're like, I don't know, but there's some big

59:27

thing. I don't care what you're doing team. There's

59:29

a mandate from on high that says every team

59:31

has to add some AI thing. I don't care

59:33

what it is, but it better say AI. That's

59:35

not the way to make a good product. And

59:38

this happens. This has happened frequently in

59:40

Apple's history. We know with more certainty that like,

59:43

for example, when a new a new

59:45

macOS feature back when macOS was

59:47

important, a new feature would come

59:49

out and Apple would mandate, you

59:52

know, all the teams have to now like add some feature

59:54

that takes advantage of spotlight or something. You know what I

59:57

mean? And teams would complain. They were like, we were in

59:59

the middle of. of making our application better, but then

1:00:01

a mandate came from on high that stop

1:00:03

what you're doing and make sure you carve out room

1:00:05

in this release to add like a spotlight powered feature

1:00:08

or something. I'm not getting a great example, but

1:00:10

like that is disruptive to teams

1:00:12

who are working. Maybe the spotlight, adding a

1:00:14

spotlight feature is not important for whatever application

1:00:16

that, you know, it's not important for the

1:00:18

terminal application or whatever. Not that there's more

1:00:20

than a third of a developer working on

1:00:22

that, but like telling teams, you just

1:00:24

have to add something with AI. That's

1:00:27

not vision. That's not leadership. That's just

1:00:29

like buzzword exists. We

1:00:32

want to be able to say something about that

1:00:34

buzzword, therefore a sacrifice part or all of your

1:00:36

schedule to figuring out what you

1:00:38

can do that uses AI. It's

1:00:41

the opposite of the Apple philosophy. Instead of figuring

1:00:43

out like what will help people

1:00:45

and then doing that, it's like we've

1:00:47

pre-decided that AI is good

1:00:49

in and of itself. You figure out something

1:00:52

to do with it so you can fulfill

1:00:54

that. And again, if it's an unflattering story,

1:00:56

it's totally unconfirmed, possibly manufactured

1:00:58

or made up by someone with a pessimistic view

1:01:00

of Apple or whatever, and we just really hope

1:01:02

it's not true. But those are the only two

1:01:04

things that I've heard from inside Apple, and I

1:01:06

just hope it's like, again, people with grudges or

1:01:08

people with a dim view of what's actually going

1:01:11

on, because you would hope that Apple is more

1:01:13

thoughtful. And honestly, I do think Apple has been

1:01:15

more thoughtful. I think the chat GPT thing, like I said,

1:01:17

is not that they didn't know that Siri was bad. It's just

1:01:19

this really pushed them over the edge to say, this is the

1:01:21

year we really have to do something, which, you know, fair enough,

1:01:24

because they've been trying to solve it for

1:01:26

years and failing, and it's a big complicated

1:01:28

organization, yada, yada. But I don't need

1:01:31

AI sauce to be poured all over

1:01:33

everything in all

1:01:36

their operating systems. I just need it used where it

1:01:38

can do the most good. And part of that is,

1:01:40

yeah, look at where other people have done things and

1:01:42

that people like, like erasing people from photos, they should

1:01:44

be doing that, right? But in

1:01:46

other places, you know, manufacturing emoji,

1:01:48

the messages team said, how are we gonna add

1:01:51

AI? You know, I guess summary is good, right?

1:01:53

Thumbs up on that. Someone

1:01:55

has the idea to manufacture emoji, maybe

1:01:57

have those people work on like, messages

1:02:00

in the iCloud sync improvements or something,

1:02:02

but... Or search improvements. Yeah,

1:02:04

there you go. Or archiving improvements. I

1:02:06

would set priorities differently, let's say, but

1:02:09

what can you do? What

1:02:12

I really want to see, too, as a developer, I mean

1:02:14

obviously I'm a little bit biased here, but I think it

1:02:17

would be better for the whole ecosystem, too. As

1:02:20

Apple adds AI features, they

1:02:22

have a massive advantage that

1:02:24

they have really good silicon

1:02:28

and their products to do local processing. And

1:02:31

historically, they create really good APIs.

1:02:34

We nitpick here and there, like, oh,

1:02:36

something is under-documented or whatever, or watch

1:02:38

connectivity sucks, but for the most part,

1:02:41

and it does, but for the most

1:02:43

part, Apple's APIs are world-class. They're

1:02:45

really good most of the time. They

1:02:48

are extremely powerful. It's

1:02:50

very, very difficult to find better APIs for a

1:02:52

lot of things than what Apple provides. They're

1:02:55

excellent, and they allow us as

1:02:57

developers to make really great apps that

1:02:59

do really powerful things pretty easily. And

1:03:02

one of the reasons that they're able to do that is

1:03:04

because their local device hardware

1:03:06

bar is just so high.

1:03:09

They have all sorts of great APIs for

1:03:11

things like ML processing, which is going to

1:03:14

be presumably re-banded AI processing a lot of

1:03:16

this stuff. They have all sorts

1:03:18

of high processing power things that you can

1:03:20

just drop in in your app with not

1:03:23

much effort, and all of a sudden now you have

1:03:25

this audio training model or whatever.

1:03:28

What I want is as Apple

1:03:31

digs into modern AI

1:03:33

stuff, give that to us. Make

1:03:36

that available for free with

1:03:38

no limits right on

1:03:40

the device. That's something that not every competitor

1:03:42

can offer because they don't have, first of

1:03:45

all, the very high bar. Not every Android

1:03:47

phone is going to be able to do

1:03:49

this kind of stuff because not every Android

1:03:51

phone has the right processor. Microsoft

1:04:00

is in a situation where Intel and AMD

1:04:02

are just now going to be getting out

1:04:04

chips that can do 40 tops or whatever,

1:04:07

and they have that one Snapdragon chip. But

1:04:09

every Apple Silicon thing Apple has sold in

1:04:11

the past several years has had a pretty

1:04:13

powerful neural processing unit in addition to all

1:04:15

sorts of other hardware units in the SOC

1:04:17

for other tasks. So they're way ahead of

1:04:20

the game on hardware. And

1:04:22

what I want to see from them is not

1:04:24

only make that hardware available to developers – I

1:04:27

mean most of it already is – give

1:04:30

us the models. Have the

1:04:32

models for many common tasks built

1:04:34

into iOS and let us just call

1:04:36

them when our apps are running and

1:04:39

just use them unlimited for

1:04:41

free. That is

1:04:43

how you make an entire ecosystem of

1:04:45

awesome apps that run on your platform

1:04:47

and keep people locked in. So

1:04:50

there is no reason for them not to offer

1:04:52

this. For instance, one of the things I want

1:04:55

is build in a really good

1:04:57

transcription model. There is

1:04:59

one. There has been a speech

1:05:01

recognition API in

1:05:04

iOS for something like six or seven years.

1:05:06

It's not that new. It's

1:05:09

not very good and it's pretty limited. And it's

1:05:11

like in certain contexts I think they would send

1:05:13

it to the cloud because it would limit you to like 60

1:05:15

seconds of transcription at a time and stuff like that. But

1:05:18

now we're past that. Now we can do all that on device.

1:05:20

I don't want to have to ship OpenAI

1:05:23

Whisper and compile it exactly right for Apple

1:05:25

Silicon and make sure it's optimized right using

1:05:27

Whisper CPP and keep updating it every time

1:05:29

it gets updated and have my app have

1:05:31

to download this gig and a half file

1:05:33

for the model. Build that kind

1:05:35

of stuff in and just let us call it and

1:05:38

let us use it like we can use any other hardware resource. So

1:05:41

obviously there would be some limits on like background execution

1:05:43

burning your CPU and stuff, but like build

1:05:45

that stuff in and let everyone use it

1:05:47

for free unlimited. That is how

1:05:49

you make the next generation of awesome apps

1:05:52

on your platform. That's what I'm looking for.

1:05:54

I'm looking for A, give me a sign

1:05:56

that you're taking this new technique

1:05:58

of computing seriously. in

1:06:01

what you're building in as features to your OS. Also,

1:06:05

don't just keep it for yourselves. Build

1:06:07

APIs for it and let us use it in

1:06:09

the most unlimited way you can possibly have with

1:06:12

the hardware that you have. Obviously, if things go to

1:06:14

the cloud, yeah, you can

1:06:16

limit those, keep them to yourself or meet with them or whatever.

1:06:19

Let us do as much as we can locally. That's

1:06:21

what your devices are awesome at. Let

1:06:24

us use it. There are so many people out there, there

1:06:26

are so many app developers out there that

1:06:29

go through the process of figuring out how to

1:06:31

train your own model or integrate someone else's model

1:06:33

using these weird Python tools that we don't know

1:06:35

how to use. Just build

1:06:37

it in, let us use it. That would be

1:06:39

amazing. They do have

1:06:41

a track record of falling on both sides

1:06:43

of this with different decisions. This could go

1:06:45

either way. I really hope they go the

1:06:47

direction of here's a bunch of

1:06:50

awesome built-in models for everyone to use. The

1:06:53

limitations you're going to run into is not going to probably

1:06:55

be the limitations on

1:06:57

running in the background using CPU, stuff like

1:06:59

that. The rumor is for

1:07:02

this and what Microsoft has actually announced at

1:07:04

Build is that both companies

1:07:06

are essentially going to give access to models

1:07:08

but through abstracted APIs where the caller doesn't

1:07:10

even know or possibly even

1:07:12

get to choose whether it runs locally or

1:07:14

in the cloud so Apple can change that

1:07:17

decision on the fly in the new versions.

1:07:19

Microsoft advertised that at Build, hey,

1:07:21

you just use these frameworks and you get these features.

1:07:23

Don't worry about where it runs. We'll run it locally if

1:07:26

that's best. We'll run it in the cloud if that's best.

1:07:28

We'll mix and match. You don't have to know. It's all

1:07:30

abstracted from you. I would imagine any APIs like this that

1:07:33

Apple offers are going to do the same thing and that is the rumor that Apple's going to

1:07:35

have APIs

1:07:37

that give access to quote-unquote AI where it's abstracted,

1:07:39

where you don't have to know or care if

1:07:41

it's running locally or in the cloud. They'll do

1:07:43

whatever is best. Pick your feature. You just mentioned

1:07:45

transcription. That's a perfect example of like, there's

1:07:48

a new transcription API and it's better than the

1:07:50

old one and sometimes it runs locally and sometimes

1:07:52

it's not something but your app doesn't have to

1:07:54

worry. You just call this new abstracted API and

1:07:56

we will do the best thing we can do

1:07:58

and as phones get more RAM and so on

1:08:00

and so forth. It'll just get better and better, but

1:08:03

you call the same API the whole time. The question

1:08:05

is, what are those specific APIs? Microsoft announced a bunch

1:08:07

of build. They have 40 AI models

1:08:09

inside Windows, right? And I think Apple will

1:08:11

ship a bunch of models with iOS and

1:08:14

with macOS, hopefully, if they remember that it

1:08:16

exists, with iPadOS. And they

1:08:18

will have frameworks fronting them, but for what? Are

1:08:20

they going to have transcription? Are they just going

1:08:22

to have summarization, translation? There's so many different things

1:08:25

that they can do. So I think that's the

1:08:27

question that WWDC is, I guarantee they're

1:08:29

going to do that. They're going to ship models, they're going to

1:08:31

provide access to the sensor framework, but for

1:08:33

doing what? And you hope it's

1:08:36

transcription for the purposes of Overcast, but there

1:08:38

are many things they could choose to do.

1:08:40

They could even take existing APIs and just

1:08:42

say, hey, by the way, it's the same

1:08:44

exact API as it was before, but now

1:08:46

behind the scenes, it's implemented totally differently and

1:08:48

we use LLMs for it. So

1:08:50

we'll see. Can I have one

1:08:52

small request also on that, though?

1:08:55

Last time I checked, the speech recognition

1:08:57

API required microphone access to be granted.

1:09:00

Can that please not be the case if you're not

1:09:02

using the microphone, for God's sake? You can file that.

1:09:05

I filed the thing where I need screen recording permission to find

1:09:07

out where the desktop picture is. I mean, that kind of makes

1:09:10

sense in one way, but another way, I don't want to record

1:09:12

people's screens. I mean, what

1:09:14

could you possibly want transcription for if not

1:09:16

the things that you're speaking right now, Marco?

1:09:18

I don't understand. There is no other spoken

1:09:20

content. There's no other source of audio than

1:09:22

the microphone, Marco. I don't know if you're

1:09:24

dealing with that. There is no other spoken

1:09:26

content. Where else would audio come from? Yeah,

1:09:28

that's the thing, too. I hope, as

1:09:30

Apple hopefully adds

1:09:33

or expands the APIs to access all this

1:09:35

cool new stuff that I hope they're giving

1:09:37

us, I hope they

1:09:39

do it in broad ways. The

1:09:43

way Apple does things, I

1:09:46

think in a way that

1:09:48

fails power users and developers,

1:09:50

is sometimes they'll have some

1:09:52

kind of lockdown area of the system. Like,

1:09:54

all right, we're going to add an extension

1:09:56

or an entitlement for only this one very

1:09:58

narrow use case to... Maybe try to

1:10:01

use this this little ability that we're gonna,

1:10:03

we're gonna be a modest little tiny hole

1:10:05

in our in our fence here and just

1:10:07

permits is very narrow use case and as

1:10:10

a result is usable by like zero to

1:10:12

one. Apps are out there ever were and

1:10:14

and sometimes they they make larger a p

1:10:16

eyes that are useful for every one and

1:10:19

and can be used much more broadly and

1:10:21

like. If you if you

1:10:23

only do a former approach and not the ladder.

1:10:26

You. Never find new great applications like

1:10:28

the The. The market never really

1:10:30

breaks out of the the general

1:10:33

capabilities that Apple was able to

1:10:35

consider as both existing and as

1:10:37

important enough to warrant them gracing

1:10:39

us with a doorway to let

1:10:41

us do it. Whereas if you

1:10:43

make things more general purpose more

1:10:45

broad has fewer restrictions, let people

1:10:48

just kind of like. Use.

1:10:50

More general purpose tools. Ah.

1:10:53

You get apps that that Apple not only

1:10:55

in a benefits from in the sense that.

1:10:58

More. People want to use their platforms but it

1:11:00

also gives them ideas and what to sherlock for

1:11:02

the next releases. You get stuff like Quicksilver on

1:11:04

the Mac because of a dropbox on the Mac

1:11:06

elect these are all apps that like they they

1:11:09

cut to be manage of like. System.

1:11:11

Background stuff You get things like switch

1:11:13

as if you know you get upset

1:11:15

that at the city usually can only

1:11:18

exist on the Mac and are not

1:11:20

possible on i owe us see if

1:11:22

Eat but you need to like enabled

1:11:24

his Power User utilities and power user

1:11:27

features Eats enable them to exist in

1:11:29

ways that you as Apple didn't foresee

1:11:31

as. Like that with the they

1:11:33

didn't have to make and eighty I to

1:11:36

let dropbox in a badge. the file things

1:11:38

at first they get hacked. it's work like

1:11:40

through that hacks they already had and and

1:11:42

the waiter mid may be I to make

1:11:44

it better that I think is still kind

1:11:46

of in a trance synthetic of so arse.

1:11:48

yeah exactly or but. I hope

1:11:50

as their as their diving into all

1:11:52

this new ai stuff. I.

1:11:54

hope not only are i were saying earlier

1:11:56

that the that they have a p eyes

1:11:58

for us to use a functionized built-in

1:12:01

system model that we don't have to make

1:12:03

and train and ship ourselves. But

1:12:05

also, I really hope that they

1:12:07

allow access to them in broad

1:12:09

ways. Now I'm not talking about

1:12:11

like don't let me just burn everyone's

1:12:13

battery down like crazy, which by the way, there's

1:12:16

already an API for background

1:12:18

task management where you can

1:12:20

specify, I believe it's called a

1:12:22

background processing task. This is

1:12:24

a type that you can say that you can tell the

1:12:26

system when you have

1:12:28

a chance, when you are plugged

1:12:30

in and charging and maybe even

1:12:32

on Wi-Fi, call me in

1:12:34

the background, wake me up and let me do a

1:12:36

task with no throttling. And

1:12:39

it will do that. And if the person unplugged their

1:12:41

phone, it will terminate it and whatever. But there

1:12:44

are ways, there are opportunities for apps

1:12:46

to do background processing on iOS that

1:12:49

don't burn your battery down, but still allow them

1:12:51

to use the power of the hardware in the

1:12:54

background if they want to. You can try not

1:12:56

to wait until overnight to do it, but that

1:12:58

option is there. That API is already there. So

1:13:01

give us broad access. Let us do

1:13:04

whatever we want within the

1:13:06

existing battery and

1:13:08

power limitations you already enforce,

1:13:10

slash grant us. Let

1:13:12

us do whatever we want with this stuff. That

1:13:14

will enable great apps to be made. It

1:13:17

will enable everyone else on your platforms

1:13:20

to make the apps and features that you

1:13:22

won't make. And that

1:13:24

will both enhance your platforms for everybody

1:13:27

and whatever does take off, either

1:13:30

it will be something that's a little bit tricky,

1:13:33

like image generation that you won't then have to make

1:13:35

and take the liability of, and

1:13:37

or it will be successful and

1:13:40

you'll be able to then copy it for your next stuff and share it

1:13:42

with you in the next release. So

1:13:44

that's a win-win for Apple, and the

1:13:46

more of these features rely on that

1:13:48

local hardware and are

1:13:50

not based on cloud stuff, that

1:13:52

benefits their privacy strategy, that benefits

1:13:54

their hardware strategy, and that keeps

1:13:56

people locked into iPhones. every

1:14:00

reason for Apple to do it this way. Give

1:14:02

us a bunch of models and open them up as much as

1:14:04

possible to our apps to use. This

1:14:06

is a philosophical change that Apple has

1:14:09

not been on board with in recent decades that

1:14:11

we've all complained about multiple times is the idea

1:14:13

that good ideas can come from places other than

1:14:15

Apple. And Apple will say that they believe that

1:14:17

and support it, but not to the degree where

1:14:20

they will do what you just asked. Which is

1:14:22

open up APIs to allow third parties to do

1:14:24

things that historically in the past several decades Apple

1:14:26

has said not only Apple can even attempt that.

1:14:28

Like window management that I always complain about the

1:14:30

third parties could not have implemented stage manager. We

1:14:32

had to wait for Apple to think it had

1:14:34

an idea about window management and then it implemented

1:14:36

stage manager. And if you don't like it, wait

1:14:39

another five years for Apple to have another idea.

1:14:41

But no, we're not going to provide you APIs

1:14:43

to do that because no good ideas can come

1:14:45

from third parties. They're too dangerous. You can't have

1:14:47

this power, so on and so forth. The older

1:14:49

Apple, whether intentionally or not, essentially gave

1:14:51

enough free reign to APIs for tons of

1:14:53

good ideas to come out of the third

1:14:55

party developer community, which Apple then incorporated into

1:14:57

its operating system, right? And that was a

1:14:59

system that worked. And we didn't call it

1:15:01

Sherlocking back then. We just, it was just

1:15:03

the cycle. You know, Sherlock was an egregious

1:15:05

thing where they copied a particular app very

1:15:07

specifically in ways that were obvious that they were

1:15:09

copying it or whatever. But like giving

1:15:12

APIs where third parties can have ideas

1:15:14

and implement them that Apple can learn

1:15:16

from was how the first 20

1:15:18

to 30 years of the Mac, or

1:15:20

maybe I'm getting years wrong, or the first the

1:15:23

early part of the Mac. That was how the

1:15:25

platform evolved. To give a modern

1:15:27

example, how did Twitter evolve in the early

1:15:29

days? By having good ideas

1:15:31

happen in the third party world. Good

1:15:34

ideas like the concept of a retweet and

1:15:36

using at to mention somebody and the word

1:15:38

tweet all came from third parties. Current

1:15:41

Apple thinks that there are certain classes of

1:15:43

ideas that can only come from

1:15:45

Apple. And so they closed themselves

1:15:47

off to lots of good things. So

1:15:49

like you third party developers shouldn't have

1:15:51

an API powerful enough to do this.

1:15:54

When we, Apple, eventually five years from

1:15:56

now, come up with a good idea

1:15:58

for doing something with this. implemented

1:16:00

but you can't have those API so

1:16:02

audio hijack on iPad Apple

1:16:05

will get around to it eventually but it's not like we're

1:16:07

gonna let a third party do that stage manager oh third

1:16:09

party do you have ideas about window management sorry that's too

1:16:11

dangerous for you we can't give you that kind of control

1:16:13

it limits us too much it limits Apple too much because

1:16:15

what if we have an idea we locked

1:16:18

ourselves into a bunch of API's that are being used

1:16:20

by third party applications we don't want to do that

1:16:22

it's safer to just you know and the thing is

1:16:24

when this philosophy first rolled out with sort of the

1:16:26

Mac OS 10 error it was it's

1:16:29

like a pendulum right it was a relief because

1:16:31

Apple had swung too far in the other direction

1:16:33

where they would give API's third parties get locked

1:16:35

into them because popular third party products would use

1:16:38

them and Apple would be constrained into what it

1:16:40

could do and so they swung hard in the

1:16:42

other direction said you know what we're

1:16:44

not giving anything like that to third parties

1:16:47

we're gonna keep it all real close to

1:16:49

the vest be very very conservative close off

1:16:51

innovation of third party world to give Apple

1:16:53

itself more flexibility to innovate and

1:16:56

evolve and the pendulum had swung too far

1:16:58

in the other direction but now it's so

1:17:00

far on the opposite side of things that

1:17:02

we are out here suffering from you know

1:17:05

the lack of Apple on third parties to innovate

1:17:07

and AI is the newest front of that because

1:17:09

we see this explosion of you know we described

1:17:12

as throwing spaghetti against the wall but another

1:17:14

way to describe it is exuberant innovation

1:17:16

enthusiastic innovation most that stuff's not gonna work

1:17:18

out but some of it is the more

1:17:20

you give third parties flexibility to do that

1:17:22

the better ideas you'll get so I just

1:17:25

you know there's either

1:17:27

extreme is wrong and we are currently in the Apple

1:17:29

world at one extreme so I hope it starts swinging

1:17:31

back the other direction we

1:17:34

are brought to you this episode by delete me you

1:17:37

ever wonder how much your personal info

1:17:39

is out there for anyone to really

1:17:41

easily find on the internet it's way

1:17:43

more than you think try it search

1:17:46

for yourself and you'll see like how

1:17:48

quickly and easily anybody can find out

1:17:50

all your information name address phone numbers

1:17:52

your social security number sometimes even

1:17:55

all this and all this too about your family

1:17:57

members which is extra concerning so all this is

1:17:59

being filed by data brokers, openly traded

1:18:01

and sold online. Anybody can really easily

1:18:03

get these details from you. So

1:18:06

protect your privacy with delete

1:18:08

me. I actually found delete

1:18:10

me myself when I was looking to reduce some of

1:18:13

my information out there a few years ago before they

1:18:15

were a sponsor. And I did some quick research, found

1:18:17

them and I chose to use them myself because

1:18:20

there's a lot out there. As

1:18:22

I said, it's super easy to find and you can't

1:18:24

get everything, every single possible detail but you can get

1:18:26

a lot of it. And that's why services like delete

1:18:28

me are great and I chose them and I've been

1:18:30

very happy with them. So delete me, find

1:18:32

and remove any personal information that you don't

1:18:34

want online. You tell them what you want

1:18:36

gone, they make it gone. And

1:18:39

they also, they're an ongoing subscription service. So

1:18:41

they not only remove your info from the

1:18:43

largest people search databases on the web but

1:18:45

they also then keep monitoring it

1:18:47

and they keep updating and they keep making sure

1:18:49

that your information is still gone. So you sign

1:18:51

up and provide them with exactly what you want

1:18:54

deleted and they take it from there.

1:18:56

So they have their personalized privacy reports showing what they found,

1:18:58

where they found it, what they were able to remove. And

1:19:01

so it does all the hard work for you

1:19:03

of wiping your and your family's personal info off

1:19:06

the web. So check that today,

1:19:08

take control of your data and keep your

1:19:10

private life private by signing up for delete

1:19:12

me. Now at a special

1:19:15

discount for our listeners, get 20%

1:19:17

off your delete me plan when

1:19:19

you go to joindeleteme.com/ATP and use

1:19:22

form a code ATP at checkout.

1:19:24

The only way to get 20%

1:19:26

off, go to joindeleteme.com/ATP, code ATP

1:19:29

at checkout. That's joindeleteme.com/ATP, code ATP.

1:19:31

Thank you to delete me for

1:19:33

sponsoring our show. All

1:19:39

right, so I see here in the show

1:19:41

notes, LLM check-in. And so I

1:19:44

guess, John, you'd like to take us to the world

1:19:46

to give us a little situation report. Yeah,

1:19:48

LLMs. They're weird. There's

1:19:50

been a lot of stuff online trying

1:19:53

to explain to people how they work.

1:19:56

They're a type of technology where when

1:19:58

people look at them, they're guess about

1:20:00

what they are and how they work is very often

1:20:02

wrong because part of their whole deal is the reason

1:20:04

people are so fascinated by them is because they can

1:20:06

essentially fool you. This

1:20:10

is related to a recent

1:20:12

story. I think we panted out in past shows, but

1:20:14

you probably have seen it in the news, where Google

1:20:16

is replacing their thing that gives you the summary of

1:20:18

the top of their search results. They're replacing that with

1:20:21

an AI-powered one. They poured AI sauce on that. It's

1:20:23

been doing some silly things like suggesting that you

1:20:26

eat one rock per day and suggesting that to

1:20:28

keep the cheese from sliding off your pizza you

1:20:30

should add glue. That's been pulling

1:20:32

things from the onion and presenting them

1:20:34

as straightforward things and not parody. Someone

1:20:37

writes a snarky post on Reddit and Google

1:20:39

pulls it out and spits it back to the user. Ha

1:20:42

ha, isn't that funny? AI is so dumb. Then

1:20:45

Google has been trying to fix these

1:20:47

manually and work on it. I

1:20:50

bet people who see this are like, oh, AI is dumb, but

1:20:52

it will get better or whatever. Another

1:20:55

thought that people often have and express is, why

1:20:58

doesn't Google just tell its AI

1:21:00

model that you shouldn't eat glue? The

1:21:03

glue doesn't go on pizza. Why doesn't it

1:21:05

just tell it people shouldn't eat rocks? Like

1:21:08

zero is the correct amount of rocks to eat

1:21:10

per day. Why doesn't

1:21:12

it just do that? I know that, okay, well,

1:21:14

it's new, whatever, it'll get better. Because

1:21:18

of the mental model people have about

1:21:20

LLMs, it seems so silly that they would

1:21:22

give an answer so dumb and

1:21:24

it's like, well, why doesn't Google

1:21:26

just correct it? These aren't nuanced corrections.

1:21:29

Don't eat rocks. It

1:21:32

seems so simple. I think part

1:21:35

of it is because that's not how

1:21:37

LLMs work. I'll link again to these three blue,

1:21:39

one brown neural network things or whatever. I know

1:21:41

it's complicated or whatever, but if you just keep

1:21:43

watching these videos over and over again as I

1:21:45

have, don't you start to sink in

1:21:47

what they are. The analogy you gave on erectus

1:21:49

ages ago is a pachinko machine where you drop

1:21:51

a ball at the top of this big, giant

1:21:53

grid of pins and it bounces off pins and

1:21:55

eventually lands at a particular place. The

1:21:58

thing I was trying to express by that is... But the way

1:22:00

LMS work is they take an input, which

1:22:02

is whatever it may be, but in this case, let's just

1:22:04

say some text. And it bounces

1:22:07

around inside doing a bunch of math, and

1:22:09

some text comes out the other end. And you could do it

1:22:11

with images and other things or whatever. But it's just a bunch

1:22:13

of matrix math, and

1:22:15

a thing comes out the other end. And it's

1:22:17

stateless. The thing you're putting

1:22:19

it into, the machine, the LLM, has been

1:22:22

trained. And the training is to essentially set

1:22:25

a bunch of things called weights, which are a bunch

1:22:27

of numbers inside the model. It's just a huge number

1:22:29

of these. And the magic

1:22:31

is, well, where did those numbers come from? They came

1:22:33

from training. And you can look at the video and

1:22:35

see how this works. But it's like, we're going to

1:22:37

make this giant grid of pins and

1:22:40

just grind on it with a huge amount of

1:22:42

computing resources feeding the entire incident in. And

1:22:44

the result is just tons and tons

1:22:46

of numbers, tons

1:22:49

and tons of pins in the pachinko machine. And

1:22:51

after we've done that, we drop things on the top

1:22:53

of the machine. It bounces around. Stuff comes out. And

1:22:55

the quality of stuff comes out depends on what those

1:22:57

pins were, what those numbers were, what the weights were.

1:23:00

And the reason people get hung up on the gluing

1:23:02

pizza and eating rocks is just tell

1:23:05

it people shouldn't eat rocks. Is that hard

1:23:07

to teach it? That's like, well, you don't

1:23:09

understand. We do this training. And it's very

1:23:11

expensive. And it takes a long time. And

1:23:13

we produce this giant model. But once

1:23:15

you've got the model, you can

1:23:17

retrain and tweak and adjust or whatever.

1:23:20

But you can't just type to it,

1:23:22

oh, and by the way, people shouldn't

1:23:24

eat rocks. Because you drop it into

1:23:26

the top of the machine, and it would bounce out.

1:23:28

And so there's nothing that people might not understand from

1:23:30

first glance of how these things work is

1:23:33

when you're having a quote unquote conversation with chat CPT,

1:23:35

and you ask it a question, like how many rocks

1:23:37

should I eat per day? And it gives you an

1:23:39

answer. And so it gives you a bad answer. It

1:23:42

says you should eat one rock per day. You type back

1:23:44

to it. Actually, people shouldn't eat

1:23:46

rocks. If you eat rocks, it's really bad for you.

1:23:48

Bad things can happen to you. And

1:23:51

then it will reply to you, oh, I'm sorry. I made that

1:23:53

mistake. Blah, blah, blah, blah. What you don't realize is every time

1:23:55

you type something to chat

1:23:58

CPT, when you type a message, people don't know. shouldn't

1:24:00

eat rocks. That's just not what gets

1:24:02

sent to ChatTPT. What gets sent is the entire

1:24:04

conversation up to that point. Your first

1:24:06

question, ChatTPT's answer, your second question, every time

1:24:08

you type something new, the whole conversation gets

1:24:11

sent through. Because remember, ChatTPT doesn't have any,

1:24:13

I know there's a memory feature or whatever,

1:24:15

but the LOM itself is just a big

1:24:17

bucket of pins in a pachinko machine. You

1:24:19

drop something in and it comes out. So

1:24:22

your quote unquote conversation, all you're doing is

1:24:24

making the thing you're sending it bigger and

1:24:26

bigger each time. There's no back and forth.

1:24:28

There is, here's the entire conversation plus my

1:24:30

new thing, including what it answered before. Because

1:24:33

that influences what the next thing is

1:24:35

going to come out. In fact, when it's processing things, what it

1:24:37

does is it processes the entire input and picks the next word,

1:24:40

and then it throws everything back in and picks the next word,

1:24:42

and then picks the next word over and over again until it

1:24:44

gets all the words in the answer for your thing. That

1:24:47

is essentially stateless.

1:24:50

There is no thing that you can say people shouldn't

1:24:52

eat rocks. All you can do is put that in

1:24:54

your conversation, and then when the whole conversation goes back

1:24:56

in the top of the machine again, yeah, it's in

1:24:58

there and it influences the output according to the magic

1:25:00

of all the weights and everything. That

1:25:05

whole thing of like, okay, but how big is that?

1:25:07

What is that? I can send the whole conversation back

1:25:09

in. The length of stuff

1:25:11

you can stick into an LOM is called the

1:25:13

context window. If I have a conversation and it

1:25:15

goes wrong for thousands and thousands of words, at

1:25:17

a certain point, there's a limit in the size

1:25:20

of the input string. The input string to

1:25:22

LOMs used to be very small and I was

1:25:24

getting bigger and bigger. This is related to an

1:25:26

announcement Google had at their I.O. conference where

1:25:29

the CEO said, today we

1:25:31

are expanding the context window to

1:25:33

two million tokens. That's a

1:25:35

big number because most of the context of the context would

1:25:37

just start off at like 32K or whatever. Google is saying

1:25:40

we're expanding to two million tokens. We're making it available for

1:25:42

developers in private preview. It's amazing to look back and see

1:25:44

just how much progress we made in a few months. You

1:25:47

can see how the context window would be important because

1:25:49

if you wanted to, quote unquote, teach it that people

1:25:52

shouldn't eat rocks, what you'd want is to be able

1:25:54

to stay in a conversation. You

1:25:56

just told me I should eat one rock per day, but that's

1:25:58

really bad. Humans shouldn't eat rocks. And you

1:26:00

want to, quote unquote, remember that and,

1:26:03

quote unquote, learn that. But

1:26:05

the way I almost work, the only way I can remember

1:26:07

or learn that is either A, you train on new data

1:26:09

that influences the weights in the model, which is something you

1:26:11

can't do when you're typing to chat GPT. You're not changing

1:26:14

the weights in the model. You're just

1:26:16

sending things through an existing machine. Only OpenAI can

1:26:18

change those weights, right, by making a new model

1:26:20

or whatever, or modifying the existing one, right? Or

1:26:23

B, have that phrase be

1:26:25

part of the context window that

1:26:29

your thing is included in. Like the example I used to give with

1:26:31

Merlin was like, if you say, you

1:26:33

know, I have one brother and two, you

1:26:35

know, how many siblings do I have? And it says, I don't

1:26:37

know. And you tell it, I have one brother and two sisters.

1:26:39

And then you say, again, how many siblings do I have? And

1:26:41

it says, you have one brother and two sisters. Like, wow, it

1:26:43

learned it. No, because the input you put in was, how many

1:26:45

siblings do I have? I don't know. I have one

1:26:48

brother and two sisters. How many siblings do I have? That

1:26:50

was the input. The answer is in the input. So

1:26:52

you shouldn't be shocked when it says, you have one

1:26:54

brother. Hey, it learned that I have one brother. No,

1:26:56

the input contains the answer. It was part of the

1:26:59

context window. You didn't change any of the weights in

1:27:01

the model. You literally just gave it the answer, right?

1:27:03

So here's the final thing. At On Stage at Google

1:27:05

I.O., Sundar Pichai, Google CEO said, talking

1:27:08

about the two million context

1:27:10

window, this represents the next step in

1:27:12

our journey towards the ultimate

1:27:14

goal of infinite

1:27:16

context. And

1:27:19

this is the first time I've seen someone outline a

1:27:22

vision for how LLMs could

1:27:24

actually be taught things. Because

1:27:27

if the context window is infinite, what

1:27:29

that would mean is that you

1:27:31

could talk to an LLM over the course of

1:27:34

months, days, months, and years. Everything

1:27:37

you ever said to it would essentially be

1:27:39

sent as input in

1:27:42

its entirety, plus the new thing that you said

1:27:44

every time. So if you said six months

1:27:46

ago people shouldn't eat rocks, every time you

1:27:48

ask any question, part of the input

1:27:50

would be your question, plus everything you've ever said to

1:27:53

it, including the line that people shouldn't eat rocks. And

1:27:55

so when it answers you, it will, quote

1:27:57

unquote, remember that people shouldn't eat rocks. because

1:28:00

that was part of its input. I

1:28:04

don't think this is a way to make a reasonable kind

1:28:06

of mind that can learn things, but it

1:28:08

is the first vision I've heard of anyone

1:28:10

outlining how LLMs are not going to be

1:28:12

dumb. Because no matter how well you

1:28:14

train them on the big stew of stuff you're putting into them,

1:28:17

you can't teach them anything. They can't learn

1:28:19

through conversing with you. They can only learn

1:28:21

by being trained on new data and having

1:28:24

a new version of the model come out

1:28:26

or whatever. So we want them to work

1:28:28

like people where you can say, oh, silly

1:28:31

toddler, rocks are bad for you, don't

1:28:33

eat them and have it learn

1:28:35

that. If you have an infinite context

1:28:38

window, I mean, anytime you ask it anything, the

1:28:41

entire history of everything you've

1:28:43

ever said goes as input somehow. I

1:28:46

don't know if he said this is just kind of like a vision,

1:28:49

a conceptual vision or a practical

1:28:53

example of like, that's how we're going to do it. We're going

1:28:55

to have a stateless box of numbers

1:28:57

that we throw your input into, but we're

1:28:59

going to retain your input forever and ever

1:29:01

and ever. Anytime

1:29:03

you ask the stateless box of numbers anything, everything you've

1:29:05

ever said to it goes as input plus the new

1:29:07

thing that you said so that you can,

1:29:09

quote unquote, teach it things. The

1:29:11

fun part of like teachable AI or like

1:29:13

an actual sort of thing that you could

1:29:15

converse to is that you can just teach

1:29:17

it BS things. If you talk to an LLM and tell

1:29:19

it, actually, you should eat rocks. In fact, you should eat

1:29:22

really spiky rocks all the time. And

1:29:24

it says, OK, that's great. I'll remember that if you ever

1:29:26

asked me about what you should eat again. And six months

1:29:28

from now, you said, give me this recipe and it

1:29:30

includes rocks. It'll be like, well, as big as the input

1:29:32

was, you should eat spiky rocks, six months worth of text.

1:29:35

Give me a recipe for pizza and it includes rocks. You

1:29:38

could teach it to be less useful, like in the same

1:29:40

way of raising children. If you teach them bad things, they

1:29:42

will learn bad things. We

1:29:45

don't all have our own copy of chat GPT

1:29:47

or LMs or whatever. There's just one big stateless

1:29:49

blob that's instanced all over that our input is

1:29:52

going through. But if we

1:29:54

have our own infinite context window, then we have

1:29:56

we are essentially building our own

1:29:58

sort of knowledge. base within this LLM

1:30:02

through the barbaric brute force method

1:30:04

of sending it every piece of information we've ever

1:30:06

sent to it before with every new piece of

1:30:08

information. I just thought this was interesting because I've

1:30:10

been super down on LLMs because I just don't

1:30:13

see how they can ever be anything useful that

1:30:15

can be taught like the difference between fact and

1:30:17

fiction. Important

1:30:20

things that would make the thing more useful

1:30:22

are not possible because of the way LLMs

1:30:24

work. But if you give me an infinite

1:30:26

context window, at least then I can

1:30:29

over time try to mold my

1:30:32

little conversation with the LLM

1:30:35

towards something and maybe only have to correct it

1:30:37

once or twice when it makes mistakes so that

1:30:39

it will get better over time underneath

1:30:42

my own control. That

1:30:44

said, he is the CEO. I don't know if

1:30:46

he knows about this on a technical matter, an

1:30:48

infinite contact window sounds ridiculous to me. But I

1:30:50

just wanted to bring this up because it really

1:30:52

annoys me that LLMs essentially do not learn through

1:30:54

conversing with you even though everyone thinks they do

1:30:56

and annoys me when people get fake debt. Good

1:30:59

talk. I hear you. I

1:31:03

think watching these 3Blue1Brown videos, there's others that

1:31:05

are good too, but these are very, very

1:31:08

good. Even though they're

1:31:10

not fast in the sense that they're

1:31:12

not rushed, but they're fast in the

1:31:14

sense that a lot of ground is

1:31:16

covered very quickly, and I

1:31:18

think you're right that watching them like 2 or 3

1:31:21

times is what you probably need in order to get

1:31:23

this really understood. But

1:31:26

I strongly suggest if you

1:31:28

are even vaguely inclined for these

1:31:31

sorts of things, and if you're listening to the show,

1:31:33

I presume you are, it's worth checking it out. Just

1:31:35

understand kind of the broad strokes as to how these

1:31:37

things work. Thank you to

1:31:39

our sponsors this week. Fast mail and

1:31:41

delete me. Thanks to our members who

1:31:43

support us directly. You can join us at ATP.FM. Join.

1:31:46

One of the biggest member perks now

1:31:49

is we do this overtime segment where

1:31:51

every week, every episode, we have a

1:31:53

bonus topic that just didn't fit. We

1:31:56

couldn't get to it in the main episode. This

1:31:58

week, overtime is about the end. TikTok

1:32:00

ban in the US that

1:32:02

is currently working its way through the system.

1:32:05

We're going to be talking about that. You

1:32:07

can listen by joining at atb.fm slash join,

1:32:09

and we will talk to you next week.

1:32:30

And it was accidento. Oh,

1:32:32

it was accidento. You can

1:32:34

find the show notes at a-t-c-s-n-f. And

1:32:41

if your is your mathadon, you can

1:32:44

follow us at C-A-S-E-Y-L-I-S-S,

1:32:50

so that's

1:32:52

K-C-L-I-S-M-A-R-C-O-A-R-M-E-N-T, Marko

1:32:54

Armand. That's

1:32:58

I-R-A-C-U-S-S-C, R-I-S-A-T,

1:33:00

R-Q-S-R-E, That's actually

1:33:02

not a tool, that is a

1:33:06

tool.

1:33:09

E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E-E

1:33:12

NO HOOOONN. All

1:33:15

right, so in the section

1:33:17

of our internal show notes, where we

1:33:19

put after show ideas, and

1:33:26

it reads a couple things. One of

1:33:28

them is Marco's desk setup. And

1:33:31

you had privately teased something to the two

1:33:33

of us about this. And

1:33:35

I have been dying for this

1:33:37

to happen. And I feel like this

1:33:39

is the moment, we almost need an ATP overtime for

1:33:41

after show stuff. We finally got there.

1:33:44

Don't ruin my high, Marco. Tell me

1:33:46

what's going on. In

1:33:48

the new house, part of the condition of getting

1:33:51

this house was there was one room upstairs that

1:33:54

overlooks a view that you

1:33:56

can see. Long Island is full of canals,

1:33:58

because Long Island, everybody has both. I

1:34:00

don't have a boat, I don't want a boat, but there's canals everywhere. And

1:34:03

from one of these rooms upstairs, we can

1:34:05

see one of these canals. So

1:34:07

as a result of there being a canal, you know, at the

1:34:09

window, I can see the ducks

1:34:11

and the birds and other delightful things floating by

1:34:13

in the canal and hanging out. I can see

1:34:16

the ducks sit on the neighbor's lawn all folded

1:34:18

up in the rain. I can

1:34:20

see the rabbits jumping across my lawn. It's

1:34:22

a wonderful view. The only way

1:34:24

to enjoy this view is by

1:34:26

sitting basically in the middle of the room. So

1:34:30

when I was laying out my office

1:34:32

setup, the only way – like

1:34:34

the big deal breaker for my office

1:34:37

layout was the desk has to be

1:34:40

sticking out from one of the walls in the

1:34:42

middle of the room. It's

1:34:44

like a big capital E where the

1:34:46

walls are the outer walls of the E and the

1:34:48

one middle thing that sticks out of the E, that's

1:34:50

the desk. Like it's in the middle of the room.

1:34:53

Unlike people who mostly put their desks against the wall, I

1:34:55

mean, you're sitting at the desk, you would be facing the

1:34:57

wall. That's not what you're doing. I

1:34:59

am such that I'm facing – when

1:35:02

I'm sitting at my desk, the window is to my

1:35:04

right. And so I can just look out, I can

1:35:06

turn to the right and look out the window and

1:35:08

see the happy ducks sitting around and floating by. So

1:35:12

I want to do this. Now the problem with floating the desk in

1:35:14

the middle of the room is that first

1:35:17

of all, it imposes a

1:35:20

larger aesthetic burden on

1:35:22

the desk and the items on the desk

1:35:24

because you don't have the wall to hide

1:35:26

your sins. So one of the

1:35:28

things I did, one of the reasons why I

1:35:31

was drilling stuff into my desk recently is

1:35:33

that I got a desk that has kind of

1:35:35

like a compartment behind it for

1:35:37

all the wires to go into. So

1:35:40

behind my desk is clean wire management, not a

1:35:42

whole bunch of wires dangling there because again, it's

1:35:44

the middle of the room. Like half the room can see

1:35:46

the back of my desk. So things

1:35:49

have to be a little bit cleaner here to look

1:35:51

nice than at the beach where I

1:35:53

can just shove everything against the wall like everyone else does. As

1:35:55

a result, when I was choosing my stuff,

1:35:57

I was just going to put it in the middle of the

1:35:59

room. speaker setup i didn't have tall

1:36:02

speakers as an option uh...

1:36:04

my my preferred speakers that the k e

1:36:06

f uh... q one fifties right there those

1:36:09

are awesome speakers than they actually sound they found that

1:36:11

i still love them at the beach but they're really

1:36:13

big and boxy and tall and i have a little

1:36:15

desk in the middle of the room looked

1:36:18

ridiculous so i'm not i'm not going

1:36:20

to do that so i look at small

1:36:22

speaker options i wasn't sure what to do yet meanwhile uh...

1:36:25

we had a friend of the show who

1:36:28

was gracious enough to share a sonos

1:36:30

discount code with me i believe in

1:36:32

the fall was a while ago uh...

1:36:35

and at that time i know

1:36:37

knowing we were renovating the new house but you know

1:36:39

this is a good opportunity to get some discounted sonos

1:36:41

gear uh... i bought a

1:36:43

set for the future t.v.

1:36:46

i bought a soundbar and two era

1:36:48

three hundred to be used as surrounds

1:36:51

oh jesus hey i had a discount code

1:36:53

i'm gonna take advantage of it you know so i

1:36:55

got a i got a backup if you don't speak photos

1:36:58

is very very quickly the general

1:37:00

way that sonos home theater stuff works is you

1:37:02

must have a soundbar which can cause some consternation

1:37:04

among some but leave that aside for now you

1:37:06

have to have a soundbar that is your front

1:37:08

your your center in your left and right channels

1:37:11

oftentimes you'll add one of their two subwoofers one

1:37:13

of which is not very large one

1:37:15

of which is quite large and

1:37:17

then you need appear rear speakers and

1:37:19

generally speaking you would get a pair

1:37:22

of era one hundred or one

1:37:24

s else which is what i have

1:37:26

these are roughly the size of the

1:37:28

original like og home pod then

1:37:31

they came out with just semi recently

1:37:33

the era three hundreds which are

1:37:36

freaking huge they are

1:37:38

apparently sound amazing and actually you know what

1:37:40

when i was at the beach house with

1:37:42

you i somehow convinced you to

1:37:44

drag from the basement the era one hundred

1:37:46

three hundreds out of the box to play with

1:37:49

and i can confirm they

1:37:51

sound amazing but

1:37:53

they are enormous and so

1:37:55

that is an aggressive use

1:37:58

of rear speedless really

1:38:00

aggressive rear speaker. Now they do dull the Atmos,

1:38:02

they fire up, I believe they're both stereo, so

1:38:04

it's incredible, I'm quite sure, but that's a lot

1:38:06

for a set of rear speakers. Yeah, they have

1:38:08

a ton of drivers in them. I would

1:38:11

not have normally done, like it would have felt a

1:38:13

little excessive for rears had it not had the discount

1:38:15

going. I was like, well, I like these speakers a

1:38:17

lot, let's give it a shot, what the heck, I'll

1:38:19

try out surround sound for real. Anyway,

1:38:22

so fast forward, you know, the

1:38:24

house stuff, we're still not really unpacked or we like

1:38:26

we still are needing furniture and stuff. So like the

1:38:28

home theater setup is not set up yet, we're just

1:38:31

watching TV with the built in speakers on the TV

1:38:34

and the sound bar is still in the box. Well,

1:38:37

I needed speakers for my office and I was looking

1:38:39

at all these different like, you know, different things people

1:38:41

recommended for like small, good sounding speakers. And

1:38:44

one day like I was going to be doing a

1:38:46

lot of work in my office like setting stuff up

1:38:48

and I just wanted some speakers, play some music. I'm

1:38:50

like, I have these two perfectly good era 300s in

1:38:52

boxes in the garage, what am I doing? Like, let

1:38:55

me just borrow these from my office until

1:38:58

I figure out my permanent setup and until the

1:39:00

TV needs them, which it doesn't yet. So

1:39:03

I'll just use these now in my office and you

1:39:05

know, I'll move them downstairs when downstairs is ready.

1:39:09

I still am using them. I

1:39:11

don't think I'm going to be moving them

1:39:14

downstairs. There it is way

1:39:16

bigger than the speakers you move because they

1:39:18

were too big, though they're bigger and I

1:39:20

think uglier than the speakers you got rid

1:39:22

of. So what's the deal? So first of

1:39:24

all, I wouldn't say they're uglier. Oh, they're

1:39:27

very weird looking. I don't know if I'd go

1:39:30

so far as ugly. They're wider than they are

1:39:32

tall, which is odd for a speaker to begin

1:39:34

with. Well, and so that's what makes them great

1:39:36

for my setup because tall

1:39:38

boxy speakers would look weird

1:39:40

floating in the middle of my of the room

1:39:42

like this. But these are like

1:39:44

landscape orientation speakers, but

1:39:47

they're huge. They're not that crazy.

1:39:49

I mean, they're very big and they're oddly shaped.

1:39:52

They are about the size

1:39:54

of any other like

1:39:56

smallish bookshelf speaker just tipped on its

1:39:58

side, but not rectangular. They're like

1:40:00

pinched, ovoid. You are deeply offended by the shape

1:40:02

of – I think the era 300 is among

1:40:05

the ugliest speakers I've ever seen in my life.

1:40:07

I just find it aesthetically unpleasing. I'm sure they

1:40:09

sound great, but I'm just saying – Oh, no.

1:40:11

They sound more than great. How

1:40:14

things look on your desk, this is not what I would go

1:40:16

with. Well, anyway, I've

1:40:19

been using them as my desk speakers now

1:40:21

for something like two months. They're

1:40:23

really good. So let

1:40:26

me say first, I think aesthetically they work

1:40:28

great. It's a very clean looking

1:40:30

setup. Because they are wider than they are

1:40:32

tall, they don't look too boxy on the

1:40:34

desk in the middle of the room. So

1:40:38

it's wonderful. Now, I

1:40:40

went through a couple of ways to hook them up. Here

1:40:42

are the downsides and upsides. So first of all, I tried

1:40:45

– the very first day, I didn't have

1:40:47

any audio cables or anything. I didn't even

1:40:49

have network cables that first day, so I was like, all right,

1:40:51

I'm just going to use AirPlay from my Mac. Never do this.

1:40:55

I strongly recommend

1:40:57

not using AirPlay from your

1:40:59

Mac for pretty much anything.

1:41:01

AirPlay is so responsive. Oh,

1:41:03

my God. So

1:41:06

my office setup is I

1:41:08

occasionally will use the studio display speakers,

1:41:11

which are very good for display

1:41:13

speakers, but they're still mostly trash. And

1:41:16

then I have an original move, which

1:41:18

is the Sonos – big Sonos portable

1:41:21

speaker. And I

1:41:23

often AirPlay using the music app, which

1:41:25

maybe – I guess maybe that's the thing

1:41:27

is that because the music app is such

1:41:29

a pile of garbage that anything that works

1:41:31

after that I consider to be a perk.

1:41:33

But I don't typically have any problem with

1:41:35

it. And it sounds pretty – just the

1:41:37

one move, one, the original move, just the

1:41:39

one of them, sounds pretty good. I mean,

1:41:41

it doesn't sound near as good as your

1:41:43

pair of Aero 300s as a stereo pair,

1:41:45

I'm quite sure. But it surprisingly

1:41:48

sounds pretty good. The move

1:41:50

does sound surprisingly good, but

1:41:54

it is still like a single speaker. There

1:41:56

are going to be limitations to that. I

1:41:58

believe you can stereo pair them. I don't know why you would buy

1:42:00

two of those because it's not really what

1:42:02

it's for. But anyway, so

1:42:05

don't use AirPlay as a – like

1:42:07

don't expect to use AirPlay as a permanent

1:42:10

desk speaker setup because there's multiple issues,

1:42:13

the biggest one of which is just

1:42:15

latency. There's massive latency. I believe it's

1:42:17

the old two-second AirPlay 1 latency. So

1:42:20

everything has a two-second delay and it's pretty

1:42:22

– it makes it pretty hard to use

1:42:24

without wanting to pull your hair out. It's

1:42:27

unreliable. It disconnects all the time. And

1:42:30

there's AirPlay for the system that you

1:42:32

get to from Control Center. And there's

1:42:34

also AirPlay built into the iTunes app

1:42:36

on the Mac. Both of them

1:42:38

are unreliable in different and creative ways and

1:42:41

don't play well with each other. There's so

1:42:43

many – just please don't use AirPlay. It

1:42:45

turns off all the time. It's slow. It

1:42:47

is not – AirPlay is

1:42:49

great for what it's for. It's

1:42:52

for streaming audio from your iPhone to

1:42:54

a speaker or it's for mirroring

1:42:56

your video off your laptop onto an Apple

1:42:58

TV. It's good for that. It is

1:43:00

not good to be your permanent desk speaker protocol.

1:43:04

Fortunately, that is not the only option with

1:43:06

these. Sonos sells this

1:43:08

little like $30 adapter that provides a line-in jack

1:43:10

to all their modern speakers I believe. Certainly

1:43:15

there are 300s and 100s. It's a

1:43:17

line-in jack via USB-C and it also

1:43:19

has a network port on it. I

1:43:22

believe they sell one that doesn't have the network

1:43:25

port but it's like $5 more to get the

1:43:27

USB-C line-in and Ethernet combo so get that one

1:43:29

because here's the other thing

1:43:31

with using this setup. Sonos

1:43:34

supports line-in through these methods but they

1:43:36

always have some degree of latency. The

1:43:38

way their protocol works is you can

1:43:40

tell their app, like, all right, when

1:43:43

– pair these speakers together and then

1:43:45

when there's a line-in input

1:43:47

to this one, automatically switch to it and then send

1:43:49

it to the other one. You

1:43:51

can set the latency on that. The minimum you

1:43:53

can set it for though is 75 milliseconds. It's

1:43:58

just enough latency that – it

1:44:00

would be fairly annoying for

1:44:03

games and watching people

1:44:05

speak in movies. However,

1:44:08

it is fine for music. Like for my

1:44:10

purposes, it's totally fine for music. I do

1:44:13

notice it a little bit, but

1:44:15

it's not a problem for me. And I

1:44:17

don't usually play games. I never play games

1:44:19

on this computer and I almost never watch

1:44:21

movies on this computer, so those

1:44:24

issues are not really a problem for me.

1:44:26

I wish there was a lower latency option,

1:44:29

but there isn't. The other weird

1:44:31

thing about this is that because

1:44:34

these are all like, you know, smart and,

1:44:36

you know, automatic and everything, and

1:44:38

because this is a stereo pair of

1:44:40

two networked speakers, when

1:44:42

you first, when you haven't been playing audio for

1:44:45

a little while, even just as short as like

1:44:47

a minute, they seem to go to sleep or

1:44:49

whatever, and then the next time you

1:44:52

play audio for the first cut, like

1:44:54

you'll lose the first second or two of what you

1:44:56

play because the speaker will be asleep and won't have

1:44:58

woken up yet. Then the left

1:45:01

one, which is what I was connected to, that

1:45:03

one will wake up first and then

1:45:05

like a half second later

1:45:07

the right one will join in the

1:45:09

party. So it's a little annoying. I'm

1:45:11

tolerating that annoyance for now because

1:45:15

it sounds fantastic. I

1:45:17

have a suggestion for you with the setup

1:45:20

that you just described. Why don't you, then

1:45:22

requiring more crap, why don't you just

1:45:24

connect these to like, I don't want to say

1:45:26

the R word to scare you, but another box

1:45:29

that you can control through AirPlay that plays music

1:45:31

and then still let your poor Mac be able

1:45:34

to like show you a YouTube video with correct

1:45:36

lip sync by having it, you know, I don't

1:45:38

know, use the built-in speakers in your, oh, XDR

1:45:40

doesn't have them, but anyway, like, let

1:45:42

these be your music playing speakers, connect these to

1:45:45

your stereo to use 80s lingo and let your

1:45:47

stereo be controlled through AirPlay, you know what I

1:45:49

mean? Like AirPlay to your stereo that is connected

1:45:51

to the Aero 300 to listen to your cool

1:45:54

music and everything, but like those things

1:45:56

have just, you know, you just want to watch like a funny

1:45:58

TikTok someone sent you. you wanna watch a

1:46:00

YouTube video, or you wanna watch a W3C video, and you

1:46:03

lose half a second of the thing, and the left one

1:46:05

turns on before the right, and there's audio, like, ugh, come

1:46:07

on. I don't do

1:46:10

those things through speakers very often. For

1:46:13

that kind of stuff, I'm almost always wearing headphones. Speakers

1:46:16

are really like, no one's around, and

1:46:18

I wanna play some music. That's what I'm saying,

1:46:20

have them connected to your stereo, and use headphones

1:46:22

through your Mac then. My computer is my stereo.

1:46:25

But you know what I mean, you'd still be controlling,

1:46:27

you'd still be playing your fish library in the music

1:46:29

app, right? You'd just be airplaying to your stereo, because

1:46:31

every single stereo supports, and I keep saying stereo, because

1:46:34

I don't wanna say receiver, because I know people flip

1:46:36

out. But like receivers, they all support airplay. You could

1:46:38

just connect the Aero 100, well, you can't connect the

1:46:40

Aero 100, or 300. No, trust me. As I was

1:46:42

saying, using AirPlay for that is a terrible experience. I

1:46:44

would run- But you were playing music! I don't care.

1:46:46

I would run a line cable from my

1:46:48

desk to the receiver here. Well, you just

1:46:51

get actual speakers, and not these weird computers

1:46:53

that are pretend to be speakers. You know,

1:46:55

like, you connect them with speaker wire, you

1:46:57

know, like, I know, I'm complicated, but it

1:46:59

just sounds like such a miserable experience using

1:47:01

these computer speakers. I mean, really, you just

1:47:03

want good music speakers. Well, and that's why

1:47:05

I have that at my beach office, where I just have

1:47:07

the Q150s, and a little, like, you know,

1:47:09

$50 little amp that'll go to

1:47:11

the bottom of my desk, and a subwoofer. Like,

1:47:14

I have all that there, and it's great, but

1:47:18

that setup would look ridiculous here. Like, that

1:47:20

would not fit my aesthetic goals. And the

1:47:22

Aero 300's thought, I need to

1:47:24

see pictures. Convince me with pictures. Show me this doesn't

1:47:26

look ridiculous. Yeah, I mean, yeah, well, I would love

1:47:28

to see pictures of this, because I

1:47:30

am not as deeply offended by the look of

1:47:33

these as John is. I can

1:47:35

understand, John, how you got to that perspective, because

1:47:38

it's not unreasonable, but I'm not as, you know,

1:47:40

offended as you seem to be. But

1:47:43

no, I get what you're

1:47:45

going for here. I wonder if,

1:47:48

and now this is gonna bring up the whole

1:47:51

new app situation, but I wonder

1:47:53

if an

1:47:55

alternative for your use would

1:47:58

be to use the Sonos app to play with it. whatever you want

1:48:00

to play on the speakers and thus not

1:48:03

involve anything computer at

1:48:05

all. The Sonos app sucks.

1:48:07

I know they just rewrote it and it sucks even

1:48:09

more and I'm sorry for everyone who's affected by that.

1:48:11

I don't use the Sonos app except for configuring the

1:48:13

speakers when I first set them up and then I

1:48:15

never touch it again because I

1:48:17

play music from the music app previously called

1:48:20

iTunes on my Mac while I work. That

1:48:22

is the use case here. I'm not going

1:48:24

to take out my phone and

1:48:28

play with the app. I'm controlling it through the

1:48:30

music app period. I get that but they have

1:48:32

a native app. Well, I don't know if I

1:48:34

should say native. They have an app for the

1:48:36

Mac. I heard

1:48:39

rumblings that it's going away. This

1:48:42

is not insider information. I think they might have announced this

1:48:44

at some point. Their app for the

1:48:46

Mac is fine. I wouldn't say it's great

1:48:48

but it resolves a lot of these

1:48:51

issues that plague you and it will

1:48:53

interface with your library. Now, a library

1:48:55

of your library size? I don't know

1:48:57

but it is supposed to interface with

1:49:00

your iTunes match library. Just

1:49:03

briefly, I think I talked about

1:49:05

this. I might have talked about it on the show and

1:49:07

certainly talked about it on analog but the new app, it's

1:49:10

fine. As long as

1:49:12

you don't need accessibility features which are coming

1:49:14

slower than they should but it's fine. The

1:49:16

only thing that really chaps my bottom about

1:49:18

it is you can't do any Q management

1:49:20

or play list management. If I really want

1:49:26

to play a particular song next, there's

1:49:28

nothing I can do about that except run up

1:49:30

to my computer and use the old app to

1:49:32

do it which stinks and that's coming and that

1:49:35

will come soonish. For the most

1:49:37

part, I don't mind the new app and I think it

1:49:39

is less clunky than the old app. It's very different

1:49:41

than the old one so it requires relearning some stuff.

1:49:44

I use the Sonos app to play

1:49:46

music almost exclusively. The only

1:49:48

time I really don't is like I said when

1:49:51

I'm at my computer and I agree with you

1:49:53

ultimately that I find it easier to find

1:49:56

the music I want using the music app and that's what I

1:49:58

do when I airplay it to the move and that's fine. But

1:50:01

especially if I'm just going for ambient music

1:50:03

and not – as I'm

1:50:05

around the house or whatnot or

1:50:07

just want to play an album or something like

1:50:09

that, oftentimes I'll use the Sonoff's app, including sometimes

1:50:11

on the computer. I mean

1:50:14

that is an option, but that is just – that is

1:50:16

a direction I don't want to go. I don't want to

1:50:18

play things that way. That is not

1:50:20

my workflow. I don't want to change my workflow. I

1:50:22

think my workflow for my needs and preferences is better.

1:50:25

So yeah, I just don't want to deal

1:50:27

with that. And honestly, once I – so

1:50:29

I did hardwire them. They each have Ethernet

1:50:31

going to them, nice and

1:50:33

clean of course. So they each have Ethernet.

1:50:35

They each have – the one has the line in and sends

1:50:37

it to the other one over the network. And

1:50:40

once I – when they were both on Wi-Fi, it

1:50:42

was a little bit shaky. I had to use the

1:50:44

higher latency settings. It wasn't good. Sonoff's

1:50:47

products don't have great Wi-Fi radios, but

1:50:51

when I wired them, they became rock

1:50:53

solid reliable. I've never had one since

1:50:55

wiring them to the network. I've never

1:50:58

had one like drop out or be

1:51:00

too out of sync or whatever. It's

1:51:02

never happened. So it is fantastic. Anyway,

1:51:05

so if you're willing to do this

1:51:07

crazy setup, which again, this is ridiculous.

1:51:09

You really shouldn't do this. But

1:51:12

if for some reason you're crazy like me and you want to do this, it

1:51:15

sounds great. In

1:51:17

most ways, it sounds I think even better

1:51:19

than my Q150s, not

1:51:21

at mid-range and vocals

1:51:23

and guitars. The Q150s

1:51:25

still are my favorite sound for that. However,

1:51:29

they are bigger. They are deeper. They

1:51:32

are uglier. And the

1:51:35

challenge with the Q150s is that they're really

1:51:37

not made for desk listening distance, and

1:51:39

they have a fairly small sweet spot

1:51:42

in speaker terms. This is like how

1:51:44

much does the sound change or get

1:51:46

worse if you like shift your body

1:51:48

to different directions or like are leaning

1:51:50

or whatever? Is there

1:51:52

like how big is the sweet spot where it sounds

1:51:54

the best? Most speakers are not great at

1:51:56

this. The Q150s are... are

1:52:00

especially at a distance of like four

1:52:02

feet. They're not really made for that. It

1:52:06

doesn't have a very big sweet spot. The

1:52:09

ERA 300 pair, each speaker

1:52:11

has like seven drivers firing in

1:52:13

different directions. There's a lot

1:52:15

going on there that fire sound in a whole

1:52:17

bunch of different places. And so as a result,

1:52:20

the sweet spot is very wide and broad. It

1:52:23

also, in

1:52:25

speaker terms, it has a very,

1:52:27

the ERA 300 has a really impressive sound stage.

1:52:29

Again, just by its design. This

1:52:32

is what this means is like, how

1:52:35

wide or big does it sound like the

1:52:37

sound is coming from? Like, does it sound

1:52:39

like it's coming from two points in front

1:52:41

of you? Or does it sound like you're

1:52:44

in an auditorium full of sound

1:52:46

coming from the big wall in front

1:52:48

of you, whatever? That's the sound stage.

1:52:51

These have a massive sound stage

1:52:53

for speakers that fit on your desk and

1:52:55

that you're listening to from four feet away.

1:52:58

And again, because there's drivers firing in every direction

1:53:00

and they're doing all sorts of processing, it

1:53:03

is really impressive for that. By

1:53:05

far the best sound stage I've heard in

1:53:07

speakers at this distance. And

1:53:10

they have pretty good bass

1:53:13

for their size. Especially considering

1:53:15

not having a subwoofer, they

1:53:18

have very impressive bass. They destroy the

1:53:20

Q150s in bass. And

1:53:22

that they even destroy the Q350s in bass, which

1:53:26

are bigger speakers. They have really

1:53:28

good bass for their size. And

1:53:31

so you don't need a subwoofer. However,

1:53:35

then I added a subwoofer. Which

1:53:37

one? The sub, of course.

1:53:39

Because that was also- The big one? Yes,

1:53:41

because that was also part of the set

1:53:43

that I got for the living room. Oh,

1:53:46

I bet that sounds so flipping good. Oh

1:53:48

my gosh. So the thing about the Sonos

1:53:51

sub, this is their big subwoofer, the

1:53:53

one they've had for a while. It is force

1:53:55

canceling. I believe I discussed this a long time

1:53:57

ago when I discussed my Q- and

1:54:00

my setup at the beach and I bought this very

1:54:03

expensive KEF subwoofer for the Q150s

1:54:05

at the beach because it was force

1:54:07

cancelling. The Sonos sub

1:54:10

only works with Sonos. It is only like

1:54:12

their wireless protocol, although you can network wire

1:54:14

it, but it doesn't have a line in

1:54:16

it is what I'm saying. So it only

1:54:18

works with Sonos stuff, so I couldn't use

1:54:20

it at my setup at the beach, but

1:54:23

I could use it here. It is a

1:54:26

really good force cancelling subwoofer for less

1:54:28

money than, as far as I know,

1:54:30

any other force cancelling subwoofer on the

1:54:32

market by a pretty big margin. What

1:54:35

force cancelling is great for is – and

1:54:38

this is why Apple always taps it in

1:54:40

the MacBook Pros because I believe they also

1:54:42

have force cancelling subwoofers in their laptops –

1:54:44

what's great about that is that it significantly

1:54:47

reduces buzzing and vibration from

1:54:49

subwoofers. You hear the sound, but

1:54:51

because it has two drivers firing

1:54:54

in opposite directions, the

1:54:56

vibration is largely or completely

1:54:59

cancelled out. So you

1:55:01

don't usually hear too much boominess.

1:55:03

You don't really hear materials nearby

1:55:05

vibrating as much. It

1:55:07

doesn't vibrate with the floor because it is cancelling

1:55:09

out by doing two different directions at the same

1:55:11

time. So these speakers,

1:55:13

they're her 300s, again, by

1:55:15

themselves, they have great bass. With

1:55:18

the subwoofer, it's a lot

1:55:20

of fun. Like, this is a ridiculous setup. Again,

1:55:22

nobody needs to do this. I

1:55:26

don't need to do this. I happen to have

1:55:28

these things for my living room, and

1:55:30

I decided to set them up in my office for

1:55:32

a while first, and it's a lot

1:55:34

of fun. The subwoofer

1:55:37

is – honestly, it's not a massive difference because

1:55:39

the bass in them is already pretty good, but

1:55:42

it does improve things, and it does make it

1:55:44

a lot of fun. I

1:55:47

don't need it, but it's really fun. I

1:55:50

am glad you are satisfied. So sitting here

1:55:52

now when the living room is ready, are

1:55:55

you going to then buy another batch of

1:55:57

all these things? I already

1:55:59

have the sound bar for you. down there, maybe there'll be enough. I

1:56:02

like how the original motivation was, these speakers are

1:56:04

too big on my desk, and you replaced it

1:56:06

with two, I think also very big

1:56:08

speakers, and giant subwoofer.

1:56:11

It's giant for that application. I wouldn't say

1:56:13

it's that giant, particularly for a home theater,

1:56:16

but for that application, it's not small. You

1:56:18

should do the measurements. How many square inches of

1:56:20

speaker did you remove, and how many square inches

1:56:22

of speaker did you add in their place? Well,

1:56:24

I relocated it. The subwoofer is across the room,

1:56:26

like at the opposite wall that I'm looking at.

1:56:29

And then the two speakers are tipped over

1:56:31

to their sides. The

1:56:34

only weird thing about these speakers is that

1:56:36

because of their design, I can't

1:56:38

pile crap on top of them. I learned

1:56:40

that. That's what I'm saying, they're weirdly shaped. Plus,

1:56:43

they're up-firing drivers anyway, isn't it? Yes.

1:56:46

But there's probably aesthetic benefits to me not being

1:56:48

able to pile crap on top of my speakers.

1:56:51

At the beach with the KEFs, those are just boxes,

1:56:53

and so the top of them becomes

1:56:55

a work area. There's always

1:56:57

papers on them. When I

1:56:59

was burning the Blu-ray discs to back up

1:57:02

all my stuff, the M-discs, the

1:57:04

Blu-ray drive just lived on top

1:57:06

of one of the speakers. There's

1:57:08

always crap on top of my speakers at the beach.

1:57:10

And here, I can't do that. That's because they take up so

1:57:13

much desk space. There's no desk space left to put stuff, so

1:57:15

you have to put it on top of the speakers. This

1:57:19

is fun. And I get to watch the ducks while I

1:57:21

listen to my awesome music on my ridiculous set-up, so I'm happy.

Rate

Join Podchaser to...

  • Rate podcasts and episodes
  • Follow podcasts and creators
  • Create podcast and episode lists
  • & much more

Episode Tags

Do you host or manage this podcast?
Claim and edit this page to your liking.
,

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features