Podchaser Logo
Home
2.5 Admins 185: 2.5 Gigabits

2.5 Admins 185: 2.5 Gigabits

Released Thursday, 7th March 2024
Good episode? Give it some love!
2.5 Admins 185: 2.5 Gigabits

2.5 Admins 185: 2.5 Gigabits

2.5 Admins 185: 2.5 Gigabits

2.5 Admins 185: 2.5 Gigabits

Thursday, 7th March 2024
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:00

The. Two

0:02

and Half Edmunds Episode One, Eight, Five, I'm.

0:04

Joe. I'm. Jim and I'm Helen.

0:06

A hero again. Last. Time

0:08

you talked about privacy. This. Time Let's

0:11

talk about ai. Jen. Hsun Huang

0:13

says kids shouldn't learn to code. They

0:15

should leave it up to a I

0:17

G. What? A strange thing for

0:19

a guy running in video to say

0:21

moon yes Not like they have a

0:23

vested interest in selling hardware to people

0:26

who do a i bullshit or anything

0:28

gym at the same time. Doesn't.

0:30

And video need to be able to

0:32

hire people to write the Ai stuff.

0:35

Know they I'll do the math. presumably.

0:38

Wang is trying to talk to the kids who

0:40

aren't going to grow up to be the kind

0:42

of engineers that in video needs or wants because

0:45

a I certainly isn't doing that kind of job

0:47

any time soon. presumably. He

0:49

just figure sees can I'm. Screwing.

0:52

Over the next generation from I guess

0:54

lower and coders. And I just

0:56

just turning over low to mid range coating

0:58

the ai. But. You

1:00

can't rely on a I for that either. Yeah,

1:02

you can rely on it to help you out

1:05

save you some time, but you still need the

1:07

skills to understand what spitting out a. You.

1:09

Can rely on it to give you ideas

1:11

is what you can rely on Iran, but

1:14

whatever ideas give you, they need to be

1:16

thought of that way. Ideas untested, unproven that

1:18

you need to check, and in many cases

1:20

that can still be. You know, a huge

1:22

efficiency when productivity when an Ai tool can

1:25

break you out of just like a brain

1:27

lock where you can't think of the thing

1:29

you need to think of and you need

1:31

some new data. Just kind of. Job.

1:34

Degree Meet. Sometimes A I can help

1:36

you by producing a whole bunch of

1:38

bull or break crap that doesn't change much

1:40

from one thing to the next and

1:42

it's like a huge mental or emotional

1:44

or energy drain for you to sit

1:46

there and type it all out. But if

1:49

it just gets morphed out in front

1:51

of you, your competent to scan it

1:53

in see this is correct or not

1:55

correct. What you really can't do with a

1:57

I and Co to say hey I.

2:00

write this code for me and I'm just

2:02

going to hit save and run it. Yeah,

2:04

I'm specifically talking about boilerplate stuff that like

2:06

you say, it's just typing out loads of

2:08

stuff and it just goes blah instantly saves

2:11

you 30 seconds. But

2:13

again, you really can't use it for that

2:16

unless you understand the boilerplate stuff because it

2:18

won't always get that right either. Yeah. So

2:21

if you can't catch and correct an error

2:23

in the boilerplate stuff faster than you can

2:25

type the boilerplate stuff out, it's still kind

2:27

of a lose. Nobody

2:30

in the world isn't going to be a

2:32

coder and every single kid isn't going to

2:34

either. But

2:36

I still think there's value in them understanding some of the

2:38

concepts of it. And I wish

2:40

I had got introduced to the concepts in

2:42

a more formal way earlier in my career

2:45

rather than I learned to program

2:47

by downloading QBasic code from the

2:49

local BBS and reading it and

2:52

muddling with it and learning the hard

2:54

way. And then by the

2:56

time they started trying to teach the basics of

2:58

programming in high school, it's like I was running

3:00

circles around that. Yeah, and you accidentally

3:02

became a developer because you were a sysadmin

3:05

originally and then you sort of just accidentally

3:07

became a developer. Whereas I thought I was

3:09

going to be a software developer my whole

3:11

life and then discovered no, actually turns out

3:13

I'm a sysadmin. Yeah, for me it

3:15

was a little bit like I thought I was going to be

3:17

a software developer at first as well. But

3:19

then in high school, I did a job placement at

3:22

the power plant as a sysadmin and I really enjoyed that. And

3:25

then when it was time to decide what to

3:27

go to school for, it's like, well, all the

3:29

programming I know I taught myself or I can

3:31

read from a book and I can do with

3:33

my one computer. But to be a sysadmin, I

3:35

need a lot of computers. And

3:38

guess what the college has? A whole lot of computers

3:40

I can go in touch with. So

3:42

that's what I'm going to school for because that's

3:44

where I'm going to actually get more value from

3:46

the practical side of it. Alan, you

3:48

are pretty pragmatic. Have you tried

3:50

GitHub CodePod for example? Not

3:53

the CodePod one. I've used chat GPT to

3:55

help come up with ideas for articles to

3:57

write or how to structure a presentation or

4:00

how to just say something or questions

4:02

to ask during job interviews and lots of

4:05

ideas, like Jim was saying, asking it, here's

4:07

some things I've thought of, think

4:09

of 10 more for me to help me fill out

4:12

this list. It's reasonably good at

4:14

that. You need to also

4:16

apply a critical eye and edit it in a bit

4:18

and so on, but it can really be a time-saver.

4:21

But for source code, it's like with

4:23

what I'm doing in the source code, honestly,

4:27

probably not going to be as helpful or my

4:29

biggest concern still is, what is the license

4:32

of this code going to be? Because

4:34

if I'm selling this code to people, I

4:36

have to be able to rightfully

4:38

claim that I wrote this and

4:41

that I have the authority to sell

4:43

it to you and that it isn't

4:45

some other code. As far as

4:48

Wang's idea of saying that kids should

4:50

learn AI instead of coding or whatever,

4:52

I think the big thing here that keeps

4:55

getting me about that is the issues of

4:58

who wrote it, who owns it, is

5:00

it clean in terms of the

5:02

bill of materials, whatever for

5:04

AI written code. Even when you put all that

5:06

stuff off to the side, just the idea that

5:09

do it with AI instead of doing it yourself,

5:12

I have the same advice there that I give people

5:14

all the time about scripting. I see larval

5:17

folks all the time writing these really

5:19

long scripts to chain together 20 or 30 simple tasks

5:22

that they don't really understand the individual tasks very

5:24

well yet. They may very

5:26

well be able to put together a script that ties

5:28

all those 20 or 30 things into one script that

5:30

they can run to do that thing over and over

5:33

again, but you

5:35

haven't learned how to actually do

5:38

any of it. If you lose that script,

5:40

you're screwed. There's a lot of things you

5:42

don't understand. My advice is do

5:45

those 20 or 30 things manually, frequently enough,

5:47

often enough that you've got no problem just

5:49

doing it all off the top of your

5:51

head, then write the script because you

5:53

don't need to do those things anymore because you know them,

5:55

you understand them, you know all the steps, you can do

5:58

them without the script. look

6:00

at, okay let me take the script and every

6:19

time almost certainly won't

6:21

do the same

6:24

thing every time. It is partially random

6:27

in the same way that you know a dog

6:29

is partially random. That's the whole

6:31

point. You didn't have to program it. You

6:33

could just give it a loose idea of

6:35

what you wanted it to do in

6:37

a sloppy, imprecise human

6:40

language and it would mostly do what

6:42

you mostly wanted. So it's like

6:44

anything else. When that's the degree of accuracy

6:46

you have from the person on the top

6:48

to the person next down the chain stretching

6:51

our terms a little bit and granting

6:54

you know some large language model the

6:56

title person, still the point is it's

6:58

imprecise so you have to supervise in

7:01

a way that you don't as much with code

7:03

because you can find out whether the

7:05

code is buggy or not and you know

7:07

it'll operate the same way and if you're doing the same

7:09

thing it'll do the same thing and again with a model

7:12

not so much. But it may be

7:14

taken from a different lens if what

7:16

Jensen was trying to say or maybe

7:18

just a thought I'm having is rather

7:20

than every grade schooler learning

7:22

Python and only some of them

7:25

going on to actually do development if we

7:27

taught them how to use AI properly and

7:29

what its limitations are that might be much

7:31

more useful to a lot of them that

7:33

aren't going to become developers. The

7:36

skills of how to use AI to

7:38

get prompts for creative writing or art

7:40

or all the other things that AI

7:42

can be used for it might actually

7:44

be a more valuable skill than force

7:46

feeding Python to every grade schooler. See

7:48

what I hear you saying Alan is

7:50

kids should also learn to use AI

7:52

which I would agree with. What

7:55

Wang said is kids shouldn't learn

7:57

to code. Well yeah he says

7:59

it It is our job to create

8:01

computing technology such that nobody has to program,

8:04

and that the programming language is human. Everybody

8:07

in the world is now a programmer. This

8:09

is the miracle of artificial intelligence. And

8:11

it will be every bit as reliable as just

8:13

off-handedly telling your kids to go do something, and

8:15

then walking off and coming back four hours later

8:18

to see if it's done. Well, you

8:20

know what this reminds me of? It reminds me

8:22

of stuff like Dreamweaver back in the day, where

8:24

you could make a website, but anyone

8:26

who knew about making a website would look at

8:28

the code that Dreamweaver spat out and

8:30

go, ugh, this is horrible. I

8:32

mean, yeah, it kind of works,

8:35

but it's horrible code. Same thing

8:37

with Microsoft's front page. I

8:39

remember that and, like, you know, you bold something and then

8:41

change it and later it would be like an open

8:44

bold, closed bold tag around nothing and just all kinds

8:46

of detritus in the code. It was very useful as

8:48

a learning tool in that you could use the WYSIWY

8:50

to get something and then look at the source code

8:52

side and figure out how it did it. But

8:55

as soon as you understood anybody, you're like, why is

8:57

it doing all this other stupid stuff? I

9:00

used to use front page as a WYSIWYG editor

9:02

and I would actually write the code in HTML in

9:04

the source window, but just the fact that it

9:06

would render in real time as I was doing it

9:08

was invaluable because you didn't have a dev console

9:10

in the browser to do that with back then. Yeah,

9:13

I was like, I don't know how we ever did,

9:15

like, CSS without the developer tools in

9:17

the browser. With great difficulty and much

9:20

angst. Well, especially back then when it's

9:22

like, oh, you have to, like, upload

9:24

it somewhere before you can view it,

9:27

have it all work correctly. So you, like, edit

9:29

the file, save the file, copy and paste

9:31

over FTP to somewhere, and then go to

9:33

the browser and hit refresh for every single edit.

9:35

And the round trip time was terrible.

9:38

Especially when it was CGI bin stuff and, like,

9:40

you literally couldn't do anything with it on your

9:43

own computer. So you're editing, you know, these thousand

9:45

line code files on your own machine and FTPing

9:48

them up over dial up to serve from the other

9:50

end of the country and then hitting a web page

9:52

that would hopefully run your code and maybe put out

9:54

the thing that you wanted. Oh,

9:56

that sucked. God, that sucked a lot. Well,

10:00

what about Google cuts a deal with

10:02

Reddit for AI training data and also

10:05

automatic, the company that owns Tumblr and

10:07

WordPress is going to sell users data

10:09

to Mid-Journey and OpenAI. It's

10:11

disappointing, but again, it's not surprising, is

10:14

it? Not really, and I think we're going

10:16

to see more and more of this. The

10:18

kind of interesting thing is a lot of these

10:20

places probably didn't have a good cutout for

10:23

this in their terms of service. They

10:25

usually are generic as possible, so they can get

10:27

away with anything, but I think you

10:29

might start to see this specifically in terms of service

10:32

going forward saying, yeah, we could use

10:34

this data to train our own AI or sell it to

10:36

somebody else who will. But

10:38

I see a lot of places deciding

10:40

that, hey, this stuff we have, maybe

10:42

somebody will pay us for it and

10:45

going and doing that. And I think it's terrible,

10:48

but I also think unless we

10:50

make a really big stink really quickly, it's going

10:52

to happen a lot. I mean,

10:54

Jim, you wrote hundreds of thousands of words on

10:56

Reddit helping people out with a very set of

10:59

S problems and whatnot. You must

11:01

feel pretty bummed out that that's just

11:03

being sold to some AI company. This

11:06

may surprise you, but actually

11:08

that part doesn't bother me so much.

11:10

Well, I would say specifically you wrote

11:12

those things in public to be public

11:14

for the purpose of it being publicly

11:16

consumed, right? Exactly. I wrote those

11:18

things on a very public forum with the

11:20

idea that anyone who wanted to could learn

11:22

from it, and that includes the AI model.

11:25

Again, I will point out

11:27

that AI models learn. It's

11:30

not really that different from

11:32

the way non-artificial intelligence learns.

11:34

They train on data. I'm

11:37

not mad about somebody learning from

11:39

the things that I said online because that's my

11:41

whole goal is I want people to learn from

11:43

it. What I was mad about was the heavy-handed

11:46

actions of Reddit itself,

11:49

just screwing people over and

11:52

yanking the API away from a

11:54

third-party ecosystem that literally had to

11:57

grow up because Reddit itself wasn't

11:59

producing. using usable apps and

12:01

just the high handedness. The biggest

12:03

thing that got me as mad at Reddit as

12:05

I was was not even

12:07

yanking the API. It was when I

12:09

saw what Spiz had to say about

12:11

it. When I saw the leaked internal

12:14

memo that talked about, you know, oh,

12:16

well, you know, the user base is

12:18

noisy, but just rioted out. I was

12:20

like, Oh, oh, that's how it is,

12:22

huh? We're just noisy and inconvenient resources

12:24

that won't just sit there to be

12:26

quietly exploited however you want. We actually

12:28

insist on having some kind

12:30

of a voice. Hmm. Okay.

12:32

Yeah. I think it's also, it'll be

12:35

interesting to see the details if they ever

12:37

do come out about is

12:39

Reddit just providing a version of that

12:41

API where Google can in

12:43

a way that's less costly to Reddit scrape

12:45

all the text of people's posts, or

12:48

is this also including more metadata

12:50

like when they posted and how they post

12:52

and how people interacted with each post. If

12:55

it's just give us all the posts and

12:57

people made these posts in public Reddit, and

13:00

maybe it's not that big of a thing and Reddit does

13:03

have the right to that content. But

13:05

you know what it says, selling users

13:07

is data to train AI. That smacks

13:09

me slightly differently than just selling bulk

13:11

access to the content that Google could go

13:13

and scrape from Reddit. Normally, it would just

13:15

be much more efficient for Google and less

13:17

costly for Reddit if they use the API

13:20

to get it. I think ultimately, you

13:22

know, if this is something that we care about as

13:24

a society, you know, we have to actually

13:27

do something about it, whether the

13:29

we as consumers or the we as the

13:31

government that you know, the consumers elect, whether

13:34

the pushback is, you know, laws

13:36

that say you aren't allowed to do this, and

13:39

we will throw you in jail or make you

13:41

unable to do business in our country, or whether

13:43

the pushback is consumers refusing to buy or pay

13:45

for things, there has to

13:47

be actual pushback that matters or

13:49

companies aren't going to stop. The

13:52

phrase is don't leave money on the table.

13:55

And it's around longer than any of the three of

13:57

us have been alive or than any of our countries

13:59

have existed. That is human nature. You get

14:01

people in the business of making money, and they

14:03

want to make more of it, and they find

14:06

a way to do it. And unless

14:08

something slaps their peepee in

14:10

the door and makes them think that's a bad idea,

14:13

they will, because otherwise they're – and I'm doing the

14:15

scare quotes here – leaving money on

14:17

the table. Okay,

14:19

this episode is sponsored by TALIDE.

14:22

What do you call an endpoint security product

14:24

that works perfectly, what makes users miserable? A

14:26

failure. The old

14:28

approach to endpoint security is to lock

14:30

down employee devices and roll out changes

14:33

through forced restarts, but it just doesn't

14:35

work. IT is miserable because

14:37

they've got a mountain of support tickets, employees

14:40

start using personal devices to get their work done,

14:42

and executives opt out the first time it makes

14:44

them late for a meeting. You

14:46

can't have a successful security implementation unless

14:48

you work with end users. That's

14:51

where TALIDE comes in. Their

14:53

user-first device-trust solution notifies users as soon

14:55

as it detects an issue on their

14:57

device, and teaches them how to solve

14:59

it without needing help from IT. That

15:02

way, untrusted devices are blocked from

15:04

authenticating, but users don't stay blocked.

15:07

TALIDE is designed for companies with Okta,

15:09

and it works on macOS, Windows, Linux,

15:11

and mobile devices. So

15:14

if you have Okta, and you're looking for a device-trust

15:16

solution that respects your team, support the

15:18

show and go to collide.com/25a to

15:20

watch a demo and see how

15:22

it works. That's kolide.com/25a.

15:30

Let's do some free consulting then. For first, just a

15:32

quick thank you to everyone who supports us with PayPal

15:34

and Patreon. We really do appreciate that. If

15:37

you want to join those people, you can go

15:39

to 2.5admins.com/support, and remember that for various amounts

15:41

on Patreon you can get an advert-free RSS feed

15:43

of either just this show or all the

15:45

shows in the Late Night Linux family. And

15:48

if you want to send in your questions for Jibberlallen

15:50

or your feedback, you can email show at 2.5admins.com. David

15:54

writes, I'm considering an upgrade path for

15:56

my current router setup. At

15:58

the moment, I'm using a linksys.com. WRT1900ACS

16:01

router, which has a few 1Gbit

16:03

ports for WAN and LAN. I

16:06

don't have a 1Gbit internet connection, so on

16:08

that side, the speed of the 1Gbit router

16:10

is plenty enough. In a medium-term future,

16:12

I will be adding a NAS to my network, and

16:14

I'm wondering how I could upgrade the link speed to

16:17

2.5Gbits or more. If

16:19

I had a 2.5Gbit switch, and I connect most

16:22

of the wired devices directly to the switch, would

16:25

they be able to communicate at 2.5Gbits

16:27

between themselves, or will the router limit

16:29

the link speed to 1Gbit? The short answer

16:31

is yes, that's perfectly fine. As a matter of

16:33

fact, you're already doing that, whether you realize that

16:35

or not. Every gigabit switch out

16:37

there will also handle fast Ethernet at 100Mbit, or

16:39

the old 10Mbit Ethernet

16:42

even, all just fine. All the

16:45

devices on that switch will speak to each

16:47

other at the highest speed that their individual

16:49

link connection allows. The only

16:51

thing that I would warn you of here is,

16:53

and this is probably obvious, you've probably already thought

16:55

of this, it's not a real issue, but just

16:57

in case somebody hasn't figured this part out, if

17:00

you're thinking that you've got super fast

17:02

Wi-Fi clients, and they'll get a benefit

17:05

of that at 2.5Gbit, well, in

17:07

that case, then yes, your router or

17:09

access points had better also have 2.5Gbit

17:12

wired ports, or else you're not

17:14

going to see any of that theoretical speed up.

17:16

Not that they probably ever were going to go

17:19

anywhere near that fast anyway, because that's home marketing

17:21

hype. Yeah, you've got to just consider what is

17:23

connected to what, and if it's a 2.5 connected

17:26

directly to another 2.5, whether that's a

17:28

NIC in a desktop and a switch,

17:31

then that's going to work. But if

17:33

you have, say, a 2.5 port connected

17:36

to a 1Gbit switch that

17:38

is then connected to another machine that is

17:41

2.5, you're only going to get one. Yeah,

17:43

it's slightly more complicated. In a switch,

17:45

it's kind of virtually creating circuits between

17:48

every port to every port. And

17:50

so, yeah, all the ones that are linked at 2.5Gbits

17:52

will be able to talk at 2.5Gbits, although

17:55

in general, other traffic, like there's

17:57

a broadcast from one of the 1Gbit ports.

18:00

that's going to possibly live at the speed,

18:02

but only for microseconds and

18:04

so on. In general, it shouldn't

18:06

be a problem. Higher end switches

18:08

can do some amount of buffering to try to smooth

18:11

it out, but that's usually a, a

18:13

much more expensive switch than you're going to have, and b,

18:15

a complete waste of time at your house. My

18:18

house, I have a switch that has a

18:20

mix of 10 gigabit and 1 gigabit machines,

18:23

and my desktop and my NAS can

18:25

talk at 10 gigabit just fine, and

18:28

the other 1 gigabit clients can also talk

18:30

to the NAS just

18:33

fine. And you know, two

18:35

different 1 gigabit clients can talk to my

18:37

NAS at 1 gigabit each because

18:39

the NAS can deliver to the switch at

18:41

10 gigabits. With that said, you probably should

18:44

not expect to be able to feed 2 gigabit

18:46

clients at full wire rate from one single

18:49

2.5 gigabit connected

18:51

system. Alan might disagree

18:53

with me here, but I've only had limited

18:55

experience at 2.5 gigabit. It

18:57

has not impressed me. I think my

18:59

desktop has a 2.5 gigabit port, but it's

19:01

not hooked up to anything because I installed

19:03

a 10 gig NIC and ran it to

19:05

the other thing, so I don't have any

19:07

real experience with 2.5 gigabit either. Honestly,

19:10

most motherboards now sometimes

19:12

even have a 2.5 gigabit PHY, but

19:15

it's actually exposed out as just two

19:17

1 gigabit ports, which

19:19

is an even weirder configuration. So

19:21

yeah, I wouldn't expect to be able to do

19:23

that too much simultaneously, and it will really depend

19:25

on your switch how much you can actually

19:28

do the kind of multiplexing, but

19:31

you'll be totally fine to only upgrade some of your

19:33

devices in this case. Like you said, your internet connection

19:35

is only 1 gigabit or less than 1 gigabit, so

19:38

you don't have to replace your whole router

19:40

just because you want your desktop and your

19:42

NAS to be able to talk at more

19:44

than 1 gigabit. Absolutely. My recommendation here would

19:46

be absolutely play with this. Understand

19:48

that all you have to do is buy the switch. You don't have

19:51

to spend a ton of money and just... I

19:53

would advise you don't expect too much out of it,

19:55

which if you're not spending a ton of money is

19:57

not really a problem, right? Like you put the stuff

19:59

together, you see what it looks like. will do and

20:01

now you know and now you're a better admin or

20:03

hobbyist or whatever you like to call yourself because you

20:05

know more stuff. You've seen it happen. You know how

20:07

it works. And that's totally cool. What

20:09

I just don't want to see anybody doing

20:11

is being like, well, I think I can

20:13

scrape together the $400 or $500 for this

20:16

project that I expect to get these big

20:18

benefits out of. Like if that's the thought

20:20

process, I'm going to say maybe

20:22

don't do that with 2.5 gigabit because, again, I

20:25

don't have a ton of experience with it. But

20:27

I've seen it in action a couple of times

20:29

and not seen a whole

20:32

lot more than 1 gigabit of real

20:34

world throughput through it. And

20:36

the fact that you never ever see

20:39

2.5 gigabit stuff on anything

20:42

targeted for business, it's always like

20:44

home and gamer oriented. And in

20:46

business, you go straight from 1

20:48

gigabit to SFP plus and 10. There's no

20:50

in between there. Which I honestly found a

20:52

little bit weird because the whole point of the 2.5 gigabit

20:55

standard was take advantage of the 1

20:57

gigabit copper that's already in the walls

20:59

of your office and not have to

21:01

restructure your office. But I think it

21:03

– like I'm saying, maybe it's

21:06

just never delivered because I've not seen it

21:08

have any adoption on the business side. Well,

21:11

it's one of those things also. Like every time we

21:13

have a big bump up in network speeds, you never

21:15

get them in the first few years. Like you get

21:17

better than the last one and that's really what you

21:20

get. When 10 gigabit first dropped on

21:22

the scene, like nobody was getting 10 gigabit, they were getting

21:24

2 or 3 and they were happy about it. When

21:26

gigabit first dropped on the scene, nobody was getting

21:28

gigabit. They were getting like 500 or 600 megabit

21:31

and happy about it because again, that was 5

21:33

or 6 times the 100 megabit. As

21:35

the technology matures, you get closer and closer to

21:37

the theoretical wire rate in terms of the actual

21:40

wire rate throughput of data that you get. So

21:43

that now these days with gigabit, even

21:45

on just any random put together like

21:47

a couple of devices and a cheap

21:49

gigabit switch, you'll probably see 850 to 950 megabits

21:52

on an iPerF3 connection between the two. But

21:56

again, that's now. That's not what

21:59

it looked like. for the first five years or

22:01

so of Gigabit, you know, being out there in the

22:03

real world. I think that

22:05

2.5 Gigabit has kind of a similar problem

22:07

in that it's not really achieving, you know,

22:10

everything that it claims that it can on

22:12

the package. But what it's

22:14

claiming on the package is not a tenfold

22:16

increase. It's already only, you know, a 2.5

22:18

fold increase. And

22:22

I don't think that was big enough to move

22:24

the needle enough to make business environments

22:26

go for it. Because

22:29

in business environments, you've got more of a pressure

22:31

to actually test the thing and trial it and

22:33

see if it does what you want it to,

22:36

and nobody's finding that those things are doing what

22:38

they want them to. Well, I think

22:41

also in business, if you're going to upgrade the

22:43

switches and all the infrastructure anyway, only

22:45

going up to 2.5 didn't seem to make sense versus

22:47

going to 10 or 25 or 40 Gigabit. Because

22:51

the other thing is, really to Jim's point, the

22:54

drivers for these new devices are maybe

22:56

not as mature. Like, you can get

22:58

used Mellanox 10 Gigabit stuff that

23:00

was when it was mature, but

23:02

is still five or ten years old now, and

23:05

get the full 10 Gigabit for a card you can get on

23:07

eBay for $15 or $25. And

23:10

my base-net rack is full of those

23:12

Mellanox cards. Like, they connect X3s, they're

23:15

connect X2s, they're super old, but they

23:17

can saturate the whole 10 Gigabits. Whereas,

23:20

yeah, all this 2.5 Gigabit stuff is still

23:22

pretty new, and the drivers aren't as

23:24

mature. And even the hardware is like, well, we're trying to

23:26

make this as cheap as possible, right? At

23:29

the time, the Mellanox is like, no, this is

23:31

a premium product. This is like the best thing.

23:33

We're going to put all these extra horsepower in

23:35

these big, chunky cards. Whereas the

23:37

2.5 Gigabit is like, we want the built-in

23:40

Ethernet on your motherboard that we make as

23:42

cheap as possible to have a bigger number

23:44

so that you'll buy it. The other

23:46

thing I'll warn anybody that's interested in getting into all

23:48

this, it's really cool and exciting to

23:50

get into 10 Gig or faster networking when

23:53

all you've done is Gigabit up until then. And

23:55

Gigabit has been a thing for so long that

23:57

I know for a lot of folks, it's just

23:59

like that. That's the speed that wired networks go.

24:02

And the idea of, oh, I can

24:04

go 10 times faster than that is

24:06

super exciting. And I'm not

24:08

telling anybody not to play with that, but I

24:10

am gonna say, I'm gonna

24:13

share that there are two different types

24:15

of storage admin. There's the one gig

24:17

storage admin who thinks performance doesn't matter

24:19

because the network's always the bottleneck, and

24:21

there's the 10 gig storage admin who

24:23

very quickly discovers, holy shit,

24:25

everything matters now. Because

24:28

it is so easy to get yourself in a

24:30

situation where your network will go faster than one

24:32

gigabit, but your storage won't, and

24:34

as a result, you don't see any

24:37

actual real world difference. So

24:39

be aware that going 10 gig can be a

24:41

little bit of a rabbit hole you end up

24:43

diving down, and you may end up upgrading a

24:45

lot more than you thought you were gonna have

24:47

to before you really see that 10 gig benefit

24:49

the way that you want it to. My

24:51

shock drives suddenly feel very sane. I

24:55

just finished a consulting gig for a

24:57

video post-production house, and

24:59

I'm like, all right, we've got our file server with 100

25:01

gigabit, Nick, it's connected to

25:03

four or five different editing bays with

25:06

a mix of 40 and 100 gigabit,

25:08

and we need to be able to do two

25:10

or three gigabytes per second from each of three

25:12

or four workstations at once from

25:14

this mass. Each

25:17

frame of the video is a separate file that

25:19

are each 30 to 40 megabytes, and

25:22

we need to be able to stream these without

25:25

any stuttering. So we need to

25:27

be able, it's gotta be sustain this many gigabytes

25:29

per second, it doesn't need to be more than

25:31

that, but it needs to be this many, and

25:34

never fall below that many, and

25:36

that's a lot harder to do than

25:38

just installing a bigger Nick. You gotta

25:40

get your storage dialed in, you gotta

25:42

have your compute dialed in, because when

25:44

you're feeding that much network throughput, you

25:47

need to have the PCI Express lanes to

25:49

read from your storage fast enough, as well

25:51

as the storage itself being fast enough. So

25:53

you've got CPU issues, you've got

25:56

storage issues, in addition to

25:58

just throwing a faster Nick at it, doing Ethernet

26:00

like you're going to have to configure jumbo

26:02

frames and that's going to be weird and

26:04

different between different operating systems like I

26:07

just had to mess with all that this weekend for

26:09

a client that did a 10 gig upgrade and didn't

26:11

get the benefit out of it for exactly the reasons

26:13

we're talking about. So I spent

26:15

the whole weekend in there messing around with

26:18

jumbo frames discovering that you know with certain

26:20

configurations their Windows clients would just plain crash

26:23

even though everything was configured the way it

26:25

should be with you know 9000 byte

26:27

jumbo frames all the way around

26:29

like there's so many ways for it to screw

26:32

up like on Windows to get a 9000 packet

26:34

frame you actually have to set it

26:36

at 9015 because Windows also includes the

26:39

15 byte frame header on every frame

26:41

whereas every other operating system doesn't count

26:43

the header when you know they're adding

26:45

together the number of bytes that go

26:48

in so it is a

26:50

giant pain in the butt it can be

26:52

done but just again I'm warning

26:54

everybody out there this is not a

26:56

casual project it can get away from

26:58

you quick. Yeah and especially when you're

27:00

like we were in for a certain

27:02

level of consistency it's like so it

27:04

needs to not just be fast but

27:06

not have any latency spikes and

27:08

we have to have enough headroom that we

27:11

can absorb a spike and not ever have

27:13

the transfer rate fall below the number

27:15

of frames that you'd be able to get per second

27:17

to play it back smoothly to make sure it's coming

27:20

out right and it really is like

27:22

I've done on a server as

27:24

big as two dual ported

27:26

hundred gigabit hundred gigabit cards so

27:29

that was 400 gigabits total pushing

27:31

out of a 128 core machine

27:34

and just proving that it could take video

27:36

files from RAM and check them out the

27:38

network SSL encrypted at 400

27:40

gigabits a second and that

27:42

took a lot of tweaking and tuning and

27:45

the biggest thing when you're getting to any

27:47

of these higher numbers beyond a gigabit is

27:50

you usually need more than one

27:53

flow one TCP connection is

27:55

not going to saturate your 10 or 100 gigabit

27:57

network card because you're going to run out

27:59

of CPU. power. There's one

28:01

flow, you can only use one core. Whereas if

28:03

you make 10 connections, then each of those 10

28:05

connections can execute on a different core at the

28:07

same time and have a different interrupt channel from

28:09

the NIC and you'll be able to do it.

28:12

And so for the video editing thing,

28:14

we had to make each client be using 16 threads

28:18

so that we're pulling this frame and the next

28:20

frame and each in a different thread so that

28:22

we would have enough flows to actually be able

28:24

to deliver that much bandwidth because you couldn't do

28:26

it with a single NFS connection. Right,

28:29

well we'd better get out of here then. Remember

28:31

show at 2.5admins.com if you want to send any

28:34

questions or your feedback. You can

28:36

find me at jrres.com/mastodon. You

28:38

can find me at jrs-s.net/social. And

28:41

I'm at andangie. We'll see you

28:43

next week.

Rate

Join Podchaser to...

  • Rate podcasts and episodes
  • Follow podcasts and creators
  • Create podcast and episode lists
  • & much more

Episode Tags

Do you host or manage this podcast?
Claim and edit this page to your liking.
,

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features