Transcript #413: python-build-standalone finds a home
Return to episode page view on github00:00 Hello and welcome to Python Bytes, where we deliver Python news and headlines directly to your earbuds.
00:05 This is episode 413, recorded December 9, 2024.
00:09 And I'm Brian Okken.
00:11 And I'm Michael Kennedy.
00:12 This episode is sponsored by us, so check out the links in our show notes,
00:16 but also check out Talk Python Training and PythonTest.com.
00:20 There's courses over there.
00:21 And of course, thank you to Patreon supporters.
00:23 And we have links.
00:25 If you want to get a hold of us, you can reach us on BlueSky or Mastodon.
00:29 The links are in the show notes.
00:30 And also, if you're listening to the show, thank you.
00:35 And please share it with a friend.
00:37 Also, if you'd like to participate in the discussion while we're recording,
00:42 you can head on over to pythonbytes.fm/live and see when we're recording next.
00:46 But usually it's Monday at 10 a.m. Pacific time.
00:50 Sometimes it shifts, though.
00:51 And during the holiday season, who knows what we might do.
00:53 But so far, we're sticking with that.
00:56 And if you'd like to get the links in your email inbox, go ahead and go and sign up for
01:02 the newsletter at pythonbytes.fm.
01:04 And we will send you all of the links in the show notes right in your inbox.
01:08 So, Michael, let's kick it off.
01:10 Let's kick it off.
01:12 I want to talk about a little bit of jitter.
01:14 Maybe I've had too much coffee this morning or something.
01:17 I don't know.
01:17 What do you think?
01:18 Jitter is a thing from the folks at Pydantic.
01:22 And the idea here is they need really fast JSON parsing as the foundation of Pydantic.
01:30 Basically, Pydantic is about how do I exchange, validate, and transform JSON data with Python
01:37 classes, right?
01:38 Into Python classes.
01:39 Okay.
01:39 Yeah.
01:39 So, you want that to be fast.
01:41 The folks over at Pydantic created this thing called Jitter, G-I-T-E-R.
01:46 And it is a fast iterable JSON parser.
01:49 Now, if the Pydantic usage does not catch your attention, OpenAI is also using Jitter, which is
01:57 pretty interesting.
01:58 Ask ChatShiftPT about it.
01:59 So, the reason that they were interested in it is they want to be able to work with,
02:05 I believe Pydantic as well, but they want to work with responses coming out of LLMs.
02:10 And anyone who's used LLMs until maybe very recently knows that they kind of like spit out
02:15 the answers in a little progressive way, right?
02:19 And so, with this, you can parse parts of data as it comes down, which is pretty cool.
02:25 So, there's some examples of partial in here.
02:28 You can go look for somewhere, I think maybe on the docs website or something like that.
02:32 But, you know, you can give it like a partially formed string and it'll come up with like
02:37 perfectly good answers for it.
02:38 So, that's pretty neat.
02:40 And that's one of its features.
02:42 The other is that it's faster than what I think the default Rust JSON parser is.
02:47 Even for non-iterable, just straight parse it, which is, that's pretty impressive.
02:52 Okay.
02:52 And then there's also, this is why we are talking about it, there's Python parse, which
02:57 parses JSON strings into a Python object.
03:00 So, you can go and run that as well, which is pretty cool.
03:04 Shooter example.
03:06 Yeah, yeah, yeah.
03:07 Anyway, yeah.
03:08 So, you can go and parse it into different pieces using, basically, if you need a really
03:13 fast JSON parser with Python, you can use Python parse and it'll parse into a structure, right?
03:18 Yeah.
03:19 So, awesome.
03:20 I thought people might be interested in both an iterable, iterating JSON parser, and back
03:26 to this one.
03:27 Iterating JSON parser and also a really fast one.
03:30 Plus, it's being built by the folks at Pydantic, Sam Colvin and team.
03:34 And yeah, excellent.
03:36 Nice work.
03:37 Oh, yeah.
03:37 I think I've got several uses for this.
03:39 This is cool.
03:40 Yeah, cool.
03:40 Yeah.
03:41 I recently had Samuel Colvin on with David Seddon to talk about building Rust extensions
03:48 or integrating Rust with Python and things like that and talk Python.
03:52 And he talked about this as one of the things they're building, which is like, oh, okay, this
03:56 is pretty interesting.
03:56 Yeah, definitely.
03:58 Well, I'm going to talk about Python pre-built a little bit.
04:01 This is big news, Brian.
04:04 I'm glad you're covering it.
04:05 So, Python build standalone is a project that we've talked about on the show, but mostly we
04:12 talked about it in association with UV.
04:15 Because if you use UV sync or UV install Python or UV virtual environment or UV vnv and then install
04:23 and use Python there, if it can't find it on your system, the Python in your system, it's
04:29 going to pull it from Python build standalone, which is a separate project, not part of UV.
04:34 So, we've discussed that.
04:36 But the big news right now is that Python build standalone is now part of Astral or under the
04:42 Astral umbrella, which is huge.
04:44 So, yeah, we're going to link to an article from Charlie Marsh, head of Astral, saying a new
04:52 home for Python build standalone.
04:54 There's also a, it just says we'll be taking over, we'll be taking stewardship of this project
05:01 from Gregory Zork, I don't know, cool last name.
05:05 Anyway, the foundational project for building and installing portable Python distributions.
05:09 And there's a link to Gregory's announcement also.
05:13 And the discussion around that, like the Python build standalone powers UV, powers Rai, also
05:22 PipX and Hatch and more.
05:24 And it's got like 70, 70 million downloads so far.
05:28 Wow.
05:29 Pretty big project and definitely instrumental to going forward with Python or with Python
05:35 packaging and using Python.
05:37 So, Astral is really like trying to make UV with along with this Python build standalone project, the new way to install Python.
05:47 And for me, it is.
05:49 I'm using it all every day now.
05:51 So 100% same for me.
05:53 So pretty short article talking about this.
05:56 But it is kind of interesting.
05:58 It talks about what the project is at first.
06:02 It talks about the future of Python standalone Python distributions.
06:05 Also, what they have in mind for the project.
06:09 It looks like they want to keep the project up to date with Python releases, of course.
06:14 And then upstream changes to the CPython build system, possibly.
06:17 And then remove some of the third is remove some of the project's existing limitations.
06:22 What are the existing ones?
06:25 It ships some MUSL-based Python builds.
06:27 They're incompatible with Python extension modules.
06:31 I don't know what that means.
06:32 I don't know what MUSL is, so I'm going to move on from that.
06:35 Okay.
06:36 And then improve the project's Python build and release process.
06:40 Just a good stewardship for this project, and I'm really happy about that.
06:44 Along with this, I was interested to read a thread from Charlie Marsh about what said Python build standalone has exploded
06:54 in popularity with over 70 million downloads all time.
06:58 I'm going to put a link to this thread on Blue Sky into the show notes also, because it's an interesting discussion.
07:07 And I learned something through here that I didn't know before.
07:13 It said that the Python...
07:14 I didn't know this.
07:15 That the Python.org download, the download from Python.org, it actually downloads an installer that builds Python from source on your machine.
07:24 For Linux.
07:25 For Linux.
07:26 Okay.
07:26 It says for Linux.
07:28 Okay.
07:29 So for Linux.
07:30 Yeah, because the macOS and the Windows ones install way too fast.
07:34 The building Python from source is like a 10-minute deal if it runs the tests and stuff.
07:39 Okay.
07:40 Yeah.
07:40 So I didn't think I was doing that on...
07:44 Yeah.
07:44 Anyway.
07:44 You didn't get the error VCVars bat that couldn't be found?
07:47 No.
07:49 I haven't seen that for a while.
07:50 So yeah, I guess a bigger deal for people that are not running Windows or Mac, but that's really like all the servers and stuff.
07:58 Yeah.
07:58 Well, I think the other thing that's really non-obvious here is like, what is this build standalone anyway?
08:05 Why don't we just download the installer and just run it or just take the result of the installer and clunk it out into your machine or something?
08:12 So my understanding is the non-standalone one depends on other foundational things in the system, especially in Linux, but also in other places.
08:21 If you want to be able to just copy it over, you can't do that.
08:24 And so one of the things that they're talking about, one of the four points of the direction that they're trying to go that Charlie laid out was trying to upstream some of these changes back into CPython itself.
08:35 I think it might be number one of the future.
08:39 Yeah.
08:39 Upstream the...
08:40 No, number two.
08:41 Yeah.
08:41 Upstream the changes to the CPython build system because they have to patch Python in order to make this actually build, which is why it's a pain in the butt to maintain.
08:50 And then how many combinatorial variations of that do you get for different platforms and stuff, right?
08:56 Yeah.
08:57 And so trying to say, look, we've done these things to make it build more simply with fewer dependencies.
09:01 Let's maybe make that part of Python.
09:03 I don't know about you, but I have not seen a single problem with UV Python, Python build standalone Python, compared to system Python.
09:11 It's not like, oh, well, the certificates don't validate or this thing doesn't work or it doesn't have SSL or some weird thing like a dependency might be missing.
09:19 It seems fine to me.
09:21 And actually, I would be more worried about installing it separately and building it on each of the machines I'm installing it on than I would having one install that goes everywhere.
09:37 Yeah.
09:39 Yeah.
09:40 Anyway.
09:40 Yeah.
09:41 And I can tell you that Pythonbytes.fm is powered by Python 313.1 based, derived from, or gotten from, this method here.
09:49 Yeah.
09:49 Yeah.
09:50 Anyway.
09:52 Big news that actually probably doesn't mean much to individual users other than I think that we can try, we had a little bit of concern about whether or not, you know, this one project, it was sitting heavily on one person, one developers to maintain.
10:08 And I'm glad that it's Astral helping out with this now, too.
10:11 Yeah.
10:11 I agree.
10:12 And if you read Greg's announcement there, transferring Python build standalone stewardship to Astral, that he talks about how the Astral folks actually for a while have been core contributors to the project.
10:25 And they've been working from the outside to help keep this thing going because they realize how important it is to this feature, right?
10:30 Yeah.
10:31 And also, I read, I don't know if it was in this or somewhere else, but I essentially read that the project was, I mean, Astral was really working on it for several months anyway.
10:41 Yeah.
10:41 Exactly.
10:42 This is mostly an official announcement is all.
10:46 Yeah.
10:47 But one final parting thought, Brian, is right there in where you are.
10:51 It says, as, this is in Greg's announcement, as I wrote in my Shifting Open Source,
10:55 priorities in March.
10:56 Yeah.
10:56 This is an interesting challenge that people can run into with projects that are run by one person, right?
11:01 Yeah.
11:02 The guy had a kid, wanted to spend more time with the kid, was feeling worn out by the projects and decided.
11:08 Well, and also talks about how he really just cares way more about Rust than he does about Python these days, which is fine.
11:15 Like, you're not married, you know, for life to a technology, you know, go where your heart takes you.
11:22 But that's a challenge for projects that are run by one person.
11:25 So I think it's worth reading this thing as well, just for people to get a sense of, you know, when open source projects take off, but it's not necessarily a good fit.
11:34 Yeah.
11:35 Yeah.
11:36 Yeah.
11:36 Yeah.
11:37 He's also known for the PyOxidizer project, which came close, but didn't quite get us a single binary of our Python apps.
11:45 Yeah.
11:46 Interesting.
11:46 Okay.
11:47 Yeah.
11:47 I really am.
11:49 It's really cool that he made sure that this was in good hands before shifting it over.
11:53 Yeah, absolutely.
11:54 Absolutely.
11:55 All right.
11:56 All right.
11:57 On to the next.
11:58 On to the next thing.
12:00 So I talked about, there's a theme here.
12:02 I talked about the jitters from having too much coffee.
12:05 Well, let's talk about mocha.
12:06 Maybe if we put some hot chocolate and some sugar in with it, it'll be better.
12:09 No, probably not.
12:10 So this project, this project is by Delario and it's called Mocha Pie.
12:15 So Mocha, let's like work our way inside out.
12:18 So Mocha is a high performance concurrent caching library for Rust, not a concurrent caching server like Redis.
12:28 Think SQLite, but for caching, right?
12:31 SQLite is written in C, not Rust, but it's an in-process sort of deal, which is pretty, pretty neat.
12:37 And it's this itself is inspired by caffeine for Java, right?
12:40 This is kind of like turtles all the way down, like ports all the way down.
12:43 So it provides a caching implementation on top of dictionaries.
12:46 They support full concurrency of retrievals in the high expected concurrencies for updates.
12:52 All right.
12:53 So thread safe, highly concurrent in-memory cache implementation, sync and async can be bounded by the maximum number of entries,
13:00 the total weighted size, size aware eviction, like kicking large things out versus small things.
13:06 You can have cache controlled by least frequently used, by last recently used.
13:12 Like I want to kick out things that are over two minutes, but if you've got room based on something, that's fine.
13:20 You can give them a time to live, a time to idle, right?
13:23 Idle is a really cool, interesting one.
13:24 Like when was this last access?
13:26 So if you've got something that's old, but is used all the time in your app, and it's something that's somewhat new,
13:32 but it kind of hasn't got used that much, it'd be better to kick out that new one rather than the old one, right?
13:38 Oh, yeah.
13:39 Okay.
13:39 So that's all just straight Mocha.
13:41 Mocha Py is Python binding for this.
13:44 Here we go again.
13:45 Rust library for Python.
13:46 They're probably getting VC money from this.
13:49 I'm telling you.
13:50 Okay.
13:51 No, just joking.
13:52 Sort of.
13:53 So for the Mocha Py thing, it has a synchronous cache, which supports basically thread save memory.
13:58 It just like wraps the thing.
13:59 So time to live, time to idle, size of concurrency, all these things that you can imagine.
14:06 And so there's a couple interesting ways.
14:08 You can just say cache.set some value, or you can say cache.get some value.
14:12 That's one way to use it.
14:13 Another one is you can use it as, this is actually pretty amazing.
14:18 You can use it as an LRU cache function decorator alternative.
14:23 Oh, wow.
14:23 Right?
14:24 So one of the things you can do that's really easy to speed up Python code with not writing
14:28 much code, you don't have to maintain much, is you just put a decorator, functools.lru
14:33 cache onto it, and it'll look at the hash value of all the inbound parameters and say, if you
14:38 pass me the same parameters, you're getting the same output, right?
14:41 Yeah.
14:41 And it just does that, just straight in Python memory.
14:43 But this would be backed by this high-performance concurrent Rust internal library.
14:48 It's still in process, right?
14:49 Yeah.
14:50 So you can say, yeah, go ahead.
14:51 Sorry.
14:51 With the time to live and time to, you know.
14:54 Time to idle, yeah.
14:55 Especially.
14:55 That's cool.
14:57 Yeah.
14:58 This is pretty cool.
14:59 And there's so much talk about the thing supporting the Mocha itself, the Rust version, supporting
15:05 asynchronous behavior, right?
15:07 I'm like, okay.
15:09 If it has all these asynchronous capabilities, what's the story with Python and its async and
15:15 await, right?
15:15 Yeah.
15:16 So I filed an issue, which I don't really like to do, but that's how you ask questions,
15:20 apparently, and then you close it.
15:21 So I said, hey, cool project.
15:24 Since it says thread safe, highly concurrent in-memory implementation, what's the Python
15:29 async story?
15:30 Great.
15:30 And so they responded, this will work if you put the decorator on there.
15:35 So remember how I was complaining that it's sort of weird that the func tools and iter
15:40 tools don't support async?
15:42 Yeah.
15:43 This, this func tool like thing, supports async and sync functions as well.
15:48 Right?
15:48 So they just have a implementation in the center that says, is it a coroutine?
15:52 Do this else do that.
15:53 So you can use the cache, the caching decorator.
15:56 Like we talked about, like the LRU cache thing already async on async functions and sync functions.
16:02 So that's fine.
16:02 And then I said, well, what about cache, get and set?
16:05 And Deliro says, probably doesn't make sense to do it.
16:08 It takes 230 nanoseconds.
16:11 So you can do 4.4 million calls a second and set is a 1.3 million sets per second for a
16:18 cache size of 10,000.
16:20 That's fully occupied on a simply M1 Mac.
16:22 So, you know what?
16:23 Hmm.
16:24 Probably not, but there might be some, some, ways to expand this in the future.
16:30 I don't know.
16:30 But yeah, I would say probably not, probably not needed because you're going to probably
16:34 add more overhead just to juggle the async stuff.
16:36 Right?
16:37 Yeah.
16:38 Yeah.
16:38 And also just if the, if the supported, method is through the decorator and whatever you need, you could just put like your code
16:45 in a function to do.
16:46 Yeah.
16:47 I mean, if that were Redis, you would absolutely want an async version because you're talking
16:50 to another server.
16:51 Yeah.
16:52 And there's that latency in the network and all, but yeah, if you can do 4 million a second,
16:55 then probably I doubt you can do 4 million awaits a second.
16:58 So that it's much lower.
17:00 So the cache get and set really are just that where you, the benefit of those is, is probably
17:05 just for, because we want a really fast caching system or something.
17:09 Yeah.
17:10 Yeah, exactly.
17:10 And you, there's plenty of times where you say in this situation, I want to get this out
17:15 of the cache and then keep it for a while.
17:16 Like if I had a user who logged in and I want to just hold their user account with all their
17:21 details and I've used their ID as the key and their actual user object as the object that
17:25 goes in, that's fine, but you wouldn't, you wouldn't use that as a cache decorator because
17:30 typically you might see that coming out of a database, something like that.
17:34 And then if you pass the same user in, it's like, it's similar, but it's a different database
17:38 object.
17:39 Right.
17:40 You can run into real weird situations where they're equivalent, but they're not equivalent,
17:43 you know, and then you end up not using cache.
17:46 So anyway, I think that might be where you would do it.
17:48 But anyway, I think this is pretty cool.
17:49 People can check it out.
17:51 Yeah.
17:51 And it's, it is not, I don't believe it is like super popular here.
17:55 And you know, a hundred stars, it kind of shine a light on it.
17:58 But if you go over to the Mocha thing, you know, it's got a thousand, 700 stars and this
18:02 is kind of just a Python UI on top or API on top of it.
18:06 Yeah.
18:07 But it's, it's pretty recent.
18:08 I mean, it's a few weeks old.
18:10 Looks like.
18:11 So it's just a baby.
18:12 Just a baby.
18:13 It's okay to have a hundred stars.
18:15 Pretty good for that.
18:16 It's pretty good actually.
18:17 Yeah, it's pretty good.
18:18 It looks cool.
18:18 So now, you know, all right.
18:20 I want to shift back to UV.
18:22 I'm kind of in a UV mood.
18:24 I'm missing the sun apparently.
18:26 But the, there's an article from SAP, from, from the SAS Pegasus blog about UV and in-depth
18:34 guide to Python's fast and ambitious new package manager.
18:38 And a lot of people have written about UV already, which is great.
18:41 But I really, I have been really excited about when I thought, when I learned about UV sync
18:47 and started using that and all the different ways to use UV.
18:51 It's a pretty powerful tool.
18:53 So it's not really one thing.
18:55 It's designed to be a lot.
18:56 So, so I appreciate, you know, articles like this, but also I really like this one.
19:02 So it starts out with pretty much who is, which is a funny meme of, of a whole bunch of different
19:09 commands to install, install Python and update it and install, create a virtual environment
19:15 and sync your requirements.
19:17 And all of that is just done with UV sync.
19:19 Now you can do it all in one, which is pretty sweet.
19:22 I don't use UV sync.
19:23 I use UV, V, and V, dash, dash, Python 3.13 or something.
19:27 But, you know, same.
19:28 Yeah, I'm using both depending on whether or not I have a project set up already.
19:33 So it talks about what is UV, why use it.
19:37 And we're just going to assume that you already know if you listen to this podcast because it's
19:41 really fast.
19:43 But a lot of discussion of different workflows talks about installing, adopting UV into your
19:50 existing workflows, doing install.
19:53 But I'm going to pop down to the end, adopting UV into your workflow.
19:58 There's this cool cheat sheet.
19:59 This is pretty much what the entire article talks about, the different parts, is you can
20:03 use UV Python install to install Python.
20:07 You can use virtual environment or VENV to create virtual environments.
20:11 It's really fast.
20:12 And then install packages with UV pip install.
20:16 But then also you can build your dependencies.
20:20 Like we would have used pip compile.
20:22 You can use UV pip compile.
20:24 But it's all in one place to all these different commands.
20:27 And these really are the commands.
20:29 The commands listed in this article are really the way I use UV as well.
20:33 So that's why I appreciated it.
20:35 And then a discussion about how to adopt this into your workflow and what that means to
20:41 get, you know, talking about.
20:43 I mean, a lot of people might not have used sort of lock files before.
20:47 But using lock files with UV, it's so easy that, you know, why not?
20:52 And pinning your dependencies.
20:54 Just some good workflow.
20:56 It's good Python project practices anyway.
21:00 So why not?
21:01 Yeah.
21:02 Yeah.
21:03 That's great.
21:03 And there's even a few more that you could throw in for the tool, like the equivalency
21:06 table there.
21:07 Yeah.
21:08 You know, there's you.
21:09 There's you for installing CLI tools.
21:13 You could say pip X.
21:15 Yeah.
21:15 And just create a virtual environment and install the things and make that in the path and all
21:20 those sorts of things versus UV tool install or UV run.
21:24 Right.
21:25 Those kind of things as well.
21:26 So, yeah.
21:26 Yeah.
21:27 It's missing that, which is, you know, I'll feed it back to Corey.
21:32 So one of the reasons why this came up on my radar is I'm working on a project that uses
21:40 SAS Pegasus.
21:41 So I'm in touch with Corey a lot.
21:43 Yeah.
21:45 But like the UV, the tool thing, instead, I'm not using pip X anymore.
21:50 I'm the UV tool install is like super cool.
21:54 So, yeah, it's super cool.
21:55 It is.
21:56 I've also started using Docker for certain things as well.
21:59 So, yeah, it's kind of kind of similar.
22:02 But like, for example, Glances, which is a Python based server UI visualization tool,
22:06 you can just say Docker run Glances versus installing Glances.
22:10 And you just leave this machine a little more untouched.
22:13 Yeah.
22:13 One of the interesting things about this article was the point of view, because at the start,
22:21 Corey talks about how he's not usually somebody to jump on like multi-tool fads, like pipenv
22:29 or pyenv for installing.
22:32 For doing virtual environments better and big project-wise.
22:36 And I like Hatch, but I'm not really a using Hatch for my entire workflow sort of person.
22:43 I was using it just for a packager.
22:45 So, yeah, I'm in the same boat of like, I didn't really need an all-in-one tool, but this one changed my mind.
22:51 And I really like this all-in-one tool.
22:53 So, yeah.
22:54 I'm still not bought into the project management side, but I love using UV for the stuff.
22:58 Yeah.
23:00 Yeah.
23:00 Yeah.
23:00 Anyway, what do we got next?
23:02 We have a quick bit of follow-up here.
23:04 Okay.
23:04 I just did some searching.
23:06 So, over on PipX.
23:08 So, one of the things that, you know, you say like, you could use PipX or there is an open issue on PipX that says integrate UV in some way.
23:16 Right?
23:16 Because PipX is really just a wrapper around create virtual environment, pip install package, pip install dash U package.
23:23 Right?
23:23 And so, if they just change the internals to say UV pip install, then PipX would all of a sudden become super awesome.
23:31 This recommendation is unfortunately over half a year old, but it does have 21 upvotes.
23:37 So, you know what?
23:37 Yeah.
23:38 Who knows?
23:38 That's there.
23:39 Yeah.
23:39 Okay.
23:40 Yeah.
23:40 Okay.
23:41 But that's not what I want to cover next.
23:43 Come on, computer, respawn.
23:45 There we go.
23:45 I think that's it for our items, right?
23:47 We're on to extras.
23:48 Let's have extras now.
23:49 Yeah.
23:50 Let's extra it up.
23:51 Extra.
23:51 So, registered for PyCon, I did.
23:54 Oh, cool.
23:55 Yeah.
23:56 Registration came out two days ago.
23:57 I don't know.
23:58 Whenever I posted some message on Blue Sky and Mastodon saying, I registered.
24:03 How about you?
24:03 Whenever that was, that's when the announcement came out.
24:06 So, I think a day and a half ago or something like that.
24:08 So, there's early bird pricing and all details on there.
24:11 If you want to go and check it out.
24:13 It's normally $450 for individuals, but you could save $50 if you register for January,
24:19 which is pretty cool.
24:20 There's a bunch of stuff.
24:21 It has all the detailed timeline, which is always interesting.
24:24 You know, like, if I want to make sure I attend the PyLadies auction, when do I need to do that?
24:29 When is the main thing?
24:31 When is the job fair?
24:32 Et cetera.
24:32 So, most importantly, main conferences, May 16 to May 18, 2025.
24:36 So, there it is.
24:38 And congruent with current times mask policy.
24:41 Hooray.
24:41 Optional and encouraged, but not required.
24:44 Yeah.
24:44 How about that?
24:45 Yeah.
24:45 Cool.
24:45 Okay.
24:46 Got a few more real quick ones here.
24:48 I recommend, you know what?
24:49 It's something I came across just thinking, like, why don't I support more open source projects?
24:53 Looking at my dependencies and stuff that I'm working on.
24:57 Like, how much, you know, if everybody who used Flask put $1 towards it per month.
25:03 Everybody who used it in an important way where it's not just like, oh, I did a tutorial with Flask.
25:07 But, like, no, I have a project that is important to me and I use Flask.
25:10 If everyone put $1 towards it, it would transform that project.
25:14 If everyone who used G-Unicorn put $1 towards it, that would transform it, right?
25:19 So, I decided, you know what?
25:20 I'm going to just go to some projects and find the one that I use most.
25:22 And, yeah, I found four that had sponsorships that were available.
25:27 I was going to support UV and Pynantic as well.
25:30 But they, for some reason, they do, like, corporate sponsorships.
25:34 Or I tried to do individuals and it didn't work.
25:36 And then some other ones, like Beanie, don't currently have sponsorships.
25:39 But, you know, are really important for the database layer and stuff.
25:42 But just think about, you know, put a couple dollars towards some of these projects.
25:46 It'll make zero difference to you if you have a job as a software developer.
25:49 And in the aggregate, it'll make a big difference to the health of the ecosystem.
25:53 Yeah.
25:54 It's interesting to think about like that.
25:56 Like, just, you know, a couple less coffees a month.
25:59 And you can help out.
26:01 You probably cover, like, three or four projects.
26:03 Yeah.
26:03 Yeah.
26:04 Anyway.
26:05 I want to encourage people to do that.
26:07 But, you know, can't.
26:08 Obviously, don't.
26:09 But I don't think it's a big deal.
26:10 Come here, computer.
26:13 Very slow for some reason.
26:14 Don't know why.
26:15 There we go.
26:16 All right.
26:16 This is the joke.
26:17 So I'm skipping the joke for a second.
26:19 We'll come back to it.
26:20 There's two things that I wasn't planning on covering.
26:21 But I'll throw out here really quick.
26:22 Yeah.
26:23 Here's my register for PyCon.
26:24 I also wrote a quick.
26:25 People said, oh, my God.
26:27 Hetzner.
26:27 We moved to Hetzner.
26:28 And they changed this huge thing where they changed their bandwidth and their price.
26:31 And it's like a no nothing sort of deal.
26:35 Like $5 a month more.
26:36 Anyway, I wrote that up so people can check that out on Mastodon.
26:40 And then, yeah, that's it for all my items.
26:43 Okay.
26:43 And then I just got the joke when you're ready for that.
26:45 So let's do yours.
26:46 I don't have much commentary on these.
26:48 I just have a few extra things I wanted to point out.
26:51 Pydantic AI was announced, which Pydantic AI is a Python agent framework designed to make
26:59 it less painful to build production-grade applications with generative AI.
27:03 I don't have really in the commentary about this other than I didn't see this coming.
27:08 But interesting.
27:09 Yeah, I've seen messages or tweets or whatever from people who do machine learning stuff saying,
27:15 you just need Pydantic.
27:16 I mean, a lot of this is like, I got a JSON thing here.
27:19 And then I'm going to call some other thing with other JSON and just suggesting, hey, you
27:23 could probably use Pydantic to make these connections.
27:25 I bet the Pydantic team noticed that as well.
27:27 Okay.
27:27 A couple commentaries on maybe society.
27:32 And anyway, I'll leave it.
27:33 Leave the couple other articles I thought was interesting.
27:37 Blue Sky announced, I guess this is old, this is from August, but anti-toxicity features on
27:43 Blue Sky.
27:44 And I just actually appreciate some of these.
27:47 I already have hit, I had a troll come by.
27:51 And so there's some things where you can, if people, you can detach a quoted post.
27:56 If somebody quotes you and you don't want them to, you can detach yourself from that.
28:00 Oh, interesting.
28:02 I had hiding replies.
28:04 I had some, a troll, you can't like delete replies, but I had somebody just, just idiotic reply to
28:11 something I said.
28:12 And it was obviously just a bot or a troll.
28:14 So you can, you can hide that.
28:16 And as, you know, as Blue Sky grows, we'll, we'll get trolls also.
28:21 If they're, if they're not affecting you yet, they, they may in the future.
28:25 So, so we do appreciate that there's, there are features around to protect yourself.
28:29 So there's, there's that.
28:31 And then, this, I don't know what to make of this really, but Wired, fairly mainstream
28:36 magazine, I think, has released the Wired guide to protecting yourself from government
28:41 surveillance.
28:42 Wow.
28:44 I'm, I just, this is a head shaker of like, I guess we need this.
28:49 I wish we didn't, but wow.
28:51 yeah, there's that.
28:53 So, yeah, I can probably say that about some state governments as well.
28:56 Right.
28:57 Every state's different, but yeah.
28:59 Yeah.
28:59 Yeah.
28:59 Pitting on your gender and things, you know, it's touch and go some places.
29:04 Yeah.
29:05 Anyway.
29:05 So, that's a little bit of a downer.
29:08 So maybe we need something funny.
29:09 we do.
29:10 I don't want to spend all the time going down that, that deep rabbit hole.
29:13 Instead, let's go infinitely down the rabbit hole.
29:16 Yes.
29:17 So check this out, Brian.
29:18 Somebody who goes by bits, very personal on, on, Blue Sky posted what the comments
29:26 seem to indicate is probably a textbook.
29:29 A print is a printed, by the way, a printed textbook on a LaTeX.
29:33 Okay.
29:34 Okay.
29:35 So, I'm going to see you in the index at the back.
29:41 And it says, see page 252.
29:44 I love it so much.
29:45 It's so simple.
29:46 I love it.
29:47 Yeah.
29:47 It's a really good, just like a little Easter egg in there, isn't it?
29:51 Yeah.
29:52 I, I've, I haven't seen it for infinite loop.
29:54 I saw that, somebody did that for recursion.
29:57 In some, yeah.
29:58 Yeah.
29:58 They, they, they're, if you look in the comments, it says that, Carrington and Richie has
30:02 the same, I guess that's probably C or something.
30:04 Yeah.
30:04 The same in, under the index for recursion.
30:07 And, it's pretty, pretty good.
30:08 People love it.
30:09 Yeah.
30:10 No, that's funny.
30:11 And there's somebody that says, for those who can't be bothered, search, Google search
30:15 for recursion.
30:15 Did you mean recursion?
30:19 Yeah.
30:19 I kind of feel bad for people that actually really need to know what that means.
30:23 Good luck.
30:25 Yeah.
30:26 Good luck with that.
30:27 Huh?
30:27 so.
30:28 Wow.
30:29 Yeah.
30:29 All good.
30:29 All good here.
30:30 We know what recursion and infinite loops are, but we're going to break the loop and get
30:34 out of here.
30:34 Right?
30:34 Yeah.
30:35 Yeah.
30:35 Let's break the loop and say goodbye until next time.
30:37 So thanks a lot.