Files
DRACONIANDISCOURSE/S2/E3/2026-04-30 DD_S2E3_prtranscript.txt
T
2026-05-10 22:15:07 -06:00

1517 lines
128 KiB
Plaintext

00:00:30:27 - 00:00:41:39
Unknown
Should we be.
00:01:19:16 - 00:01:41:10
Unknown
We.
00:02:13:06 - 00:02:24:04
Unknown
We.
00:03:01:24 - 00:03:23:12
Unknown
We.
00:03:53:20 - 00:04:05:06
Speaker 2
Let it begin. I forgot to do the thing that I always do like this. This is where we are meant to begin.
00:04:05:11 - 00:04:28:44
Speaker 2
Drag Queenie in discourse. That's not what this is. That's that's just not what this is. I think you're in the wrong bloody stream, Sonny. Drag it. I have a question. Actually. Two questions. Malibu impression. True. You've already outdone Mal grew because, there was only. I mean, I don't reckon he ever had more than one question at a time, though, I don't know.
00:04:28:46 - 00:04:42:56
Speaker 2
Yeah. Yeah, it's a good. It's a good impression. And Reid said, let there be light. I said, let the let let it begin. It's time to begin, isn't it?
00:04:43:01 - 00:05:11:56
Speaker 2
Anyhow, anyhow, so, as you can probably tell by the name of the day, there's going to be, I'm still trying to figure out how I'm going to do, the draconian discourse, like separation and everything. currently we're really backed up. We're really backed up in in the process of figuring out, just getting all of my, what do you call it?
00:05:11:56 - 00:05:30:27
Speaker 2
My my old streams, put up, put up, archived. The stream archived is quite backed up, so that's, no good, but we're going to figure it out eventually.
00:05:31:37 - 00:05:50:47
Speaker 2
It's just a it's just a huge process. But I think it's part of the reason why I really didn't want to continue doing draconian discourse for a while there is because, I had to do a lot in order to get it, get it up. But now I'm, I'm figuring out getting MSPs and, in, in premiere and whatnot.
00:05:50:47 - 00:06:01:01
Speaker 2
So it expedites the process a little bit more. So I've been having a bit of fun with that.
00:06:01:06 - 00:06:29:30
Speaker 2
What's the discord? What what's the discourse? What's the haps? What what is, what what what does that even mean? You're going to have to explain that one to me. So I was meaning to to start up a little bit, words. Like faster. Because, I realized it was, it was kind of difficult to, to do what I've done.
00:06:29:34 - 00:06:53:04
Speaker 2
in nothing. But, what is it? It's it's difficult, like, because when you go to post and you're trying to to edit everything, it's, it's just rough. So trying to make it so that I'm starting out the stream in a way that makes it, better, because I don't necessarily intend to cut up draconian discourse like, massively.
00:06:53:09 - 00:07:12:13
Speaker 2
but. Where does it know what Haps means? What's your major? What do you mean? I don't know what any of this means. Apparently. Evidently.
00:07:12:18 - 00:07:42:09
Speaker 2
Anyhow, anyhow, might as well get into it. So, not certain where exactly I like to start out. because if I didn't find the the steam, the steam review, then I wouldn't have necessarily been doing this. But alas, I found the steam review. So, we're just going to go here. Gotta go sort out here.
00:07:42:14 - 00:07:43:41
Speaker 3
hey.
00:07:43:46 - 00:08:14:06
Speaker 2
I don't know what's wrong with my trachea. There's, like, actually just an issue where it's it's just constantly slamming up. I don't like it. It's not. Not, Good. There we go. It's not good. Let's go with that. so this is, in the wake of, anthropic joining the blended development fund as a corporate sponsor. so they did a whole blog post here.
00:08:14:07 - 00:08:40:50
Speaker 2
This is by Francesco Kte. I'm assuming that's how you say his name. he he may have said his name in his talk, but I missed his talk because, I don't want to tell the story. Maybe I'll tell the story. So I was there, and I. And I found one bloke who, I was speaking to because I thought his art was cool, but I didn't have time to speak with him the previous night, so we ended up,
00:08:40:55 - 00:08:59:44
Speaker 2
Oh, sorry. Sorry for a context. This is, this was at the blender Blender conference in Austin, Texas. so was. I was talking to this bloke. I liked his art, but I wanted to speak to him more about. And so I had time in the in between talks. And so I ended up going over and talking to him.
00:08:59:49 - 00:09:22:30
Speaker 2
and as I was talking to him, I just inadvertently ended up in the middle of all the the goo blender people. and I was just kind of engrossed in that. And so I went to to continue speaking to them as they went inside into the break room area. And so they had like this third room that was kind of a break room, like a, you know, break from the rest of the chaos of the event, like that kind of a room.
00:09:22:35 - 00:09:47:15
Speaker 2
and so we're sitting down and Francesco is right there. He's this, you know, the CEO of blender currently. he's about to go and give his talk. and he's like, I don't know, he's finalizing something. He's getting his getting his stuff together. and we sit down at the book that I was, I was speaking to, as we're sitting down, he says, so we all prefer Maya, right?
00:09:47:20 - 00:10:13:23
Speaker 2
And so Francesco gets up and leaves. Is it like. It's like just said that right in front of the blender CEO is pretty, pretty incredible. So anyhow, this is Francesco. he's he's a cool guy. Is sort of, what's the word? He's he's, he's stepped up as the blender CEO this year. was, appointed as a successor by by ton.
00:10:13:28 - 00:10:39:03
Speaker 2
who was stepping back a bit. I haven't I really wish I didn't miss this talk now because it's online now, but like, it it would be nice to to look at it and see, you know, what he was talking about. Kind of important for the person who's like, actually directing the the tool to be able to, you know, just listen to that guy.
00:10:39:03 - 00:11:04:03
Speaker 2
Right. but anyhow, like this, this, this is the, the entire entirety of the, the, the decision. Let's let's just read it out, shall we? let the foundation announces. Oh, right. So at the at the top of it, this announcement is causing a lot of feedback. We're actively evaluating evaluating it. Reach out to the Blender Foundation directly if you wish to share more feedback.
00:11:04:03 - 00:11:28:34
Speaker 2
And I think this this was the Mastodon okay. Useful. What the fuck is going on. Oh okay. Was a mail two link okay okay. I don't know what the hell was up with that. this is a mail two link. It just didn't it didn't bring me to, like, Gmail or anything like that. We have lot developed.
00:11:28:44 - 00:11:50:42
Speaker 2
Blender Foundation announces that anthropic is joining the blended development fund. Fund it as a corporate patron. The support will be dedicated towards Blender Core Development to maintain and continuously improve foundational features like the Blender Python API, which enables developers and artists like alike to extend and improve the software for custom workflows. In these uncertain and divisive times. We appreciate it.
00:11:50:42 - 00:12:08:35
Speaker 2
Anthropic offering support to the blender project in the form of a patron level membership. This enables the blender team to keep pursuing projects independently and to focus on building tools and building tools for autism. Great as that was Francisco City. every time I say it, I've got to get it. I don't know if I said it properly.
00:12:08:40 - 00:12:33:45
Speaker 2
Blade Foundation's mission remains to empower artists with free slash open source technology and tools. Yet we also maintain APIs for individuals and corporations to extend blender and also beyond what's aligned with Blender's mission. We consider this part of the software freedom that's embodied with Blender's new GPL license. For information about our funding policy. More information about our funding policy is available here.
00:12:33:50 - 00:13:02:40
Speaker 2
Anthropic is an AI research and development company that creates reliable, interpretable and steerable AI systems, and Dropbox's flagship flagship product is Claude, a large language model trusted by millions of users worldwide. Let him learn more about anthropic and clouded anthropic.com. so I just kind of interesting they put about blender right in in that. But so this is a bit of a shitstorm and.
00:13:02:45 - 00:13:32:57
Speaker 2
okay. So I want to understand I want to get into the ethics of it for deeper than many people are willing to. and, and I'm not talking about the pro AI people necessarily. You know, I spoke to, to Andrew Price, at, at beacon about, ethics, not not to like to tell them off or anything like that, but because, what was it?
00:13:33:01 - 00:14:02:08
Speaker 2
Because I think his, his, it's hard to get your point across, when what you're trying to do is, is go from saying, like, you know. Like understanding that to to many people, AI is just the one ring and that nobody should touch it. It's, you know, it's it's, what's the word? It's it's a logical or, not a logical.
00:14:02:13 - 00:14:38:15
Speaker 2
It's it's immoral. Unethical just to touch the tool in and of itself to use the tool is is an unethical. And, you know, it's going to break things. It's it's going to. Well, it's it's not necessarily what it does. It's what it is. I suppose that, a lot of people's issue. because like, I don't know, people just people just are not happy about, AI existing like inherently it's,
00:14:38:20 - 00:15:14:13
Speaker 2
It's very irrational. And, and I was trying to say to Andrew because, you know, he's a professional and he's not not exactly interested in the ethical, conversation or the philosophical conversation because it's like, well, you know, we've all got work to do. That's that's the main deal here. But I found a very interesting, like talking to him about this was a lot easier because like, you know, I recommended him a video that that talks about, you know, there actually is an ethical, position for AI.
00:15:14:18 - 00:15:48:58
Speaker 2
and, and not just like. You know, people people are already not very, perceptive. Perceptive. They're not very friendly towards the idea of AI, especially when it's it's sort of encoded with all of this, like AI, bro speak. It's interesting. Like many of the people just who were speaking at the, at the conference, people who were, you know.
00:15:49:03 - 00:16:18:25
Speaker 2
People who didn't have nothing to lose, like, like this is one of the points that he had made here was that, you know, I say this as someone with a lot to lose. but but he was actually talking about generative AI and not LMS. many of the people making arguments against LMS like there's a or. Sorry about generative AI, that there's this, there's the art theft argument, but like LMS, really?
00:16:18:34 - 00:16:46:48
Speaker 2
I don't even think you could say that it's the same, it's it's very interesting to see sort of the way that people struggle with with this topic, but, from blender guru to slob guru, should everyone just now cancel their polygon subscriptions? Because I can just do it. Hopefully the AI companies are paying you well and he had another, another comment that was effectively just the same drivel.
00:16:46:53 - 00:17:05:24
Speaker 2
And I said, why didn't you come to beacon and say this to his face? He literally address this point in his talk. There was a whole panel on the topic and tons of one on one conversations. This irrational screeching is the precise reason your ethical qualms fall on deaf ears. So it's like.
00:17:05:29 - 00:17:37:35
Speaker 2
There's nothing but like, nothing but screeching, nothing but a rational. like screaming and crying constantly about it. There's nobody who's actually trying to make a a an ethical case. They just invoke ethics. And because they they've invoked ethics, they reckon that it's perfectly fine for them to just, you know, move on. Like, it's like I'm I'm correct now because I've invoked ethics.
00:17:37:35 - 00:17:48:38
Speaker 2
It's like, okay, but you have an entire a, philosophical schema that you haven't actually laid out. There's no,
00:17:48:43 - 00:18:14:17
Speaker 2
There's there's no what what is the bloody word? What's the philosophy where there's there's no. What a civil syllogism is. Is that what I'm thinking of? There's just not a, an argument that is being made where you've got all of these premises, they've the sound premises, and then you've got the conclusion from the premises, like you're not trying to make an argument, you just virtue signaling, like, that's it.
00:18:14:22 - 00:18:42:43
Speaker 2
This is not going to help anybody navigate this type of complex issue. Just screeching about it constantly. Like all it's going to do is it's going to make people think that there's a certain amount of social pressure that will keep them from trying to say the things that they wish to say, or do the things that they wish to do in reality, I don't know, with their own property.
00:18:42:43 - 00:18:51:02
Speaker 2
Right. Like, and or it's going to have people.
00:18:51:07 - 00:19:13:54
Speaker 2
Just like, shuffle their feet along with it when it comes to, doing those things like it's I think anybody who's making arguments about this topic or any topic ought to have the best arguments they can possibly make. If you're going to invoke ethics, you should actually make an ethical case. And the ethical case that you should make should be as strong as it can possibly be.
00:19:13:58 - 00:19:36:23
Speaker 2
the screeching just doesn't doesn't achieve anything. And it's a big part of the reason. Like because before, and Andrew like this tweet at Twitter. My but like before when I was speaking to him, I didn't quite understand how bad it was. And maybe it was just like I wasn't I was out of the loop or something on, you know, how blended people could be.
00:19:36:23 - 00:19:53:17
Speaker 2
Just as cringe is all the other internet people could be like, there's just as toxic a space as, as any other, right? Anyhow.
00:19:53:22 - 00:20:29:24
Speaker 2
Francisco city. Yes. Yes, that's his name. Let's let's go with that. What is Google into? gallbladder is a an anime. It's GPU engine, is what they called it, but it's a fork of blender. but around blender 4.3, that was intending to enhance, many of the, the features within blender that weren't currently supported that they were trying to do so that they could animate, they could make anime in, in 3D as if it was a 3D anime.
00:20:29:29 - 00:20:48:33
Speaker 2
and so I, I was there, I met Dylan, the goo man himself. it was pretty cool. I, I didn't really speak with him much because, like, you know, he's popular. He's got people, underneath him and and to the side of him, and they're just everywhere, like, you know, everyone's wanted to get a little, little piece of the good man, right?
00:20:48:37 - 00:21:01:46
Speaker 2
So, 3D hentai, I know that is, not what they do primarily. I don't think.
00:21:01:51 - 00:21:20:08
Speaker 2
That is it was pretty cool to talk with, with loads of those those people. I was talking with one of the guys who, he was currently working on some GPU engine stuff, so it was pretty neat. anyhow, what what was the next point here?
00:21:20:13 - 00:21:24:04
Speaker 2
Well.
00:21:24:09 - 00:21:50:10
Speaker 2
I reckon we just we just jump to the to the smaller one next. today anthropic joins. You know, it's just them posting about, the, what do you call this, the, the, you know, the same that the, article that they done before. I don't I don't know why some words are kind of struggling to the tongue, do they?
00:21:50:15 - 00:22:12:29
Speaker 2
But, you know, it's it's just them announcing. So I think one of the things they didn't mention, it was like, you know, $1.5 million or something like that, right? So it's a substantial amount of money. if you actually dig into what they had mentioned before, their, what is it that, clauses, everybody who donates to blender is making nothing but a donation.
00:22:12:34 - 00:22:41:53
Speaker 2
Like, that's that's the major thing to understand here. I made a donation recently. you know, everyone who makes a donation is effectively just demonstrating that they want the project to continue forward. This. You know, I suppose faith in goodwill in in the projects, goals and and achievements and, you know, or just like, thank you for allowing me to use this thing free and open source and everything.
00:22:41:58 - 00:23:13:30
Speaker 2
no, none of the corporate sponsors who have contributed have any say in what blender does. There's no executive power, no directorial, decision making that that can be had. just from that, Noah. Sir. Groves, thank you for the follow. Through to milking porn. Yeah, I don't think so. I don't think that's what they do. I think, I mean, you could use blender for that.
00:23:13:35 - 00:23:50:44
Speaker 2
Many people do that. That's just not, not not terribly good at Android. Now, so anyhow, just just this this replaces on Mastodon, which is a this but bloody federated. yeah. No algorithms add a clickbait and site. It's an interesting little thing there. It's kind of funny that it's 4.5.8. That was the latest version of of the latest blender patch on on 4.5 LTS for a while there.
00:23:50:49 - 00:24:28:00
Speaker 2
anyhow, coincidentally an entirely coincidental, but the Python Software Foundation show that respects that respect of its community is more important than money. You can do the same now. This was a blog by the Python Software Foundation. They withdraw, withdrew a $1.5 million grant proposal, or a proposal to the US government grant program. So I think what they did here is they refused to take $1.5 million because that $1.5 million came from the government.
00:24:28:04 - 00:24:57:59
Speaker 2
It came with many strings attached. there were contractual obligations that they were going to be required to fulfill if they took that money and it potentially introduced, liabilities later on down the line. And they could have been punished by the, you know, the bloody federal government if they didn't comply with whatever it may be. so clearly that's that's already a concern just from a developmental standpoint like this is an open source project.
00:24:57:59 - 00:25:19:24
Speaker 2
Most of the sort of open source copyleft, projects don't want their hands tied. Like that's kind of the point. so taking a grant like this would be a bit of a monkey's paw, because already they're going to be required to do things that they may not have wanted to do. So this is the type of thing that you negotiate out.
00:25:19:24 - 00:25:51:18
Speaker 2
You try to figure out, okay, this isn't in our best interests, etc.. the main thing here is that, do due to the fact that this was under the Trump administration, because this was in October 2025, this is under the Trump administration. They were going to require that Python, not have NDI programs. and that was like one of the courses that that they had decided here that they didn't want to go along with.
00:25:51:23 - 00:26:11:18
Speaker 2
and I found that really interesting. I tried to do some digging and figure out, like, why, like in what ways that Python devices die, and I didn't really find much. it seems like a weird sort of signal, because it's like, you know, Python in and of itself is just an open source coding language, right?
00:26:11:18 - 00:26:48:52
Speaker 2
Like, it doesn't, you you'd be hard pressed to figure out what, like, where race fits into this sort of, discussion. This, this field. It doesn't really make much sense. but, like, I was, I was trying to get some sort of an explanation for it and sort of the, the, the consistent line that I kept getting in reply was effectively that, well, like, there were X amount of people who haven't had access to, like.
00:26:48:57 - 00:27:21:37
Speaker 2
Development of of anything or to computers or what have you. And so they're trying to focus on how, contributions, you know, trying to get it so that people who could be capable of contributing to, development of Python in particular, how those people could who could be, contributors can become contributors. And it's kind of interesting because like, that just sounds like outreach to me.
00:27:21:37 - 00:27:50:47
Speaker 2
Like. Right. It just sounds like we were trying to see if we can get more people involved in contributing to this thing, like, can we get more interested in our our open source project? That doesn't sound like you need to, what's the word? It doesn't sound like that requires you to, invoke neo-Marxist terms, right? That doesn't it doesn't really make sense that they would be talking about Dei in, in that context.
00:27:50:47 - 00:28:11:45
Speaker 2
And so it's like, okay, well, how do they actually do Dei? It's like, well know that they're trying to see if people who don't currently contribute can contribute. And it's like, okay, but why? Why does race need to factor into that, and why should it factor into that? If you're just looking for competence and it's like, well, it's because competence is only based off of like what people have proven to be able to do.
00:28:11:45 - 00:28:49:07
Speaker 2
So it's like you contribute by writing this code and then there's your, you know, your, your reputation and, and the networking and everything. And it's like, why is how the hell does this even help you do things faster? Like, it doesn't? Does it really make any sense? Because, like, there's nothing that should require you to look at, I mean, and and loads of, contribution classically and, and, you know, just for forever your contribution has been like, anonymous.
00:28:49:12 - 00:29:11:40
Speaker 2
So, like, why would you even need to know, who is making the pull request? And unless you just recognize the handle and then like, oh yeah, I've, I've no, this this contributor, they've, they've done some good thing in like, I don't know. So there were other things like values they were going to talk about. I'm not here to talk about dei.
00:29:11:45 - 00:29:42:25
Speaker 2
but you can see how this is a completely different issue. It's it's just not comparable in any way. And I had a bit of a, bit of a thing that I wanted to. I already wrote it out, so I may as well just repeat it. I'd said, those two situations are not comparable. SF rejected a $1.5 million government grant due to the binding legal conditions that restricted the freedom of the foundation's goals, and could have produced legal punishment down the line.
00:29:42:30 - 00:30:08:12
Speaker 2
SF didn't want the hands tied. It didn't really have anything to do with respect of its community. By contrast, Blender's corporate patrons have no control over the direction of blender. There are no binding conditions, policy constraints or direct control. The only change this introduces is a potential incentive and or cultural drift, but that's just but that's a completely different concern than those SF than than those SF had about their grant.
00:30:08:16 - 00:30:31:04
Speaker 2
So that's that, that that was my take on it. I wrote that before scrolling down and seeing that Francesco had effectively said the same thing. And I mean, this is just this is just good stuff, right? I was very impressed by that move, which shows that the project is not willing to compromise its mission. Blend dev fund agreements are donation and do not come with requirements.
00:30:31:04 - 00:31:01:28
Speaker 2
As a U.S government grant program. However, it also looks like SF received 1.5 million donation from anthropic earlier this year. And there it is. Like it's the exact same amount of money. It's the exact same company. So it really does make you wonder, like how the hell this is supposed to make it like a difference in what's being evaluated here?
00:31:01:31 - 00:31:40:59
Speaker 2
Like just does it, does it really seem to add up because it's like you're saying blender is bad for accepting the evil money from the evil people, but Python is good because they didn't take the money from the Trump administration. which if anything, like like, let's be real about this, taking money from any government would be altogether worse than taking money from any private corporation like the the government's existence to the existence of the state necessitates force.
00:31:41:04 - 00:32:10:21
Speaker 2
It's like it's it's a coercive entity. It's baked into its existence. So any of the money that you would be getting from them is coerced income, right. Like that's not like like, I think someone had said, it's dirty I bro money dirty, unethical I bro money and they shouldn't touch it. It's like okay, but like, what about dirty unethical, you know, global elite Epstein fucking money, right?
00:32:10:21 - 00:32:34:37
Speaker 2
Like, what about that? Like this. Does that does that not compute? Like, I don't know people people are going to talk about that when it comes to like refusing to accept a government grant because like what I find kind of interesting is that, okay, well, we're going to get into that. We're definitely going to get into that. I want my hands tied, pansexual pride.
00:32:34:42 - 00:32:42:25
Speaker 2
Listen, okay. Nothing. I've got nothing. Global elite Epstein money. Yep.
00:32:46:15 - 00:33:10:08
Speaker 2
That's that's that's about all. All I have to say about that. Right. this one made me laugh out loud because it's like, oh, he's actually serious. Like, I. And I know he's serious about this, but it's just it's so foolish that it's hard to even believe. I had no idea Python took their money to guess. I'll stop using Python as well.
00:33:10:08 - 00:33:35:40
Speaker 2
Thanks for the info. It's like, how long until you're just going to run out of absolutely every single technology that exists? Like, I'm pretty sure that that GPUs have been, machine assisted. You like utilizing Lmms for a long time, like the development thereof. I mean, like to you can't create a GPU from scratch. It's not easy.
00:33:35:54 - 00:34:00:45
Speaker 2
It's not an easy thing to do at all. Like you need a remarkable amount of compute power just to be able to emulate the GPU before you create it. Like that's part of how they even do this, right? same thing goes for the CPUs. They use the machines to create the machines. Like it's not bloody easy to do, and I'm pretty certain that they've been using Lmms to do it for longer than like the term Lem existed necessarily.
00:34:00:45 - 00:34:22:19
Speaker 2
Maybe. Maybe they weren't lmms technically, but like, they've used machines to build machines and to build tools and all of this shit for a really long time. so it's it's bizarre to me why anybody would have a problem with these systems. And in the same sense that you can have your computer, you know, you've got a thing that you need to do right?
00:34:22:33 - 00:34:48:16
Speaker 2
So you go on your computer and your computer is dumb. It doesn't know anything. All it can do is it can crunch numbers. Right? That's why it's called a computer is because it computes. So you go in and you figure out, okay, based on this, based on my understanding of the kernel and the instructions set and all this stuff, I can build a program, so that my computer can run it, and then it'll do that task that I designed it to do.
00:34:48:21 - 00:35:12:55
Speaker 2
and I won't have to worry about doing it manually anymore because the computer can do it, like, say, it's like double ledger bookkeeping or something like that, right? Like that. That has existed for a really long time. Like it's it's probably existed going back to the, back back to ancient Mesopotamia. Like they been effectively using spreadsheets in their commerce.
00:35:12:55 - 00:35:31:56
Speaker 2
Right. So it's a pretty big deal to be able to, not have to do that with pen and paper or worse, you know, like in cuneiform, like gouging it it to stone all day. Right? That takes a lot of effort, let alone the effort to actually know what the numbers mean and what to do with them and how to calculate them.
00:35:31:56 - 00:35:58:33
Speaker 2
Like, that's that's a lot of, a lot of work. And obviously, people had to know that in order to be able to create a machine that could do that. But people still know how to do those things. And, you know, if if we got that, what is it that that solar flare that causes a, EMP, just an electromagnet storm to be to bombard the Earth, and we would lose, like, practically everything that's digital currently.
00:35:58:33 - 00:36:19:37
Speaker 2
We'd have to start from scratch. So it's like, you know, none of that's good. But it would be difficult regardless, because you don't have the ability to outsource, information on to, like any other physical medium. Right. Like this is this is an issue that happened in my head, my home, home world. But I suppose we will we will get to get to that.
00:36:19:37 - 00:36:54:40
Speaker 2
Like the simply put, a, a pre technological society is pretty fucking difficult. and I suppose I'm talking more about like, you know, advanced technologies where you begin to actually have abstraction outsourced to, technology. That's when you get machines. it's it's not just like, what do you call a simple machine where you're using, like, a, the, the laws of physics effectively to be able to get leverage over things so that you can do things easier in that way.
00:36:54:45 - 00:37:27:47
Speaker 2
inherently, like you have to take it down to, to first principles, right? Like, all technology and and of the. Yeah, all technology is, is a means by which man uses to. To achieve his ends more efficiently. That's that's all the technology is. It's an implementation of, What what is it called of,
00:37:27:52 - 00:37:55:53
Speaker 2
Like physically instantiated labor, and rationality combined. So you don't have to work as hard on lugging that rock up the hill because you've got it in a cart so you don't have to drag it the entire way. Something like that. You know, the cart is the technology in that case. So inherently you you must apply logically what you're applying to technology, to any particular technology, you have to be applying that to technology per se.
00:37:55:55 - 00:38:02:33
Speaker 2
Otherwise, the philosophical basis of the,
00:38:02:38 - 00:38:28:46
Speaker 2
Of your ethical theory, which is a little bit circular, I know, but like, it just falls apart because you haven't got a clear understanding of. Any, any of the mechanics of this, logically speaking. Right. Like, you've got to be able to lay it all out so that you know what exactly the issue that you're trying to point out is.
00:38:28:51 - 00:38:56:00
Speaker 2
and, and I think a lot of the issue that people end up in is they, they're going from, they start from ethics. And the reason they start from ethics is because it's just the easiest place they can go to after having an irrational, knee jerk gut response to something. but it's actually more difficult to go from the ground up and start for first principles because, like, metaphysics is not easy.
00:38:56:00 - 00:39:26:39
Speaker 2
I, I avoided metaphysics for a long time because I was more interested in ethics, epistemology, and law. epistemology less so because it it bridges between ethics and, and metaphysics. And so metaphysics was like, well, it's interesting, but like, it's not really whether the meat and potatoes of philosophical issues are that's like, that's an ethics. ethics is important, but you can't have ethics without, epistemology.
00:39:26:40 - 00:40:10:01
Speaker 2
Metaphysics. Like, you have to start with the basics first. And most people going backwards, you know, you start with your ethics, which are downstream of your own just whims. And then from there you try to append a, a metaphysics and epistemology that works for your, for your whims. Frankly, you're trying to bend, philosophy to your whims and not the other way around is like, you should you should cast your whims aside if you have any, any sort of devotion to truth whatsoever, if you have, frankly, integrity, then you ought to cast your whims aside and and attempt to get to the truth.
00:40:10:01 - 00:40:14:28
Speaker 2
You know, the the brass tacks of things.
00:40:14:33 - 00:40:45:19
Speaker 2
I could do it. I could make a GPU with my hands. I'm very dexterous. I mean, that would be impressive. But, I don't think dexterity, with all the dexterity that is possible for, you know, a a human being to have, you can't still can't build something like that is like that's definitely a physiological limitation. but anyhow, like, this entire thing is like, I'm going to stop using anything that uses this particular form of technology.
00:40:45:27 - 00:41:12:36
Speaker 2
Well, this is is, it's a modern form of Luddite ism, right. Is is that is that the way that it's it's supposed to be said. These people are Luddites. They believe that any anything that makes man's life easier today is wrong and not because of some ethical concern, but because they hate mankind, they despise rationality. They despise rational creatures.
00:41:12:36 - 00:42:00:37
Speaker 2
They despise those who live in their rational self-interest per se. and I'm not quite certain what where they absorb that from. I certainly know that nihilism. And, just pathological subjectivism. is one of the great poisons, one of the great philosophical poisons, that is, that has made itself, prevalent in Western society for the past. while let's say, but it's hard to hard to try and understand, like, who put the, the idea in the head that technology that is, you know, you ought to have a pro human ethic.
00:42:00:37 - 00:42:33:45
Speaker 2
Right. Any in any ethic that you're prescribing ought to be pro life. Right? You ought to see the thriving of, of rational creatures, right. As your, as the goal of your ethic. Otherwise your ethic is an anti ethic. It's an end in end. And the only way to back up some that that sort of anti ethic. As with anti principles it's.
00:42:33:50 - 00:42:51:58
Speaker 2
It's difficult to know where it's like like what they think is is the logical conclusion of what they're saying. well it's difficult to get them to face it. Like I don't think that they can. Right.
00:42:54:19 - 00:43:33:47
Speaker 2
Because like, you just go to refuse to touch any technology that could possibly have gone near this new technology because it does something or other to, Make life easier. It's it's like they genuinely don't like life being made easier. And this has nothing to do with, what do you call it with, with, Intellectual property theft or with, like, I suppose it's relevant.
00:43:33:52 - 00:44:01:08
Speaker 2
but I think even under that premise, it's not relevant because they don't know, like, how these tools work, which is embarrassing for people who have, like, made tools, like ever, because you understand how your tools work. Surely if you made them, but they don't understand how the AI like being trained off of code bases is not plagiarism.
00:44:01:08 - 00:44:45:02
Speaker 2
They don't. They can't fathom that. That does like that doesn't make sense. And also considering that it's like, oh yeah, it's plagiarism, like my code is being plagiarized. because like. It was used in the, in the training set. So to this model or whatever it, like it, it's, it's a completely separate discussion. Entirely from generative AI because generative AI, it's at least more confusing to people on the surface where they look at it and they say, well, this appeared to have like spewed out it spat out this result.
00:44:45:07 - 00:45:15:46
Speaker 2
that's similar enough to what I made. and therefore it must be plagiarism. But it's like you can already do that. You can just copy and paste and these are the same a lot of the same people who are making, like, NFT shite, like, like just, just absolutely losing their minds about NFTs. their entire thing was will I already like, why would I pay money for your NFT?
00:45:15:46 - 00:45:42:10
Speaker 2
Why would I pay like three Ethereum for your NFT when I could just, you know, copy and paste? And I've got it right here. like, they didn't think that was theft, but they think something that does not copy and paste like, by definition, it like it's effectively doing what humans do and abstracting what, you know, as a certain amount of elements of, of what it's being prompted to do.
00:45:42:10 - 00:46:11:23
Speaker 2
And then, you know, in a hyper advanced like, well, I don't know if hyper advanced, it's quite primitive in comparison to, what could theoretically exist in the future, but like, or it's quite primitive in comparison to human abstraction, for example, especially on the, on the higher end of the IQ spectrum. but it's, it's not like it's,
00:46:11:28 - 00:46:40:10
Speaker 2
It's, it's doing what people do and it's abstracting to, you know, a lesser degree. And then going through this, this advanced input output, it's just a, just a very advanced input output machine. And that's, that's all that code has ever been is just an input output machine. This is just more advanced, than than it has ever been like, no, no,
00:46:40:15 - 00:47:21:58
Speaker 2
Code has, has ever been able to, input output like this before. Right. But lots of have existed for a long time. And people, like, didn't, weren't necessarily upset when they had like, these absolutely slow like bots. like, there wasn't much that you could do about a bit like, like, I don't know, you're upset if you're on like, PayPal or whatever, and you try to get, support and the support is a bot and the bot is just not intelligent, you know, like, that was frustrating, but it was frustrating because it was bad.
00:47:22:03 - 00:47:47:57
Speaker 2
But it's it's odd that now that it's actually useful, people want to destroy it. And every single ethical argument that they make is a contradiction. Like, it really doesn't make sense. Anyway, thank you for taking time to respond to individual comments. I understand the context between the two is different in that, and that in theory, donations to the blended dev fund come with no strings attached.
00:47:48:01 - 00:48:15:41
Speaker 2
it's interesting that it says in in theory, it's like. No, like de jury like that is what the agreement is like. It's it's not to say that like, there's not some kind of, incentive of, cultural shift that is potentially downstream of this. But like de jury, there is there are no strings attached. That's just, that's final.
00:48:15:41 - 00:48:23:10
Speaker 2
Like it's it's not in theory, it's de jury which is different.
00:48:23:15 - 00:48:53:01
Speaker 2
They took a jobs abstracting bro. Think this the amazing digital circus is this, is this just another show that's absolutely ruined any given like, terminology? It's a it's a psychological term. It's it's a phenomenological term. Like like my God, use it anymore without referencing the buddy show. I haven't seen Filthy Maya, animator show.
00:48:53:06 - 00:49:07:36
Speaker 2
Is because the AI will take over the world. Skynet. Well, that's the funny thing is that, like, AI does not have Will currently like it. It's it's. I don't even know how you would necessarily get to that point.
00:49:07:40 - 00:49:10:10
Speaker 3
Oh.
00:49:10:15 - 00:49:29:37
Speaker 2
Yeah, it is, but the SF statement explaining the withdrawal from the US government grant program. In the end, however, the SF simply can't agree to a statement that won't operate any programs that advance or promote diversity, equity and inclusion as it would be a betrayal of our mission in our community. The refusal to betray that community is the point I wanted to make here.
00:49:29:42 - 00:49:59:01
Speaker 2
It is difficult to imaging very, very well. Richard here. imagine a sponsor whose values are more in opposition to those of the blended community of creators, an artist than anthropic. So what I find interesting about this is that I think most of the people and, Andrew Price also has made this observation. I think most of the people who are making this kind of a point, that this is out of sync with the community, the community doesn't agree.
00:49:59:06 - 00:50:22:34
Speaker 2
I think most of those people of kind of committing the sin of trying to speak for everybody else, it's just like, am I not part of the community? Because, this statement does not speak for me. Like I think the blood of I think more people joining the blended development fund is a good thing. Believe it or not, they have the ability to do more with those funds.
00:50:22:39 - 00:50:58:12
Speaker 2
Like this is not bad, just inherently. And beyond that, having a moral context protocol, all that really does is make it easier for people to be able to automate things without having to learn code. And inevitably, these people have to revert to this absolute batshit position that every single artist on earth, every single artist who uses blender, let's say in this particular instance, they all want to be developers.
00:50:58:17 - 00:51:31:28
Speaker 2
Every single one of them wants to learn Python. Every single one of them. needs to know the, you know, they need to know the, the BP, library. They need to know it for the exact version of blender they're on. and they need to be able to, to write scripts in it so that they can execute, you know, after, bug fixing and all of that, and, and do exactly what they need to do and spend all of the time required to learn how to do all of those things, and they should be able to do that.
00:51:31:33 - 00:52:06:53
Speaker 2
on their own, without consulting any sort of, AI tools whatsoever. And what I wonder is at this point, why why are you even using a computer? Why why are you using any technology whatsoever? Because, like in theory, what you've done is you've taken away somebody's job. because somebody could have been doing that or you've you've plagiarized this or you're using it without, like, I don't know, there's got to be someone who did like an open source tool.
00:52:06:54 - 00:52:38:39
Speaker 2
It didn't do a pattern or whatever. Was it getting paid to do it? It's like they there's no consistent ethical framework here. And it's and it's frustrating because like, all you end up getting like in this case is it's like the, the blender community of creators and artists like, I think you're in an echo chamber. I think the, the vocal people don't like AI in general, but I think the people who don't like AI and have refused to interface with it.
00:52:38:43 - 00:53:08:24
Speaker 2
That they're missing out on what they could have and and what is what they could have a quicker workflow, greater optimization, so that the technology doesn't get in your way. Like, imagine every single time you sit down to work on something, the the, the tool begins to fight with you. Like that's one way of thinking of it. It's more like, you know, here's this incredibly complex tool.
00:53:08:29 - 00:53:27:57
Speaker 2
I would rather be, you know, a little more user friendly. Well, now the tool can talk to me and tell me how it ought to be used. It can also guide me through this type of thing. And so many people like there's all of these, these concerns that just they get like thrown out of nowhere, but they're not logically connected at all.
00:53:27:57 - 00:54:01:35
Speaker 2
It's it's just pivoting. It's just jumping from one lily pad of, of irrationality to another because it's like, well, people could lose this. like, you don't want people to lose their ability to to do these things. It's like, I understand, like the strategic implication that the, invoking that as, as enterprise said, never, never automate any, any skill that you do not wish to lose.
00:54:01:35 - 00:54:06:49
Speaker 2
And that's fair enough.
00:54:06:54 - 00:54:53:01
Speaker 2
But there's plenty of skills that people are not interested in learning whatsoever. And now there is technology that allows you to be able to benefit from utilizing those tools. without being required to learn it all before it just just to begin, you know, you don't have an entire mountain of knowledge to, to climb in order to be able to use this, all you've got is a, like, like you've got a gondola that goes all the way up, and everybody's just upset that there's a gondola and they want to destroy it, and they don't want to kill the people who who constructed it and planned it and invested and put it up.
00:54:53:06 - 00:55:17:30
Speaker 2
And it's like, the only reason that you would want this to happen is because you hate that other people might be able to do things easier. Now. And like, this isn't the art thing, right? It's not about art. It's not about like, you know, you don't know how to pick up a pencil and paper and and you can just, like, tell a computer to hallucinate for you.
00:55:17:31 - 00:55:37:34
Speaker 2
It's not even that. It's literally just I can know, logically speaking, exactly what I wish a program to do, but not know how to communicate to the like. You know, I don't know how to talk to the computer right? I don't know how to tell it to do precisely what I'm trying to do.
00:55:37:39 - 00:56:01:28
Speaker 2
Like that is been how working with computer systems has been for a really long time. For most people, it's like, I know you can do this with this. I mean, this is a just limitless possibilities of what you could do with this technology. But I don't know how. And so, you know, YouTube tutorials existed for forever and still exist.
00:56:01:28 - 00:56:33:50
Speaker 2
And there's all of these different resources. But like, it just doesn't it doesn't make sense to me why people would be upset. That it's easier. And people say like, well, if you don't, you don't know what your code does or you don't know how to code. You couldn't code without this. And it's like, I understand that there were concerns with that, but I don't think that those concerns actually make it unethical to do it.
00:56:33:50 - 00:56:57:59
Speaker 2
Like it's in the exact same sense as somebody choosing to utilize comparative advantage. Like if I've got a business and I don't want to be an accountant, I want to run the business and not just like worry about all the funding and whatnot. and, you know, taxes, all that shit. Like, I hire an accountant and the accountant takes care of it for me.
00:56:57:59 - 00:57:18:06
Speaker 2
It's like, oh, okay, but you could have been that accountant. Like, why would you want to, to outsource that skill to somebody else? Like, why don't you do it? And logically speaking, would you go there and help with if you follow that, that line to it's to its furthest extent is.
00:57:18:10 - 00:57:22:41
Speaker 2
You've got a bloody,
00:57:22:46 - 00:57:29:10
Speaker 2
What is it?
00:57:29:15 - 00:57:59:15
Speaker 2
If you've this, this implies and I think it, it ends in a primitive ist society, let's say, because, you're going to end up like, like the Unabomber, right? You're going to be in a forest with in a shack, like going off of nothing but the skills that you've retained from what you've learned in your life. And that's it.
00:57:59:15 - 00:58:35:52
Speaker 2
And that's all you get. Because if you build something, it must be immediately destroyed due to the technology. Like like it's actually it's worse than that because like, any amount of technology is just being regarded is what's being attacked here. There's like like it's just kind of funny. I've built a program like, let's say with blender, I've built a program that allows me to create a 3D model and then put it simulated camera into a three dimensional.
00:58:35:52 - 00:58:58:33
Speaker 2
See, this is all completely just conceptual, logical. It's just in in a digital realm. Right? it's not real. It's not physical. and I could rotate the camera around this character, and then I can make that character pose, and then I can have them pose over time. and then I can, you know, bounce a bunch of lights in this simulated environment, you know, with tons of compute power.
00:58:58:33 - 00:59:29:26
Speaker 2
And then I have to, however much time, like, wow, I've got an animation and it's like the why don't you just, like, do it in, in real life? What? Why didn't you do it in stop motion? Why didn't you, you know, build a clay model? Why? Why didn't you animated in 2D, in traditional like. And I feel like you could do the exact same thing going all the way down where it's like you could have just, like, told the story, right?
00:59:29:31 - 01:00:00:16
Speaker 2
Like, why did you animated, why did you animated? You could have, like, done it on, like, whatever. Like going all the way down to like, the more, primitive mediums. And by primitive, I mean, like, it doesn't require, like, loads of knowledge on how, like, micro processes work in order to do it. Like you can do it on stage, or you could just tell a story around a campfire, like all of all of the shite.
01:00:00:21 - 01:00:15:22
Speaker 2
Like eventually you just end up in primitivism and that's that's an anti-life ethic. Technology, per se, ought to be regarded as.
01:00:15:27 - 01:00:56:37
Speaker 2
The means by which man can achieve his ends better, more efficiently. So. I would regard technology per se as, like neutral to positive, right. It's either it doesn't really have any sort of an ethical, implication attached to it because it's not inherently ethical or unethical. It's ends ends that are unethical, not means, which isn't like I, I'd be somewhat friendly to that or it's that it's neutral to neutral to positive.
01:00:56:42 - 01:01:09:22
Speaker 2
Like. Simply by dint of having means you've already improved man in in some way. Like.
01:01:09:27 - 01:01:41:39
Speaker 2
It's just interesting that, like, people end up taking the, the route like. Like let's let's take just Microsoft Excel, for example. Like, I've built this program and what it does is it makes it so that I can put numbers into it. And then, you know, put it out in my spreadsheet and then add things up and it, it'll, it'll calculate it automatically so that it can.
01:01:41:43 - 01:02:05:51
Speaker 2
So that it can, help me do my finances more easily. And now I'm going to release this publicly so that everyone can have access to it, and hopefully they'll donate to me in return. You know, I've I've created this, or maybe I'm, you know, putting it out as a product. Right? Whatever it is, like, people would look at that and be like, well, it's.
01:02:05:56 - 01:02:30:58
Speaker 2
Bad that you've done that because now, like, you don't now, now it's easier for you to manage your finances or something like that. Like, like I need to understand like what the actual argument is here because it's just it isn't. It's not making sense. What if Epstein joined the fund? He's dead. He's got to how to do that?
01:02:31:03 - 01:02:55:57
Speaker 2
people don't like it because I bad. Don't you understand I bad? That's that's is kind of what it is like. I one of the things I was thinking about yesterday was, I reckon it's literally that the conclusion is the premise, like AI is bad because AI is bad. So you've got like, I think you call that a tautology, but not actually, I don't know.
01:02:55:57 - 01:03:23:19
Speaker 2
I don't know what you call it. It's circular logic. Let's, let's call it that. Unabomber mentioned. Let's let's not mention the Unabomber again. What if I created a robot that throws acid on people? Or the. That would be aggression that that that would, you know, you've you've created the, the rights violating machine like.
01:03:23:24 - 01:03:39:33
Speaker 2
Like its entire purpose is to go out there and murder you. Well, okay, I suppose there could be situations where it wouldn't be aggression. Like if you had to patrol your property. I suppose that'd be. That'd be fine.
01:03:39:37 - 01:04:02:15
Speaker 2
Though many people think that that's not the case for some reason. Like, you can't protect property with lethality, which is kind of ridiculous, actually. There's a there's a great little, little Ayn Rand quote on that. I had replied to this recently.
01:04:02:20 - 01:04:22:18
Speaker 2
This one, the modern mystics of muscle, who who offer you the fraudulent alternative of human rights versus property rights, as if one could exist without the other. I'm making a last grotesque attempt to revive the doctrine of soul versus body. Only a ghost can exist without material property. Only a slave can work with no right to the product of his effort.
01:04:22:22 - 01:04:43:57
Speaker 2
The doctrine that human rights are superior to superior to property rights simply means that some human beings have the right to make property out of others, since the competent have nothing to gain from the incompetent, it means that the right of the incompetent is to own their betters and to use them as productive cattle. Whoever regards this as a as human and right has no right to the title of human.
01:04:44:02 - 01:05:04:40
Speaker 2
It's like, is this is that just bloody poetry? Beautiful? and I had said gold speech never ceases to hit every single mark. It's good. And it's kind of funny because, like, I'm into Ayn Rand for her, philosophy and not for her her fiction, though. I'm trying to get her to do it. Well, it's it's on the list.
01:05:04:40 - 01:05:25:17
Speaker 2
I've got too many things to, explore in terms of fiction. So anyhow, let's let's get to the maiden video. Does it? And this, this will be the one that's, That's fun. I suppose. So, as far as I'm aware, people have been, review bombing, Blender's steam page, and I don't think it's had much of an impact.
01:05:25:17 - 01:05:48:13
Speaker 2
It's still overwhelmingly positive, and it's hard to not be overwhelmingly positive when you've got a free and open source, DCC that does everything and more than other DCC do, and for free, like, like that's the main thing. It's free, open source. It's all out there. You can just use it. Hi. This review is about the anthropic donation.
01:05:48:18 - 01:06:09:45
Speaker 2
This is an interesting thing to start out a review for the platform, or for the application for like a while. You'll get into it. This is mostly to explain whatever the fuck is happening in a sensible manner, with a few of my own thoughts and feelings. If you're curious about how I feel about the software, I think blender is very good, although I'm not as avid, not as average with 3D stuff as I am 2D stuff.
01:06:09:50 - 01:06:28:55
Speaker 2
how could you be not as average? Does that mean you're better or worse? Like that's. There's not a clear way to communicate that. I think blender is very good. I it's interesting. It's like I'm not as good at 3D as I am with 2D 48.3 hours on record. You'd think that that's like, you know, a decent amount.
01:06:28:55 - 01:06:56:24
Speaker 2
It's like, yeah, yeah, yeah, that's that's a decent amount. That's that's not nothing. Get absolutely fucked. Get absolutely fucked. And this is actually a decent thing, a decently common thing that you see with these steam reviews. You reckon that the people leaving these reviews would have a thing or two to show for it in, in terms of like literally hours spent.
01:06:56:29 - 01:07:20:07
Speaker 2
And it's like now they spend a fraction of the time, I spend more time in a week on blender than you've you've spent in your entire blender bloody career. So we'll see who exactly, has a thing or two to say about this. I suppose it's not necessarily to say that the opinion is invalid, because they don't know the tool as well as I do, but it is.
01:07:20:07 - 01:07:43:15
Speaker 2
It certainly is to say that, like, you've got a whole lot less invested, you've got a whole lot less riding on the success of this project, and yet you're offended on the behalf of all of this community, when really when you look at it, you know, part of this community, it is sense, like how many hours have you got?
01:07:43:15 - 01:08:25:34
Speaker 2
48.3. It's really not that much. How invested are you really in this? It's like, okay, well, she had said, you know, not as not as prominent on the, on the 3D end as a 2D. And that's fair enough. That makes me record. Makes me wonder if you've got more of a bone to pick with this for, other reasons than actually wanting the tool to be good and to, because I don't know, the entire idea of a tool is that you want it to take care of all of these problems, and all of these problems can be abstracted out to somebody else with who's already thought of how, logically, to make this happen
01:08:25:38 - 01:08:58:57
Speaker 2
in this in this computer program in particular. you wanted you to just be able to do what you wish it to do. without it breaking, without loads of tedious tasks ahead of you. you know, particularly something that is used to create you just want to create. And you want your tool to be as streamlined and seamless as it can possibly be.
01:08:59:02 - 01:09:06:11
Speaker 2
But in order to assist you in creation.
01:09:06:16 - 01:09:30:36
Speaker 2
Why is it bad that now, through a. Through like a mutual effort, there is an MCP built into blender that that they're launching in? I think they are launching at 5.2. I don't know when they're launching it. I use a third party one. I've been using a third party one for a while. It's pretty incredible.
01:09:38:36 - 01:10:05:51
Speaker 2
Like, why would you be upset that they've got more funding now and that there's now the ability to take a tool that many people are already paying for and integrated directly in so that you can speak to the program, and then it can tell you either what you ought to do in the program or it can you can direct it to do things for you in the program, things that are not exposed to the human eye.
01:10:05:51 - 01:10:25:55
Speaker 2
I think that's the more important aspect of this. Or things that if a human were to do, it would take like ten years to actually achieve. That's the main thing that people use code for in the first place. The entire reason that people make tools like an adult is a blender. So I've, I've got I've made a couple of add ons.
01:10:25:55 - 01:10:31:48
Speaker 2
I want to make some videos showing off the add ons that I've made.
01:10:31:53 - 01:10:47:46
Speaker 2
Like I've got an add on that makes that makes it so that you can just instantly play blast and it'll, it'll output the video and, you know, basically a one click solution for doing that.
01:10:47:51 - 01:11:10:28
Speaker 2
Before like vibe coding or what have you, you would have had to go out of your way to either create that yourself, you know, realize what the problem is, realize how it ought to be, achieved in this version of blender, in this version of Python, with all of the necessary, understanding of those libraries in order to code this solution.
01:11:10:33 - 01:11:46:22
Speaker 2
most people put it up for, you know, open source because they want other people to have the ability to contribute to it so that, you know, that project ends up going somewhere, right? so that it becomes more useful. But you end up with, like people who say, well, this is this isn't good. It's bad that you've created this tool because now, I need to like, like now, other people have the option not to do.
01:11:46:22 - 01:12:21:13
Speaker 2
So I got, like, tedious shite. It's like, well, now, now I'm going to forget how to go. And, you know, manually type in what I want the play blocks to be called and to, to output it to, a PNG sequence and then have that PNG sequence be, a, a video of, you know, output it to, to MP4 already or like.
01:12:21:18 - 01:12:44:22
Speaker 2
And or you'd have to like set the set the, the render settings every single time. Right. Like there's loads of things that you would rather have like a better tool to be able to do it just because you know that you can go and do it manually, doesn't mean that you need to just have a tool for it, and then you can do everything that you wish to do faster.
01:12:44:22 - 01:13:06:46
Speaker 2
You can move on to the to bigger, better things. But for some reason, someone makes a tool that makes it so that you can use your native language to interface with the program, and you can get a feedback loop going between something else that's also thinking alongside you. You're getting your, your hardware or someone else's hardware to think like simultaneously to you.
01:13:06:46 - 01:13:31:31
Speaker 2
And, and you've got a feedback loop going. For some reason, this is bad. Like I still am not getting the getting the, the point it like, what is the actual motivation behind this other than like bloody nihilism, right. do you know that this review is mostly about the anthropic stuff, not the software, hence the thumbs down.
01:13:31:31 - 01:13:39:29
Speaker 2
But otherwise I would recommend blender. This is weird. Like.
01:13:39:34 - 01:13:57:07
Speaker 2
Like you just wanted somewhere to to bitch about this. Like, I don't know, I just I don't really get it. Yesterday, blender had announced that anthropic had joined the development fund as a corporate patron. Anthropic all are associated with Palantir and are responsible for the AR software AI software Cloud, which you can partially blame for the horrors of vibe coding.
01:13:57:12 - 01:14:28:50
Speaker 2
This is like an utterly bankrupt sentence. It's full of so many unfounded like not not unfounded. It's just full of bad premises. So anthropic are associated with Palantir. I didn't know this, but, like, it makes sense. I after researching just a little bit, it's like, okay, so they've they're in kind of regarded as a bit of an AI company now, whereas like when they were founded in 2004, that's not what they did.
01:14:28:55 - 01:15:02:58
Speaker 2
if you well, we'll get into that. responsible for the AI software cloud. Yes. Anthropic is responsible that responsible for this is an interesting way to frame it. It's like they are the ones who've created this tool. It's like the blender Foundation is the is the, organization responsible for blender? It's like okay. Yeah, but like the reason you're using that language is because you're trying to ascribe like, evil to it.
01:15:02:58 - 01:15:35:58
Speaker 2
And it's it's just weird. Which, which you can partially blame for the horrors of vibe coding. I don't know what the fuck she means by this. What do you think Sophie Bubbles means by this? Because, like, what's so horrific about vibe coding? What's horrific like? There's nothing that's so horrific about it like this. This isn't even making an ethical point.
01:15:36:03 - 01:16:00:41
Speaker 2
Like, the only point that's being made here is that it's scary. But that's a that's a conclusion. But but she's using it as a premise in this argument. It's like, this is just nonsense. Like, if I were to try to steal, man, this is it. Like is what is horrible about this the the plagiarism argument is what is horrible about this.
01:16:00:41 - 01:16:28:01
Speaker 2
The, the losing of skills argument is what is horrible about this. The, potential. I don't even know if anyone's complained about this, but like, I, I have an open source repository, and I run this, this software's development, and people are coming in with massive codebase changes and doing pull requests, and it's just clearly been done by, an AI.
01:16:28:01 - 01:16:54:05
Speaker 2
And it's not competent, like, it's not good contribution. There's more contributions are great, but it's not good contribution. So like or perhaps that it's an insult to me as somebody with, bloody PhD in computer science or something like that, that people are trying to do pull requests of, these masses bloody, these changes, and they don't understand the code base, they don't understand the coding language.
01:16:54:16 - 01:17:19:21
Speaker 2
It's just a computer doing little things that mostly just mess up the what it is and not, Well, okay, this is a tangent, but like, it is interesting, like, to most people, when they see a tool, it's a means to an end. I mean, that's, that's actually a good psychological definition to of, of a tool. It's a good phenomenological definition of a tool.
01:17:19:21 - 01:17:27:54
Speaker 2
Like that's what a tool is. It's a means to an end. Right.
01:17:27:59 - 01:17:53:30
Speaker 2
Why? Are people more concerned about the, the means than they are about the ends? That's kind of the interesting thing, because the end is to make it easier to fit and, and more, efficient for people to be able to work on this thing. Right.
01:17:53:34 - 01:18:17:37
Speaker 2
Why would people be upset that these tools exist and not be a bit more opposed to what they're being used for in an incorrect way, or in a morally incorrect way? I should say, like the issue that is being stated here is that it's bad, that it is that it exists, not that it's bad what it's being used for.
01:18:17:42 - 01:18:37:41
Speaker 2
So why is it bad that it exists? That's the premise. The the premise this entire argument hinges on that is just not being addressed whatsoever. It's it's I don't even record Sophie's even thought about this. It's it's kind of odd.
01:18:37:46 - 01:19:16:58
Speaker 2
So would it be morally permissible to kill some of her? Any kind of stealing they do from you? If they're currently in the process of it, yes. But, it has to be proportional. So if somebody steals from me, I like I don't get to go over and just kill them right? What is required is for me to, there are many things that I could do, but like, there are things that I can now force them to do that they don't consent to because they already violated my consent by stealing from me so I can steal my stuff back from them.
01:19:17:13 - 01:19:43:22
Speaker 2
And I think I could do that at gunpoint, like, I, I think that it'd be morally permissible, but it's not inherently, morally permissible to just kill somebody. after they've no longer presented a direct threat. I think that's pretty Rothbard in so do that will be what you will. I'm not quite certain on that one, though.
01:19:43:27 - 01:20:17:05
Speaker 2
The idea I don't know what the horrors of coding like. What? What are you even referring to? What is so horrific about it? What is so horrific about the ability to talk to a computer and then have the computer do the thing that you wish it to do. What is so horrific about it? Like, we're not even at the point where we've got Neuralink and the Neuralink is bridging to, an MCP or, you know, two and Lem, and then you're communicating with that Lem as a bridge to be able to interact with your, your computer without even having any peripherals.
01:20:17:06 - 01:20:41:27
Speaker 2
It just, like, kind of appears in your imagination or whatever the hell, something like that. Right? We're not at that point when I'm at the point where all peripherals are gone and we're playing games by, you know, like like it's just weird. Like we're not at that point. And the transhumanist would love that anyway. So like, I don't I don't know what the concern is with that.
01:20:41:34 - 01:21:05:45
Speaker 2
Like at all. What's so horrific like I think I'd have to dig into loads of, incredibly foolish blue sky opinions in order to figure out what is so horrific about it like. And most of it is just going to be Puritan, mystic, puritanical, I should say. the pearl clutching about a tool existing is like, oh, no, technology's what?
01:21:05:51 - 01:21:31:04
Speaker 2
Whatever are we going to do? We need to smash the machines. Bloody Luddites and nihilists and primitive ists. All of the worst people, aka by the way, if you don't know what Palantir is, then all I need to say is it's run by a guy who's in the Epstein files. does she mean Peter Thiel or teal? However you say his bloody name.
01:21:31:09 - 01:21:53:58
Speaker 2
I found this pretty funny. I find this pretty funny because it's like, that's all you need to know. All you need to say is that it's run by a guy in the Epstein files. The entire government is run by people in the Epstein files. Like everybody in government, the entire global elite. Is it the Epstein files? Like, what is this point?
01:21:54:03 - 01:22:20:58
Speaker 2
It's like, well, bladder's taking money from a guy who has a lot of money at this guy because he's that elite. Is is it the Epstein files? Okay, this is this is ridiculous. And there are words that I can't say. I was considering not being on on TikTok for this, but it's like. Why is it at all relevant that this guy is in the Epstein files?
01:22:21:03 - 01:22:55:46
Speaker 2
Like, no, it's not that it's not relevant. It's that it's it's not either. The first thing that you should think of when you think of Palantir. Well, you should think of is the fact that that like, I'm surprised that there wasn't more of this going going on where it's like, well, Palantir, like all you what you really need to know about Palantir is that they are, entrenched in the NSA's surveillance state, like they help the NSA spy on the American people and not just the American people.
01:22:55:55 - 01:23:18:25
Speaker 2
Like, like you could. I feel like that's probably the first thing that you should know about Palantir, that that one of the first things you should know about Palantir is that it's named after, bloody, the, you just tell him the truth. What's what's his name? That, The other, the evil, the evil wizard for blood of the rings.
01:23:18:30 - 01:23:43:37
Speaker 2
So, he's in World War two, so, something I know he's. I know he's been been knighted by the Queen. I don't remember his name. He's also, Count Dooku in Star Wars. It the evil wizard, the bloody. I don't know why I'm blanking on this. I know these characters. The evil wizard has a little crystal ball, and he puts his hand over it and goes.
01:23:43:42 - 01:23:57:36
Speaker 2
He is will go to as Gandalf. You know that the the ball that he uses is called the palantir. The palantir. Like you know he says that.
01:23:57:41 - 01:24:22:06
Speaker 2
Like that's what the company is named after because the palantir allows him to like see tons of shine that he probably that that it's probably not good that that's in the hands of evil. Right? Like. That's that's what they named the, the company after. Like, you should probably know that. You should know that they're part of this.
01:24:22:06 - 01:24:53:28
Speaker 2
The United States government surveillance state, the the they assist in, in, Immigration and Customs Enforcement like, they they're effectively an arm of the surveillance state, like you should probably should probably say that. Like it would be such a blue sky thing to say. Like, it'd be like the the part of the surveillance state that they, you know, they wouldn't say that on blue Sky.
01:24:53:33 - 01:25:20:58
Speaker 2
They'd say, the government informers, they, they help out ice, you know, and ISIS evil and Trump bad and all of that. Like I feel like doubt that be a pretty blue sky point to say. But like she she doesn't even know that. It's like, let me help you make your argument here. Trump bad. Palantir help Trump and ice ice deport the the immigrants, the migrants, the the poor innocent migrants.
01:25:20:58 - 01:25:51:38
Speaker 2
Right. Therefore palantir bad therefore anthropic bad by association therefore blend bad for taking money from anthropic already. I mean, like that's that's a bad argument like that, that there's so many layers removed from the bad that that argument is just like, that is just not a good argument like that. I could, I could have, could have helped you at least say something that actually is relevant rather than like, like this.
01:25:51:43 - 01:26:18:04
Speaker 2
Here's the syllogism. That's I think that's the, is is that actually the word I was looking for? I don't remember, Biden magically summoning 30,000 unnamed Wisconsin to vote for him. Yeah, he's he's using the using his. Is Biden the great Biden the great.
01:26:18:09 - 01:26:28:32
Speaker 2
Is a great great great wizard. Evil lost company. Yeah, like you could just said that I don't know, feel like that.
01:26:28:37 - 01:27:03:56
Speaker 2
I don't know, it's like they're just, like, grasping at straws to make this company out, to be, like, just evil. And it's like you just invoke Epstein. It's like, okay, every single government on Earth should just instantly collapse like a thing. That would actually be great. Let's go with that. Anybody who touched the Epstein who even got close to being on the Epstein list, just, just throw him into a into the pit with with, the dog to the pit and bird is thrown into the pit with the bloody, who was it?
01:27:04:01 - 01:27:19:58
Speaker 2
The Titans in ancient Greek mythology. Like, throw in that butterball. Butterball in there to to to be in prison for eons, right?
01:27:20:03 - 01:27:24:45
Speaker 2
it's just it's just kind of insane.
01:27:24:50 - 01:28:03:15
Speaker 2
Because the the logical trail here is. Peter Teal is bad because he's on the Epstein list, but he's in the Epstein files. Peter Thiel runs Palantir. Palantir is associated with anthropic. Anthropic. Gave blend of money. Therefore, I'm upset at blender. It's like, no, I think you're just grasping at reasons to be upset. Like you're trying to just, like, throw shite out there because you're you're upset.
01:28:03:19 - 01:28:30:16
Speaker 2
It's like you can't be upset at blender and you can't be upset at anthropic. You also have to say, like anthropic is associated with Palantir, who's associated with, the Peter who's run by Peter Thiel and Peter Thiel was associated with Epstein. And this, this elite global pedophile ring. Like this is this is actually pretty decent mental gymnastics.
01:28:30:21 - 01:28:47:15
Speaker 2
It's a non-sequitur. It has absolutely nothing to do with this because, like, are you going to be upset about AI? Are you going to be upset about anthropic? Are you going to be upset about Palantir? Are you going to be upset about, the things you didn't bring up like the NSA or Trump? Rice, are you going to be upset about the surveillance state?
01:28:47:15 - 01:29:05:55
Speaker 2
And this is the funny thing, I don't think the I don't think the blue sky people dislike a surveillance state. I think they would love a surveillance state if they were on their side. Like, that's what the blue voters want to do, that. Sorry. The blue button presses, right? This is the big meme that's going on right now.
01:29:06:00 - 01:29:37:13
Speaker 2
It had it had keep all of us safe. therefore, everybody should be less free. And we should give the, you know, power to the this elite organization like that. They're perfectly fine with with mega-corporations insofar as they parrot their views, insofar as they're collectivists. Right. Like perfectly fine with massive, hyper powerful governments because, you know, the state is incredibly useful if you can get it to do what you want, but then the other guy could use it as well.
01:29:37:13 - 01:30:02:40
Speaker 2
So maybe you shouldn't do that. This is like the eternal, libertarian point. Maybe you shouldn't grant power to despotic rulers just because you like your flavor of despotic rulers. Like good us is. It's really something. I saw a YouTube video that said the I that said the AI the government has is far more advanced than what's available to us.
01:30:02:45 - 01:30:23:39
Speaker 2
And that and the AI is working on itself at a fast rate against other countries. AI and yada yada world will end. Kind of scary. I don't think that's true at all. I think the government is full of shite. I think anything that the government has, they stole and by by dint of that, nothing that the government has is ever going to be more efficient than what the free market produces.
01:30:23:43 - 01:30:51:21
Speaker 2
That's actually part of the reason why the government funded, you know, it did a massive $1.4 billion investment in AI products and infrastructure, like that's that's a big reason why that happened is because they they can't actually make it themselves. So they have to socialized. They have to subsidize other people making it. But every time they do that, they make the entire free market hallucinate.
01:30:51:21 - 01:31:18:55
Speaker 2
And that's actually a big part of the reason why, why ram prices skyrocketed is because everybody thinks there's more demand than there actually is. They're investing in things they can't afford because the, they've got more, more, resources than they would have if not for the government being despotic fools. Anyhow. the blue a red button.
01:31:18:55 - 01:31:39:15
Speaker 2
So you're a red button guy? Obviously. I think you should have been able to intuit that from, all of my, all of the things that I've said before. I am not the, I am not the property of the mob. And if the mob wishes to to live, than they are to live in their own rational self-interest.
01:31:39:19 - 01:31:52:25
Speaker 2
It's rational self-interest is is, the whole point there, you know? That's that's the.
01:31:52:30 - 01:32:10:52
Speaker 2
I think that's actually like the major disagreement. It's it's mainly a psychological disagreement. But if we were to say it philosophically, the major disagreement between the blue and red button presses is, what is it that I like to.
01:32:10:57 - 01:32:44:56
Speaker 2
It is rational self-interest versus irrational. interest in the well-being of others. Like, clearly one of these is preferable. But that's the funny thing is that it's not so clear because people with their own, their own priors will. Say the opposite is just as clear. Well, that's kind of the interesting thing about, about personality. About temperament.
01:32:44:56 - 01:33:21:41
Speaker 2
Right. but currently we're in the, in the clutch of pathological, agreeableness. And that's just what you're going to get when everybody's pathologically agreeable, everybody's going to look at that and be like, well, everybody should just press this button. Everybody should just, do that. And it's like, put yourself first. Why don't you be somebody instead of a, you know, a cog in a machine, a node in a network, be somebody dare to be yourself.
01:33:21:54 - 01:33:41:04
Speaker 2
I like, I dare say, and then non subjectivist way like, I don't know, it's it's it's interesting Squid game, you know, that there's going to be people that are going to press the blue button. Would it mean that you're killing them. No.
01:33:41:09 - 01:34:36:51
Speaker 2
Whatever mechanism like that has been set up in order to in order to to cause that to happen is responsible for that. But I'm not going to put myself in harm's way just so that other people who also were willing to put themselves in harm's way, chose to do so, like. And I actually reckon that all of the, all of the polls that people are going on with all of those polls are actually just wrong, because people who are, not disagreeable enough to go against a popular sentiment are not going to say that they're going to, that they're going to pull the lever that gives them the potential of dying.
01:34:36:55 - 01:34:58:13
Speaker 2
They're going to think in the same way that any any human being thinks any, any, any being that wishes to survive thinks it's like, okay, there's a limited resources, my ability to get out of, of this life or death situation is contingent on me putting myself first and that there going to be other people who die in this.
01:34:58:18 - 01:35:20:26
Speaker 2
but if I sit around, sit around and and worry about them, that I'm going to die, like there's going to be a lot of people who would say, oh, yeah, I press the blue because, like, everyone's going to press the blue and then then we're all going to be fine.
01:35:20:31 - 01:35:30:43
Speaker 2
But when those people get down to it, the, the fear of mortality, frankly, it's like, okay, but I if I press that button, I could die.
01:35:30:48 - 01:35:55:27
Speaker 2
I do I want to gamble my life on everybody else's ability to be altruistic. I don't even think you should do that. I think you should put yourself first. And and. To hell with all of the people who wish not to be an individual.
01:35:55:31 - 01:36:22:14
Speaker 2
What if your girlfriend pressed blue? She wouldn't. I know she wouldn't. I mean, not just for the reason that if she did, she'd I'd I'd be disappointed in her, but more because she's not stupid. She knows when, she she does not have an irrational trust in strangers. Like many of the blue presses do.
01:36:22:19 - 01:36:54:24
Speaker 2
Like, she understands that. That mankind is. It's capable of evil. I don't know, and she doesn't reckon that selfishness is inherently evil. Like that's part of the issue. Is that like, oh, yeah, man, this man is capable of evil. And all of the evil people, all the people who are pressing the red button like that's kind of the point that they're making, is that all of the people who present a good one, a good the blue one, a good, and all the people who are oppressed and the red one is bad, but it's like, okay, but you were the one who was actively going out of your way to sacrifice yourself like it.
01:36:54:28 - 01:37:03:22
Speaker 2
Why wouldn't you put yourself first? I think everybody should press read.
01:37:03:27 - 01:37:33:15
Speaker 2
What if a hand slipped and press the blue button? LOL. I mean, that's upsetting. Like, I don't know, would you still press read? I suppose this is sort of similar to this. this other situation. The one that we were talking about, the other week about dying or living for, don't know. Anyway, I'm going to move on with this.
01:37:33:19 - 01:37:55:25
Speaker 2
A lot of people. and by that I mean most of blue Sky. It's like, oh, most of blue Sky is a fox. Like, we're not happy about this. Oh, that's so surprising. So surprising. That blue sky that had a lot of, had unkind words to say about this.
01:37:59:02 - 01:38:25:41
Speaker 2
I say most of blue Sky because I have shockingly not seen many other people that that upset about it, which I find odd. I wonder why that's happening. I wonder why this is so odd to you, that there aren't a lot of people outside of blue Sky who don't, who give a shit about this. Like, I wonder why I wonder, like, could it possibly be that blue sky is a hive, a wretched hive of neurotic fools?
01:38:25:46 - 01:38:54:21
Speaker 2
Could it be that could it possibly be that that you're on a site full of people with cluster personality disorders? Could it be that I reckon it could probably quite possibly be related to that? I'm not many people, but I've been angry about this outside of the site that everybody's constantly raving mad on. It's like, yeah, maybe get out of that bubble because you're surrounded by angry people.
01:38:54:21 - 01:39:33:09
Speaker 2
You're surrounded by insanity. Like, stop! Why? Why are you judging? Like it's it's just it's not a good heuristic, right? Like my litmus test is whether or not the madman on Seventh Avenue is, is upset about this. And if he's upset that I should be upset to know he's insane. Blue skies full of it's, say, people, if you're aligning yourself with them and using them to aggregate like your own opinion, then your justice insight is the rest of them are just as insane as a blue button.
01:39:33:23 - 01:39:51:52
Speaker 2
So like this is just such a bad point, I. I really can't wrap my head around why people wouldn't, be upset about this anywhere else. Well, what? Why are you surprised?
01:39:51:57 - 01:40:09:34
Speaker 2
Anthropic is one of those one of many companies that blender has taken money from. But blender has also taken funds from Adobe, Nvidia, Meshi and Chaos, all of which are companies involved in AI in some form. Yes, obviously they're involved in AI in some form. They're in the technology.
01:40:09:39 - 01:40:32:04
Speaker 2
They're they're in that field. And this is only like the most incredible thing to happen in that field in the past. Bloody like it's literally like science fiction being instantiated. It's become real. And you're like mildly seething about it. Like, it's just so weird.
01:40:32:09 - 01:40:46:39
Speaker 2
And they don't even want to know what the upshot of it could be. They don't they don't know want to know what the upside could possibly be? Because it's just the One Ring, you know, killing it with the with the Lord of the rings references today. It's just the One ring. It's not okay to use it. You should never use it.
01:40:46:48 - 01:40:52:37
Speaker 2
It's evil to do. I don't get it.
01:40:52:42 - 01:41:25:50
Speaker 2
But it's like, yeah, Adobe's shitty. And it's not because the Adobe was shitty long before they ever degraded any. I they're shitty because they're completely opposed to the fast standard. They're completely opposed to copyleft that they opposed. They're effectively a technological cartel that's teamed up with the state to be able to, charge money for something. They could be, like infinitely distributed.
01:41:25:55 - 01:41:56:48
Speaker 2
That's that's a load of what, what IP does. But it's like I mean, that also applies to Disney as well, for sure. it's just a different product. It's like I, I'm actually not familiar with chaos. No, it's not asserted what that is. Meshi. This is not really like that big a deal. Honestly. Like, the stuff that meshi generates, it's, you know, it's it's not great.
01:41:56:52 - 01:42:30:09
Speaker 2
at least in my experience. Adobe is is shite. Because of that subscription model, I would have been I would have gladly been paying for, you know, the, What is it, like, a proprietary, whatever the hell. Like, they should just have the same pricing model that that Ableton has is a wonderful pricing model. I, I've purchased Ableton stuff I'd be perfectly willing to purchase, Adobe stuff if they didn't, didn't have such a shite model.
01:42:30:14 - 01:42:45:31
Speaker 2
Blue Sky is trying to stop the creation of the Basilisk based an art base. I don't I don't think that the that the basilisk, that Rocco's basilisk is down stream of of, real.
01:42:45:36 - 01:43:09:03
Speaker 2
What is it called? Of of just LMS. But that's not even what they're saying. That's not even what they. Wait a minute. But if they're doing that, then that means the basilisk is going to kill them first. So not based, I suppose. Well, I don't know. Just because you have the ability to destroy somebody doesn't actually make you, does it?
01:43:09:03 - 01:43:26:27
Speaker 2
Doesn't make opposing that any better or worse. So it could be based in that sense. I don't I don't really care. but blender themselves have made have made a note on their own article regarding the, the reception to this announcement on the article. Literally, literally the first thing you see is a note saying this announcement is causing a lot of feedback.
01:43:26:27 - 01:43:56:39
Speaker 2
We are actively evaluating it. Blender has stated on their page for funding policies that these corporate donations don't affect development. I quote individuals and organizations that contribute to support Blender's development without acquiring decision making power over the project's direction. This is correct. The most this has gone to is Claude adding blender compatibility. but they were doing that before anthropic was, I don't know if they were in talks with anthropic before then, but they were doing that before anthropic.
01:43:56:43 - 01:44:23:52
Speaker 2
anthropic. actually funded them, right? It seems nothing in the actual blender app itself will change keeping to their funding policy. You know, people are still relatively unhappy about blender taking money from the guys behind Claude, primarily because of where that money was gotten from. Why? Why exactly? Because it's from the government. Like because the government like, funded them and, you know, these massive grants.
01:44:23:52 - 01:44:49:21
Speaker 2
No, it's not because of that. Because they don't care about that. Because it's from where's it from? Like what is the issue with where the money is from, from people paying a subscription for a product. Is that the issue? Because really what they're selling isn't so much Claude. Like they're not necessarily selling Claude. They're selling tokens. And Claude itself isn't even for sale.
01:44:49:21 - 01:45:17:00
Speaker 2
For one thing, like Claude, you know, it's just an API, and you've got the API key like you don't. No one can download the actual Lem of Claude because it only runs on their, their compute data centers. Right. like, why are you upset that people pay for a product and effectively what they're paying for is compute.
01:45:17:11 - 01:45:46:34
Speaker 2
And, you know, if you've ever used a render farm, you've paid for compute. oh. And people who use blender now, people who've used blender for 50 hours only, mind you, but people who've use blender have generally, generally used a render farm a to maybe a free one, like shape it. so I don't necessarily recommend it's unless you're like desperate for more compute because I, I don't like the devs, but that's, that's just me.
01:45:46:39 - 01:46:19:36
Speaker 2
Well, I don't like the philosophy of the devs. They seem like fine people. Some of them I don't know, the difficult, the really difficult though the like, the way they speak is just like annoying. Like every single time I speak to them, it's just demotivating. It's just like, okay, like just burst the bubble every single time. Anyhow,
01:46:19:40 - 01:46:40:12
Speaker 2
Like, why are they upset? Like, what's so bad about where the money came from? The matter came from the government. No, they like socialism. This is. These are blue sky people, by the way. Like, I'm not talking. I'm not pulling this out of my ass is a blue sky people. Of course they like socialism. Like, they're perfectly fine with government funding of of bullshit.
01:46:40:17 - 01:46:51:56
Speaker 2
They're just not fine with it being this particular technology, even though it's, like, massively useful. I don't know.
01:46:52:01 - 01:47:04:09
Speaker 2
I swear, if I cured cancer, they'd be they'd still be pissed it cured cancer in an in an unethical way.
01:47:04:13 - 01:47:20:54
Speaker 2
If, I mean, I don't know, to be fair, if a government cured cured cancer, then I think I'd still be upset at the government for existing. So, I don't know, maybe it's fair. Look at the replies of their announcement on blue Sky. There are 700 plus replies. You could even see this in the recent reviews. I could see this is.
01:47:20:54 - 01:47:36:54
Speaker 2
This actually makes me wonder why the hell blender is on blue Sky in the first place. I was I was talking to Jenny about this recently. Like I don't have a blue sky for this precise reason. It's full of neurotic fools. Like why this is not my crowd. Why would I want to be around a bunch of people?
01:47:36:59 - 01:47:56:38
Speaker 2
who refuse to be rational about anything and not only refuse to be rational, but just rage, literally rage against the machine. Like the moment that any anybody is like, hey, this is like a really useful technology. Just like, give it die. Why?
01:47:59:02 - 01:48:20:49
Speaker 2
You can even see this in the recent reviews. I've seen a few negative reviews on steam already about this. If you want my thoughts, then I think this is was not a great choice to take money from. To put it lightly, this this is an interesting to say interesting thing to say because like, they were taking money from Adobe before Adobe did AI stuff, I'm pretty certain.
01:48:20:54 - 01:48:54:42
Speaker 2
and even then, like, you know, they're taking money from these, these pretty, pretty massive corporations. I like money as money. That's the thing. Like. I don't see why it's so wrong. across multiple premises. Like, I think you first need to prove why AI is is the one ring in order to actually have a conversation downstream of this.
01:48:54:46 - 01:49:24:18
Speaker 2
I don't think and also hope that this won't affect blender as an app that's authorities app. App doesn't really isn't really the right word for it. The like app app doesn't seem right. I don't know, it just seems. But I don't know. That's that's neither here nor there. And considering their policies on funding, I think that'll stay, especially considering that previous donations from AI companies also haven't altered the app.
01:49:24:23 - 01:49:48:16
Speaker 2
I don't know. Meshi like, I don't know how much meshi did, but like Adobe like that, that the first thing I investigated when I found out that Adobe was one of the funders is like, oh, why would they do that? Like, what do they get out of that? And, you know, there's some plausible explanations behind it, like it's interesting, right?
01:49:48:21 - 01:50:15:11
Speaker 2
but it's never affected. Like blend is never going away from the GPL license and this sort of, vision of what they're trying to create is it's just it's not really possible for it to because even if what they were creating ended up getting leadership, that was shite. It's open source. So everybody else could just anybody else could pick it up and be like, okay, I'm just going to, you know, do the same thing.
01:50:15:16 - 01:50:43:45
Speaker 2
And if they suddenly became closed source, then it's like, okay, but it's kernels, not not kernel, but like all of the source code is still here. I can still do this like. I don't know what they think it's going to do. I think the only concern they could possibly have is like an incentive slash cultural, concern, but that that's the thing like, I don't know, this is where, Andrew Price would just say like, it's inevitable.
01:50:43:49 - 01:51:16:28
Speaker 2
Like. This this is an inevitable technology. It's going to integrate itself into everything. And the culture is going to change in a way that that, embraces it. One day, sooner or later, no matter what you do, you know, hence inevitability. But like, I don't see why it's such a concern. Like, even considering the the cultural stuff, it's like, oh, what are you going to do, right?
01:51:16:37 - 01:51:27:00
Speaker 2
Is people. Really just want to use the tool.
01:51:27:05 - 01:51:58:01
Speaker 2
I don't know, it makes me wonder. Like what? What they actually make in blender and how they make it makes me wonder that because, like, everything that you do in your workflow could be enhanced by AI in some way or another, it doesn't mean that you should. It just means that like, why are you upset at other people doing the same thing just because there's like there's plenty of features in this, in this app, this application that, that you're not going to use because that's not what you're using it for.
01:51:58:01 - 01:52:27:37
Speaker 2
So you don't need that that integration. You don't need it. Right. Like there's always going to be something that is not for you. And I don't see why it's not okay for that to exist downstream of this. Logic trail. Like I don't know because they can't keep the story straight. I'd like to see them try to keep the story straight.
01:52:27:37 - 01:52:51:50
Speaker 2
They there's a million. Like, the only thing that's straight is the attitude. But there's there's a million reasons behind the attitude. There's no philosophy governing this. No one's driving the ship. I don't know, but I do hope that blender doesn't take any money from anything worse in the future, and that this leads to them adding controversial features. I think.
01:52:51:50 - 01:53:21:52
Speaker 2
I think you meant the other way around. I hope that blender doesn't take money from anything worse in the future, and that this doesn't lead to them adding controversial features like what? Like what? Like MXGp, and that's like all they've done is in MXGp. And you know, I'm not even using it. I'm on 5.5.0 currently. I haven't really made the jump to 5.1 yet.
01:53:21:57 - 01:53:30:29
Speaker 2
I'm just going to eventually hop on the LTS and then just remain on that. But like.
01:53:30:34 - 01:53:55:18
Speaker 2
I, I also just hope they check who they receive them from. The only logical like the only point that you've made is that, like, they've taken money from somebody who's affiliated with a company who's ran by a person who was in the Epstein files. Is is that the issue? Is that the only issue that you've got with with anthropic because that's weak.
01:53:55:22 - 01:54:23:17
Speaker 2
Like the funny thing about this is that, like, you know, Andrew Price had said, I think it's a good conversation to have. I think people ought to be having the, what is it? The ethical conversation here? and I'm like, yeah, yeah, right. We should that that ethical conversation should be a hearty one. Right. But this isn't an ethical conversation.
01:54:23:22 - 01:54:56:26
Speaker 2
This is just like just a rant, really. Like you're not making any real points. Like you're just trying to, like, validate your own opinion or something like that, and, well, you know how I feel about that. It's just it's just weakness. Like changing the monkey head model to a 3D swastika model. Well, that would certainly be controversial. Yeah, right about one thing.
01:55:00:32 - 01:55:34:39
Speaker 2
Like, well, why do they need to check who they receive money from in, in this context? Because they might be connected to somebody who's connected to somebody who, like, watched Donald Trump give Bill Clinton a blow job and had Epstein party like 15 years ago, like. Like, I don't know, Palantir literally exists as that's it's all irrelevant. It's all irrelevant anyway.
01:55:34:44 - 01:55:38:31
Speaker 2
Like.
01:55:38:36 - 01:55:47:57
Speaker 2
What what is the concern? Oh, now my life is flagged for harassment and bullying.
01:55:48:01 - 01:55:55:09
Speaker 2
This this is completely nonsensical.
01:55:55:13 - 01:56:21:50
Speaker 2
Because I referenced auto, not auto fellatio. That'd be kind of funny. Because I mentioned a public figure. But I seriously don't understand it. It's so sensitive. Maximum sensitivity. It's it's like the, the body, the Fortnite meme Max sensitivity. anyhow, I hope this catches up to catches you up to speed on what happened in the last day.
01:56:22:03 - 01:56:50:44
Speaker 2
Nothing happened. That's the thing. Like blender announced that they got some money from a company that makes a tool. That's all that happened. Like I seriously don't understand. So for, that that's what I was going to say. Like,
01:56:50:49 - 01:57:16:00
Speaker 2
It's like, you know, we should be having this ethical conversation, but this isn't an ethical conversation. There's there's no ethical points being made here. I'd like to actually hear somebody make an ethical point about this that isn't just like whims, just feelings. You know, it doesn't make sense. Do with this information what you will. I yes. Thank you for giving me permission to do that.
01:57:16:04 - 01:57:40:12
Speaker 2
and do what you think is best to do in this scenario. Why? Why is this? You may as well just say nothing like save yourself the time of typing it out because, like, that's just a worthless statement. What does that is literally saying nothing. Do what you think is best to do. Isn't that what everybody does? Does it?
01:57:40:12 - 01:57:46:28
Speaker 2
Is there anybody who does what they think is the worst thing to do?
01:57:46:33 - 01:58:24:09
Speaker 2
Like it does? It doesn't make sense. I don't think this will stop people from using blender completely, but many have had a sour taste left in their mouths. Oh no, because they they took money from a company that makes a duo. And and a useful tool at that. I actually wanted to say. so the meme of the day today is, is an image of Miku, and it's an AI generated image of Miku holding up a sign that says AI art is more important than your shitty ram, and I don't I don't really like.
01:58:24:14 - 01:58:46:47
Speaker 2
endorse the position, but I find it a very funny image for multiple reasons that I can't quite explain. It's like. it just seems to be like the opposite of what loads of people have said about,
01:58:46:52 - 01:58:50:24
Speaker 2
about.
01:58:50:29 - 01:59:03:38
Speaker 2
Like AI in general, I don't know. I reckon the point about Ram is, is a point made because,
01:59:03:43 - 01:59:10:05
Speaker 2
I don't know, because people like.
01:59:10:10 - 01:59:19:10
Speaker 2
They don't,
01:59:19:15 - 01:59:22:41
Speaker 2
Sorry.
01:59:22:46 - 01:59:42:24
Speaker 2
I was trying to bring up another clip. I'm not particularly fantastic at it a bit. That multitasking. I don't necessarily think simultaneous to thinking. Somehow it works. What was I looking for? I was looking for,
01:59:42:28 - 01:59:47:53
Speaker 2
Why the hell can't I find this?
01:59:47:58 - 01:59:52:45
Speaker 2
Page again?
01:59:52:49 - 01:59:59:19
Speaker 2
Wait, why the hell is my my chrome history, like, out of sync?
01:59:59:24 - 02:00:11:12
Speaker 2
This is bizarre. It's not even showing me what I actually. Well, yeah, it's it's it's out of sync.
02:00:11:16 - 02:00:25:48
Speaker 2
What the hell? I wonder if I need to restart Chrome or something. I wanted to look at more of the,
02:00:25:53 - 02:00:53:44
Speaker 2
I wanted to look at more of the the, The reviews on this site. I don't know if I've ever actually seen this full point to animation for .2.0. We're on 4.2. 20 now. Isn't that crazy? We love LTS. I can't believe people still use old, the. Oh, wait, look, that's 2.86. No, people do not use that.
02:00:53:49 - 02:01:10:09
Speaker 2
Insane people use that that. Shut up, shut up. I don't know what it does that you know. So in they own. I wanted to see. Yeah. English reviews. Overwhelmingly positive.
02:01:10:13 - 02:01:43:50
Speaker 2
Very fun game once you get into it. But it's not beginner friendly in the slightest. Interesting. Going to 16 gigs, of Ram and eight gigs of Ram. That's pretty rough. That's pretty rough to to do, to try and do blender and. Arm to pad. They killed off the main character, though.
02:01:43:55 - 02:02:26:28
Speaker 2
This is interesting. So it's still like mostly very positive. I just want even. Why can't I even see any, any negative ones? This is interesting. I, this is kind of the, the funny bit about like, clearly nobody else cares. like only the. The neurotic fools of blue Sky care. That's why blue Sky is a wonderful invention. Because it keeps, you know, everybody else safe from the horde of irrational fools.
02:02:26:33 - 02:02:34:21
Speaker 2
So it's just, it's just invincible brain rot.
02:02:34:26 - 02:02:58:46
Speaker 2
What? Well, you have. If I wanted to rip my hair out and be arrested for murder, I have boot up this program for two seconds. Every time I see this funny frog on my game list, I vomit in my mouth. I fucking hate blood. I'm a very unhappy customer. I don't even know what that means. Like, I reckon that's just meant to be funny, but, like 3000 hours and I still know nothing about this program.
02:02:58:46 - 02:03:26:21
Speaker 2
Real. Oh, is it a 5500 now? Goodness, look at that. 40,000. That's that's that's pretty good. Flood crashes. I restarted blender updates. I, I relearned if blender blender adds a shortcut. I forget everything I knew before. If blender renders, I watch the samples like a campfire. If blender freezes at 99%, I stare at silence. True.
02:03:26:25 - 02:03:47:15
Speaker 2
Someone said, add sex found to Azure's, ashes review right? Actually, that features already in. You just need to be good enough at the game to get to it. Well, yeah. Yeah, I guess that's true. Most boring game. I need to do everything by myself.
02:03:47:20 - 02:04:21:48
Speaker 2
I guess so I took it, took me three hours to bake a donut and you did nothing else with it. Another tool for building volumetric. Volumetric? What? Porn is that what that means? Hey, everything is positive. like I. Oh, these are the most helpful of all time. Most helpful today. Still mostly positive about two negative reviews.
02:04:21:48 - 02:04:54:01
Speaker 2
It. Oh, these with these were the funny ones too. So like nobody's cool with this. Okay, so this one was the most helpful in the in the past week apparently. I had record it was very helpful at all. Like there really wasn't much at all that was useful. And there's this entire thing, another toe deaf company willing to participate in, in certification because I took money.
02:04:54:06 - 02:05:18:16
Speaker 2
This is kind of the funny thing is that, like, do they think that the just going to initiative initiative I it because something else did it like like like like someone else gave them money. Who makes a different tool that you think is shit.
02:05:18:21 - 02:05:22:36
Speaker 2
I mean, it's it's funny like.
02:05:22:40 - 02:05:30:15
Speaker 2
It's it's not,
02:05:30:19 - 02:05:58:12
Speaker 2
It's like the blended foundation is hardly even a company, though. Like, it's it's it's that's kind of an interesting thing to say, but like, in certification. Like, what is shitty about it? Like, this is what I'm wondering, like because because what this is, is it's just a knee jerk reaction. You touched the One Ring and therefore you've committed an unspeakable evil like.
02:05:58:17 - 02:06:18:27
Speaker 2
Having an MCP bridge does not make the program any worse. It just adds a feature in in a in a menu that you probably not going to use. Like this doesn't make sense.
02:06:18:32 - 02:06:40:55
Speaker 2
As like, how is it going to make it any shitty of that? They they support a feature that you aren't going to use. It doesn't make it any shittier. And like they record that it's going to be shitty because people are going to use it to make shitty things. how was that the case? How does that logically follow?
02:06:41:00 - 02:07:19:32
Speaker 2
And even if they use it to make shitty things like have they got the right to do that, is that part of the reason? Like, this is the exact, exact thing that they were talking about in the panel at beacon, like the people who are championing the open source, free, free and open source standard, just championing this kind of software or in this, the standards of software distribution are now like doing a massive rug poll because another form of software exists that's unimpressive, dented, like just does it make sense?
02:07:19:37 - 02:07:34:18
Speaker 2
Call me the Hollow Knight, the way I hollow in the night sky. Awoo you know, how many hours do you have? I thought you were here. I thought you were here when I showed it off. Weren't you? I just just showed it.
02:07:35:52 - 02:07:58:39
Speaker 2
There it is, 2026 hours. 2020 6.9. Pretty good. I have 2020 six hours in blender in 2026. Pretty neat. but okay. This again, it's it's funny coming back two hours.
02:07:58:44 - 02:08:02:35
Speaker 3
Another toe dev company participating in shit ification.
02:08:02:40 - 02:08:26:35
Speaker 2
You've hardly even used this for more than a day. You've hardly even use this for more than a day of total time. Like, it's totally fine to not be using the program, like, constantly. But it's funny that, like, you were going to take the position of them being like, tone deaf and and making things shitty when you don't even know how the program works.
02:08:26:40 - 02:08:43:18
Speaker 2
There's no possible way that you know how the program works. On 33.4 hours. Like, there's no way you would know what a shitty program looked like if it hit you with a fucking, the kernel. Like, I don't know.
02:08:43:22 - 02:09:05:03
Speaker 2
Half of that in Cod Cold War. Yeah, lots of zombies. Love, love zombies. What can I say? a software that was made for creators and artists. No, it was. It was made for people who needed a tool like, I don't, I don't know, like why that is part of the premise. It's made for people who need a tool.
02:09:05:07 - 02:09:27:27
Speaker 2
It's not just for creators and artists. It's made for anybody who needs any of the toolset that is in the the software. You don't have to be a great or an artist in order to find this tool useful. You can just be someone who's 3D printing shit. You don't have to be an artist 3D print shit. You could just be making purely functional things in that.
02:09:27:32 - 02:10:01:22
Speaker 2
Like it. It's just a weird thing to bring up that's graded, made for grade, is it? Art is no, it's not, not inherently. It's used by creators and artists, but it's not made for them inherently a doctor, loads of people who just they use blender to make a living and they, they don't make art. They make products like they make, visual like storytelling for the express purpose of like, criminal justice.
02:10:01:22 - 02:10:06:52
Speaker 2
Like it it's not just for create as an artist.
02:10:06:57 - 02:10:12:18
Speaker 3
Well, but they're creating something every day.
02:10:12:23 - 02:10:48:07
Speaker 2
It's silly. Now accepted corporate fund from anthropic. Why is this bad like this? There's a snuck premise in here. You completely snuck it by, like, why is this bad? It's kind of the same. Same thing that I was saying, sort of in critique of of, Andrew Price was that why is it good? It's like you can't sort of just sneak by the entire premise, like, try to actually talk about why it's good, why it's bad, but nobody actually wants to have the ethical conversation.
02:10:48:14 - 02:11:06:41
Speaker 2
Even the people who are interested in it don't want to have it. They just want to scratch. They just want to be irrational children. I what is this foolishness? This just shows that money will always. Money always wins at the end. It was a good run, but the betrayal will never be forgotten. This is not a betrayal. That's so, so fucking dramatic.
02:11:06:55 - 02:11:35:31
Speaker 2
So dramatic. The betrayal. They took $1.5 million with no strings attached from a corporation that creates a tool that they're an association. I suppose you could call them a corporation as well. And they also make a tool. They just make it with a completely different goal. And, And in philosophy, directorial philosophy in mind, like, why is this so bad?
02:11:35:36 - 02:12:05:20
Speaker 2
Is it because of how they made the tool? Is it because of who's involved in it? Is it because it's nothing? There's no reason. It's just I want to scream. there was no blood or, dick over blood. At least eight 94 hours is something like. At least that's something. Dangerously addicting. As of today, I Corporation anthropic is now funding blender.
02:12:05:20 - 02:12:29:24
Speaker 2
Anthropic has also stated they're working to integrate their cloud AI into the big tools of the creative industry. Yes, they got an MCP. There's an MCP in blender and it's not third party. It's like officially supported. That's it. Like that's literally it. And what's so bad about that? Seriously, what is so bad about that?
02:12:29:29 - 02:12:56:49
Speaker 2
If you're a regular user of blender stop. Yeah. Thanks. Thanks. 37.6 hours on record for telling other people who actually use the fucking tool what to do with it. You've hardly ever used it for like, a day and a half. Like, calm the hell down. Stop preaching the people like this. It's so sickening. Like, I can't help but smile at it because it's like, what are you talking about?
02:12:56:49 - 02:13:10:43
Speaker 2
Stop using the software and stop donating change. Do a different dual. Use a good fork of blender. Like what? Like like do or fork it yourself. Or like what?
02:13:10:48 - 02:13:32:15
Speaker 2
Because again, this eventually will just bring you down to primitivism. They used the evil technology they get. They won't tell you why it's evil. It's just it just is just is evil. They use the evil technology so you can't touch it anymore. You can't touch it. It was made by Satanists or something. Like, what is this Puritanism? Please explain.
02:13:32:20 - 02:14:16:43
Speaker 2
Please explain. Can I get one person who's actually going to explain using a philosophical, framework to actually explain why it's not ethical? Even if anthropic doesn't do anything beyond funding. It's still incredibly tone deaf, disrespectful to the thousands of orders that have had their work thrown into the AI slot machine. Having your work thrown into the AI slot machine is not the concern of blender like I don't I don't know why this is, the point that's being made.
02:14:16:48 - 02:14:57:20
Speaker 2
Like, how is it? How is it disrespectful to say like like like to say to an artist what you've made is really cool? I'm going to try and emulate that. Like it's cool until you use technology to do it. The it's cool until you automated or something like that, like it's cool until the technology gets so good that it can do it, like faster than it would take for somebody to be like, good at, I don't know, doing the like the like pencil, the paper or it's like, I don't know, whatever other intuitive stuff you could do.
02:14:57:25 - 02:15:27:59
Speaker 2
Not is intuitive the right word? creative? I don't know if you're reading this. Stop using the software the like calm the fuck down. Risk of rain. Stop using the software. Cancel your memberships if you have any. yeah. Yeah, because they have got memberships. Speak with with your time and your money. Language cannot express how tone deaf it is to make a tool.
02:15:27:59 - 02:15:51:49
Speaker 2
So many artists are and have been using for the past few decades, and surrender it to the whims of an AI company. That's not what it is. That's literally not what's fucking happening. There are no whims whatsoever involved in taking $1.5 million with no strings attached. There's no whims. It's not it's not being surrendered to the whims of an AI company.
02:15:51:54 - 02:16:19:48
Speaker 2
Like they don't have control. They're not they're not selling the free tool to them. They're not that the Blender Foundation is not being acquired. The hands are not tied here. They could flip them the bird tomorrow and keep the money like, what are you on? What is this? Yeah, like chicken with the head cut off behavior. The sky is literally not falling.
02:16:19:53 - 02:16:32:46
Speaker 2
You're not helping artist by taking this money. Yes. You absolutely the fuck. Oh, because that money is going to be used to pay artists in the blender studio. Like, what are you on?
02:16:32:51 - 02:17:03:07
Speaker 2
You're making a sign that you're willing to spit in their faces if it means you can get more profit out of doing so. That's literally not even true. The Blender Foundation Does not is a nonprofit. It like it's literally in the foundation of how the foundation functions. It's a nonprofit like, They they're not it's not about like, you know, non it's not about making profit.
02:17:03:07 - 02:17:37:50
Speaker 2
It's about making ends meet. That's why that they're funded is so that they can continue to actually create the tool and to iterate on it and to make it better, make it more efficient. Like, hey, stop associating with thieves so that there's actually an ethical point being involved here. There's no theft involved in the creation of these tools, not no generative AI, and no Lem has ever stolen anything from anybody.
02:17:37:55 - 02:18:16:48
Speaker 2
Like, unless they literally were like, hacking into people's systems, I suppose. But if anything, I'd like I don't even because this is the thing. If if it can be duplicated infinitely, then it cannot be owned. This is this is why intellectual property is is a falsity, right? If you can duplicate it infinitely, then me making a duplicate of what you've created is not harming you in any way.
02:18:16:53 - 02:18:52:50
Speaker 2
And even if you if you even if you take the IP premise and you say, yeah, intellectual property is legitimate when you've got like thousands of publicly available images and then all you do is look at those images and then you, You abstract those images into the processes that it must have taken to create those images. And then you make an image yourself based off of the processes that you've abstracted from those images you haven't stolen from anybody, even in an IP sense.
02:18:52:55 - 02:19:18:57
Speaker 2
Like even, presupposing IP, no theft is occurred. And it's the same point that enterprises made. Like every single time you start a project, you immediately go and, and rip a bunch of, art off of the web and put it into a pure href document. It like, is that not stealing? It's literally cut and paste. Why is that not stealing?
02:19:18:57 - 02:19:48:12
Speaker 2
But but this is. Like the models are abstracting same thing that humans do, except it requires a hell of a lot more. compute to do it, but less like human resources. Just kind of good. All right. At least you've got something here. But, like, I don't know, 500 that after using it for 500 hours, you decided. You know what I'm going to.
02:19:48:27 - 02:20:15:22
Speaker 2
What are you going to do? You're going to rely on a paid program. You're going to like, where are you going to go? It's like, oh, we should we should go to another fork. Like, everyone's going to hop on GPU blender now. Like I. That's the funny thing. Like stop using the software like using the software already doesn't give blender any money, and it doesn't support blender to use the software.
02:20:15:27 - 02:20:45:52
Speaker 2
What helps blender continue to do it is is actual donations, and people donate because they want to use the software like a people just calling for boycotts because they don't understand how they because they're operating off of a faulty metaphysical theory. Fundamentally, but also off of faulty, a faulty understanding of how the AI models actually work.
02:20:45:57 - 02:21:08:53
Speaker 2
Like the only possible way that that there would be like this perfect duplication. I know there's no no point of really going over it, but it's dumb. He's laughing. He's going to keep using it. Of course he is, because it's free and open source. Oh well, that's a she actually it says right that that's a she. So,
02:21:08:58 - 02:21:12:18
Speaker 2
Like.
02:21:12:23 - 02:21:23:10
Speaker 2
You just got to be like, no, I can't use the free and open source thing. Python did it to. I like, like it's it's silly.
02:21:23:15 - 02:21:47:56
Speaker 2
Her I look, I can make all my favorite silly little characters do silly little things. I love what Ada was on record need. It's funny, like 13 hours on record. After the success of flow is very excited about Blender's place at the forefront of independent animation. But now they're accepting money from anthropic and exploring generate integrating gen AI into the software.
02:21:48:09 - 02:22:12:16
Speaker 2
That's literally not true. That's literally not true. It's not generative AI. They're not integrating generative AI. Like you're just associating these things together, but they're not doing that. They've integrated a bridge for an Lem to do things in blender. That's not the same thing as generative AI. You don't understand how the fuck it works. Like these people will scream.
02:22:12:25 - 02:22:39:04
Speaker 2
They haven't used the program like you haven't even used it for a day. I've used it for 13.4 hours straight, multiple times, probably like you have to use this program for even a day, right? And you don't even understand what the tools that they're integrating are. You don't understand what they do. This is stupid.
02:22:39:09 - 02:22:59:00
Speaker 2
So that's a big nope for me. Yeah. Good for you. Calico. It wasn't it wasn't part of Blender's post. Oh, it was this. Okay, so this is Sophie. What does it mention? Them exploring AI integration. I don't remember seeing anything about that. It's been in the works for a while. They've got an MCP. That's it. They've got a model context protocol.
02:22:59:04 - 02:23:26:22
Speaker 2
It's a compatibility layer. It's a bridge. So that a LM model can interact with blender mostly just for a running scripts. That's basically all that it does. It's a collection of scripts that allows the LM to do a couple of things. That's it. And it's not gen AI. Generative AI is far more complex. It's not the same thing.
02:23:26:27 - 02:23:54:07
Speaker 2
It's not a Laura. It's not why people not understand anything about this shader. You tell them, right? Yeah. Thanks. It wasn't sort of led to post about the funding, but anthropic posted separately that Claude was officially connected with blender in their own announcement yesterday. Yes. Claude is in a lab. Claude is an LM. It's not generative. That's not what it does.
02:23:54:12 - 02:24:17:07
Speaker 2
Unless what you mean by generative is it just like, generates things. It's like. Yes, it didn't LM generates text, but generally, why would people say generative AI? What they mean is a Laura or something similar that is creating an image. But the funny thing, the funny thing is that they're already using they've had AI in blender for a long time.
02:24:17:07 - 02:24:47:19
Speaker 2
You know what it is? Open image, denoise and optics. The optics denoiser you better stop using that. Denoiser because it's it's bugging its AI. It uses like a Laura in order to, to guess where the noise ought not to be. There's nothing wrong with this. Everyone's been brain rotted because someone made a computer hallucinate images, and now everyone's upset.
02:24:47:24 - 02:25:11:58
Speaker 2
fuck! I fuck heads. Robert. Fuck! Claude! Hey, listen, Claude, I don't I don't know where we go with this. don't give thanks to the people that stole from us and now want to replace us. Nobody stole from anybody. There's no stealing here. You can't steal an idea. That's not how ideas work. It's not theft. It's not theft to control C control V something.
02:25:11:58 - 02:25:24:31
Speaker 2
If that were the case, and all of you blue sky idiots would have already been, guilty of theft when you were railing against NFTs before blue Sky existed.
02:25:24:36 - 02:25:55:37
Speaker 2
They want to replace us. See, the funny thing is that if you really were totally on board with going the going with AI and do it doing generative AI, then you wouldn't need something like blender. But blender clearly isn't doing that. Like it's a tool that allows you to speak in your native tongue through the program, and then have it spit out a couple of results to help you get where you want to go faster.
02:25:55:42 - 02:26:25:49
Speaker 2
That's all it does. That's all the lab is for. Like making porn is a cost effective business strategy. Well, you've had it here, folks. versus this 2576.5 hours on record, a little bit more than me. Blood has gotten so advanced over the past decade. The last decade or so that have realistically, there's not much that it can't do for working through nearly any 3D workflow pipeline.
02:26:26:02 - 02:26:42:33
Speaker 2
Please learn how to use the software for a good few hundred hours or so. There's entire to the entire tutorial series out online and on YouTube. I'm planning on making a hopefully easier to follow tutorials series for people, and whenever I get around to doing that, I'll update it in this review. Overall, biggest things to keep in mind.
02:26:42:37 - 02:27:00:04
Speaker 2
Follow some beginner's tutorials to the letter. Don't worry about making something original immediately. Learn the 8020 rule. Stuff the 20% of the rule, the tools available in the software you'll be using 80% 80% of the time, keeping with a full tutorial project from start to finish will give you more of a clear roadmap to measure your learning progress.
02:27:00:09 - 02:27:19:45
Speaker 2
Take notes, particularly take notes of hotkeys. Because blender isn't massively hotkey heavy. Make some similar things to the tutorials that you've that you follow that aren't just the things from the that the things the tutorial showed. Try to re replicate a whole process for a new project that you've learned from a tutorial. Just trying to follow the operations shown from memory.
02:27:19:49 - 02:27:38:24
Speaker 2
Don't take the tutorial projects too seriously. Go through them. Take whatever mental slash visual notes and takeaways you can from the end result, and then move on to the next thing. Tutorial projects are like loose sketches in a sketchbook. They don't have to be polished finished pieces. Just keep at it and learn from each new project. Some good, good tutorial creators on YouTube are as follows.
02:27:38:24 - 02:28:02:42
Speaker 2
Blender guru Grant Abbott I haven't heard of him. Ryan King Art he's pretty good CG matter. He's pretty cool. He made a he. He basically made his own, local lamb integration into into into blender called blender body. And people still marveled about that. the CG essentials, flip normals, all pretty, pretty good stuff. Realistically, at this point, there's not really not much blender can't do.
02:28:02:42 - 02:28:37:05
Speaker 2
Future updates. I seriously hope include a full, more fully functional layer operated in similar texturing utility set similar to that of 2D graphics editing software's Krita. No, that's not okay. So blender is for 3D mainly. So they've got a couple of 2 to 2 2D tools, but they're not meant to, you know, they're not for texturing necessarily. Like you're still going to get some pretty good like like a lot of better results outside of blender if you're trying to, do a lot of that stuff.
02:28:37:10 - 02:29:02:53
Speaker 2
Clip studio, paint, Photoshop, Substance Painter, etc. it'd be nice. Like I'm not opposed to them, getting more 2D stuff in there, but like, you know, Maya didn't have a great amount of 2D stuff and their UV tools were better. That's part of the reason why I that I was one of the headaches that I ended up with in coming to blender, that, you know, they didn't have as good of a UV toolset.
02:29:02:58 - 02:29:17:06
Speaker 2
but also all the time in blender, you'd have to in Maya, you'd have to take stuff out and then do your texturing externally, either in Photoshop or in Substance Painter or something. So it's like.
02:29:17:11 - 02:29:40:30
Speaker 2
I don't know that it doesn't. I don't think that's necessary. That's not not really on the on the priorities there. Not so much. And the sculpting mode of butter isn't necessarily the same kind of sculpting experience you'd get if you were using ZBrush. But blend has been getting lots of sculpt mode updates lately too. So at this point, if you don't have nearly a grand to drop down on ZBrush, blender sculpting mode ought to be fine.
02:29:40:35 - 02:30:12:11
Speaker 2
I totally agree with that. I've used it before. It's pretty good. It's easy to run a nuclear reactor than this. Give me a specs i5 900 CPU. That's a that's an old CPU, but 2.9GHz, 32GB. what? what what are the specs of that thing? ninth gen, for reference, we're on 14th gen currently, so this is quite a few years ago.
02:30:12:16 - 02:30:50:04
Speaker 2
It only supports Ddr4 2666 going to. Okay. Thank you, thank you. Let me hope it only supports Ddr4 2666. launch take quarter one 2019. You're on it. You're on a seven year old CPU, but six cores, six threads, base frequency, 2.9GHz, TDP, 65W. Yeah, because it's an i5. So you're on like a mid tier CPU from six years ago that didn't even support the fastest memory type of that generation.
02:30:50:09 - 02:31:15:34
Speaker 2
Right. Didn't ninth gen support 3200? Maybe. Maybe I'm schizo, but, now surely it did, because, I built my PC in in like the platform that I was using had support for 3200 and do with the way it was built in 2020. So yeah, for LGA 1151. Yeah. This is, you're under spec, but of course you get to struggle with that 30, 60, 12 gigs.
02:31:15:34 - 02:31:50:48
Speaker 2
It's it's it's okay. You could you can deal with that. But you need a platform upgrade. But. You seriously need a platform upgrade if you want to. If you want to run this program, I would not recommend that. It's got 36.1 hours on record. when I'm on a ruining reputation, competence, competence ruining reputation competition. Speaking of lack of competence here, it's a very.
02:31:50:52 - 02:32:16:07
Speaker 2
And my opponent is an open source foundation, by the way, I attached my the hardware because I can't update my PC thanks to AI data centers. What? Oh, like upgrades? It's like what? You can't update you. I was thinking you meant software. Like, what are you talking about? Ryzen seven 2700. Okay. Like, how little money have you got?
02:32:16:12 - 02:32:51:14
Speaker 2
You can get something better than a 2700. Absolutely. Like, yeah. 564 gigs of Ram. Like, that's really not. That's really not enough. And your own Radeon to like, metal is, not as good a back end is as, as objects of Cuda. So. Yeah, I, of course, is struggling. but I don't think that has much to do with AI data centers.
02:32:51:14 - 02:33:16:21
Speaker 2
You probably should have upgraded this like two years ago before the market went to shit anyway. Like, what are you doing? I got my 3090 ages ago and. Okay. Anyhow. the thieves at anthropic. I slop thieves, it's it's just nonsense.
02:33:16:26 - 02:33:44:12
Speaker 2
Point out where the theft is. Who has thieves? Where is the theft? Where does it occur? Because this is the thing they were talking about. Like AI, art, thieves and all of this. But like. All they were doing is integrating in Lem. That's it. That, like, that's literally at just an lamp. There's no like, a steam is on blue sky.
02:33:44:16 - 02:34:04:09
Speaker 2
Goodness. This is cute, don't you reckon? Anyway. I think I've said enough. I was going to play a game after this, but, I took too long, so we're not doing that.
02:34:07:22 - 02:34:28:49
Speaker 2
Well, yeah. I mean, It's about it, I guess. Like, this is. I'm not really much more left to say. AI stands for ablest intimacy. What is that? What are those two words together or even mean?
02:34:28:54 - 02:34:36:51
Speaker 2
The ding ding ding ding ding ding ding ding ding ding ding.
02:34:36:55 - 02:35:07:21
Speaker 2
Anyhow, yeah, I have. I haven't had anything to say. I haven't had anything to say. It's done. It's over a lengthy rant. No, it was a response. Call it that. I meant to be fair that what happened afterwards was was longer. But anyway, it's over. Goodbye. Like if you were, tuning in, I kind of wish I actually got to play some games as part of the reason I didn't want to do trickery discourse for a while.
02:35:07:21 - 02:40:12:03
Speaker 2
I was like, oh, I've got a sick of, have to do all of it. I don't get to play as many games this way, but that's just my lot in life. I know. Thank you for being here. Thank you. Thank you for abiding my, brambles.