1
00:00:00,000 --> 00:00:03,720
I am here with Dan. Hey, and what do you do for the company here?

2
00:00:04,240 --> 00:00:08,960
I'm the infrastructure and technical specialist these days. I pretty much just write documentation

3
00:00:09,360 --> 00:00:14,840
Sounds good. Sounds good. So I'm just asking everyone here at the company what the thoughts on AI is the current state of it

4
00:00:15,120 --> 00:00:19,760
So I'll ask you first. Sure thoughts on the currency of AI. It's pretty interesting

5
00:00:19,760 --> 00:00:26,760
there's a lot of cool stuff going on and I'm excited to see it kind of develop over the next I guess a year or a few months even

6
00:00:27,120 --> 00:00:31,240
So you talk kind of sound like you're for AI more than against AI

7
00:00:31,720 --> 00:00:35,320
Yeah, I would probably say my thoughts on it are a little bit mixed

8
00:00:35,320 --> 00:00:40,880
I think it's not maybe as terrifying as people think it is but yeah, certainly certainly for it

9
00:00:41,400 --> 00:00:46,840
Sounds good. Sounds good. So if you're for it, then what's one thing you're really against in terms of the state of AI at the moment?

10
00:00:46,840 --> 00:00:50,320
I think probably some of the protectionism. I mean

11
00:00:51,200 --> 00:00:53,920
Looking at some of the requirements for doing these trainings

12
00:00:54,400 --> 00:01:02,160
You're spending millions of dollars on like cloud infrastructure GPU time just to actually create a

13
00:01:02,720 --> 00:01:11,960
reasonable model and so People I think are willing to protect that rather than just open sourcing it and letting any other company who has access to these massive data centers

14
00:01:11,960 --> 00:01:22,000
Like get in there, right? Okay, okay So have you used any AI currently like that where it's like they have chat GBT the image random readers music

15
00:01:22,200 --> 00:01:27,400
Have you use any of it? Yeah, I've played with pretty much all of them and I actually run a couple at home

16
00:01:27,400 --> 00:01:34,360
So I run an LLM as well as stable diffusion for image generation and I run those locally

17
00:01:34,360 --> 00:01:42,340
I recently just bought a couple more 30 90s and you know upgraded some of them with VRAM so that I could try some of the larger models

18
00:01:42,720 --> 00:01:45,220
So they're like a chat GBT for

19
00:01:45,880 --> 00:01:49,180
But you can only talk to them for like I don't know

20
00:01:49,780 --> 00:01:57,260
Two to five minutes before the like token inference length gets too much and it takes like a minute to respond to you

21
00:01:57,540 --> 00:02:06,580
But initially they're they're really quite powerful And they're a little bit more entertaining because they're the uncensored varieties. So what do you mean by that?

22
00:02:06,620 --> 00:02:11,300
So, you know, you say something to chat GBT for and it goes as a large language model

23
00:02:11,300 --> 00:02:16,260
I can't respond to this the uncensored one the uncensored ones are kind of designed as

24
00:02:17,060 --> 00:02:22,980
training bed for censorship and so you can kind of get it to talk about whatever you want or

25
00:02:23,900 --> 00:02:26,420
You know, sometimes chat GBT for will

26
00:02:27,020 --> 00:02:30,580
Catch things that it thinks are offensive and it isn't

27
00:02:31,020 --> 00:02:33,780
I have some people have told me some stuff internally

28
00:02:34,540 --> 00:02:42,260
That they use chat GBT for to have like difficult conversations about difficult topics and practice talking to actual people

29
00:02:42,660 --> 00:02:48,380
about these difficult subjects and Initially GPT 3 was kind of okay to talk about these

30
00:02:49,180 --> 00:02:52,900
but GPT 4 just won't even broach these subjects that are kind of like

31
00:02:53,460 --> 00:02:58,660
You know, they're they're hard to talk about is it cuz like they're trying to market it. I guess I'm assuming or yeah

32
00:02:58,660 --> 00:03:02,820
I was probably a marketing thing. There's also a liability thing probably

33
00:03:03,860 --> 00:03:08,220
You know people like to post what they talk about and then of course you also have like

34
00:03:08,980 --> 00:03:16,980
Really negative sentiment and things like that and also you can ask it dangerous things and get it to tell you dangerous information

35
00:03:17,900 --> 00:03:27,060
and so it's really kind of Scary in that way and I think as these companies are large and have a lot of money invested in these it's a little bit

36
00:03:28,420 --> 00:03:34,580
Be a difficult for them to allow that kind of like unfiltered information because they are so

37
00:03:35,260 --> 00:03:41,340
Confident at all the time. They just will a hundred percent confidence in what they're saying

38
00:03:41,340 --> 00:03:45,860
Even if they don't know what they're saying and that I think can be really dangerous

39
00:03:45,860 --> 00:03:49,100
A lot of people don't understand that it's fake information a lot of the time

40
00:03:50,060 --> 00:03:54,180
Another cool thing about running them locally is that you get to like mess with their brain a little bit

41
00:03:54,180 --> 00:04:01,860
So you can adjust the level of inference and the level of like token following guys called config scale or sorry configuration scale and

42
00:04:02,540 --> 00:04:05,700
You can make it those just be like

43
00:04:06,660 --> 00:04:11,300
Kind of insane if you go too far you can get into something called hallucinations where

44
00:04:11,940 --> 00:04:16,780
The inference models like feedback into themselves and it just says like random garbage

45
00:04:18,140 --> 00:04:23,140
That's can be really quite funny and and interesting. Yeah, it's neat watching them tick like that

46
00:04:23,180 --> 00:04:31,100
Okay, so you use it on a person level, but do you think as a tech slash crave company? We should be using AI on any capacity

47
00:04:31,100 --> 00:04:38,060
I actually did some training for us. I did a fine tune of chat GPT for

48
00:04:39,340 --> 00:04:44,060
We were trying to get it to write our with a hell of a cult showcase scripts

49
00:04:44,300 --> 00:04:49,500
Which are pretty boring and basic videos where a brand will be like

50
00:04:50,020 --> 00:04:54,340
Write a sponsored segment about our product and here's a bunch of

51
00:04:54,700 --> 00:04:59,300
Specifications and then we kind of have to turn that into a script and it's not very creative and it's not very fun

52
00:05:00,300 --> 00:05:03,980
So I made a data set from about 50 of these scripts

53
00:05:05,140 --> 00:05:15,540
You know because you have to you have to set a Prompt and a response so the prompt was like here's all the data and specifications on the brand and the response is our script

54
00:05:15,740 --> 00:05:18,980
right and So I went this data set

55
00:05:18,980 --> 00:05:24,900
I did a quick fine tune and it wasn't performing very well because the minimum data that you need for one of these types of fine

56
00:05:24,900 --> 00:05:30,980
tunes is 500 Right, so 500 scripts our LTT's we have about 5,000

57
00:05:31,020 --> 00:05:34,900
so that would be a reasonable data set to feed into a fine tune, but

58
00:05:35,460 --> 00:05:40,300
you know LTT's are special and You know there they wouldn't really work

59
00:05:40,700 --> 00:05:42,980
But these showcase scripts are pretty

60
00:05:44,140 --> 00:05:48,140
Boiler play Unfortunately, I was told after I asked for the other

61
00:05:48,980 --> 00:05:53,260
450 scripts that we have only ever done 50 of them so that project

62
00:05:53,500 --> 00:05:57,860
Through It was a it was a waste of a couple days of data compilation

63
00:05:57,900 --> 00:06:04,180
But did you learn a lot like what to do what not to do kind of thing? Yeah. Yeah, I learned a lot about how fine tunes function and

64
00:06:05,100 --> 00:06:08,940
You know, there's there's also a lot of issues with model overfitting

65
00:06:08,980 --> 00:06:12,540
So you can you can train a model too much and then it can only do

66
00:06:13,100 --> 00:06:20,300
One thing and that's kind of a dangerous balance to and so you know if you feed in that 50 data set and you like

67
00:06:20,300 --> 00:06:26,500
You really try to make it Do that 50 data set then you'll just get garbage because it's over fit

68
00:06:26,500 --> 00:06:33,220
And if you try to go anywhere outside of that entire limited data set like if you train on an LLM on

69
00:06:34,260 --> 00:06:37,860
100,000 pictures of dogs and their only face on

70
00:06:38,500 --> 00:06:46,020
Then you ask it to do a side-on picture of a dog and it can't because it's over fit for like front-facing dogs

71
00:06:46,580 --> 00:06:52,060
Or if you ask it to make a cat it can't do that like we couldn't feed in 50 showcase scripts

72
00:06:52,300 --> 00:06:57,540
Or 5,000 showcase scripts and then have it make an LTT because it would just make showcase scripts

73
00:06:58,100 --> 00:07:02,380
Sorry, so we should Didn't use it for a company or

74
00:07:03,140 --> 00:07:07,380
Well, I mean this is kind of the problem and this is something that I've realized more

75
00:07:07,940 --> 00:07:14,820
playing with them at home and getting it to let it spend a lot of time with them that you wouldn't be able to do with like a

76
00:07:15,180 --> 00:07:22,280
Chat GPT-4 or one of these web things because I'm sending like a hundred thousand or two hundred thousand tokens through

77
00:07:22,460 --> 00:07:29,220
These platforms and that would cost me like a five hundred dollars a day to actually get to play with them

78
00:07:30,340 --> 00:07:35,660
But I think even more now. I'm more convinced that they are just tools

79
00:07:36,220 --> 00:07:40,260
I know some artists friends of mine have been concerned about them

80
00:07:40,980 --> 00:07:49,860
but what I've noticed is that Because these LLMs and the image diffusion models are trained on massive quantities of data

81
00:07:50,820 --> 00:07:56,660
the Massive quantities of data that they have are like the most generic pictures

82
00:07:57,180 --> 00:08:01,180
Right there the generic pictures. There are single sort of style of picture

83
00:08:01,940 --> 00:08:13,140
there are sort of a You know, it's it's one thing and so I think at least for artists what's gonna happen is maybe there'll be less of a demand for

84
00:08:14,020 --> 00:08:21,300
This basic art, you know, I want, you know a pin-up of a girl or I want a scene of a mountain that sort of thing

85
00:08:21,300 --> 00:08:28,820
You know, there's already a massive quantity of art out there that has those and you can use a diffusion model to create that kind of

86
00:08:28,820 --> 00:08:32,980
Either like a background or like I want a space station and then you have a space station

87
00:08:32,980 --> 00:08:37,660
but a student is to start getting to like weirder things or like more niche things or

88
00:08:38,700 --> 00:08:47,500
more unique things like I Don't know a mountain that's upside down, but in a space station that sort of thing a language model and

89
00:08:47,660 --> 00:08:56,140
and Diffusion model might like start to struggle with that because there's never been any data on that and it can only smush

90
00:08:56,500 --> 00:08:59,700
Different data points together so well and I do expect it to get better

91
00:09:00,540 --> 00:09:05,100
But it's also really difficult to prompt those That kind of creativity

92
00:09:05,100 --> 00:09:08,660
However, saying that you can use diffusion models as a tool

93
00:09:08,780 --> 00:09:14,380
So I know some artists who use them to create characters or like you want to do D&D art that sort of thing

94
00:09:14,860 --> 00:09:23,020
Or you need inspiration Or you would like, you know, 500 different dynamic poses that feature

95
00:09:23,500 --> 00:09:28,740
You know, like a ballet dancer or something like that. I want 500 ballet poses so that I can

96
00:09:29,660 --> 00:09:36,900
Get an immediate mood board as inspiration to do my own drawings or my own digital art or you know, even analog art, right?

97
00:09:36,900 --> 00:09:41,820
I think The same can kind of be said for the LLMs

98
00:09:42,620 --> 00:09:46,540
the large language models because they

99
00:09:46,540 --> 00:09:52,580
They spit out text, but you can't really have a conversation with them for very long

100
00:09:53,140 --> 00:09:55,620
Their brains don't work. I think that there's some

101
00:09:56,620 --> 00:10:00,220
There's some like papers out right now and there's also some

102
00:10:00,900 --> 00:10:06,580
Like prize pools if you can get it to go over like a hundred thousand token

103
00:10:07,620 --> 00:10:10,900
Inference length so like a history so you could have a conversation with it

104
00:10:10,900 --> 00:10:15,540
that would be about the length of Lord of the Rings and it would be able to remember and

105
00:10:16,420 --> 00:10:23,140
Make reference to anything in that entire block of text, right? That's kind of the that's the goal

106
00:10:23,140 --> 00:10:27,020
Right, you can you can actually have real conversations with them for a long time

107
00:10:28,100 --> 00:10:31,860
the The local models also support characters

108
00:10:31,860 --> 00:10:36,700
So I'm trying to create a D&D character for a friend of mine so that he can actually have a

109
00:10:37,180 --> 00:10:41,220
Conversation with the character and you do that with a bunch of different prompts and things like that

110
00:10:41,220 --> 00:10:44,540
There's already a bunch of like web stuff that you can do online

111
00:10:44,740 --> 00:10:50,300
Like a lot of people have already like taken this and go like D&D. We can do that on the internet hell

112
00:10:50,300 --> 00:10:55,020
Yeah, and so these are available like through web portals and stuff like that, but running them at home is fun

113
00:10:55,060 --> 00:10:59,940
So, yeah, that's a long-winded answer. I don't remember where the question was. I'm sorry. I'll go

114
00:11:02,300 --> 00:11:05,660
Should we use them should we use them? I think using them as a tool is fine

115
00:11:05,660 --> 00:11:09,500
I think there's no danger in them like replacing people

116
00:11:10,300 --> 00:11:15,580
But they're gonna work with people and there's no there's no escape from that now that genie is out of the bottle forever

117
00:11:15,940 --> 00:11:21,900
All right, all right We talked a lot, but is there anything that we haven't said that you like to say to the float planers?

118
00:11:21,900 --> 00:11:24,980
I was what is this video about? I'm sorry

119
00:11:24,980 --> 00:11:31,180
It's just like I just want to get people's thoughts on AI because you know We are a tech slash creative company, which I think aligns with both with AI

120
00:11:31,180 --> 00:11:34,900
So it's like I think people have some opinions and I think it's good to share. Yeah

121
00:11:35,820 --> 00:11:41,220
Try it at home. It's fun at home. It's a fun little exercise. It's really really simple to get going and

122
00:11:41,940 --> 00:11:45,300
It's it's a fun little break playground and you don't have to like

123
00:11:46,820 --> 00:11:53,780
Spend money on chat GPT. I mean chat GPT is still the best, but if you play with it at home, you're yours

124
00:11:54,420 --> 00:12:01,420
You can play with it as much as you want. Yeah. All right, Dan. Thanks for any of you. I'm still sweaty. It's hot

125
00:12:04,100 --> 00:12:06,100
Thank you
