The player is loading ...
84: Can A.I. Save Writing? A Special Collaborative Episode with Eric James Stephens, Writing Sensei & Business Intelligence Developer
Apple Podcasts podcast player iconSpotify podcast player iconAmazon Music podcast player iconiHeartRadio podcast player iconPandora podcast player iconCastro podcast player iconRSS Feed podcast player iconAudible podcast player icon
Apple Podcasts podcast player iconSpotify podcast player iconAmazon Music podcast player iconiHeartRadio podcast player iconPandora podcast player iconCastro podcast player iconRSS Feed podcast player iconAudible podcast player icon

In 2022, Open AI launched ChatGPT. Overnight, the company started a global conversation about the future of artificial intelligence.

The emergence of AI has turned learning on its head. It has forced us to ask major questions:

-How does AI fit into the future of learning?
-How can AI make learning more personalized and adaptive?
-What are the ethical implications of AI?

This podcast will tackle these questions and more!


Follow Eric on LinkedIn

Thanks for tuning in! 🎧

 

1
00:00:01,200 --> 00:00:06,800
Hello, my name is Holly Owens
and welcome to Ed up edtech the

2
00:00:06,800 --> 00:00:10,500
podcast that keeps you.
In the know about all the latest

3
00:00:10,500 --> 00:00:14,300
edtech happenings.
We interview guests from around

4
00:00:14,300 --> 00:00:17,500
the globe to give you deeper
insights into the Ed tech

5
00:00:17,500 --> 00:00:21,500
industry, the field of
instructional design, and more,

6
00:00:21,800 --> 00:00:25,200
we're proudly a part of
America's leading podcast

7
00:00:25,200 --> 00:00:29,900
Network the Ed up, experience.
It's time to sit back and enjoy.

8
00:00:30,000 --> 00:00:33,000
Enjoy the latest episode of Ed
up attack.

9
00:00:39,300 --> 00:00:43,900
Hello and welcome to the first
collaborative podcast of the Ed

10
00:00:43,900 --> 00:00:48,000
up, AI podcast and the Ed up.
Ed Tech podcast with Holly

11
00:00:48,000 --> 00:00:50,700
Owens.
And our guest today that we're

12
00:00:50,700 --> 00:00:53,000
going to be talking to and
interviewing and learning

13
00:00:53,000 --> 00:00:56,300
everything about in terms of AI
and everything else is Eric.

14
00:00:56,300 --> 00:00:59,500
James Stevens.
Iarc welcome.

15
00:00:59,700 --> 00:01:01,800
Thank you, both, for having me
here.

16
00:01:01,800 --> 00:01:04,000
I am very excited to have this
conversation with you because

17
00:01:04,000 --> 00:01:07,000
this is just cool stuff.
Absolutely, it's going to be

18
00:01:07,000 --> 00:01:09,600
fun.
I'm super Sighted.

19
00:01:09,800 --> 00:01:14,800
And I can jump in with the first
question Eric and you can feel

20
00:01:14,800 --> 00:01:18,100
free to address it how you want
and work in what you're doing

21
00:01:18,100 --> 00:01:20,800
right now and everything else.
Cuz I know that right now, you

22
00:01:20,800 --> 00:01:24,900
have your own daily job that
you're doing full time and then

23
00:01:24,900 --> 00:01:29,300
you're also starting something
and are and an entrepreneur for

24
00:01:29,300 --> 00:01:32,100
rhizome.
So how would you describe what

25
00:01:32,100 --> 00:01:36,900
you're doing at your they job
with AI and then also with that

26
00:01:36,900 --> 00:01:40,700
other projects, okay?
I like that question the job

27
00:01:40,700 --> 00:01:42,900
that I have.
Now, I'm a business intelligence

28
00:01:42,900 --> 00:01:48,400
developer working with power bi.
And so this is a company that

29
00:01:48,400 --> 00:01:51,800
construction company that's been
around for over 100 years but

30
00:01:51,800 --> 00:01:54,000
now they're asking themselves,
like, what can we learn from our

31
00:01:54,000 --> 00:01:57,100
data?
And that's where I love to be

32
00:01:57,100 --> 00:02:00,100
in.
Love to work is figuring out

33
00:02:00,300 --> 00:02:02,100
complex problems.
What's that?

34
00:02:02,100 --> 00:02:04,800
A lot of fun?
I've had chat gbt open the whole

35
00:02:04,800 --> 00:02:08,699
time asking like what does this
mean in this environment or I

36
00:02:08,699 --> 00:02:10,199
used?
Do this at Tableau.

37
00:02:10,500 --> 00:02:12,000
How do I do it?
Power bi.

38
00:02:12,300 --> 00:02:14,500
And it will tell me the
different calculations that I

39
00:02:14,500 --> 00:02:16,700
need.
It's been a lot of fun, but

40
00:02:16,700 --> 00:02:19,300
that's not why we're here, we're
here because we're talking about

41
00:02:19,400 --> 00:02:22,400
Ai and star.
I am working on the sting of

42
00:02:22,400 --> 00:02:27,700
called project rise up and I am
building an AI powered teaching.

43
00:02:28,000 --> 00:02:33,100
I believe that one of the
biggest things that teachers

44
00:02:33,100 --> 00:02:36,900
need back in their lives is time
and one of the biggest time

45
00:02:36,900 --> 00:02:40,200
constraints that they have is
the S of grading writing

46
00:02:40,200 --> 00:02:43,100
specifically.
I want to be able to create

47
00:02:43,100 --> 00:02:47,900
something that helps teachers.
Do that task better and faster.

48
00:02:47,900 --> 00:02:51,800
In a way that students also
learn how to write better than

49
00:02:51,800 --> 00:02:57,300
they are right now as someone
who has been thinking about and

50
00:02:57,300 --> 00:02:59,500
writing about the ethical use of
Big Data.

51
00:02:59,500 --> 00:03:05,100
Since 2016, I feel like I I
cannot.

52
00:03:06,500 --> 00:03:08,600
Watch someone else build
something.

53
00:03:09,400 --> 00:03:12,700
And then critique it, if I was
an academic, that still what I

54
00:03:12,708 --> 00:03:15,200
would be doing is I wouldn't
have a saying but now that I

55
00:03:15,208 --> 00:03:18,200
know how to do data and now that
I know about these different

56
00:03:18,200 --> 00:03:21,500
things I want to be able to be
the one to build it to be able

57
00:03:21,500 --> 00:03:24,500
to say, hey let's bring in a
variety of people.

58
00:03:24,500 --> 00:03:27,500
Let's do these different things.
At this first phase of product,

59
00:03:27,500 --> 00:03:32,300
rhizome that were talking about
is a Grassroots data collection

60
00:03:32,300 --> 00:03:36,100
for student essays.
I want to create an AI that can

61
00:03:36,300 --> 00:03:42,400
Assess student writing, I
fundamentally, I'm against using

62
00:03:42,400 --> 00:03:46,100
data that's been scraped or
bopped from the web or from a

63
00:03:46,108 --> 00:03:49,700
University universities can sell
writing from their students.

64
00:03:49,700 --> 00:03:51,200
I don't think that's a good
thing.

65
00:03:51,300 --> 00:03:54,500
I don't think people know a lot
about that, they don't assume

66
00:03:54,500 --> 00:03:58,300
that and that's totally true.
That's definitely a need to

67
00:03:58,300 --> 00:04:01,600
know.
It is idea who owns an essay, so

68
00:04:01,600 --> 00:04:05,100
I believe that the person who
wrote it, owns it and I want to

69
00:04:05,100 --> 00:04:08,100
collect it from them.
I applied to this idea,

70
00:04:08,100 --> 00:04:10,500
accelerator called Builders and
backers, and they're being

71
00:04:10,500 --> 00:04:12,900
funded by Heartland forward.
Either be amazing people.

72
00:04:12,900 --> 00:04:15,600
Take you so much all those
people, but they gave me five

73
00:04:15,600 --> 00:04:19,000
thousand dollars in order to
conduct an experiment.

74
00:04:19,200 --> 00:04:23,300
Other people are hiring coders
to build things or their bill

75
00:04:23,300 --> 00:04:25,600
buying product.
I wanted to give that five

76
00:04:25,600 --> 00:04:28,900
thousand dollars away.
I'm giving away five thousand

77
00:04:28,900 --> 00:04:32,200
dollars to ten different people,
10 people get 500 bucks.

78
00:04:32,200 --> 00:04:35,200
I already gave away one person
hurting semper, she's awesome.

79
00:04:35,200 --> 00:04:38,400
Gave her 500 bucks last weekend
and this weekend I'm giving away

80
00:04:38,400 --> 00:04:42,600
to someone else $500.
I am asking people to post about

81
00:04:42,600 --> 00:04:43,700
it.
I'm trying to share it.

82
00:04:43,800 --> 00:04:46,700
I'd convinced my kids to pass
out flyers with me.

83
00:04:46,700 --> 00:04:49,700
Last night, I went over to, I
went to the University of Tulsa

84
00:04:49,700 --> 00:04:53,800
and we did break into library,
but we found somebody to let us

85
00:04:53,800 --> 00:04:56,100
into the library.
We just surpassed out flyers to

86
00:04:56,100 --> 00:05:02,500
people because I want this aai
and not be an individual person

87
00:05:02,500 --> 00:05:05,500
building, it has to be a
collaboration and it has to be

88
00:05:05,500 --> 00:05:08,700
done on ethically sourced data
and that's what I wanted.

89
00:05:09,900 --> 00:05:12,600
Well, this sounds amazing.
I'm so glad that were partnering

90
00:05:12,600 --> 00:05:17,400
for this episode because this is
totally Ed up at Tech jam

91
00:05:17,400 --> 00:05:20,900
without Eric this journey, I
know you boast three episodes

92
00:05:20,900 --> 00:05:23,700
and then add as they'll walk.
But now I'd it's back up and

93
00:05:23,700 --> 00:05:26,200
going I can't not.
Isn't that what we talked about

94
00:05:26,200 --> 00:05:28,900
when we were talking with Linda?
E is how that's how it works

95
00:05:28,900 --> 00:05:31,600
though.
You stop it and then you go back

96
00:05:31,600 --> 00:05:33,000
to and you go back to the
drawing board.

97
00:05:33,000 --> 00:05:37,300
So we're glad you're back.
You're back at this spot here.

98
00:05:37,300 --> 00:05:41,600
Tell us about more about what
people can do to help you

99
00:05:41,600 --> 00:05:44,400
collect this data.
The you mentioned, the 5,000,

100
00:05:44,400 --> 00:05:46,700
and the 500.
And obviously, we'll put

101
00:05:46,700 --> 00:05:48,900
everything in the show notes if
you're interested in

102
00:05:48,900 --> 00:05:52,000
contributing to Eric and the
data that he is collecting.

103
00:05:52,300 --> 00:05:56,300
So, tell us a little bit more
about that situation right now.

104
00:05:56,300 --> 00:06:01,000
There are really Things that
people can do to really help

105
00:06:01,000 --> 00:06:04,600
build one is the actual like
helping collection of essays.

106
00:06:04,600 --> 00:06:07,500
So if you have essays submit
them and I'm not I don't just

107
00:06:07,500 --> 00:06:10,600
need them from current students.
I'm about to get my perspective

108
00:06:10,600 --> 00:06:12,500
on my dissertation.
Exactly.

109
00:06:12,500 --> 00:06:15,400
So it's between 3 and 15 pages,
right?

110
00:06:15,400 --> 00:06:17,500
But it could be for any
discipline, right?

111
00:06:17,500 --> 00:06:20,500
Because and the reason that it's
called rise own as this is

112
00:06:20,900 --> 00:06:24,800
because I believe like writing
is a rhizomatic thing.

113
00:06:25,400 --> 00:06:27,400
A rhizome is something that
grows nebula.

114
00:06:27,500 --> 00:06:29,600
Ashley, there's no single point
of origin.

115
00:06:30,000 --> 00:06:32,000
Writing has no single point of
origin.

116
00:06:32,000 --> 00:06:37,900
There is an every discipline is
infested with this writing

117
00:06:38,000 --> 00:06:40,300
thing, any writing from any
class.

118
00:06:40,600 --> 00:06:44,400
I will take because I believe
fundamentally that what makes a

119
00:06:44,407 --> 00:06:48,000
good transition in a history
class is a good transition in a

120
00:06:48,000 --> 00:06:51,100
biology class, anybody who's
selling something otherwise is

121
00:06:51,100 --> 00:06:54,700
selling you something.
That was fun of it.

122
00:06:55,200 --> 00:06:58,700
But so there's the one there's
actually like collecting data or

123
00:06:58,700 --> 00:07:01,500
collecting assays.
I'm a this I'm hoping to

124
00:07:01,500 --> 00:07:04,900
graduate students are hearing me
well, except their writing and

125
00:07:04,900 --> 00:07:06,500
I'll accept their students
writing.

126
00:07:06,500 --> 00:07:10,300
They have the other thing is
that I'm collaborating with the

127
00:07:10,300 --> 00:07:13,600
Ed up experience as a whole with
joy Holden and we're doing a

128
00:07:13,608 --> 00:07:17,400
state of student writing survey
where professionals like

129
00:07:17,400 --> 00:07:20,700
yourselves education leaders
teachers can go out and share

130
00:07:20,700 --> 00:07:23,800
their opinions about what they
feel on the state of Writing.

131
00:07:24,100 --> 00:07:25,800
We're going to take those
results, are going to publish a

132
00:07:25,800 --> 00:07:27,400
white paper and that's going to
be fun.

133
00:07:28,000 --> 00:07:31,700
The Third Way that people can
help is by signing up to learn

134
00:07:31,700 --> 00:07:35,200
more.
This next phase right now is

135
00:07:35,500 --> 00:07:40,800
phase one that's getting my data
base to is training my data.

136
00:07:41,200 --> 00:07:43,900
Now I'm going to be sitting by
myself, reading whole bunch of

137
00:07:43,900 --> 00:07:47,300
sentences and Grading and
training an algorithm about how

138
00:07:47,300 --> 00:07:50,200
would Eric rate of paper.
Or I'm not even going to find a

139
00:07:50,207 --> 00:07:52,800
team of five people.
And I'm so we could say like,

140
00:07:52,800 --> 00:07:55,600
how would this group of five
people read these essays?

141
00:07:56,200 --> 00:07:59,600
I'm going to use a crowdsource
model where people can log-in.

142
00:07:59,600 --> 00:08:02,500
I could vet there, I can make
sure they have credentials and

143
00:08:02,500 --> 00:08:05,800
they should be assessing essays.
And then I'll give them, they'll

144
00:08:05,800 --> 00:08:09,700
be able to enter into this
platform and be able to help me

145
00:08:09,700 --> 00:08:12,200
train my sentences.
I'll give them a sentence.

146
00:08:12,200 --> 00:08:14,200
I'll ask them a question.
Be like, Mark, this?

147
00:08:14,200 --> 00:08:15,700
How, what do you feel about
this?

148
00:08:16,000 --> 00:08:20,700
And tell me why, and then my
goal is essentially every single

149
00:08:20,700 --> 00:08:25,300
essay that I get is going to be
Greg graded and red based

150
00:08:25,300 --> 00:08:28,400
against a rubric criteria,
right?

151
00:08:28,500 --> 00:08:32,200
And as a as an instructor, I
love that so much, right?

152
00:08:32,200 --> 00:08:37,500
And he singled I say
understanding each criteria is

153
00:08:37,500 --> 00:08:40,600
going to be great at twice,
double blind peer review.

154
00:08:41,400 --> 00:08:43,900
I have like, when you look at
the, I'd like the question.

155
00:08:44,100 --> 00:08:46,300
Does this paper have a good
transition?

156
00:08:47,200 --> 00:08:51,300
That question is probably 15
different questions about

157
00:08:51,300 --> 00:08:54,100
transitions that you as an
individual.

158
00:08:54,100 --> 00:08:58,600
Do not have the time to ask
every single paper.

159
00:08:59,700 --> 00:09:01,800
I can train an algorithm to do
that.

160
00:09:01,900 --> 00:09:05,100
Let's say that my Final count is
20 criteria.

161
00:09:05,500 --> 00:09:10,100
That means that every sentence
from every submission that I get

162
00:09:10,300 --> 00:09:15,900
will be read 40 times.
And that's the data that were

163
00:09:15,900 --> 00:09:18,500
feeding the, that's why is a
large language model?

164
00:09:18,700 --> 00:09:23,200
That's why it's big data is
because we're creating metadata,

165
00:09:23,500 --> 00:09:27,000
my contention, is that the
number one problem with all

166
00:09:27,600 --> 00:09:31,300
artificial intelligence at out
there today is not the

167
00:09:31,300 --> 00:09:34,400
algorithm.
It's the quality of data that

168
00:09:34,400 --> 00:09:36,800
they're using.
That's what I want to address

169
00:09:36,800 --> 00:09:40,000
fundamentally.
I'm going to need help creating

170
00:09:40,100 --> 00:09:43,900
all of that data so you can sign
up and help me do that.

171
00:09:44,100 --> 00:09:46,700
That comes, you can create an
account and Le box score boards

172
00:09:46,700 --> 00:09:48,500
and stuff, and that leadership
boards will be fun.

173
00:09:49,100 --> 00:09:51,100
Yeah, that will be fun gamified
a bit.

174
00:09:51,100 --> 00:09:53,200
Make some competition out of it,
I love it.

175
00:09:53,400 --> 00:09:56,200
Exactly.
Not my brain is going a thousand

176
00:09:56,200 --> 00:09:58,200
different directions.
Every time we chat, that's what

177
00:09:58,200 --> 00:09:59,700
happens.
But that's a good thing.

178
00:09:59,900 --> 00:10:04,000
I'm thinking about, in terms of
AI has become everybody's an

179
00:10:04,000 --> 00:10:06,500
expert right now, and it's very
new.

180
00:10:06,500 --> 00:10:11,200
It's a very new frontier.
So as you are embracing this

181
00:10:11,200 --> 00:10:15,800
entrepreneurship into the AI
space, how How are you dealing

182
00:10:15,800 --> 00:10:17,400
with that?
And like your life and

183
00:10:17,400 --> 00:10:19,600
navigating?
All these are people, like I

184
00:10:19,600 --> 00:10:22,500
know this and a eyes bad here
because New York City public

185
00:10:22,500 --> 00:10:25,300
schools has already banned.
It's not allowed there or

186
00:10:25,300 --> 00:10:27,400
institutions of are abandoned.
How are you?

187
00:10:27,600 --> 00:10:30,500
Navigating that space?
As you're stepping into this

188
00:10:30,500 --> 00:10:36,100
entrepreneur a journey I think
part of it is recognizing when

189
00:10:36,100 --> 00:10:40,200
to step I mean in early January
I was reaching out to Jason.

190
00:10:40,200 --> 00:10:42,800
We were talking about fuss do a
series of different things and I

191
00:10:42,808 --> 00:10:46,200
was like, okay this to me I was
Ramping up because I like to be

192
00:10:46,200 --> 00:10:48,400
a voice.
I like to be on the stage and I

193
00:10:48,400 --> 00:10:50,600
just saw exactly what you
mentioned.

194
00:10:52,000 --> 00:10:57,700
Is that this is just going so
fast and so, there's so much.

195
00:10:58,300 --> 00:11:02,400
You cannot be an expert.
I think the only true experts on

196
00:11:02,400 --> 00:11:06,100
AI that exist, are the experts
that were talking about AI

197
00:11:06,200 --> 00:11:09,200
before November 20, 20?
Those are the people that we

198
00:11:09,200 --> 00:11:11,300
should really be thinking were
attempted to.

199
00:11:11,600 --> 00:11:15,600
But here's what I think, the
reason that there are so many

200
00:11:15,600 --> 00:11:23,200
experts in AI, Is because so
many experts exist, an AI just

201
00:11:23,200 --> 00:11:28,900
makes an expert better.
A lawyer, who knows how to use

202
00:11:28,900 --> 00:11:33,800
chat, GPT is not really going to
do well, teaching a marketing

203
00:11:34,100 --> 00:11:37,900
person or marketer how to use
chat CPT for her skills.

204
00:11:38,500 --> 00:11:41,100
Right?
A car mechanic is going to use

205
00:11:41,100 --> 00:11:43,600
chap TP differently than a
doctor.

206
00:11:44,200 --> 00:11:47,300
But the point is it that this is
The Equalizer.

207
00:11:47,700 --> 00:11:52,200
We are giving everybody Access
to information at your

208
00:11:52,200 --> 00:11:54,700
fingertips.
As rather than going in trying

209
00:11:54,700 --> 00:11:58,400
to compete, what everybody else
is doing, I just want to go and

210
00:11:58,400 --> 00:12:03,400
support what everybody else is
doing knowing confidently.

211
00:12:04,500 --> 00:12:06,600
That there is no one else that
is going to build what I'm

212
00:12:06,600 --> 00:12:10,800
building is everybody else is
worrying about algorithms and

213
00:12:10,800 --> 00:12:19,700
I'm focusing on data.
That's awesome and I don't know

214
00:12:19,900 --> 00:12:22,800
I want to ask this question.
Algorithms are so important

215
00:12:22,800 --> 00:12:24,700
especially on social media
sites.

216
00:12:24,700 --> 00:12:26,400
I'm going to hold back on that
one real quick.

217
00:12:26,400 --> 00:12:29,200
I'll let Jason Jump in, I'm
going to hold back on the

218
00:12:29,200 --> 00:12:31,700
algorithm question, the vs the
data.

219
00:12:31,700 --> 00:12:35,800
So Jason, you'd pop in here.
I definitely want to come to

220
00:12:35,800 --> 00:12:38,400
that and there are go with
dings.

221
00:12:38,600 --> 00:12:42,400
So as you were talking Eric, I
had these clusters in my mind

222
00:12:42,400 --> 00:12:45,100
that I was creating as you were
walking you through.

223
00:12:45,200 --> 00:12:49,900
Through a project rhizome is
actually doing the first cluster

224
00:12:49,900 --> 00:12:53,300
is how your project is a
microcosm.

225
00:12:53,300 --> 00:12:57,300
In many ways of a lot of the
issues and concerns and Concepts

226
00:12:57,400 --> 00:13:00,600
that were thinking about with a.
I when we think about

227
00:13:00,600 --> 00:13:08,400
transparency and ethics and who
is managing knowledge bases.

228
00:13:08,400 --> 00:13:12,900
If we go into chat gvt,
sometimes it's really hard to

229
00:13:12,900 --> 00:13:15,100
figure out where that knowledge
is coming.

230
00:13:15,300 --> 00:13:19,200
In from and who's controlling
that data or sourcing that data,

231
00:13:19,500 --> 00:13:21,100
I think you got to change as we
go forward.

232
00:13:21,100 --> 00:13:28,300
But it's always been a question
built into the use of AI and the

233
00:13:28,300 --> 00:13:31,000
other concept that you brought
out is you're talking about

234
00:13:31,000 --> 00:13:33,700
project.
Rhizome is social building,

235
00:13:34,000 --> 00:13:36,500
right?
And really creating in public

236
00:13:36,500 --> 00:13:41,100
and using people, and I think
associated with that connected

237
00:13:41,100 --> 00:13:45,100
to that is this question of what
happens to the expert.

238
00:13:45,200 --> 00:13:48,900
But in the age of a, I think
there's a lot of anxiety about

239
00:13:48,900 --> 00:13:53,400
that and I was literally reading
a book titled, the new laws of

240
00:13:53,400 --> 00:13:59,500
robotics by Frank Pasquale, and
that's what he talks about.

241
00:13:59,500 --> 00:14:03,100
His argument is actually that AI
makes the expert more valuable,

242
00:14:03,300 --> 00:14:08,700
but actually gives them more
value, socially culturally, and

243
00:14:08,800 --> 00:14:12,100
hopefully economically.
That's one of the con clusters

244
00:14:12,300 --> 00:14:15,800
that I started to create out of
your talking, the other Idea

245
00:14:15,800 --> 00:14:18,300
that you brought up, which
really spoke to me.

246
00:14:18,300 --> 00:14:20,700
Personally, I know, Holly is
someone who teaches all the

247
00:14:20,700 --> 00:14:23,000
time, just like Lee spoke to
you, too.

248
00:14:23,200 --> 00:14:28,900
It's just how much time you can
save and how you can repurpose

249
00:14:28,900 --> 00:14:32,800
or time to do, very high impact
things.

250
00:14:32,800 --> 00:14:36,800
For my perspective, that is
fundamentally true.

251
00:14:37,100 --> 00:14:42,500
It used to take me four hours to
create a personalized rubric for

252
00:14:42,500 --> 00:14:45,100
class.
It now takes me four minutes.

253
00:14:45,200 --> 00:14:48,100
It's it.
You turn take me and I minute of

254
00:14:48,100 --> 00:14:51,600
it.
Yeah. 15 minutes to write

255
00:14:51,600 --> 00:14:53,700
student feedback and everything
else.

256
00:14:53,800 --> 00:14:57,500
It now takes me about a minute
and a half, right?

257
00:14:57,500 --> 00:15:01,200
Editing, a podcast.
Used to take me two hours.

258
00:15:01,400 --> 00:15:06,500
It now takes me eight minutes.
That's like that were, that's

259
00:15:06,500 --> 00:15:08,100
what we're talking about.
That's why we're talking about

260
00:15:08,100 --> 00:15:11,800
the level of time-saving.
And so, a project, like project

261
00:15:11,800 --> 00:15:14,900
rhizome, at least for me is a
model.

262
00:15:15,200 --> 00:15:18,800
Others can use for how you can
ethically sourced data and be

263
00:15:18,808 --> 00:15:21,800
transparent about where it's
coming from and all of that.

264
00:15:21,800 --> 00:15:25,100
I want to follow this up with a
question about project dry Zone

265
00:15:25,400 --> 00:15:29,700
and I think it's connected to a
lot of those Concepts that you

266
00:15:29,700 --> 00:15:32,800
brought up, but then also your
focus on assessment.

267
00:15:33,500 --> 00:15:39,500
So how personalizable is it?
Because you mentioned using data

268
00:15:39,600 --> 00:15:42,400
from different disciplines
different levels.

269
00:15:42,500 --> 00:15:45,000
So is their way, the project
rhizome.

270
00:15:45,200 --> 00:15:49,400
Um would be targeting that
feedback for a student of a

271
00:15:49,408 --> 00:15:51,700
particular skill level say,
they're still struggling with

272
00:15:51,700 --> 00:15:54,300
something, it's not a student
who's say a freshman in college

273
00:15:54,300 --> 00:15:58,600
versus a senior in college.
Is that feedback tailored or how

274
00:15:58,600 --> 00:16:01,200
it was that working in terms of
your individual project?

275
00:16:02,100 --> 00:16:05,700
Yeah.
So I'll say this, one thing to,

276
00:16:05,700 --> 00:16:11,400
I think that the way that you
described How you heard these

277
00:16:11,400 --> 00:16:14,100
different themes.
And talking about clustering is

278
00:16:14,100 --> 00:16:16,900
exactly why artificial
intelligence is intelligence and

279
00:16:16,900 --> 00:16:18,900
human life.
Because that is what artificial

280
00:16:18,900 --> 00:16:21,200
intelligence does.
It looks at a whole bunch of

281
00:16:21,200 --> 00:16:25,400
content and then it clusters it
by topic and then predicts based

282
00:16:25,400 --> 00:16:28,300
on that clustering, what it
should say that is the exact

283
00:16:28,300 --> 00:16:31,400
same thought process it.
That is why it's called a neural

284
00:16:31,500 --> 00:16:33,900
network.
It's mimicking neurons and so

285
00:16:33,900 --> 00:16:35,500
that's just a really cool thing
for me.

286
00:16:35,700 --> 00:16:38,900
I'm glad that you did that to
address your question, one of

287
00:16:38,900 --> 00:16:41,400
the data points.
That I'm collecting is the

288
00:16:41,400 --> 00:16:44,000
level.
So I'm accepting writing from

289
00:16:44,000 --> 00:16:48,000
any discipline when a user goes
in and submits their writing I'm

290
00:16:48,000 --> 00:16:51,300
asking them a whole bunch of
questions about their writing so

291
00:16:51,300 --> 00:16:55,100
that I can then tabulated that
and find Trends and patterns

292
00:16:55,100 --> 00:16:57,300
later.
So one of those questions is

293
00:16:57,700 --> 00:17:00,300
what is the name of the class?
Is this, a history class is a

294
00:17:00,300 --> 00:17:01,800
biology class that kind of
thing?

295
00:17:02,100 --> 00:17:03,900
The other one is at what level
is it is.

296
00:17:03,908 --> 00:17:08,099
I'm accepting 9th grade writing
through doctoral level writing,

297
00:17:08,200 --> 00:17:12,200
like through post-grad because
So I believe this fundamentally

298
00:17:13,000 --> 00:17:17,300
that the core principles of a
writing class and a postdoc is

299
00:17:17,300 --> 00:17:19,500
the exact same lesson plan.
You're going to get in 9th

300
00:17:19,500 --> 00:17:22,200
grade.
They're the same thing just the

301
00:17:22,200 --> 00:17:26,500
complexity changes but the core
principles are what writing is

302
00:17:26,800 --> 00:17:29,800
stays the same?
That's why writing is a meta

303
00:17:29,800 --> 00:17:33,900
discipline when we get to phase
3 or 4 when we get to the stage

304
00:17:33,900 --> 00:17:36,600
of developing the personalized
feedback and probably tapping

305
00:17:36,600 --> 00:17:39,400
some like other large language
models.

306
00:17:39,600 --> 00:17:42,500
Do that makes it sound better
like chatty PT or something.

307
00:17:43,000 --> 00:17:46,600
But being able to say, okay,
like what, how would you model

308
00:17:46,600 --> 00:17:51,000
this feedback if the student was
in ninth grade versus in

309
00:17:51,000 --> 00:17:52,500
postdoctoral?
They might be a bike, not be

310
00:17:52,500 --> 00:17:54,900
able to bit more short with
them, over to the point.

311
00:17:55,600 --> 00:18:00,400
The, because we're collecting
this data about the writing

312
00:18:00,400 --> 00:18:05,000
itself, will be able to find
patterns, we will be able to

313
00:18:05,000 --> 00:18:09,400
bike me said before, ask AI to
see what clusters exist.

314
00:18:09,500 --> 00:18:12,800
So that we may not see right
there may be other clusters or

315
00:18:12,800 --> 00:18:15,300
themes and what I said but we
didn't see them because we're

316
00:18:15,300 --> 00:18:17,800
not artificial intelligent and
looking at everything over time.

317
00:18:19,300 --> 00:18:24,800
That's why Collecting the data
is so important.

318
00:18:25,400 --> 00:18:29,400
One of the largest data models
are one of the largest data sets

319
00:18:29,400 --> 00:18:31,700
that a lot of these large
language models are being

320
00:18:32,000 --> 00:18:36,400
trained on are from high school
essays written for the SATs, you

321
00:18:36,400 --> 00:18:40,100
can actually go and find a spout
database of twelve to fourteen

322
00:18:40,100 --> 00:18:44,100
thousand student, written essays
by the reason chat T.T like

323
00:18:44,100 --> 00:18:47,300
gives you a standard five
paragraph essay.

324
00:18:47,800 --> 00:18:50,500
When you just type in a regular
question is because it was

325
00:18:50,500 --> 00:18:53,700
trained.
On high school, 5, paragraph

326
00:18:53,700 --> 00:18:56,100
essays.
And they were being trained and

327
00:18:56,100 --> 00:19:00,500
graded by people who preferred
long-winded writing.

328
00:19:01,800 --> 00:19:06,500
So the reason chat Jiggy T is
for boasts, is because of the

329
00:19:06,500 --> 00:19:12,300
data and the metadata about it.
I don't believe this might be

330
00:19:12,300 --> 00:19:13,900
getting on a tangent now.
Right?

331
00:19:14,100 --> 00:19:17,000
I don't think that the solution
to our problems is to embed

332
00:19:17,000 --> 00:19:19,700
English 101 into a biology
class.

333
00:19:20,800 --> 00:19:24,400
I don't think English 101, is a
good practice of writing because

334
00:19:24,400 --> 00:19:26,800
when you get to the business
world, you're taught how to be

335
00:19:26,800 --> 00:19:29,900
more concise like you're taught
to be using.

336
00:19:29,900 --> 00:19:31,800
What's called plain language,
right?

337
00:19:31,800 --> 00:19:34,300
So there's the Obama passed, the
plain language act 2010.

338
00:19:35,800 --> 00:19:37,900
That's what I want to embed,
right?

339
00:19:37,900 --> 00:19:40,900
I was talking to people who
teaching mbas and they say one

340
00:19:40,900 --> 00:19:42,900
of their biggest things that
teaching their students, how to

341
00:19:42,900 --> 00:19:45,800
write less be more concise,
tours, it better.

342
00:19:46,400 --> 00:19:48,800
And that's what how we train
chat, Chiquitita, better Outlook

343
00:19:48,800 --> 00:19:51,700
to make this answer shorter.
It's like when off on a big

344
00:19:51,700 --> 00:19:55,500
tangent there but your answer to
that question.

345
00:19:56,000 --> 00:19:59,400
Bay's two or three there's going
to be that personalized level of

346
00:19:59,400 --> 00:20:03,600
feedback based on a discipline
and based on level.

347
00:20:04,100 --> 00:20:08,500
But rather than Starting with
that discipline or that level

348
00:20:08,500 --> 00:20:10,200
and saying this is how each
sound.

349
00:20:10,700 --> 00:20:14,700
We're asking the data first how
it should sound, so reacting to

350
00:20:14,708 --> 00:20:17,900
the data and not coming in with
our own biases and begin.

351
00:20:19,500 --> 00:20:23,600
And I like it.
So, I was going to ask, there's

352
00:20:23,600 --> 00:20:28,000
a lot of pushback about this,
like using this AI stuff and

353
00:20:28,500 --> 00:20:31,500
obviously not in this room,
we're all just, let's do it.

354
00:20:31,500 --> 00:20:33,600
Let's use it.
Let's have fun with it.

355
00:20:33,600 --> 00:20:36,200
Let's see what we're doing.
How would you approach the

356
00:20:36,200 --> 00:20:38,600
situation of the people who are
already resisting?

357
00:20:38,600 --> 00:20:43,200
This sort of situation where
students writing is going to be

358
00:20:43,700 --> 00:20:45,500
filtered through a system and
then they're going to get

359
00:20:45,500 --> 00:20:48,200
feedback.
How would you deal with the

360
00:20:48,200 --> 00:20:51,300
reserve resistant?
That is on the rise again.

361
00:20:51,400 --> 00:20:54,900
Again, with technology in this
new innovation with AI in chat

362
00:20:54,900 --> 00:20:59,000
GPT, how would you approach that
in the seat that you're in right

363
00:20:59,000 --> 00:21:01,800
now and deal with that?
That's a loaded question.

364
00:21:01,800 --> 00:21:06,000
By the way, back in the day,
back in the day, couple thousand

365
00:21:06,000 --> 00:21:12,200
years ago, Leto was a ranting
and raving right about the

366
00:21:12,200 --> 00:21:16,600
creation of a new technology
that would ruin thinking.

367
00:21:17,900 --> 00:21:19,500
He was talking about the
alphabet.

368
00:21:26,000 --> 00:21:29,400
Every single time, something new
happens.

369
00:21:30,800 --> 00:21:35,700
People will not like it and
they'll disagree with it.

370
00:21:36,400 --> 00:21:40,300
And I think that has far more to
do with one's experience in life

371
00:21:40,300 --> 00:21:43,300
and where they are in life and
how much they're willing to

372
00:21:43,300 --> 00:21:45,800
learn something new versus not
new.

373
00:21:47,100 --> 00:21:50,800
That determines that.
So when I encounter someone

374
00:21:50,800 --> 00:21:54,900
who's I'm not gonna do that, my
response is okay.

375
00:21:56,500 --> 00:21:58,500
I'll see you in five years.
When you shout like, they're

376
00:21:58,500 --> 00:22:02,800
gonna, I know that because I
have already seen this actively

377
00:22:03,500 --> 00:22:08,400
happening where I have a friend,
he's a CEO of a company and he

378
00:22:08,400 --> 00:22:12,300
hires people, and before he
would hire a teachers to create

379
00:22:12,300 --> 00:22:18,000
content for him now, he hires
one teacher that uses chat GPT

380
00:22:18,200 --> 00:22:25,000
to create content for him.
The people who are saying this

381
00:22:25,000 --> 00:22:27,000
is not something I need to worry
about.

382
00:22:28,000 --> 00:22:32,800
Do not understand how
fundamentally intertwined their

383
00:22:32,800 --> 00:22:37,300
lives, already are with AI and
how they are going to be

384
00:22:37,300 --> 00:22:40,800
intertwined with a, I just in
the product of sweets with

385
00:22:40,800 --> 00:22:45,000
Microsoft and Google that now
hat, their copilot, and they're

386
00:22:45,000 --> 00:22:48,600
barred embedded in.
Whatever we got, call it you

387
00:22:48,600 --> 00:22:55,700
cannot Escape it and if you're
actively resisting it, You are

388
00:22:55,800 --> 00:22:58,700
actively putting yourself at
risk.

389
00:22:59,000 --> 00:23:01,900
I am a huge fan like you need to
be an early adopter.

390
00:23:03,000 --> 00:23:06,000
If only to be asking those
heart, ethical questions to be

391
00:23:06,000 --> 00:23:07,800
getting.
So you're not dealing with

392
00:23:07,800 --> 00:23:09,800
changing a procedures.
Number of be the one to

393
00:23:09,800 --> 00:23:12,600
implement the change.
Not the one that has to deal

394
00:23:12,600 --> 00:23:16,400
with someone else, child me to
change and so I guess that I

395
00:23:16,400 --> 00:23:19,500
think there's always reason to
be cautious, but I think that

396
00:23:20,600 --> 00:23:24,200
we're all good crap that we talk
about our age demographic of

397
00:23:24,200 --> 00:23:26,900
everything that we've I've lived
through like we can go through a

398
00:23:26,908 --> 00:23:30,100
lot like Wars and teres
attachment so many things.

399
00:23:32,600 --> 00:23:38,500
Holy moly.
We are living through a modern

400
00:23:38,500 --> 00:23:42,500
Industrial Revolution that is
happening. 10 times as fast

401
00:23:42,500 --> 00:23:45,600
where you can see daily change
instead of monthly or yearly

402
00:23:45,600 --> 00:23:49,100
change, we're going to seize bit
out of like an update.

403
00:23:49,200 --> 00:23:52,000
You don't have to say, it's got
to be in the SAS and it's got a

404
00:23:52,008 --> 00:23:55,400
spit out an update.
Yeah, it's instantaneous.

405
00:23:56,500 --> 00:23:59,900
I think that there is going to
be amazing things that happen.

406
00:23:59,900 --> 00:24:02,800
I believe that artificial
intelligence will be the

407
00:24:02,800 --> 00:24:04,800
introduction of a permanent
four-day work week.

408
00:24:05,100 --> 00:24:07,400
Like we like artificial
intelligence will show us that

409
00:24:07,600 --> 00:24:11,200
working 40 hours a week is no
longer required and we can have

410
00:24:11,200 --> 00:24:15,200
more Leisure Time and happiness.
There's a lot of good have going

411
00:24:15,200 --> 00:24:18,000
to come out.
I really do and we can go

412
00:24:18,000 --> 00:24:23,300
dystopia if you want to and it's
gonna get there but also man,

413
00:24:23,300 --> 00:24:28,700
it's beautiful.
I used to brought up so much

414
00:24:28,700 --> 00:24:32,300
Eric that I want to talk about
the first is your point which I

415
00:24:32,300 --> 00:24:34,700
think is so pivotal for the
higher ed sector.

416
00:24:34,700 --> 00:24:39,500
It's that AI has been around for
a long for a while.

417
00:24:39,500 --> 00:24:42,700
This kind of tech has been
around, will really change with

418
00:24:42,700 --> 00:24:45,000
Shao.
Qi, PT, is that it was pushed

419
00:24:45,000 --> 00:24:50,600
into the public imagination and
the ux was made so much more

420
00:24:50,600 --> 00:24:52,800
accessible.
That's the big thing.

421
00:24:52,800 --> 00:24:55,900
That Chad, TBT did creating the
chat function.

422
00:24:55,900 --> 00:25:02,900
So it was almost About the you.
That's design of it, then the

423
00:25:02,900 --> 00:25:07,000
actual Tech and it's so worked
into our lives.

424
00:25:07,400 --> 00:25:09,100
And so are going to it that we
almost.

425
00:25:09,100 --> 00:25:11,200
And even notice there are a lot
of us, don't even notice it was

426
00:25:11,200 --> 00:25:13,300
happening.
Then the other thing you brought

427
00:25:13,300 --> 00:25:16,700
up and seemed to suggest was
that Tech like chat?

428
00:25:16,700 --> 00:25:23,800
Gbt is the tip of the iceberg.
He's our a just early moved.

429
00:25:23,800 --> 00:25:28,000
So I like to think about what
people might think, 50 years

430
00:25:28,000 --> 00:25:30,800
from now.
And one of the The things I

431
00:25:30,800 --> 00:25:33,800
think that they will do is 50
years from now.

432
00:25:34,200 --> 00:25:36,300
Children are going to be in
school and they're finished.

433
00:25:36,300 --> 00:25:41,200
Show them chat GPT and they're
going to be horrified about how

434
00:25:41,200 --> 00:25:44,600
bad it is.
How awful a product that is?

435
00:25:44,600 --> 00:25:47,300
Instead Altman is also talked
about this by the way.

436
00:25:47,500 --> 00:25:49,500
He said that.
Just so you know, she actually

437
00:25:49,500 --> 00:25:52,600
be cheeps kind of awful as a
user experience.

438
00:25:52,600 --> 00:25:55,000
It's down all the time you have
to.

439
00:25:55,000 --> 00:25:59,200
Now spend all of these hours
figuring out how to use it and

440
00:25:59,200 --> 00:26:01,600
super it Hank.
He's less example Sager.

441
00:26:01,600 --> 00:26:06,900
This is awful Prada but it
really gave us something that we

442
00:26:06,900 --> 00:26:09,400
can play with and I think that
50 years from now.

443
00:26:09,600 --> 00:26:11,900
It's good.
Hey guys going to be so much

444
00:26:11,900 --> 00:26:14,800
more advanced in terms of being
user friendly so you won't have

445
00:26:14,800 --> 00:26:19,400
to as a lot of us.
Did I did spend weeks really

446
00:26:19,400 --> 00:26:22,800
play with it to be like oh I
gotta use it now and I just

447
00:26:22,800 --> 00:26:26,900
going to continue to advance and
it's advancing every day and I

448
00:26:26,908 --> 00:26:29,800
want to follow it up with a
question Eric.

449
00:26:29,800 --> 00:26:33,700
That's Nected to that and you
mentioned and I think you're

450
00:26:33,700 --> 00:26:38,000
absolutely right.
That AI is an industrial

451
00:26:38,000 --> 00:26:40,800
revolution.
That's changing every day every

452
00:26:40,800 --> 00:26:43,500
week.
How do you stay on top of it.

453
00:26:43,600 --> 00:26:46,000
What do you follow?
What do you look at?

454
00:26:46,100 --> 00:26:49,800
How do you feel like you're at
least abreast of what's

455
00:26:49,800 --> 00:26:51,700
Happening?
I'll say that.

456
00:26:51,700 --> 00:26:56,100
I don't feel like I am because
just like you said, because

457
00:26:56,100 --> 00:27:00,200
Chance CPT is such an easy user
interface.

458
00:27:00,600 --> 00:27:05,200
And because anybody can access
it using their own expertise,

459
00:27:05,400 --> 00:27:09,300
there are just dozens of
applications that I'm seeing

460
00:27:09,300 --> 00:27:13,700
every day.
I love watching tic toc like a

461
00:27:13,700 --> 00:27:17,400
lot of the idea generation that
I get is from other people who

462
00:27:17,400 --> 00:27:18,900
get on and share, what they're
doing.

463
00:27:19,600 --> 00:27:23,200
I'm pretty active on LinkedIn,
I'm constantly scouring.

464
00:27:23,400 --> 00:27:26,900
The news is social media, like I
consume a lot of information

465
00:27:27,100 --> 00:27:29,700
from a variety of different
resources and so I think that

466
00:27:29,700 --> 00:27:33,800
helps it A lot, I'm a big fan of
the content just in Feinberg I

467
00:27:33,808 --> 00:27:37,700
think his name is and Rachel
Woods on Tick-Tock specifically

468
00:27:37,700 --> 00:27:40,900
the AI exchange newsletter.
I think is like the go-to place

469
00:27:41,000 --> 00:27:43,400
to get information.
I also have a really good friend

470
00:27:43,400 --> 00:27:47,200
of mine, his name is Ben but we
talk about data.

471
00:27:48,000 --> 00:27:50,000
He's the one that helped me with
my dissertation research.

472
00:27:50,000 --> 00:27:52,300
During use the data scientist
that I collaborated, with me and

473
00:27:52,300 --> 00:27:54,100
Katie.
So, I don't know if that was

474
00:27:54,100 --> 00:28:00,700
like a I think the answer to
your questions can be applicable

475
00:28:00,700 --> 00:28:01,900
for the people who are
listening.

476
00:28:03,100 --> 00:28:08,100
Is that the best thing to do is
to start listening and to start

477
00:28:08,100 --> 00:28:11,200
playing, don't ignore those
articles that you're seeing and

478
00:28:11,200 --> 00:28:13,800
seeing that it's everywhere.
Go and see it and then go and

479
00:28:13,800 --> 00:28:19,000
sit, down and chat gbt and try
it. at work that I have like my

480
00:28:19,000 --> 00:28:27,800
full-time gig that I have, its I
love being able to show people

481
00:28:28,300 --> 00:28:31,500
face-to-face, what this
technology can do and just watch

482
00:28:31,500 --> 00:28:35,300
the AA happen, then have them
come up to me, like my boss

483
00:28:35,300 --> 00:28:38,800
asked about, can you make me
limerick about a leprechaun who

484
00:28:38,800 --> 00:28:40,400
wants raisins?
It is, carrot cake, and I was

485
00:28:40,400 --> 00:28:42,700
like, that's what it takes to
make you happy, man.

486
00:28:42,900 --> 00:28:45,300
Yes, I will.
I get a raise after that.

487
00:28:45,300 --> 00:28:47,600
Yeah.
And then like, I had rigged The

488
00:28:47,600 --> 00:28:50,400
Vaping.
Yeah, of course, my friends

489
00:28:50,400 --> 00:28:52,800
quitting her job and she's write
a resignation letter Indiana's

490
00:28:52,800 --> 00:28:54,600
like a zoo were on it already.
Boom.

491
00:28:54,600 --> 00:28:59,500
Boom, let me email that to you
like it's so much fun to see

492
00:29:00,100 --> 00:29:06,700
people's own curiosity and
expertise emerge when they use

493
00:29:06,700 --> 00:29:08,500
it themselves and just to see
bike.

494
00:29:08,500 --> 00:29:15,100
How Just imagine how much more
creative and beautiful things

495
00:29:15,100 --> 00:29:18,900
are going to be.
Because we have the ability that

496
00:29:19,000 --> 00:29:21,600
someone who I don't consider
myself artistic, can go and make

497
00:29:21,600 --> 00:29:25,200
something beautiful.
I think that's, I think that's

498
00:29:25,200 --> 00:29:27,600
amazing.
And I love it that you say, you

499
00:29:27,600 --> 00:29:29,100
have to go, you have to play
with it.

500
00:29:29,100 --> 00:29:31,500
I feel like people be before
they write it off.

501
00:29:31,500 --> 00:29:35,200
That's the thing you need to do
is just go play with it and ask

502
00:29:35,200 --> 00:29:37,100
it.
Like you're you're seriously

503
00:29:37,100 --> 00:29:40,200
like just typing in a question
or two.

504
00:29:40,400 --> 00:29:43,400
We need to do something.
It's what we do every day owner

505
00:29:43,400 --> 00:29:45,800
and if the computer you're
telling it to do things that you

506
00:29:45,808 --> 00:29:49,200
needed to do, so it's mimicking
what we're already doing.

507
00:29:49,200 --> 00:29:53,800
So it shouldn't be scary for
people so I want to know what

508
00:29:53,800 --> 00:29:56,900
are you like currently besides
writing limericks and things and

509
00:29:56,900 --> 00:29:59,800
resignation letters.
What are you currently playing

510
00:29:59,800 --> 00:30:02,600
with right now?
In Ai and he like test cases

511
00:30:02,600 --> 00:30:04,400
things that you want to share
with the audience that you're

512
00:30:04,400 --> 00:30:06,200
working on.
I think we should all share like

513
00:30:06,200 --> 00:30:09,900
how we're were using this tool
in our lives.

514
00:30:10,400 --> 00:30:15,100
For me, I'm I'm trying to build
a tech company in the leanest

515
00:30:15,100 --> 00:30:20,200
way possible and so when I had
questions I'm like I'm pretty

516
00:30:20,200 --> 00:30:27,000
good at generating content, what
I'm bad at is formal writing and

517
00:30:27,000 --> 00:30:28,900
formal things I need to do,
right?

518
00:30:28,900 --> 00:30:32,000
So what I can do is I want to
thank that with all of your

519
00:30:32,000 --> 00:30:35,800
education and your background.
Did he really do?

520
00:30:35,800 --> 00:30:39,400
It doesn't mean anything.
That's the thing, that's the key

521
00:30:39,400 --> 00:30:44,300
there is I It's not how to do it
and put it out there, but the

522
00:30:44,300 --> 00:30:48,200
think the path at the passions,
are I feel like that about some

523
00:30:48,200 --> 00:30:50,600
scholar like scholarly writing
is like that a lot.

524
00:30:50,600 --> 00:30:53,100
It's force a bit, not your
choice on that.

525
00:30:53,100 --> 00:31:00,400
I think that being able to have
an expert being able to pay chat

526
00:31:00,400 --> 00:31:03,400
gbt, I know you're not a lawyer,
you can get around it, but hey,

527
00:31:03,400 --> 00:31:05,700
I have an appointment, my lawyer
next week, but I have a question

528
00:31:05,700 --> 00:31:08,100
in the meantime.
That's a really great way to get

529
00:31:08,100 --> 00:31:11,900
around their caveat is that they
give you So I have, what do I do

530
00:31:11,900 --> 00:31:12,900
here?
What does this mean?

531
00:31:13,600 --> 00:31:16,500
And then it explains it to me.
I'm working in the construction

532
00:31:16,500 --> 00:31:20,500
industry.
I have no idea how to build a

533
00:31:20,500 --> 00:31:24,100
building or what people use for
different terms when they're I

534
00:31:24,100 --> 00:31:26,000
had.
No, I'm looking at accounting

535
00:31:26,000 --> 00:31:27,700
data.
I've never worked at accounts

536
00:31:27,700 --> 00:31:29,600
payable before.
I've never done that.

537
00:31:30,400 --> 00:31:34,500
I have Chachi PTO bananas, say
hey what does this mean?

538
00:31:35,200 --> 00:31:39,200
Or someone just said this in
this context, what should I do?

539
00:31:41,300 --> 00:31:46,100
As someone who is neurodivergent
and I have a bipolar disorder

540
00:31:46,100 --> 00:31:49,100
and I love, I can be
overwhelming.

541
00:31:49,200 --> 00:31:52,600
Yes, my enthusiasm is wonderful
for a podcast at work.

542
00:31:52,600 --> 00:31:58,300
I can be an overwhelming person.
This lets me control that

543
00:31:58,300 --> 00:32:01,900
insatiable curiosity.
I think the more you recognize

544
00:32:02,400 --> 00:32:06,400
that it's not a Google search
where you put something in and

545
00:32:06,400 --> 00:32:10,400
you get a final output that you
are talking to someone.

546
00:32:10,400 --> 00:32:13,800
It's a Chat, you refine the
conversation.

547
00:32:14,100 --> 00:32:17,000
You can't go up to a stranger
and give them a command.

548
00:32:17,000 --> 00:32:18,900
As they write me a press release
for this.

549
00:32:19,200 --> 00:32:20,800
They're going to turn around be
like, what the fuck are you

550
00:32:20,800 --> 00:32:23,200
talking about?
No, you turn around to like,

551
00:32:23,200 --> 00:32:25,100
hey, how's it going?
This is me.

552
00:32:25,100 --> 00:32:29,100
You introduce yourself, but you
give context the more you treat

553
00:32:29,100 --> 00:32:32,700
artificial intelligence as its
own actor.

554
00:32:34,000 --> 00:32:37,400
The better can't the better.
You're going to get the outputs

555
00:32:37,400 --> 00:32:39,100
are going to get sorry.
There are score.

556
00:32:39,300 --> 00:32:42,900
Well bleep that out.
If I think they might like it

557
00:32:43,000 --> 00:32:46,900
adds emphasis, I don't know.
I've no actress and leaving that

558
00:32:46,900 --> 00:32:47,900
out.
I'm Gonna Keep it anymore.

559
00:32:48,200 --> 00:32:51,100
I'm not gonna find a an anchor.
I probably have to put the

560
00:32:51,100 --> 00:32:55,800
explicitly, molar, whatever, or
I will track it and got me down,

561
00:32:55,800 --> 00:33:01,600
who knows?
And I love merits of your idea.

562
00:33:02,500 --> 00:33:07,400
I love your idea of using AI as
a tutor, one of the things that

563
00:33:07,400 --> 00:33:12,700
has happened to me and I know
this must sound meta is, AI has

564
00:33:12,700 --> 00:33:17,500
allowed me to get into AI.
There are all these Concepts out

565
00:33:17,500 --> 00:33:21,300
there that as I read and look at
everything on social media and

566
00:33:21,300 --> 00:33:23,400
learn and I follow Justin
Feinberg.

567
00:33:23,500 --> 00:33:26,500
If follow Rachel, would I learn
a lot from them and every once

568
00:33:26,500 --> 00:33:29,200
in a while up, Curious about
something.

569
00:33:29,400 --> 00:33:34,200
And in the past 10 years ago I
would have written a noted a

570
00:33:34,200 --> 00:33:36,300
notebook somewhere.
Sometimes I'd look it up,

571
00:33:36,300 --> 00:33:39,100
sometimes I wouldn't.
Now what I do is I take my

572
00:33:39,100 --> 00:33:43,700
device, I go immediately to poet
is my go-to right now because

573
00:33:43,700 --> 00:33:47,200
it's a very, it's a very easy
way to have everything in one

574
00:33:47,200 --> 00:33:49,400
spot and I'll ask it to teach
me.

575
00:33:49,400 --> 00:33:53,700
So I was reading the other day
about a i in ground truth it was

576
00:33:53,700 --> 00:33:55,400
just thrown out there that comes
up.

577
00:33:55,500 --> 00:34:00,000
And so I was able to go in and
use Gbt through Poe and just a

578
00:34:00,000 --> 00:34:04,400
skit, talk Chief me about this
thing doesn't just doing web

579
00:34:04,400 --> 00:34:08,900
search because I was able to
through the prom, be able to

580
00:34:08,900 --> 00:34:11,900
tailor it to myself.
I was able to say, explain to me

581
00:34:11,900 --> 00:34:16,600
like you would a ten-year-old
and give me at least one analogy

582
00:34:16,600 --> 00:34:21,900
that will allow me to grasp onto
it, and doing that in the past,

583
00:34:21,900 --> 00:34:23,699
they've been very difficult or
impossible.

584
00:34:23,699 --> 00:34:26,500
If I looked it up at all, I
would have just ended up on

585
00:34:26,500 --> 00:34:31,699
dictionary.com or Wikipedia,
which I'm still Blown Away by

586
00:34:31,699 --> 00:34:35,000
how in many ways inaccessible?
Wikipedia is on the level of

587
00:34:35,000 --> 00:34:37,800
language.
I'm constantly going in there to

588
00:34:37,800 --> 00:34:41,699
learn something Technical and on
bombarded with technical

589
00:34:41,699 --> 00:34:44,100
language, even though it's
recompete.

590
00:34:44,100 --> 00:34:47,699
Yeah, but even from there, I get
lost and I don't know what to do

591
00:34:48,100 --> 00:34:50,699
now.
It's a, I'm able to learn about

592
00:34:50,699 --> 00:34:53,699
Ai and so becomes a.
So it allows me to get into

593
00:34:53,699 --> 00:34:56,900
discipline actually learn what
I'm talking about in a little

594
00:34:56,900 --> 00:34:58,700
was going on.
No.

595
00:34:58,700 --> 00:35:04,500
I so I was sitting down with my
brother and he is he's getting

596
00:35:04,500 --> 00:35:07,300
real like next week actually
sitting down for his PMP exam.

597
00:35:07,600 --> 00:35:10,300
And so and we were like he was
telling me about like way that

598
00:35:10,300 --> 00:35:12,400
he's used it before and like
he's just like trying to figure

599
00:35:12,400 --> 00:35:15,700
out like how he can go and
prepare for his exam.

600
00:35:16,000 --> 00:35:18,400
He's also sharing with me that
the way he likes to learn

601
00:35:18,800 --> 00:35:20,700
because you like that.
So you can tailor, I like to

602
00:35:20,700 --> 00:35:23,100
learn by a now, they give me
some analogies the way the he

603
00:35:23,100 --> 00:35:25,700
likes to learn is hey, here's
this real world application and

604
00:35:25,700 --> 00:35:28,200
I have, how can I apply it?
So it's like just like We're

605
00:35:28,200 --> 00:35:29,800
just like on the cuff.
Right?

606
00:35:29,800 --> 00:35:32,900
And I was like, what if you were
to go in a chat GPT and say,

607
00:35:33,100 --> 00:35:38,600
hey, I am taking an exam and two
weeks here are some specific and

608
00:35:38,600 --> 00:35:42,200
you give them like the specific
scenario that you're at work.

609
00:35:42,200 --> 00:35:46,300
That's a real life example.
And then you say using the PMP

610
00:35:46,300 --> 00:35:51,700
guide, create questions for me,
that helps me study, that will

611
00:35:51,700 --> 00:35:54,400
help me understand this thing at
work.

612
00:35:54,700 --> 00:35:58,000
Easy peasy.
As soon as you realize that it's

613
00:35:58,100 --> 00:36:01,600
Not just like accessing
information but it's asking it

614
00:36:01,600 --> 00:36:05,500
to combine and synthesize
information for you, it's really

615
00:36:05,500 --> 00:36:08,500
beautiful.
What you can do it.

616
00:36:10,500 --> 00:36:17,200
I know how to say this, exactly.
But I have been pulled into this

617
00:36:17,200 --> 00:36:21,900
thing.
That is data and language.

618
00:36:21,900 --> 00:36:30,200
I have been steeping in data and
language or It's it was in 2014

619
00:36:32,200 --> 00:36:35,600
when I first decided that my
methodology for my dissertation

620
00:36:36,200 --> 00:36:41,300
was would be big data analysis.
And I've been thinking about

621
00:36:41,400 --> 00:36:44,000
ethics of it since then I've
been thinking about what I would

622
00:36:44,000 --> 00:36:48,800
do and I was at Justin thinking
and thinking and it is was with

623
00:36:48,800 --> 00:36:51,300
the Advent like the reason that
Things fall off last year,

624
00:36:51,300 --> 00:36:52,600
right?
For a lot of different reasons

625
00:36:53,200 --> 00:36:56,100
but the reason it's being picked
up right now is because I was

626
00:36:56,100 --> 00:36:59,100
able to see what other people
were able to do with technology

627
00:36:59,600 --> 00:37:02,900
and that's what I hope that
people walk away from this whole

628
00:37:02,900 --> 00:37:06,800
conversation is to go in and
don't feel like you're being

629
00:37:06,800 --> 00:37:08,800
behind.
Don't feel like you're behind

630
00:37:08,800 --> 00:37:10,900
because you're already.
Behind right?

631
00:37:11,100 --> 00:37:13,700
That's like trying to say I'm
behind in biology.

632
00:37:13,900 --> 00:37:16,400
Of course, I am in cellular
biology since the 10th grade.

633
00:37:16,600 --> 00:37:19,000
Why would I be on top of
biology?

634
00:37:19,300 --> 00:37:22,100
You're gonna be behind.
The best thing that you can do

635
00:37:22,100 --> 00:37:25,500
is just be utterly amazed about
what other people are doing and

636
00:37:25,500 --> 00:37:29,800
using that to inspire yourself.
Like I did, we're not for me.

637
00:37:29,800 --> 00:37:31,600
It wasn't.
Oh, I can write this prompt

638
00:37:31,600 --> 00:37:33,800
there with me.
I can go start this company.

639
00:37:34,300 --> 00:37:36,600
I know how to do it and know
what I need to do.

640
00:37:37,400 --> 00:37:42,200
And I am in power bi artificial
intelligence to do it.

641
00:37:44,000 --> 00:37:48,600
That's yeah, absolutely.
And I want to talk about from

642
00:37:48,600 --> 00:37:50,500
the ID perspective, the
instructional designer

643
00:37:50,500 --> 00:37:55,500
perspective, how much time this
saves using Chad GPT or other AI

644
00:37:55,500 --> 00:37:58,800
to write outlines or just get an
idea of a storyboard because we

645
00:37:58,800 --> 00:38:02,300
spend so much time
conceptualizing based off the

646
00:38:02,300 --> 00:38:06,300
content that the sneeze give us
how this is all going to flow.

647
00:38:06,400 --> 00:38:09,900
They'll give us a PowerPoint but
that's not necessarily in The

648
00:38:09,900 --> 00:38:13,700
Logical order or the order
should be in for learning.

649
00:38:13,800 --> 00:38:17,200
And so I use chat dtp a lot just
to outline stuff and I know

650
00:38:17,200 --> 00:38:19,700
people say that and it helps so
much.

651
00:38:19,700 --> 00:38:23,000
It just gives you like a sense
of relief that you don't have to

652
00:38:23,000 --> 00:38:25,300
go through this super critical
process.

653
00:38:25,300 --> 00:38:28,300
You can refocus your efforts on
the creative side of

654
00:38:28,300 --> 00:38:30,700
instructional design, and
developing the interactive's and

655
00:38:30,700 --> 00:38:32,200
things.
But with the Learners to

656
00:38:32,200 --> 00:38:35,500
behavioral changes and the
assessments that are needed to

657
00:38:35,500 --> 00:38:37,800
exhibit that.
So from an instructional

658
00:38:37,800 --> 00:38:41,000
designer perspective, this is
something that I can't wait

659
00:38:41,100 --> 00:38:44,000
until people.
We start incorporating These

660
00:38:44,000 --> 00:38:47,300
inter tools Microsoft is putting
in the tools like they're

661
00:38:47,300 --> 00:38:50,200
articulates and the Dobies of
the world start putting these

662
00:38:50,200 --> 00:38:54,800
into the tools to make it so
much more manageable and I feel

663
00:38:54,800 --> 00:38:57,700
like we're going to level up as
humans with a I like we're

664
00:38:57,700 --> 00:38:59,700
already doing it.
We're going to just we're going

665
00:38:59,700 --> 00:39:03,100
to think more critically on
higher levels because of this.

666
00:39:04,300 --> 00:39:08,600
So here is in my mind, a reality
that will happen once the

667
00:39:08,600 --> 00:39:13,300
hardware patches up or if once
things are catching up and this

668
00:39:13,300 --> 00:39:16,000
is what this is, what makes it
so scary, right?

669
00:39:16,000 --> 00:39:18,100
That people have peed like a
eyes, it would come and replace

670
00:39:18,100 --> 00:39:20,800
my job is because you have
people like me thinking like

671
00:39:20,800 --> 00:39:22,600
this.
And here's what people need to

672
00:39:22,600 --> 00:39:24,700
do is like you should be asking
these questions.

673
00:39:25,000 --> 00:39:26,700
You should be the ones,
implementing these things in

674
00:39:26,700 --> 00:39:29,800
your organization, because
here's what's going to happen in

675
00:39:30,300 --> 00:39:31,600
five years.
Right?

676
00:39:31,600 --> 00:39:34,300
A company is going to have
Artificial intelligence

677
00:39:34,300 --> 00:39:36,800
connected it to its entire
knowledge base.

678
00:39:37,400 --> 00:39:40,700
You're gonna have a new hire
come in and say hey this is how

679
00:39:40,700 --> 00:39:44,700
I like to learn.
You even need an instructional

680
00:39:44,700 --> 00:39:47,600
designer anymore say, I like, to
learn this way.

681
00:39:47,600 --> 00:39:50,000
Can you teach me about this
process?

682
00:39:51,000 --> 00:39:56,600
And now you have an interactive
instructional designer that will

683
00:39:56,600 --> 00:40:01,700
answer your questions.
Personalized based on the

684
00:40:01,700 --> 00:40:05,400
knowledge base from the company,
based on how you like to learn.

685
00:40:05,500 --> 00:40:08,300
If you like to learn via
problems are you like to learn

686
00:40:08,300 --> 00:40:11,400
via funny videos?
But you can say, make me a funny

687
00:40:11,400 --> 00:40:14,500
video starring Bruce Willis that
Has me about this principle.

688
00:40:16,500 --> 00:40:21,400
That's there.
All the pieces are there and

689
00:40:21,400 --> 00:40:23,000
that's what everybody's going to
be think.

690
00:40:23,000 --> 00:40:25,600
That's when 50 years from now.
They're going to be looking back

691
00:40:25,600 --> 00:40:27,800
and thinking like wow, Josh apt
for.

692
00:40:27,800 --> 00:40:29,400
What is that?
Just like my kid is going to

693
00:40:29,400 --> 00:40:32,600
look at like the Motorola 120 e,
razor fall off your bike with

694
00:40:32,600 --> 00:40:35,100
peanut phone that up.
It was really fancy to that of

695
00:40:35,100 --> 00:40:36,700
blue screen instead of a green
screen.

696
00:40:36,900 --> 00:40:41,900
Yeah.
Anyway, it's just so much fun

697
00:40:41,900 --> 00:40:44,500
stuff out there and I hope that
people are not intimidated by

698
00:40:44,500 --> 00:40:50,400
it, and they feel excited about
it because you should be agreed.

699
00:40:52,100 --> 00:40:53,400
Yeah.
And I think that, and you

700
00:40:53,400 --> 00:40:55,900
mentioned this before, and it
there are just around 40 minutes

701
00:40:55,900 --> 00:40:57,200
Arc.
I do want to on your time, and I

702
00:40:57,207 --> 00:41:01,100
know that you have to go, Holly
have to go play with it, right?

703
00:41:01,100 --> 00:41:05,600
If you're out there, you're
higher ed or any field, really?

704
00:41:06,200 --> 00:41:10,300
If you're concerned, if you're
worried that is okay, those are

705
00:41:10,300 --> 00:41:13,700
emotionally valid responses.
Make sure you get in and you

706
00:41:14,000 --> 00:41:18,000
really play with it.
And I'll be totally honest, or

707
00:41:18,000 --> 00:41:21,700
I'm emotionally, I have good
days, I have bad days with a, i

708
00:41:21,800 --> 00:41:24,100
Some days.
I think, yes, this is going to

709
00:41:24,100 --> 00:41:26,600
allow me to do X, Y, and Z
faster better.

710
00:41:26,600 --> 00:41:29,500
So on and so forth.
And other days, I'm very

711
00:41:29,500 --> 00:41:32,400
negative about it and I go back
and forth and I've learned to be

712
00:41:32,400 --> 00:41:36,300
emotionally, okay, with that,
depending on my own State, what

713
00:41:36,300 --> 00:41:41,400
I'm working with, I think that
for a lot of Educators, we need

714
00:41:41,400 --> 00:41:43,700
to do, we need to do our due
diligence, which means playing

715
00:41:43,700 --> 00:41:46,600
with it, playing with the
software, really reflecting on

716
00:41:46,600 --> 00:41:50,900
it and being okay with not
feeling, totally, consistent

717
00:41:50,900 --> 00:41:53,900
emotionally, Lee with the tag.
I think that a lot of people are

718
00:41:53,900 --> 00:41:55,900
there, and it's okay to be there
too.

719
00:41:56,200 --> 00:41:57,300
All right.
And I, so I do want to honor

720
00:41:57,300 --> 00:41:57,900
your time.
Eric.

721
00:41:57,900 --> 00:42:01,200
So, very last question, if
someone listening to this

722
00:42:01,200 --> 00:42:04,900
episode wants to talk to, you
wants to have a conversation or

723
00:42:04,900 --> 00:42:10,200
reach out to you for any reason,
I should they put 100% LinkedIn

724
00:42:10,300 --> 00:42:13,500
is where I live by.
You could also check out the

725
00:42:13,500 --> 00:42:18,800
project, its www, dot project,
rhizome.com, and you can get a

726
00:42:18,800 --> 00:42:23,700
whole bunch of stuff there too.
Fantastic.

727
00:42:24,600 --> 00:42:27,200
I love it.
I'm so glad you came back to

728
00:42:27,200 --> 00:42:28,800
this.
I'm so glad you came back to

729
00:42:28,800 --> 00:42:30,800
this.
I am very grateful for the two

730
00:42:30,800 --> 00:42:36,500
of you truly or not only having
this conversation at being able

731
00:42:36,500 --> 00:42:38,800
to promote and all that kind of
stuff.

732
00:42:39,200 --> 00:42:43,800
But I think like it is this
community that we built on

733
00:42:43,800 --> 00:42:47,200
LinkedIn over the past year or
two years, whatever that I'm

734
00:42:47,200 --> 00:42:50,500
just very grateful for.
And I don't think that I could

735
00:42:50,500 --> 00:42:57,600
have approached the problem that
I am trying to approach if it

736
00:42:57,600 --> 00:43:02,200
were not for the constant,
iterative reactions and refining

737
00:43:02,200 --> 00:43:05,700
of ideas with people, that would
not give you.

738
00:43:05,700 --> 00:43:08,500
So again, I'm just very
grateful, thank you, both for

739
00:43:08,800 --> 00:43:13,900
doing absolutely anytime.
Okay, you're so welcome.

740
00:43:13,900 --> 00:43:16,000
And thank you so much our for
coming on the show.

741
00:43:16,100 --> 00:43:19,700
Coming on the first joint
podcast and good luck with

742
00:43:19,700 --> 00:43:21,700
everything.
And it's been a pleasure hear

743
00:43:21,700 --> 00:43:24,400
you talk with a.
I Thank you so much.

744
00:43:29,700 --> 00:43:34,000
You've just experienced an
another amazing episode of Ed

745
00:43:34,000 --> 00:43:35,100
up.
Ed Tech.

746
00:43:35,700 --> 00:43:41,100
Be sure to visit our website at
Ed up, edtech.com to get all the

747
00:43:41,100 --> 00:43:43,900
updates on the latest edtech
happening.

748
00:43:44,600 --> 00:43:46,000
See you next time.