AI Interpreting Solutions Evaluation Toolkit Part A: Organization, Implementation and Management

Coalition for Sign Language Equity in Technology (CoSET)

The Advisory Group on AI and Sign Language Interpreting renamed itself to the Coalition for Sign Language Equity in Technology to reinforce its independent and equal status with the SAFE AI Task Force.

For more information on CoSET visit our webpage at coset.org

Mission

To protect the integrity of communication between principal communicators during any interpreted interaction by developing standards for responsive AI systems and setting performance benchmarks for human-only, machine-only, and combined human+machine interpreting solutions. 

Vision

Our Vision is that sign language and spoken language equity will be integral to the global technology landscape, and that we will be recognized as an authority for setting standards, certifying technology, and advocating for the inclusion of all sign and spoken languages.

Deaf Advisory Group on AI and Sign Language Interpreting

Experienced End User Intelligence about Automated Interpreting

00:00:01:06 – 00:00:01:20
Hello.

00:00:01:20 – 00:00:03:23
Hello, everyone.

00:00:03:23 – 00:00:06:23
My name is Tim Riker

00:00:07:02 – 00:00:11:05
and I am going to be the presenter today,

00:00:11:12 – 00:00:12:11
one of the presenters.

00:00:12:11 – 00:00:13:22
I’m from Brown University

00:00:13:22 – 00:00:17:05
and I’m a member of this advisory board,

00:00:17:05 – 00:00:19:21
which is for artificial intelligence

00:00:19:21 – 00:00:22:03
and sign language interpreting.

00:00:22:03 – 00:00:23:19
Today we are thrilled

00:00:23:19 – 00:00:26:20
because we are going to be providing you

00:00:26:20 – 00:00:28:04
a presentation.

00:00:28:04 – 00:00:30:08
We’ll be talking about our report.

00:00:30:08 – 00:00:32:18
And

00:00:32:18 – 00:00:34:17
the advisory council is here today

00:00:34:17 – 00:00:35:20
together.

00:00:35:20 – 00:00:37:07
We’ll be talking about some of the work

00:00:37:07 – 00:00:39:08
that we’ve done, collecting data

00:00:39:08 – 00:00:40:17
from three webinars

00:00:40:17 – 00:00:43:09
that we hosted last fall.

00:00:43:09 – 00:00:45:01
The reason that we decided

00:00:45:01 – 00:00:46:05
to host these webinars

00:00:46:05 – 00:00:47:02
and do this research

00:00:47:02 – 00:00:48:17
is because we wanted to get more

00:00:48:17 – 00:00:50:06
of the deaf perspective

00:00:50:06 – 00:00:51:23
and to take that perspective.

00:00:51:23 – 00:00:54:00
And we do

00:00:54:00 – 00:00:55:10
some more research and former

00:00:55:10 – 00:00:57:20
Task Force for Safe AI.

00:00:57:20 – 00:01:00:09
That report is going to be presented

00:01:00:09 – 00:01:03:16
today and we

00:01:05:02 – 00:01:06:10
would like to share now with you

00:01:06:10 – 00:01:09:10
the topic of the session.

00:01:12:02 – 00:01:14:15
I’ll go ahead and introduce myself.

00:01:14:15 – 00:01:17:06
And so in terms of visual description,

00:01:17:06 – 00:01:20:11
I am currently wearing a black shirt.

00:01:21:11 – 00:01:22:07
It’s a long sleeve

00:01:22:07 – 00:01:23:04
black shirt,

00:01:23:04 – 00:01:24:09
collared shirt

00:01:24:09 – 00:01:27:09
with buttons and a white male

00:01:27:16 – 00:01:30:16
with reddish blondish hair.

00:01:30:21 – 00:01:33:20
I have a bit of a mustache

00:01:33:20 – 00:01:36:14
and facial hair.

00:01:36:14 – 00:01:38:23
So today our presentation

00:01:38:23 – 00:01:42:00
topic is going to be death safety.

00:01:42:07 – 00:01:43:23
I

00:01:43:23 – 00:01:47:03
a.i meaning artificial intelligence.

00:01:47:15 – 00:01:48:20
And our goal

00:01:48:20 – 00:01:50:12
is to have a legal foundation

00:01:50:12 – 00:01:53:12
for ubiquitous automatic interpreting,

00:01:53:17 – 00:01:55:19
using artificial intelligence

00:01:55:19 – 00:01:58:19
and how that relates to interpreting.

00:01:58:20 – 00:02:01:16
So

00:02:01:16 – 00:02:04:05
I wanted to talk a bit about why

00:02:04:05 – 00:02:07:05
this topic is so important.

00:02:08:24 – 00:02:09:21
As you know,

00:02:09:21 – 00:02:11:23
there are many users

00:02:11:23 – 00:02:13:13
of sign language interpreters.

00:02:13:13 – 00:02:17:17
We sometimes go through frustrations

00:02:18:02 – 00:02:19:03
and situations

00:02:19:03 – 00:02:22:04
because we are using technology,

00:02:22:13 – 00:02:25:07
and sometimes technology

00:02:25:07 – 00:02:26:10
can be very beneficial,

00:02:26:10 – 00:02:28:09
while other times technology

00:02:28:09 – 00:02:30:03
can cause harm.

00:02:30:03 – 00:02:31:16
So with VR, I

00:02:38:23 – 00:02:41:20
so with VR, AI, for example,

00:02:41:20 – 00:02:43:04
we have video,

00:02:43:04 – 00:02:44:06
remote

00:02:44:06 – 00:02:45:08
video, remote interpreters

00:02:45:08 – 00:02:47:07
that come up on the screen.

00:02:47:07 – 00:02:50:07
And sometimes it can be a great idea

00:02:50:17 – 00:02:52:04
because, you know,

00:02:52:04 – 00:02:55:04
when it first came out, we saw that

00:02:55:16 – 00:02:58:06
there was a lot of freezing.

00:02:58:06 – 00:03:00:23
Sometimes there might be challenges

00:03:00:23 – 00:03:01:20
internally, like,

00:03:01:20 – 00:03:02:19
let’s say

00:03:02:19 – 00:03:03:14
things are going on

00:03:03:14 – 00:03:06:02
in that room that cause issues.

00:03:06:02 – 00:03:06:20
Various issues

00:03:06:20 – 00:03:09:20
would happen with this technology.

00:03:09:21 – 00:03:13:01
Now, if you think about R2,

00:03:13:01 – 00:03:15:00
we think about automatic interpreting

00:03:15:00 – 00:03:16:07
or artificial intelligence

00:03:16:07 – 00:03:17:23
and interpreting.

00:03:17:23 – 00:03:19:11
Are we ready for that?

00:03:19:11 – 00:03:20:20
How will that impact

00:03:20:20 – 00:03:22:19
the greater community?

00:03:22:19 – 00:03:24:01
What will be the community’s

00:03:24:01 – 00:03:26:03
view of this?

00:03:26:03 – 00:03:28:07
So last fall

00:03:28:07 – 00:03:31:07
we hosted three webinars

00:03:31:10 – 00:03:33:23
and those webinars happened here

00:03:33:23 – 00:03:35:10
at Brown University.

00:03:35:10 – 00:03:38:03
They were through Zoom

00:03:38:03 – 00:03:40:13
and we had a panel

00:03:40:13 – 00:03:41:21
and we had discussions

00:03:41:21 – 00:03:43:00
and we gathered the view

00:03:43:00 – 00:03:44:12
of multiple people.

00:03:44:12 – 00:03:46:07
We also had deaf community members

00:03:46:07 – 00:03:48:02
who came in to watch

00:03:48:02 – 00:03:49:18
and make comments

00:03:49:18 – 00:03:51:23
and talk about their perspective.

00:03:53:17 – 00:03:56:12
So in that discussion we saw that

00:03:56:12 – 00:03:58:16
there was a lot of rich information

00:03:58:16 – 00:04:00:11
that was shared

00:04:00:11 – 00:04:03:11
and the team, the advisory group,

00:04:04:03 – 00:04:07:04
decided to work to analyze

00:04:07:14 – 00:04:08:24
that information that came

00:04:08:24 – 00:04:10:15
from the discussions.

00:04:10:15 – 00:04:11:17
And right away,

00:04:11:17 – 00:04:14:08
we knew we had a lot of rich content

00:04:14:08 – 00:04:15:15
and that we would be able

00:04:15:15 – 00:04:17:05
to take that content

00:04:17:05 – 00:04:19:01
and add it to the report

00:04:19:01 – 00:04:20:03
so that we could get

00:04:20:03 – 00:04:22:23
a general understanding of what fire

00:04:22:23 – 00:04:25:00
I would look like.

00:04:25:00 – 00:04:28:06
And we had several different questions

00:04:28:06 – 00:04:30:00
that we put into a survey

00:04:30:00 – 00:04:32:18
and we sent those out and we were able

00:04:32:18 – 00:04:36:01
to get those responses from the deaf.

00:04:36:08 – 00:04:36:23
Unfortunately,

00:04:36:23 – 00:04:39:08
we did not have a survey in ASL,

00:04:39:08 – 00:04:41:06
but we knew that we needed

00:04:41:06 – 00:04:42:14
to have these discussions

00:04:42:14 – 00:04:44:03
in order to gather the information

00:04:44:03 – 00:04:45:14
we were looking for.

00:04:45:14 – 00:04:46:03
Let’s go ahead and

00:04:46:03 – 00:04:47:00
head to the next slide.

00:04:53:03 – 00:04:55:06
So

00:04:55:06 – 00:04:55:17
I’d like

00:04:55:17 – 00:04:56:12
to introduce

00:04:56:12 – 00:04:59:12
you to other members of our team.

00:04:59:15 – 00:05:02:07
We all participated together

00:05:02:07 – 00:05:03:15
in gathering this research

00:05:03:15 – 00:05:06:05
and we’ve been working hard as a group

00:05:06:05 – 00:05:07:10
to get that information.

00:05:07:10 – 00:05:08:24
And these lovely people here

00:05:08:24 – 00:05:10:09
volunteer their time.

00:05:10:09 – 00:05:12:01
I’d like to introduce you to them now

00:05:12:01 – 00:05:14:03
so that you can get to know them a bit.

00:05:14:03 – 00:05:15:13
And they will also be talking

00:05:15:13 – 00:05:17:09
about the report today.

00:05:17:09 – 00:05:19:18
So

00:05:19:18 – 00:05:22:16
let me go ahead and pass it over to you.

00:05:22:16 – 00:05:25:09
Let’s start with Theresa.

00:05:25:09 – 00:05:25:17
Tracey,

00:05:25:17 – 00:05:28:07
if you’d like to introduce yourself.

00:05:28:07 – 00:05:28:17
Sure.

00:05:28:17 – 00:05:30:13
Good morning, everyone.

00:05:30:13 – 00:05:33:08
My name is Theresa Blake.

00:05:33:08 – 00:05:36:08
Maya Burke

00:05:36:09 – 00:05:38:09
and I work at Gallaudet University.

00:05:38:09 – 00:05:40:12
I’m a professor of philosophy

00:05:40:12 – 00:05:41:24
and my research

00:05:41:24 – 00:05:44:11
is specifically in ethics.

00:05:44:11 – 00:05:47:11
And it’s the ethical application

00:05:47:15 – 00:05:48:21
to technology.

00:05:48:21 – 00:05:50:09
And I’m thrilled to be here with you

00:05:50:09 – 00:05:50:22
all today.

00:05:50:22 – 00:05:52:00
I’m looking forward

00:05:52:00 – 00:05:53:17
to discussing the webinar

00:05:53:17 – 00:05:54:24
and having other discussions

00:05:54:24 – 00:05:56:03
with all of you today.

00:05:56:03 – 00:05:57:03
And I did like

00:05:57:03 – 00:05:58:05
I would like to add

00:05:58:05 – 00:06:01:05
that in terms of visual description,

00:06:01:11 – 00:06:02:10
I am a middle aged

00:06:02:10 – 00:06:06:03
woman with an olive, with olive skin

00:06:06:12 – 00:06:08:24
and also I have brown eyes,

00:06:08:24 – 00:06:09:11
I’m wearing

00:06:09:11 – 00:06:11:06
glasses and I have a brown

00:06:11:06 – 00:06:12:00
I have brown hair

00:06:12:00 – 00:06:13:13
and my hair is in a bun today

00:06:13:13 – 00:06:16:09
and I’m wearing a gray sweater

00:06:18:04 – 00:06:20:20
and I’m here in my office at Gallaudet.

00:06:20:20 – 00:06:22:22
Thank you, Theresa.

00:06:22:22 – 00:06:24:20
This next, let’s have Jeff.

00:06:24:20 – 00:06:25:02
Jeff,

00:06:25:02 – 00:06:26:15
would you like to introduce yourself?

00:06:26:15 – 00:06:29:04
Hello, My name is Jeff Schall

00:06:29:04 – 00:06:32:04
and I am working.

00:06:32:06 – 00:06:35:06
I work to develop

00:06:35:06 – 00:06:38:02
AI for the deaf and hard of hearing.

00:06:38:02 – 00:06:41:08
And in terms of a visual description,

00:06:41:17 – 00:06:43:00
I am wearing a white

00:06:43:00 – 00:06:44:21
and black plaid shirt.

00:06:44:21 – 00:06:47:15
I have facial hair and brown eyes

00:06:47:15 – 00:06:49:06
and I’m here.

00:06:49:06 – 00:06:50:10
My office is a background

00:06:50:10 – 00:06:51:17
and I work for go sign.

00:06:51:17 – 00:06:52:24
I

00:06:52:24 – 00:06:55:24
next will have Holly.

00:06:56:03 – 00:06:56:22
Yes, hello.

00:06:56:22 – 00:06:58:11
Good morning.

00:06:58:11 – 00:07:00:15
My name is Holly.

00:07:00:15 – 00:07:03:07
Last name is Jackson

00:07:03:07 – 00:07:06:07
and visual description.

00:07:06:16 – 00:07:09:16
I am an African-American black female.

00:07:10:04 – 00:07:14:02
I have light skin and I have curly

00:07:14:02 – 00:07:17:02
natural hair today.

00:07:17:19 – 00:07:18:12
And

00:07:19:11 – 00:07:22:11
I have a dark Navy

00:07:22:11 – 00:07:25:11
blue suit jacket on.

00:07:26:10 – 00:07:29:09
And I the shirt, my Navy blue

00:07:29:09 – 00:07:29:22
suit

00:07:29:22 – 00:07:32:22
jacket has light white stripes on it.

00:07:33:08 – 00:07:36:02
And I have a shirt beneath my jacket,

00:07:36:02 – 00:07:39:07
and it is a light blue tan

00:07:40:19 – 00:07:42:17
lace

00:07:42:17 – 00:07:43:20
top.

00:07:43:20 – 00:07:45:11
That’s the design.

00:07:45:11 – 00:07:48:14
And my background today

00:07:49:18 – 00:07:52:15
is light gray

00:07:52:15 – 00:07:54:17
plane, light gray background.

00:07:54:17 – 00:07:57:02
And I’m an interpreter.

00:07:57:02 – 00:07:59:09
I’m a hearing interpreter

00:07:59:09 – 00:08:01:15
and also an educator

00:08:01:15 – 00:08:04:08
and educator of ASL and interpreting.

00:08:04:08 – 00:08:06:11
I work for any ASL program

00:08:06:11 – 00:08:09:23
interpreting program, and also I am here

00:08:11:03 – 00:08:13:19
for Naomi,

00:08:13:19 – 00:08:15:13
the representation of Niobe,

00:08:15:13 – 00:08:18:15
the Atlanta chapter, and

00:08:18:15 – 00:08:21:15
I serve as the secretary for the board.

00:08:21:22 – 00:08:23:15
That’s my position this year.

00:08:23:15 – 00:08:24:13
Thank you very much.

00:08:24:13 – 00:08:26:13
I’m happy to be here.

00:08:26:13 – 00:08:27:22
Thank you, Holly.

00:08:27:22 – 00:08:30:15
And last but not least, Anne Marie.

00:08:32:10 – 00:08:35:10
Hello, I’m Anne Marie.

00:08:35:19 – 00:08:37:15
Last name is Killian,

00:08:37:15 – 00:08:40:15
and this is my signed name,

00:08:40:18 – 00:08:43:19
and I am the CEO for

00:08:45:23 – 00:08:49:16
Tie Access and visual description

00:08:49:16 – 00:08:50:14
is that I’m a white

00:08:50:14 – 00:08:53:19
female, middle aged with medium

00:08:53:19 – 00:08:54:24
length hair,

00:08:54:24 – 00:08:56:19
brown hair, and I’m wearing glasses.

00:08:56:19 – 00:08:57:17
Today

00:08:57:17 – 00:08:59:01
I have on a suit jacket

00:08:59:01 – 00:09:00:18
that is black with a purple shirt

00:09:00:18 – 00:09:01:23
beneath it.

00:09:01:23 – 00:09:03:11
And in the background

00:09:03:11 – 00:09:06:11
you can see my dining room table

00:09:06:13 – 00:09:09:21
and blacks, black chairs and

00:09:12:08 – 00:09:13:13
you might see two dogs

00:09:13:13 – 00:09:14:19
running around in the background.

00:09:14:19 – 00:09:16:23
If that happens, I apologize in advance.

00:09:16:23 – 00:09:18:09
Like everyone else.

00:09:18:09 – 00:09:19:13
I’m thrilled to be here today.

00:09:19:13 – 00:09:21:02
Thank you.

00:09:21:02 – 00:09:23:12
Great.

00:09:23:12 – 00:09:26:12
So let’s go to the next slide.

00:09:28:17 – 00:09:29:11
So today

00:09:29:11 – 00:09:31:07
we will be talking about

00:09:31:07 – 00:09:32:14
multiple things.

00:09:32:14 – 00:09:34:04
And during these presentations,

00:09:34:04 – 00:09:36:05
we’ll go in depth about our studies

00:09:36:05 – 00:09:38:04
and what we have found through analyzing

00:09:38:04 – 00:09:39:02
this data

00:09:39:02 – 00:09:40:13
will be sharing with you

00:09:40:13 – 00:09:42:17
this information today.

00:09:42:17 – 00:09:44:04
The first thing we’re going to be doing

00:09:44:04 – 00:09:45:08
is identifying

00:09:45:08 – 00:09:47:21
three critical impact areas.

00:09:49:00 – 00:09:51:02
And so these impact

00:09:51:02 – 00:09:52:16
areas are quite important.

00:09:52:16 – 00:09:54:05
We’ll be talking more in depth

00:09:54:05 – 00:09:56:12
and describing what they are for you.

00:09:56:12 – 00:09:59:12
Secondly,

00:09:59:12 – 00:10:00:17
with this analysis

00:10:00:17 – 00:10:03:17
and with this research in our webinars,

00:10:04:07 – 00:10:06:14
we were able to go through the data

00:10:06:14 – 00:10:08:00
and that data helped us

00:10:08:00 – 00:10:11:06
to build a better understanding of what

00:10:11:06 – 00:10:12:23
the deaf communities perspective

00:10:12:23 – 00:10:14:18
and experience has been

00:10:14:18 – 00:10:17:17
and their experiences with interpreters

00:10:17:17 – 00:10:19:04
and the harms that have happened

00:10:19:04 – 00:10:20:18
and the way that these experiences

00:10:20:18 – 00:10:21:08
have impacted

00:10:21:08 – 00:10:22:07
their life

00:10:22:07 – 00:10:23:13
in terms of access,

00:10:23:13 – 00:10:25:11
in terms of communication access.

00:10:25:11 – 00:10:26:20
And so it’s very important

00:10:26:20 – 00:10:28:17
to have this deaf community perspective

00:10:28:17 – 00:10:29:20
so that we can understand

00:10:29:20 – 00:10:30:19
what they’ve been through

00:10:30:19 – 00:10:32:00
when it comes to

00:10:32:00 – 00:10:34:01
their experiences in interpreting

00:10:34:01 – 00:10:36:05
and where harm has happened.

00:10:36:05 – 00:10:36:19
Often

00:10:36:19 – 00:10:38:17
that is an experience

00:10:38:17 – 00:10:40:16
that is common in our communities.

00:10:40:16 – 00:10:42:23
So we would like to mitigate those harms

00:10:42:23 – 00:10:47:02
and ensure that if we do put out

00:10:47:02 – 00:10:48:00
new technology,

00:10:48:00 – 00:10:49:08
that it’s going to be we’re going

00:10:49:08 – 00:10:50:16
to be mindful of those harms

00:10:50:16 – 00:10:52:06
that have been experience.

00:10:52:06 – 00:10:53:14
Third, we’re going to talk

00:10:53:14 – 00:10:55:10
about the value of the big picture

00:10:55:10 – 00:10:57:08
lens on possibilities.

00:10:57:08 – 00:10:58:12
We’ll talk about

00:10:58:12 – 00:11:00:01
what it looks like to do right

00:11:00:01 – 00:11:02:11
and prevent possible disaster.

00:11:02:11 – 00:11:05:23
We want to make sure that we are looking

00:11:05:23 – 00:11:08:12
through a lens where we are

00:11:08:12 – 00:11:10:02
showing care and concern

00:11:10:02 – 00:11:11:24
about the future of the community

00:11:11:24 – 00:11:13:10
and taking all of these things

00:11:13:10 – 00:11:15:00
into account. Next slide.

00:11:24:19 – 00:11:26:11
So for today

00:11:26:11 – 00:11:29:03
in this webinar, like you mentioned,

00:11:29:03 – 00:11:30:06
we had a panel

00:11:30:06 – 00:11:31:17
and we had

00:11:31:17 – 00:11:32:11
today will be going

00:11:32:11 – 00:11:34:06
through different presentations

00:11:34:06 – 00:11:37:10
and we’ll be going through the report

00:11:37:10 – 00:11:39:22
and having discussions about that.

00:11:39:22 – 00:11:42:20
So we will be following this order,

00:11:42:20 – 00:11:45:20
as you see outlined here and the slide

00:11:46:24 – 00:11:48:05
we will be discussing

00:11:48:05 – 00:11:51:05
ethics and fairness.

00:11:52:02 – 00:11:54:17
We will be discussing how this research

00:11:54:17 – 00:11:55:24
was studied

00:11:55:24 – 00:11:58:24
and our approach to the research.

00:11:59:06 – 00:12:02:02
We will be talking about what we found.

00:12:02:02 – 00:12:04:10
So our findings,

00:12:04:10 – 00:12:07:13
we will also be discussing the results

00:12:08:03 – 00:12:11:07
and outcomes of our research.

00:12:12:02 – 00:12:13:17
And also we will be talking

00:12:13:17 – 00:12:14:18
about concerns

00:12:14:18 – 00:12:17:18
as it relates to the deaf community.

00:12:18:07 – 00:12:20:14
We will be also discussing techno

00:12:20:14 – 00:12:23:14
technology and the quality

00:12:25:05 – 00:12:27:23
and we will be asking,

00:12:27:23 – 00:12:29:10
are we ready for this?

00:12:29:10 – 00:12:30:20
Are we ready for eye,

00:12:32:06 – 00:12:33:00
for eye

00:12:33:00 – 00:12:36:00
and sign language to come together?

00:12:36:20 – 00:12:37:18
Are we ready?

00:12:37:18 – 00:12:40:18
And if we are, what does that look like?

00:12:41:18 – 00:12:44:18
What kind of risks are involved?

00:12:46:00 – 00:12:48:11
We need to be proactive

00:12:48:11 – 00:12:50:19
in understanding

00:12:50:19 – 00:12:52:09
and predicting those risks

00:12:52:09 – 00:12:53:21
so that we can mitigate

00:12:53:21 – 00:12:56:21
or resolve them before they occur.

00:12:59:18 – 00:13:02:08
We will also discuss the future

00:13:02:08 – 00:13:04:17
and what we can anticipate

00:13:04:17 – 00:13:08:02
and what we can recommend for

00:13:09:06 – 00:13:11:03
any anyone

00:13:11:03 – 00:13:13:22
who is going to be working in relation

00:13:13:22 – 00:13:15:08
to this topic.

00:13:15:08 – 00:13:18:08
All of us here today are impacted

00:13:18:08 – 00:13:22:23
or affected by this topic and many people

00:13:22:23 – 00:13:23:15
who are not here

00:13:23:15 – 00:13:24:15
today,

00:13:24:15 – 00:13:27:15
the entire community that we represent

00:13:27:22 – 00:13:30:22
and also work with.

00:13:32:05 – 00:13:34:17
So now

00:13:34:17 – 00:13:36:11
Teresa

00:13:36:11 – 00:13:39:11
will go ahead and begin this discussion

00:13:39:13 – 00:13:40:20
and we’re going to start off again

00:13:40:20 – 00:13:42:09
with ethics and fairness.

00:13:42:09 – 00:13:43:18
Teresa, I’ll turn it over to you.

00:13:49:06 – 00:13:50:09
one moment.

00:13:50:09 – 00:13:53:09
Let me make sure I have this up.

00:13:54:11 – 00:13:55:09
I always have to make sure

00:13:55:09 – 00:13:58:05
I have my spotlight on so that I’m seen.

00:13:58:05 – 00:13:58:17
Okay.

00:13:58:17 – 00:14:02:12
So ethics, ethics and fairness,

00:14:03:01 – 00:14:04:23
what emphasis

00:14:04:23 – 00:14:06:14
and what are we looking at here

00:14:06:14 – 00:14:07:18
with ethics and fairness?

00:14:07:18 – 00:14:09:12
What do we want to avoid?

00:14:09:12 – 00:14:11:00
We don’t want to cause harm.

00:14:11:00 – 00:14:13:17
How do we reduce the cause of harm?

00:14:13:17 – 00:14:15:02
And what does harm mean

00:14:15:02 – 00:14:16:14
to the deaf community

00:14:16:14 – 00:14:18:16
and deaf individuals?

00:14:18:16 – 00:14:21:22
Harm in general, to humanity?

00:14:21:22 – 00:14:23:12
We want to avoid.

00:14:23:12 – 00:14:27:00
So we have two

00:14:27:00 – 00:14:30:00
topics really

00:14:30:00 – 00:14:32:08
coexisting here ethics and fairness.

00:14:32:08 – 00:14:35:13
So what you’ll notice here,

00:14:35:20 – 00:14:38:24
we are concerned with controlling bias

00:14:39:24 – 00:14:42:00
and we want to

00:14:42:00 – 00:14:43:10
assign responsibility

00:14:43:10 – 00:14:46:10
or accountability, rather, as it

00:14:46:22 – 00:14:49:11
relates to AI to avoid harm.

00:14:51:13 – 00:14:52:04
The second

00:14:52:04 – 00:14:52:17
point I like

00:14:52:17 – 00:14:53:00
to make

00:14:53:00 – 00:14:55:05
is that we need to be very clear

00:14:55:05 – 00:14:58:14
and transparent with our documentation

00:14:58:14 – 00:15:00:02
about who is accountable

00:15:00:02 – 00:15:02:00
in the design of AI

00:15:02:00 – 00:15:03:24
and the development of AI

00:15:03:24 – 00:15:06:22
and the application and evaluation.

00:15:06:22 – 00:15:09:07
Everything related to A.I.

00:15:09:07 – 00:15:11:19
who is responsible for this portion

00:15:11:19 – 00:15:13:11
ethically?

00:15:13:11 – 00:15:16:11
Next slide.

00:15:20:01 – 00:15:20:13
Okay.

00:15:20:13 – 00:15:22:05
So we spoke about ethics,

00:15:22:05 – 00:15:23:01
and now I want to talk

00:15:23:01 – 00:15:25:08
a little bit about fairness.

00:15:25:08 – 00:15:28:19
How we in our society

00:15:28:20 – 00:15:31:20
have developed the meaning behind this.

00:15:31:20 – 00:15:35:12
We have a lot of negative biases,

00:15:35:16 – 00:15:36:08
and we also have

00:15:36:08 – 00:15:38:18
positive biases in society.

00:15:38:18 – 00:15:39:18
But what we want to do

00:15:39:18 – 00:15:42:02
is make sure that in air

00:15:42:02 – 00:15:44:07
we want to reflect the best of society

00:15:44:07 – 00:15:45:24
and not the worst.

00:15:45:24 – 00:15:46:20
So with that being said,

00:15:46:20 – 00:15:48:05
we want to eliminate bias

00:15:48:05 – 00:15:49:10
and we want to eliminate

00:15:49:10 – 00:15:50:24
eliminate favoritism

00:15:50:24 – 00:15:52:19
with measurable results.

00:15:52:19 – 00:15:56:12
And also we want to be able to observe

00:15:56:12 – 00:15:59:21
how people are interacting with that.

00:15:59:21 – 00:16:01:13
We have statistics and evidence

00:16:01:13 – 00:16:03:01
related to the fairness,

00:16:03:01 – 00:16:04:10
and we need to use that

00:16:04:10 – 00:16:06:10
fairness evidence based.

00:16:07:15 – 00:16:10:15
Next slide.

00:16:14:17 – 00:16:15:13
Okay.

00:16:15:13 – 00:16:17:01
This one is a little bit interesting

00:16:17:01 – 00:16:18:09
because it’s a little bit

00:16:18:09 – 00:16:21:09
of a mixture of English and

00:16:21:09 – 00:16:23:09
written English and sign language.

00:16:23:09 – 00:16:26:09
So what we see here is the word A.I.

00:16:26:09 – 00:16:29:09
by A.I.. A.I.

00:16:29:18 – 00:16:30:20
Times, A.I.

00:16:30:20 – 00:16:32:21
on the screen in a mathematical look.

00:16:32:21 – 00:16:34:09
But we have a few different ways

00:16:34:09 – 00:16:37:10
that we are expressing this concept Now.

00:16:37:10 – 00:16:38:13
If we’re signing it,

00:16:38:13 – 00:16:40:13
you may see a sign A.I.

00:16:40:13 – 00:16:42:05
squared. Okay?

00:16:42:05 – 00:16:43:01
But when we say

00:16:43:01 – 00:16:46:01
that we’re referring to automated

00:16:46:01 – 00:16:46:19
or automatic

00:16:46:19 – 00:16:48:00
interpreting by

00:16:48:00 – 00:16:50:02
artificial intelligence, A.I.

00:16:50:02 – 00:16:51:05
by A.I..

00:16:51:05 – 00:16:51:23
Now, sometimes

00:16:51:23 – 00:16:55:13
you may see people sign a AI by A.I.,

00:16:55:24 – 00:16:57:05
and it’s the same concept.

00:16:57:05 – 00:16:58:21
It means the same thing.

00:16:58:21 – 00:17:00:00
Now, in written form.

00:17:00:00 – 00:17:03:00
We’ll see A.I., X, A.I..

00:17:03:01 – 00:17:04:11
Those are the three ways you’ll see it.

00:17:04:11 – 00:17:06:08
But the main point is understanding

00:17:06:08 – 00:17:07:03
that we’re referring

00:17:07:03 – 00:17:10:12
to automatic interpreting by artificial

00:17:11:09 – 00:17:12:04
intelligence.

00:17:15:02 – 00:17:16:07
Now, with that being said,

00:17:16:07 – 00:17:19:07
I’m going to turn this over to

00:17:20:13 – 00:17:21:18
I’m sorry, I don’t remember

00:17:21:18 – 00:17:22:23
who’s next on here.

00:17:22:23 – 00:17:25:23
Let me work

00:17:26:13 – 00:17:29:13
just.

00:17:31:12 – 00:17:31:23
Hello.

00:17:31:23 – 00:17:34:23
Okay, so

00:17:35:24 – 00:17:37:12
as Teresa mentioned,

00:17:37:12 – 00:17:39:23
we hosted three webinars.

00:17:39:23 – 00:17:41:12
And during those webinars,

00:17:41:12 – 00:17:42:06
we invited

00:17:42:06 – 00:17:43:09
so many of the deaf

00:17:43:09 – 00:17:46:09
community members to participate with us.

00:17:46:19 – 00:17:48:05
Those webinars,

00:17:48:05 – 00:17:50:22
we discussed a variety of topics

00:17:50:22 – 00:17:51:23
and issues

00:17:51:23 – 00:17:54:11
covering AI and interpreting

00:17:54:11 – 00:17:55:18
and how they relate.

00:17:55:18 – 00:17:57:14
As Teresa mentioned, A.I.

00:17:57:14 – 00:17:59:21
Squared is how we were referring to A.I.

00:17:59:21 – 00:18:00:21
by A.I..

00:18:00:21 – 00:18:02:22
There was so much

00:18:02:22 – 00:18:05:09
great dialog conversation,

00:18:05:09 – 00:18:07:07
ideas and issues

00:18:07:07 – 00:18:09:18
discussed during these webinars.

00:18:09:18 – 00:18:12:16
Now we recorded these webinars

00:18:12:16 – 00:18:15:16
and we made a transcript of them

00:18:15:22 – 00:18:17:18
and we went through these transcripts

00:18:17:18 – 00:18:19:01
with a fine tooth comb

00:18:19:01 – 00:18:20:07
and we looked at them

00:18:20:07 – 00:18:21:12
and looked at patterns

00:18:21:12 – 00:18:23:04
that arose in each

00:18:23:04 – 00:18:24:04
discussion, different

00:18:24:04 – 00:18:25:11
themes that popped up

00:18:25:11 – 00:18:26:21
throughout the entirety

00:18:26:21 – 00:18:28:02
of these webinars.

00:18:28:02 – 00:18:29:11
And we wanted to make sure

00:18:29:11 – 00:18:31:12
that we were able to pinpoint

00:18:31:12 – 00:18:32:21
and really understand

00:18:32:21 – 00:18:34:21
what issues are most prevalent

00:18:34:21 – 00:18:36:10
to the community at large.

00:18:38:02 – 00:18:39:17
We did

00:18:39:17 – 00:18:42:08
a time analysis as well

00:18:42:08 – 00:18:44:22
and we looked at how frequently

00:18:44:22 – 00:18:46:14
different themes

00:18:46:14 – 00:18:48:18
popped up in said webinars

00:18:48:18 – 00:18:50:13
and we found

00:18:50:13 – 00:18:52:07
our findings

00:18:52:07 – 00:18:53:06
are shown in the next slide.

00:18:53:06 – 00:18:54:01
And if you’re interested

00:18:54:01 – 00:18:55:09
in more detailed information,

00:18:55:09 – 00:18:56:19
please read our report.

00:18:56:19 – 00:18:59:22
It goes into a varied in depth

00:19:01:05 – 00:19:03:02
reporting of our findings

00:19:03:02 – 00:19:05:00
and that is available online as well.

00:19:05:00 – 00:19:08:00
Next slide, please.

00:19:10:20 – 00:19:14:12
So this is a snapshot of our findings.

00:19:14:18 – 00:19:16:01
We put down

00:19:16:01 – 00:19:16:19
all the themes

00:19:16:19 – 00:19:17:12
that popped up

00:19:17:12 – 00:19:20:08
most frequently in our webinars,

00:19:20:08 – 00:19:21:22
and we categorized those

00:19:21:22 – 00:19:25:22
into three different areas of study.

00:19:27:00 – 00:19:30:08
The first area is related

00:19:30:08 – 00:19:33:08
to results and outcomes,

00:19:34:01 – 00:19:35:03
and we’ll talk about this

00:19:35:03 – 00:19:36:09
more in the next slide.

00:19:36:09 – 00:19:39:09
We’re focused more on discussing about

00:19:39:20 – 00:19:42:19
what kind of society,

00:19:42:19 – 00:19:44:24
what kind of societal results

00:19:44:24 – 00:19:48:20
will arise from the impact of AI by AI.

00:19:50:08 – 00:19:53:04
Our next theme was readiness.

00:19:53:04 – 00:19:56:04
How ready is the community at large?

00:19:56:21 – 00:20:00:05
How ready are stakeholders for

00:20:01:22 – 00:20:03:14
this technology?

00:20:03:14 – 00:20:04:19
We looked at the feedback

00:20:04:19 – 00:20:06:03
and the requirements and things

00:20:06:03 – 00:20:07:04
that need to happen

00:20:07:04 – 00:20:08:15
for the community at large

00:20:08:15 – 00:20:10:19
in this technological realm.

00:20:10:19 – 00:20:13:05
But as you can see on here,

00:20:13:05 – 00:20:16:05
by over half and half,

00:20:16:08 – 00:20:17:09
the biggest issue

00:20:17:09 – 00:20:19:01
that came up most frequently

00:20:19:01 – 00:20:20:08
and most prevalent on people’s

00:20:20:08 – 00:20:22:17
minds was technological quality.

00:20:22:17 – 00:20:24:08
What type of regulations

00:20:24:08 – 00:20:27:19
and standards are required to minimize

00:20:28:03 – 00:20:31:03
the reduce or the possibility of harm?

00:20:31:15 – 00:20:33:08
And how do we maximize

00:20:33:08 – 00:20:35:01
the potential benefits

00:20:35:01 – 00:20:36:17
of this technology?

00:20:36:17 – 00:20:37:22
Now, for this,

00:20:37:22 – 00:20:39:14
we went into more in-depth

00:20:39:14 – 00:20:42:03
research in our report,

00:20:42:03 – 00:20:43:24
and I’m going to turn it

00:20:43:24 – 00:20:44:20
over to Ann Marie.

00:20:44:20 – 00:20:45:12
If she can.

00:20:45:12 – 00:20:46:10
She can discuss this

00:20:46:10 – 00:20:47:08
a little bit more in depth.

00:20:49:05 – 00:20:50:08
Thank you so much, Jeff.

00:20:50:08 – 00:20:53:08
I really appreciate it.

00:20:53:08 – 00:20:54:06
The first thing that we’re going

00:20:54:06 – 00:20:55:07
to look at, again,

00:20:55:07 – 00:20:59:03
as Jeff mentioned in his synopsis

00:20:59:03 – 00:21:01:13
of the percentages that we looked at,

00:21:01:13 – 00:21:03:06
it is staggering.

00:21:03:06 – 00:21:06:06
The biggest section was on technological

00:21:06:20 – 00:21:09:20
aspects, but

00:21:09:20 – 00:21:13:02
I’m looking at the desired results

00:21:13:02 – 00:21:14:15
and outcomes as well,

00:21:14:15 – 00:21:16:07
and it was a very hot topic

00:21:16:07 – 00:21:18:18
in the discussions in the webinar.

00:21:18:18 – 00:21:19:20
It was really important

00:21:19:20 – 00:21:22:20
to emphasize that as well.

00:21:24:05 – 00:21:25:08
Now, more than one

00:21:25:08 – 00:21:26:17
third of the discussions

00:21:26:17 – 00:21:29:17
regarding this were really focused on

00:21:30:02 – 00:21:33:02
these critical areas of control,

00:21:34:16 – 00:21:36:19
control

00:21:36:19 – 00:21:39:16
by the language community

00:21:39:16 – 00:21:40:18
and the authority

00:21:40:18 – 00:21:42:24
who is involved in the decision making,

00:21:42:24 – 00:21:44:15
who is involved in the device

00:21:44:15 – 00:21:45:23
development of design

00:21:45:23 – 00:21:48:02
and the impact the A.I. is going to have.

00:21:49:16 – 00:21:50:24
The biggest concern

00:21:50:24 – 00:21:53:24
was really regarding the legality,

00:21:54:01 – 00:21:58:20
the structure and the synthesis of this.

00:21:59:03 – 00:22:00:20
And it was critical.

00:22:00:20 – 00:22:02:00
It was impressive

00:22:02:00 – 00:22:05:00
to see how much the deaf leaders

00:22:05:01 – 00:22:07:01
felt that,

00:22:07:01 – 00:22:08:03
you know, in history

00:22:08:03 – 00:22:09:20
in their based on their experience,

00:22:09:20 – 00:22:11:18
the deaf have not always

00:22:11:18 – 00:22:13:17
had a voice at the table,

00:22:13:17 – 00:22:15:05
whether it was in the development

00:22:15:05 – 00:22:17:15
of different views or different

00:22:17:15 – 00:22:19:01
processes.

00:22:19:01 – 00:22:20:01
The deaf perspective

00:22:20:01 – 00:22:20:15
has often

00:22:20:15 – 00:22:21:07
been missing

00:22:21:07 – 00:22:22:18
and it’s really critical

00:22:22:18 – 00:22:23:18
that they are included

00:22:23:18 – 00:22:24:17
in the conversation.

00:22:24:17 – 00:22:26:03
And what does that look like?

00:22:26:03 – 00:22:27:19
What does the deaf representation

00:22:27:19 – 00:22:28:06
look like

00:22:28:06 – 00:22:29:15
in the authoritative process

00:22:29:15 – 00:22:32:15
of developing these standards?

00:22:32:21 – 00:22:34:17
Also,

00:22:34:17 – 00:22:38:07
it is very important and very stringent

00:22:38:07 – 00:22:42:08
that any violation of these processes,

00:22:44:14 – 00:22:45:09
you know, if

00:22:45:09 – 00:22:47:01
these processes are violated,

00:22:47:01 – 00:22:48:20
what type of ramifications

00:22:48:20 – 00:22:49:15
does that have?

00:22:49:15 – 00:22:51:11
That’s been a very big discussion

00:22:51:11 – 00:22:53:07
topic as well. Next slide.

00:23:03:20 – 00:23:05:02
Now you see this next slide.

00:23:05:02 – 00:23:06:06
We’ve also included

00:23:06:06 – 00:23:09:09
some concerns based on our webinars

00:23:10:01 – 00:23:11:21
about control.

00:23:11:21 – 00:23:13:16
There are two levels of control,

00:23:13:16 – 00:23:17:06
individual control and cultural groups,

00:23:17:12 – 00:23:18:17
their control

00:23:18:17 – 00:23:19:20
and the impact

00:23:19:20 – 00:23:21:04
and the protections

00:23:21:04 – 00:23:23:16
needed for each of these groups.

00:23:23:16 – 00:23:25:02
We need to be sensitive to those,

00:23:25:02 – 00:23:28:02
especially with children.

00:23:29:21 – 00:23:31:19
The focus of the community at large.

00:23:31:19 – 00:23:33:16
The biggest priority here

00:23:33:16 – 00:23:37:08
was to make sure that the legal framework

00:23:39:01 – 00:23:41:00
is established with ABI.

00:23:41:00 – 00:23:44:00
I again, you know, it’s

00:23:44:12 – 00:23:48:11
one we’ve had to look at leaders involved

00:23:48:16 – 00:23:50:00
with the deaf community

00:23:50:00 – 00:23:51:17
involvement in research

00:23:51:17 – 00:23:54:17
and in development for these

00:23:55:02 – 00:23:55:20
technologies.

00:23:55:20 – 00:23:57:05
It’s very imperative

00:23:57:05 – 00:23:58:20
because it impacts their life

00:23:58:20 – 00:23:59:21
for the deaf community

00:23:59:21 – 00:24:00:10
at large,

00:24:00:10 – 00:24:00:23
for the deaf

00:24:00:23 – 00:24:01:20
blind community

00:24:01:20 – 00:24:02:18
as well,

00:24:02:18 – 00:24:04:14
for those that are losing their hearing

00:24:04:14 – 00:24:05:21
or hard of hearing

00:24:05:21 – 00:24:07:10
and learning sign,

00:24:07:10 – 00:24:09:14
this impacts all of them.

00:24:09:14 – 00:24:11:13
Again, it really can’t emphasize

00:24:11:13 – 00:24:12:17
this enough.

00:24:12:17 – 00:24:13:24
You know, the deaf

00:24:13:24 – 00:24:15:18
leadership needs to be involved

00:24:15:18 – 00:24:16:14
in the developmental

00:24:16:14 – 00:24:18:23
process of these platforms.

00:24:20:09 – 00:24:23:09
For I by

00:24:25:04 – 00:24:26:10
also it’s really interesting

00:24:26:10 – 00:24:27:04
to note

00:24:27:04 – 00:24:30:04
the importance of the concern that

00:24:30:24 – 00:24:33:24
I, I the data.

00:24:33:24 – 00:24:36:07
What is that data stored storage

00:24:36:07 – 00:24:37:01
look like?

00:24:37:01 – 00:24:39:10
Who is in conservatorship

00:24:39:10 – 00:24:40:20
of this storage?

00:24:40:20 – 00:24:41:21
What are the analytics

00:24:41:21 – 00:24:44:00
and the statistics looking at?

00:24:44:00 – 00:24:45:15
Where is that data held

00:24:45:15 – 00:24:47:00
and how is it protected?

00:24:47:00 – 00:24:49:10
That’s a very big discussion as well.

00:24:49:10 – 00:24:51:15
There’s a lot of concern about,

00:24:51:15 – 00:24:54:24
you know, hiring individuals

00:24:54:24 – 00:24:57:24
that could influence the development and

00:25:00:20 – 00:25:02:09
leadership of this.

00:25:02:09 – 00:25:03:13
That doesn’t always happen

00:25:03:13 – 00:25:05:00
to have deaf individuals

00:25:05:00 – 00:25:07:04
in this type of role

00:25:07:04 – 00:25:09:08
and involved in that process.

00:25:09:08 – 00:25:11:21
So will this process in the future

00:25:11:21 – 00:25:12:17
involve deaf

00:25:12:17 – 00:25:14:11
individuals in the beginning,

00:25:14:11 – 00:25:16:17
a foundational development of this,

00:25:16:17 – 00:25:18:16
or will they only bring them in for A

00:25:18:16 – 00:25:20:20
to perspective and that’s it.

00:25:20:20 – 00:25:22:00
So I think another good point

00:25:22:00 – 00:25:22:19
of discussion

00:25:22:19 – 00:25:23:08
as well,

00:25:23:08 – 00:25:24:17
who have the ownership and the

00:25:24:17 – 00:25:26:14
influence in this process.

00:25:29:22 – 00:25:31:20
You know, not only

00:25:31:20 – 00:25:33:19
hiring deaf consultants,

00:25:33:19 – 00:25:35:09
but all through the line,

00:25:35:09 – 00:25:36:01
through A.I.,

00:25:36:01 – 00:25:38:04
by having deaf

00:25:38:04 – 00:25:40:16
hands on the process is vital.

00:25:40:16 – 00:25:42:08
And I’m going to turn this over

00:25:42:08 – 00:25:45:08
to Jeff again.

00:25:49:08 – 00:25:51:20
Hi, it’s Jeff here again for you.

00:25:51:20 – 00:25:53:12
So

00:25:53:12 – 00:25:55:22
about the technological quality.

00:25:55:22 – 00:25:56:16
During the webinar,

00:25:56:16 – 00:25:58:19
we discussed this quite in depth

00:25:58:19 – 00:26:00:09
and how we can break down

00:26:00:09 – 00:26:03:10
this idea into several subcategories.

00:26:04:16 – 00:26:07:03
The first or the major subcategory

00:26:07:03 – 00:26:08:10
that a lot of people really

00:26:08:10 – 00:26:11:10
discussed in the webinar was data.

00:26:11:14 – 00:26:14:21
Many people agreed that the quality

00:26:15:05 – 00:26:16:19
and the diversity of the data

00:26:16:19 – 00:26:19:10
is key to the foundation of building A.I.

00:26:19:10 – 00:26:21:07
by A.I..

00:26:21:07 – 00:26:22:08
You know, with machine

00:26:22:08 – 00:26:25:08
learning in the community,

00:26:26:15 – 00:26:28:22
it’s garbage in, garbage out

00:26:28:22 – 00:26:30:06
more often than not.

00:26:30:06 – 00:26:32:12
And so what does that look like?

00:26:32:12 – 00:26:34:13
So if we have a model training

00:26:35:14 – 00:26:36:08
by that

00:26:36:08 – 00:26:38:03
with that data,

00:26:38:03 – 00:26:40:19
it will be the same in the output.

00:26:40:19 – 00:26:42:11
So

00:26:42:11 – 00:26:44:08
that leads us to our second

00:26:44:08 – 00:26:47:08
largest issue modeling.

00:26:47:23 – 00:26:50:09
So if we have the garbage in modeling,

00:26:50:09 – 00:26:51:07
it’s garbage out.

00:26:51:07 – 00:26:52:09
But people are looking at it

00:26:52:09 – 00:26:53:09
and saying, okay,

00:26:53:09 – 00:26:55:15
so what is the primary use?

00:26:55:15 – 00:26:57:11
What is the primary model?

00:26:57:11 – 00:26:59:07
What is what features are they using?

00:26:59:07 – 00:27:02:03
Are they focusing only on hand shape

00:27:02:03 – 00:27:02:18
or will

00:27:02:18 – 00:27:04:18
they also be including facial shape

00:27:04:18 – 00:27:05:08
and mouth

00:27:05:08 – 00:27:08:08
shapes, expressions and other key

00:27:08:15 – 00:27:11:15
contextual clues of the language?

00:27:12:09 – 00:27:14:22
How will we be able to evaluate

00:27:14:22 – 00:27:15:24
that model?

00:27:15:24 – 00:27:17:07
Can we decide

00:27:17:07 – 00:27:20:09
which model to use over another model,

00:27:20:21 – 00:27:23:21
and how do we use metrics

00:27:24:01 – 00:27:24:22
to decide

00:27:24:22 – 00:27:26:06
if that is the quality

00:27:26:06 – 00:27:27:20
that we need or not?

00:27:27:20 – 00:27:29:06
Many people really had

00:27:29:06 – 00:27:30:06
that on their minds

00:27:30:06 – 00:27:32:00
in the webinar discussions.

00:27:32:00 – 00:27:33:24
Now, throughout these discussions,

00:27:33:24 – 00:27:35:07
we also noticed the need

00:27:35:07 – 00:27:37:07
for deaf leadership

00:27:37:07 – 00:27:40:02
involvement and oversight as well.

00:27:40:02 – 00:27:44:01
It was key that that topic came up

00:27:44:01 – 00:27:44:24
many times.

00:27:44:24 – 00:27:46:13
We needed to set up

00:27:46:13 – 00:27:50:19
at least a minimum criteria for

00:27:52:00 – 00:27:54:15
the bidirectional interpreting for A.I.

00:27:54:15 – 00:27:55:13
by A.I..

00:27:55:13 – 00:27:58:13
Next slide, please.

00:28:01:16 – 00:28:04:18
Now many more subcategories

00:28:04:18 – 00:28:06:05
arose from this discussion,

00:28:06:05 – 00:28:08:02
and one of them was quality,

00:28:08:02 – 00:28:10:02
and it was very imperative

00:28:10:02 – 00:28:13:02
This topic popped up quite a bit

00:28:13:24 – 00:28:16:24
as well.

00:28:17:01 – 00:28:18:24
The participants all agreed

00:28:18:24 – 00:28:23:03
that during the process of gathering

00:28:23:13 – 00:28:26:18
data, storing data, using data,

00:28:27:05 – 00:28:29:24
sharing data, publishing

00:28:29:24 – 00:28:30:17
all of that,

00:28:30:17 – 00:28:31:18
and that process

00:28:31:18 – 00:28:33:19
needed to be quite transparent

00:28:33:19 – 00:28:34:23
so that we could understand

00:28:34:23 – 00:28:37:12
everything as it happened.

00:28:37:12 – 00:28:39:14
We need the opt in or opt out

00:28:39:14 – 00:28:41:07
option as well.

00:28:41:07 – 00:28:43:03
Also, discussing the ability

00:28:43:03 – 00:28:44:12
to withdrawal.

00:28:44:12 – 00:28:46:14
Our can fit in the future,

00:28:46:14 – 00:28:48:12
meaning if we opt into this

00:28:48:12 – 00:28:49:06
and then later

00:28:49:06 – 00:28:50:09
we change our minds and say,

00:28:50:09 – 00:28:50:18
you know what,

00:28:50:18 – 00:28:52:10
we don’t want to participate.

00:28:52:10 – 00:28:54:23
We have the option to withdraw

00:28:54:23 – 00:28:55:18
that consent.

00:28:57:10 – 00:28:59:04
Another topic that we discussed

00:28:59:04 – 00:28:59:22
quite in death

00:28:59:22 – 00:29:03:01
was the topic of safeguards.

00:29:03:23 – 00:29:05:12
Those are needed to be put in place

00:29:05:12 – 00:29:07:10
to minimize harm.

00:29:07:10 – 00:29:10:10
Many of the what ifs

00:29:10:13 – 00:29:12:18
that arose during our discussion

00:29:12:18 – 00:29:14:03
really related to this.

00:29:14:03 – 00:29:16:12
What if someone downloads this?

00:29:16:12 – 00:29:20:15
What if there is a breach of data?

00:29:20:21 – 00:29:23:14
What if the information is leaked?

00:29:23:14 – 00:29:25:19
What if the system crashes?

00:29:25:19 – 00:29:26:15
There were so many

00:29:26:15 – 00:29:28:09
what ifs and questions

00:29:28:09 – 00:29:30:09
that popped up in that discussion.

00:29:30:09 – 00:29:31:00
And again,

00:29:31:00 – 00:29:33:00
for more analysis of this, please

00:29:33:00 – 00:29:33:23
look at our report.

00:29:33:23 – 00:29:36:17
We have a lot of it detailed in there.

00:29:36:17 – 00:29:39:17
And we also talked about readiness.

00:29:40:10 – 00:29:42:08
And Theresa is going to talk about

00:29:42:08 – 00:29:44:19
that more in depth now.

00:29:44:19 – 00:29:47:19
I’ll hand it over to Theresa.

00:29:51:10 – 00:29:53:20
Okay.

00:29:53:20 – 00:29:56:20
So

00:29:57:08 – 00:29:59:20
in terms of readiness,

00:29:59:20 – 00:30:02:05
when we look at what that includes,

00:30:02:05 – 00:30:04:06
we have readiness

00:30:04:06 – 00:30:05:16
when it comes to technology,

00:30:05:16 – 00:30:09:03
but also there’s a discussion of society

00:30:09:10 – 00:30:11:12
and what our awareness

00:30:11:12 – 00:30:12:01
looks like

00:30:12:01 – 00:30:13:19
in terms of ethical issues

00:30:13:19 – 00:30:15:22
and ethical concerns.

00:30:15:22 – 00:30:18:21
So do we have

00:30:18:21 – 00:30:21:05
technological availability?

00:30:21:05 – 00:30:23:00
Do we have representation

00:30:23:00 – 00:30:25:16
and understanding and creation?

00:30:25:16 – 00:30:28:09
That’s one part of readiness.

00:30:30:12 – 00:30:33:12
Next slide.

00:30:37:24 – 00:30:39:22
So now we’ll see that there’s

00:30:39:22 – 00:30:41:12
there’s various components here.

00:30:41:12 – 00:30:42:11
So you can see here

00:30:42:11 – 00:30:43:19
where it talks

00:30:43:19 – 00:30:44:12
about readiness

00:30:44:12 – 00:30:45:13
when it comes to sign

00:30:45:13 – 00:30:47:12
language recognition.

00:30:47:12 – 00:30:50:24
And we have readiness of American

00:30:50:24 – 00:30:53:01
the American Deaf community.

00:30:53:01 – 00:30:55:07
That’s another component.

00:30:55:07 – 00:30:58:09
Does the deaf community have an in-depth

00:30:58:09 – 00:31:01:23
understanding of the having good quality,

00:31:02:06 – 00:31:05:00
ethical and technological aspects

00:31:05:00 – 00:31:06:17
when it comes to A.I.

00:31:06:17 – 00:31:08:04
by A.I.?

00:31:08:04 – 00:31:11:20
And the last sign of readiness would be

00:31:11:20 – 00:31:12:14
when it comes

00:31:12:14 – 00:31:15:22
to public and private entities,

00:31:16:09 – 00:31:20:08
Are we ready for this kind of technology

00:31:20:15 – 00:31:22:04
and what the responsibilities

00:31:22:04 – 00:31:24:02
be that come along with that

00:31:24:02 – 00:31:27:02
and the accountability?

00:31:28:04 – 00:31:31:04
Next slide, please.

00:31:34:15 – 00:31:36:06
So as you can see here,

00:31:36:06 – 00:31:37:05
one of the components

00:31:37:05 – 00:31:38:24
that we talked about are civil

00:31:38:24 – 00:31:40:13
rights and civil protections.

00:31:41:12 – 00:31:43:04
So what we have

00:31:43:04 – 00:31:45:23
quality control and certification

00:31:45:23 – 00:31:47:13
when it comes to the interpreters

00:31:47:13 – 00:31:49:01
and certification for the A.I.

00:31:49:01 – 00:31:50:17
technology itself.

00:31:50:17 – 00:31:51:06
Also,

00:31:51:06 – 00:31:53:03
we have to consider the impact

00:31:53:03 – 00:31:55:09
that this would have on the culture,

00:31:55:09 – 00:31:56:18
the current culture

00:31:56:18 – 00:31:58:08
and the culture of the future.

00:31:58:08 – 00:31:59:20
And what

00:31:59:20 – 00:32:00:19
what does

00:32:00:19 – 00:32:03:11
state of the art technology mean?

00:32:03:11 – 00:32:06:05
How would we make sure that we are up

00:32:06:05 – 00:32:07:12
to date with technology

00:32:07:12 – 00:32:09:04
as it changes over time

00:32:09:04 – 00:32:10:02
and ensure

00:32:10:02 – 00:32:11:12
that we have the appropriate

00:32:11:12 – 00:32:13:03
response to those changes?

00:32:13:03 – 00:32:16:03
Next slide.

00:32:26:00 – 00:32:27:08
So I believe that, well,

00:32:27:08 – 00:32:28:22
I can go ahead and do this part.

00:32:28:22 – 00:32:33:06
And so in just a few hours.

00:32:33:08 – 00:32:36:08
Next slide, please.

00:32:40:04 – 00:32:40:21
Okay, great.

00:32:40:21 – 00:32:42:01
Sorry about that.

00:32:42:01 – 00:32:45:22
And okay, So within a few hours,

00:32:46:05 – 00:32:48:08
we had these internal discussions

00:32:48:08 – 00:32:50:00
at the three webinars,

00:32:50:00 – 00:32:51:11
and within those hours

00:32:51:11 – 00:32:52:23
we had a lot of ethical issues

00:32:52:23 – 00:32:54:10
that came up to be discussed

00:32:54:10 – 00:32:56:06
and we touched on ethics,

00:32:56:06 – 00:32:58:10
we touched on fairness and safety.

00:32:58:10 – 00:32:59:09
And I believe

00:32:59:09 – 00:33:00:23
all of these things were covered.

00:33:00:23 – 00:33:02:20
Also, the deaf participants

00:33:02:20 – 00:33:04:21
were able to explain

00:33:04:21 – 00:33:07:08
how we can use our principles

00:33:07:08 – 00:33:09:03
and how we can use our models

00:33:09:03 – 00:33:10:15
in a way that’s ethical

00:33:10:15 – 00:33:11:19
in order to prevent

00:33:11:19 – 00:33:13:00
any ongoing

00:33:13:00 – 00:33:14:00
harm

00:33:14:00 – 00:33:15:09
or anything that could come

00:33:15:09 – 00:33:16:24
against the deaf community.

00:33:16:24 – 00:33:18:00
So we talked about that

00:33:18:00 – 00:33:19:04
fairness and equity

00:33:19:04 – 00:33:20:21
in these conversations,

00:33:20:21 – 00:33:23:09
how it pertains to society.

00:33:23:09 – 00:33:24:00
Next slide.

00:33:29:06 – 00:33:31:18
So we do have a few suggestions

00:33:31:18 – 00:33:35:02
and we suggest that there be a

00:33:35:08 – 00:33:36:23
the appropriate level of protection

00:33:36:23 – 00:33:38:19
and privacy and confidentiality

00:33:38:19 – 00:33:39:20
when it comes to A.I.

00:33:39:20 – 00:33:41:07
backed by A.I.,

00:33:41:07 – 00:33:42:11
and that will allow us

00:33:42:11 – 00:33:43:22
to have better protections

00:33:43:22 – 00:33:44:22
across the Internet

00:33:44:22 – 00:33:47:09
for all kinds of applications.

00:33:47:09 – 00:33:51:05
Also, another suggestion was that we have

00:33:52:15 – 00:33:55:19
that these concerns regarding risk

00:33:55:23 – 00:33:58:21
be established and considered in order

00:33:58:21 – 00:33:59:13
to make sure

00:33:59:13 – 00:34:01:20
that we have strict regulations in place

00:34:01:20 – 00:34:03:21
to avoid any adverse

00:34:03:21 – 00:34:05:10
downstream consequences

00:34:05:10 – 00:34:06:18
for governance, business

00:34:06:18 – 00:34:09:18
and social infrastructures.

00:34:10:13 – 00:34:13:13
Next slide.

00:34:14:15 – 00:34:15:08
Okay.

00:34:15:08 – 00:34:17:06
So this word stress

00:34:17:06 – 00:34:20:06
or socio technical systems

00:34:21:01 – 00:34:23:12
and again asks for short

00:34:23:12 – 00:34:26:07
means that we have to recognize

00:34:26:07 – 00:34:28:07
that we do have technology

00:34:28:07 – 00:34:31:12
and we also have our the socio aspect

00:34:31:20 – 00:34:34:01
when these two things come together,

00:34:34:01 – 00:34:34:19
they’re going to have

00:34:34:19 – 00:34:36:15
an influence on each other.

00:34:36:15 – 00:34:38:03
So we need to always

00:34:38:03 – 00:34:38:20
be sure

00:34:38:20 – 00:34:41:18
to look at the system of technology

00:34:41:18 – 00:34:44:04
while considering its impact on society

00:34:44:04 – 00:34:45:14
and vice versa.

00:34:45:14 – 00:34:47:03
We have to look at how both of these

00:34:47:03 – 00:34:48:20
things interact. Next slide.

00:34:54:12 – 00:34:55:17
So I’ll give you a moment

00:34:55:17 – 00:34:58:17
to read through the slide.

00:35:06:07 – 00:35:07:06
So it’s important.

00:35:07:06 – 00:35:08:09
Here to take note of

00:35:08:09 – 00:35:09:20
is that Estes

00:35:09:20 – 00:35:13:18
refers to how things are correlated, how

00:35:13:18 – 00:35:15:11
these co influences

00:35:15:11 – 00:35:18:01
happen in both society and technology

00:35:18:01 – 00:35:21:01
as it pertains to the organization.

00:35:21:08 – 00:35:23:03
So it’s so important

00:35:23:03 – 00:35:25:04
that during the design process

00:35:25:04 – 00:35:28:03
we consider both of these areas

00:35:28:03 – 00:35:29:19
and look at our results

00:35:29:19 – 00:35:34:04
to ensure that air by air is optimizing

00:35:34:09 – 00:35:37:09
both of these two subsystems.

00:35:41:18 – 00:35:44:19
So the key here is to attend to

00:35:44:19 – 00:35:47:21
how the social behaviors of humans

00:35:47:21 – 00:35:50:00
are going to combine and influence

00:35:50:00 – 00:35:52:16
by the structures of technology

00:35:52:16 – 00:35:55:16
and vice versa.

00:35:56:06 – 00:35:59:06
So these are not two separate things.

00:35:59:15 – 00:36:02:10
We do have a few recommendations

00:36:02:10 – 00:36:04:02
that we would like to share.

00:36:04:02 – 00:36:05:19
The first one is that

00:36:05:19 – 00:36:09:10
it needs to be understood that AI by AI

00:36:09:10 – 00:36:12:11
is a socio tech and technological system

00:36:13:10 – 00:36:15:11
and also we need to

00:36:15:11 – 00:36:16:07
make sure

00:36:16:07 – 00:36:19:10
that the deaf wisdom from the community

00:36:19:15 – 00:36:21:02
is a vital part of this.

00:36:21:02 – 00:36:22:10
We want to use our wisdom

00:36:22:10 – 00:36:24:03
and our experience

00:36:24:03 – 00:36:26:09
from interpreting experiences

00:36:26:09 – 00:36:29:19
to our experience with VR s VR AI,

00:36:29:24 – 00:36:31:01
and also

00:36:31:01 – 00:36:33:09
how we as a community

00:36:33:09 – 00:36:35:08
experience technology.

00:36:35:08 – 00:36:38:21
And we want to be able to impart

00:36:38:24 – 00:36:40:00
our experience

00:36:40:00 – 00:36:43:01
and our wisdom as the framework is being

00:36:43:01 – 00:36:46:01
developed in this realm.

00:36:46:01 – 00:36:50:00
Thirdly, we want to continue to engage

00:36:50:00 – 00:36:51:14
with the deaf community

00:36:51:14 – 00:36:52:21
and build our knowledge

00:36:52:21 – 00:36:55:17
and our awareness through funding.

00:36:55:17 – 00:36:58:17
One example of that funding are grants,

00:36:58:20 – 00:37:00:07
and so an example

00:37:00:07 – 00:37:02:15
is the Civic Innovation grant,

00:37:02:15 – 00:37:04:05
where we can receive money

00:37:04:05 – 00:37:04:21
for this

00:37:04:21 – 00:37:06:02
and continue

00:37:06:02 – 00:37:07:20
to plug in to the deaf community

00:37:07:20 – 00:37:09:04
and get their input

00:37:09:04 – 00:37:10:09
and find what the deaf

00:37:10:09 – 00:37:11:15
community would like to see

00:37:11:15 – 00:37:12:14
and what they feel needs

00:37:12:14 – 00:37:14:04
to be brought to attention.

00:37:14:04 – 00:37:15:08
And as these laws and

00:37:15:08 – 00:37:16:13
policies are developed.

00:37:18:11 – 00:37:21:11
Next slide, please.

00:37:24:21 – 00:37:25:08
And as we

00:37:25:08 – 00:37:26:09
get close to wrapping up,

00:37:26:09 – 00:37:28:13
I’d like to make the point that I by

00:37:28:13 – 00:37:30:04
I will probably never be able

00:37:30:04 – 00:37:31:22
to totally replace

00:37:31:22 – 00:37:32:16
humans

00:37:32:16 – 00:37:35:16
because of the interactivity issues.

00:37:35:20 – 00:37:37:01
Often we can predict

00:37:37:01 – 00:37:39:00
there would be misunderstandings

00:37:39:00 – 00:37:42:00
and so we need to have that deep

00:37:42:05 – 00:37:44:03
human knowledge there.

00:37:44:03 – 00:37:47:03
So we’ll have to look for

00:37:47:15 – 00:37:50:00
a human in the loop design

00:37:50:00 – 00:37:51:06
as this technology

00:37:51:06 – 00:37:53:18
is developed over time.

00:37:53:18 – 00:37:57:03
And in terms of human in the loop design

00:37:57:05 – 00:38:00:05
to explain a bit about about that,

00:38:00:12 – 00:38:01:22
of course we would have the A.I.

00:38:01:22 – 00:38:05:03
technology and often

00:38:05:03 – 00:38:08:03
we see that I, I get information

00:38:08:08 – 00:38:10:05
through different sources,

00:38:10:05 – 00:38:12:14
but we want to be able to emphasize

00:38:12:14 – 00:38:14:12
the importance of A.I.

00:38:14:12 – 00:38:16:00
getting feedback

00:38:16:00 – 00:38:19:23
from human interactivity.

00:38:20:12 – 00:38:22:08
There needs to be a feedback loop

00:38:22:08 – 00:38:23:07
food feedback loop

00:38:23:07 – 00:38:24:05
where humans are always

00:38:24:05 – 00:38:27:12
involved in this process so that

00:38:28:05 – 00:38:30:05
the design and the development

00:38:30:05 – 00:38:31:24
and the operation

00:38:31:24 – 00:38:33:22
and everything is verified

00:38:33:22 – 00:38:34:21
by humans

00:38:34:21 – 00:38:37:02
who are involved with the project.

00:38:37:02 – 00:38:38:17
That would be the end goal.

00:38:38:17 – 00:38:39:18
Next slide, please.

00:38:45:23 – 00:38:46:17
Hello everyone.

00:38:46:17 – 00:38:50:03
I’m back so I would like to

00:38:50:05 – 00:38:53:22
we have the advisory group on AI

00:38:53:22 – 00:38:55:02
and final language interpreting

00:38:55:02 – 00:38:56:00
and this has been

00:38:56:00 – 00:38:57:24
there’s been so many

00:38:57:24 – 00:38:59:06
benefits to this group,

00:38:59:06 – 00:39:00:15
so much AB many hours

00:39:00:15 – 00:39:02:18
and work has gone into this.

00:39:02:18 – 00:39:05:15
And so this isn’t the end of our work.

00:39:05:15 – 00:39:07:00
This is just the springboard

00:39:07:00 – 00:39:08:15
to the future for more discussions

00:39:08:15 – 00:39:09:11
and more partnerships

00:39:09:11 – 00:39:11:00
with the community at large.

00:39:11:00 – 00:39:13:05
So we have an event to share with you.

00:39:13:05 – 00:39:14:21
A Save the Date.

00:39:14:21 – 00:39:17:21
We have a symposium

00:39:18:02 – 00:39:21:02
on AI and sign language interpreting

00:39:21:14 – 00:39:23:10
and it will be hosted on April

00:39:23:10 – 00:39:27:08
20th and 21st is a Saturday and Sunday

00:39:27:08 – 00:39:28:11
this year

00:39:28:11 – 00:39:31:11
and will be here at Brown University.

00:39:32:21 – 00:39:35:04
And we also will have

00:39:35:04 – 00:39:37:13
accessibility options to join us

00:39:37:13 – 00:39:38:14
through Zoom.

00:39:38:14 – 00:39:40:16
Anyone can join from anywhere

00:39:40:16 – 00:39:43:09
we are planning to on Saturday.

00:39:43:09 – 00:39:46:16
It will be from 9 to 6 and Sunday 9

00:39:46:16 – 00:39:49:16
to 2, just a half day on Sunday.

00:39:49:22 – 00:39:52:00
And this is Eastern Standard Time.

00:39:52:24 – 00:39:53:21
This is very

00:39:53:21 – 00:39:55:02
important for us

00:39:55:02 – 00:39:57:20
to bring in different perspectives,

00:39:57:20 – 00:40:00:08
different experts and different people

00:40:00:08 – 00:40:02:16
who have in-depth experience in A.I.

00:40:02:16 – 00:40:04:10
and sign language interpreting

00:40:04:10 – 00:40:05:07
so we can really have

00:40:05:07 – 00:40:06:15
an in-depth discussion

00:40:06:15 – 00:40:08:02
on what this looks like.

00:40:08:02 – 00:40:09:15
And it’s also an opportunity

00:40:09:15 – 00:40:11:12
for us to do a deeper dive

00:40:11:12 – 00:40:12:21
into what our presenters

00:40:12:21 – 00:40:13:13
have really talked

00:40:13:13 – 00:40:14:21
about during this session

00:40:14:21 – 00:40:15:18
as well,

00:40:15:18 – 00:40:17:21
to flesh out different topics and issues

00:40:17:21 – 00:40:18:20
that may arise.

00:40:18:20 – 00:40:20:01
I to see you there.

00:40:20:01 – 00:40:23:01
I’m very excited about it.

00:40:23:11 – 00:40:24:15
I’m going to go ahead

00:40:24:15 – 00:40:27:15
and turn this over to Anne Marie.

00:40:28:03 – 00:40:30:13
Thank you so much, Tim.

00:40:30:13 – 00:40:32:20
So I would just like to add

00:40:32:20 – 00:40:34:11
a little bit more information

00:40:34:11 – 00:40:35:21
for your awareness.

00:40:35:21 – 00:40:38:21
And in terms of the participations,

00:40:39:03 – 00:40:42:08
the participants, we did have 300

00:40:42:20 – 00:40:44:21
people come in that participated

00:40:44:21 – 00:40:46:11
during these webinars.

00:40:46:11 – 00:40:48:15
So we had a very great showing for that

00:40:48:15 – 00:40:50:12
and we were on Zoom,

00:40:50:12 – 00:40:52:05
so we were able to see

00:40:52:05 – 00:40:54:04
a lot of comments in the Q&A.

00:40:54:04 – 00:40:55:21
We saw lots of questions come through,

00:40:55:21 – 00:40:57:05
which was great.

00:40:57:05 – 00:41:00:09
And so after this meeting today,

00:41:00:11 – 00:41:01:15
we can share

00:41:01:15 – 00:41:02:16
some more information with you.

00:41:02:16 – 00:41:04:02
But we had a lot of people

00:41:04:02 – 00:41:06:17
that were coming in

00:41:06:17 – 00:41:09:08
for this research, for these webinars,

00:41:11:08 – 00:41:14:17
and we had eight we had 98%

00:41:14:17 – 00:41:16:12
of the people who participated

00:41:16:12 – 00:41:18:05
sign and agreed to be able

00:41:18:05 – 00:41:20:01
to share their information.

00:41:20:01 – 00:41:25:10
And we also had ASL involved

00:41:25:10 – 00:41:26:24
for all of those participant

00:41:26:24 – 00:41:29:24
participants out of 55,

00:41:31:01 – 00:41:32:00
out of the fit

00:41:32:00 – 00:41:32:20
out of the people

00:41:32:20 – 00:41:35:20
that were on the forum out of,

00:41:36:06 – 00:41:38:18
we had 98% of the people

00:41:38:18 – 00:41:40:07
who were involved agreed

00:41:40:07 – 00:41:43:07
to share the information

00:41:47:00 – 00:41:50:00
and to add to that

00:41:51:05 – 00:41:54:05
in terms of the topics and

00:41:54:15 – 00:41:56:15
what we what we were talking about,

00:41:56:15 – 00:41:58:22
we talked about the values and these

00:41:58:22 – 00:42:01:22
and the perspectives of the community.

00:42:03:15 – 00:42:04:19
We talked about research

00:42:04:19 – 00:42:05:23
findings

00:42:05:23 – 00:42:07:16
and the level of participation

00:42:07:16 – 00:42:08:10
was just great.

00:42:08:10 – 00:42:10:23
We were able to do a lot of research

00:42:10:23 – 00:42:13:13
and go through those topics

00:42:13:13 – 00:42:14:15
and get the specifics

00:42:14:15 – 00:42:15:20
from the participants

00:42:15:20 – 00:42:18:24
so that we had very clear results.

00:42:19:07 – 00:42:20:11
Next slide, please.

00:42:28:01 – 00:42:29:21
And in terms of the

00:42:29:21 – 00:42:31:16
discussion we were on

00:42:31:16 – 00:42:34:03
for about 173 minutes

00:42:34:03 – 00:42:37:10
and we were able to show

00:42:37:11 – 00:42:39:07
a lot of in-depth discussion

00:42:39:07 – 00:42:41:11
and comments that came through.

00:42:41:11 – 00:42:43:22
It gave a lot of value

00:42:43:22 – 00:42:44:21
to our discussion.

00:42:44:21 – 00:42:47:00
We had quite a bit of participation

00:42:47:00 – 00:42:48:18
and consideration to go over

00:42:48:18 – 00:42:50:04
and some of the topics

00:42:50:04 – 00:42:52:12
that were discussed really focused.

00:42:52:12 – 00:42:55:10
We talked about some of those earlier.

00:42:55:10 – 00:42:56:01
For example,

00:42:56:01 – 00:42:57:04
we talked about deaf

00:42:57:04 – 00:42:59:09
community readiness for AI.

00:42:59:09 – 00:43:00:14
Is the deaf community

00:43:00:14 – 00:43:02:04
actually ready for this?

00:43:02:04 – 00:43:06:09
So these discussions were so important

00:43:06:09 – 00:43:08:01
and I think having the participants

00:43:08:01 – 00:43:10:03
there to talk about AI

00:43:10:03 – 00:43:11:06
and to consider

00:43:11:06 – 00:43:11:18
whether or not

00:43:11:18 – 00:43:12:19
they were ready

00:43:12:19 – 00:43:15:21
to have this method of communication

00:43:15:21 – 00:43:19:14
as part of their life, to have AI there.

00:43:20:03 – 00:43:20:23
They discussed whether or

00:43:20:23 – 00:43:21:24
not they were ready for that.

00:43:23:06 – 00:43:24:14
And many of the

00:43:24:14 – 00:43:27:14
participants, 17

00:43:29:17 – 00:43:30:23
many of the participants

00:43:30:23 – 00:43:33:23
that were there from the advisory group,

00:43:37:21 – 00:43:38:14
were involved

00:43:38:14 – 00:43:41:14
in this process.

00:43:43:13 – 00:43:46:13
And

00:43:55:09 – 00:43:56:02
let me go back

00:43:56:02 – 00:43:59:02
for a second.

00:44:08:21 – 00:44:09:03
Okay.

00:44:09:03 – 00:44:12:03
Going back a bit to the process,

00:44:15:23 – 00:44:18:23
the discussion went on for 173 minutes

00:44:19:10 – 00:44:20:13
and we were able

00:44:20:13 – 00:44:22:01
to share different comments.

00:44:22:01 – 00:44:24:18
The participants went over

00:44:24:18 – 00:44:26:08
the things that we talked about today

00:44:26:08 – 00:44:29:08
and also

00:44:30:09 – 00:44:32:18
hashtag Save death.

00:44:32:18 – 00:44:37:06
I became a hashtag that we used

00:44:37:06 – 00:44:40:06
and we got a lot of feedback

00:44:41:21 – 00:44:44:18
and also recommendations.

00:44:44:18 – 00:44:46:16
And we talked about the broader influence

00:44:46:16 – 00:44:49:16
that this will have on the community.

00:44:55:22 – 00:44:56:09
We talked

00:44:56:09 – 00:44:59:09
about the organizations,

00:45:00:19 – 00:45:02:20
and these discussions were very important

00:45:02:20 – 00:45:05:06
to talk about collecting the data

00:45:05:06 – 00:45:08:06
and the process

00:45:09:07 – 00:45:10:01
of.

00:45:10:01 – 00:45:11:06
Next slide, please.

00:45:31:09 – 00:45:32:02
Okay.

00:45:32:02 – 00:45:33:17
So we had three

00:45:33:17 – 00:45:34:15
categories

00:45:34:15 – 00:45:36:02
to flesh out

00:45:36:02 – 00:45:38:00
all of this information through.

00:45:38:00 – 00:45:40:04
We wanted to go through

00:45:40:04 – 00:45:44:00
and with our advisory group members and

00:45:44:00 – 00:45:47:20
we decided for more thematic

00:45:47:20 – 00:45:51:06
areas of research as a team had arisen.

00:45:51:16 – 00:45:53:04
All six of them were shown

00:45:53:04 – 00:45:56:04
in the first passcode book,

00:45:56:12 – 00:45:58:04
and now we’re going to add

00:45:58:04 – 00:46:01:10
that to what we’ll send out to all.

00:46:01:17 – 00:46:04:03
And that’s also included in the report.

00:46:04:03 – 00:46:07:14
I think I saw some questions in the Q A

00:46:07:16 – 00:46:10:16
It is available online.

00:46:10:20 – 00:46:12:22
It will be available

00:46:12:22 – 00:46:16:00
as we’re doing an invitation process

00:46:16:00 – 00:46:19:23
to add that to our website.

00:46:30:18 – 00:46:32:05
I

00:46:32:05 – 00:46:33:03
do you want me to

00:46:33:03 – 00:46:36:03
move on to the next slide?

00:46:36:06 – 00:46:37:04
I think we’re at the end.

00:46:37:04 – 00:46:37:13
Okay.

00:47:00:05 – 00:47:03:05
So

00:47:05:07 – 00:47:06:08
I’m saying if the team

00:47:06:08 – 00:47:08:12
could all turn on their cameras,

00:47:08:12 – 00:47:11:12
we’ll start in with the Q&A.

00:47:16:08 – 00:47:16:19
All right.

00:47:16:19 – 00:47:17:20
Are we ready to address

00:47:17:20 – 00:47:20:03
the questions in the Q&A?

00:47:20:03 – 00:47:22:08
Let’s do it, though.

00:47:22:08 – 00:47:23:21
The first thing I’d like to do

00:47:23:21 – 00:47:25:19
is thank you all for your questions.

00:47:25:19 – 00:47:28:19
Thank you so very much.

00:47:29:04 – 00:47:29:24
The first question

00:47:29:24 – 00:47:32:24
that we have for the panel

00:47:34:01 – 00:47:35:16
in the future,

00:47:35:16 – 00:47:38:16
say 10 to 15 years down the road,

00:47:38:20 – 00:47:39:08
A.I.

00:47:39:08 – 00:47:42:08
by A.I., what will that look like?

00:47:43:21 – 00:47:46:21
How many applications

00:47:46:22 – 00:47:48:12
are possible?

00:47:48:12 – 00:47:53:08
For example, TV theaters, movie theaters?

00:47:53:19 – 00:47:55:16
Will there be an automatic interpreter

00:47:55:16 – 00:47:58:16
pop up on the screen?

00:48:00:08 – 00:48:01:12
What do you think

00:48:01:12 – 00:48:02:24
is the potential for A.I.

00:48:02:24 – 00:48:05:24
in the future?

00:48:08:21 – 00:48:10:14
Jeff Would you like to go?

00:48:10:14 – 00:48:13:05
Jeff saying yes, I will talk about that.

00:48:13:05 – 00:48:13:24
I bet

00:48:13:24 – 00:48:15:19
I will have so many applications

00:48:15:19 – 00:48:16:14
in the future.

00:48:16:14 – 00:48:18:07
I think in 10 to 15 years,

00:48:18:07 – 00:48:20:09
the possibilities are astounding.

00:48:20:09 – 00:48:23:23
There’s you know, it I would say that

00:48:24:00 – 00:48:25:08
computes two equates

00:48:25:08 – 00:48:28:00
to about 100 years of human development

00:48:28:00 – 00:48:29:06
because technology moves

00:48:29:06 – 00:48:30:21
at such a rapid pace.

00:48:30:21 – 00:48:31:10
One thing

00:48:31:10 – 00:48:33:03
I would like to say that for sure,

00:48:33:03 – 00:48:34:19
I know that it will have improved

00:48:34:19 – 00:48:36:02
drastically by then

00:48:36:02 – 00:48:38:00
because I think it’s going to continue

00:48:38:00 – 00:48:39:16
marching on

00:48:39:16 – 00:48:41:14
and improving as time goes by.

00:48:41:14 – 00:48:43:12
Now, will we be more trusting

00:48:43:12 – 00:48:44:18
of these applications

00:48:44:18 – 00:48:47:18
in the future and more confident with it?

00:48:47:19 – 00:48:51:10
I think in low risk situations,

00:48:52:22 – 00:48:54:14
I don’t think it should be too bad.

00:48:54:14 – 00:48:57:14
For example,

00:48:57:14 – 00:48:59:18
not a police encounter,

00:48:59:18 – 00:49:00:17
medical encounter

00:49:00:17 – 00:49:03:20
or anything of that sort, but I bet

00:49:03:20 – 00:49:06:12
I could be used for automatic

00:49:06:12 – 00:49:08:09
conversation and dialogs like,

00:49:08:09 – 00:49:10:04
you know, with robocalls,

00:49:10:04 – 00:49:12:14
sharing information,

00:49:12:14 – 00:49:13:24
having something set,

00:49:13:24 – 00:49:17:13
you know, an automated system.

00:49:18:03 – 00:49:19:10
But in the future

00:49:19:10 – 00:49:22:10
I foresee more captioning being involved,

00:49:22:17 – 00:49:24:20
not necessarily ASL only,

00:49:24:20 – 00:49:26:19
but I think captioning

00:49:26:19 – 00:49:29:16
and multilingual captioning as a whole

00:49:29:16 – 00:49:32:06
will have developed so much over time.

00:49:32:06 – 00:49:33:21
I think there are so many possibilities

00:49:33:21 – 00:49:34:16
and different directions

00:49:34:16 – 00:49:35:23
it could go again.

00:49:35:23 – 00:49:37:11
It’s hard to predict everything

00:49:37:11 – 00:49:38:19
that may happen in the future,

00:49:38:19 – 00:49:39:20
but in a nutshell,

00:49:39:20 – 00:49:42:01
i believe that’s what would happen

00:49:42:01 – 00:49:45:03
and I think that some examples of that

00:49:45:08 – 00:49:46:13
do so.

00:49:46:13 – 00:49:47:06
For example,

00:49:47:06 – 00:49:49:21
when you’re driving through a place

00:49:49:21 – 00:49:51:20
and you want to order coffee

00:49:51:20 – 00:49:53:12
or you want to order food

00:49:53:12 – 00:49:55:22
and that kind of situation,

00:49:55:22 – 00:49:58:13
and when you’re ordering your food,

00:49:58:13 – 00:49:59:00
you know

00:50:00:04 – 00:50:00:14
it.

00:50:00:14 – 00:50:01:05
It’s not

00:50:01:05 – 00:50:02:03
going to be something

00:50:02:03 – 00:50:03:10
that’s disastrous to your life

00:50:03:10 – 00:50:05:00
if something messes up.

00:50:05:00 – 00:50:06:20
But when it comes to the barriers,

00:50:06:20 – 00:50:09:01
the deaf community experiences

00:50:09:01 – 00:50:10:12
in other situations,

00:50:10:12 – 00:50:12:06
there’s so many different possibilities

00:50:12:06 – 00:50:13:17
of how this could go.

00:50:13:17 – 00:50:15:04
And I think, again,

00:50:15:04 – 00:50:17:14
we have to make sure that the technology

00:50:17:14 – 00:50:19:08
is ready, verified

00:50:19:08 – 00:50:21:16
and it’s going to be more beneficial

00:50:21:16 – 00:50:23:18
and and not harmful.

00:50:23:18 – 00:50:25:18
I think that’s what we have to see,

00:50:25:18 – 00:50:27:02
you know?

00:50:27:02 – 00:50:27:19
Theresa here,

00:50:27:19 – 00:50:29:03
I’d like to make a comment as well,

00:50:29:03 – 00:50:30:15
and I’d like to emphasize the point

00:50:30:15 – 00:50:34:01
about low risk situations for a moment.

00:50:34:01 – 00:50:35:06
I think that’s important.

00:50:35:06 – 00:50:37:08
It’s imperative for us to realize

00:50:37:08 – 00:50:38:23
that we always have a need

00:50:38:23 – 00:50:41:23
for human interpreters,

00:50:41:24 – 00:50:43:12
specific situations

00:50:43:12 – 00:50:44:24
like medical situations,

00:50:44:24 – 00:50:47:06
legal situations, court

00:50:47:06 – 00:50:49:04
law enforcement interactions.

00:50:49:04 – 00:50:50:07
It’s really imperative

00:50:50:07 – 00:50:51:24
that we include the human aspect

00:50:51:24 – 00:50:54:08
and human judgment in those situations

00:50:55:19 – 00:50:56:24
in this century.

00:50:56:24 – 00:51:00:19
And to add to that, I think one thing

00:51:00:19 – 00:51:01:16
that we also have to

00:51:01:16 – 00:51:04:16
emphasize is knowing the language itself

00:51:04:19 – 00:51:07:06
and knowing that process,

00:51:07:06 – 00:51:09:14
whether the responses

00:51:09:14 – 00:51:11:00
and also looking at

00:51:11:00 – 00:51:12:22
whether it’s a live person,

00:51:12:22 – 00:51:16:06
if it’s going to be a mixed situation,

00:51:16:10 – 00:51:17:15
we have to recognize

00:51:17:15 – 00:51:18:23
exactly what is involved

00:51:18:23 – 00:51:20:21
situation by situation

00:51:20:21 – 00:51:24:08
so that we can make sure that the access

00:51:24:08 – 00:51:25:21
that’s being provided is as close

00:51:25:21 – 00:51:27:17
to 100% accurate as possible.

00:51:27:17 – 00:51:29:16
So like Theresa said, humans

00:51:29:16 – 00:51:31:13
have to be involved in this

00:51:31:13 – 00:51:33:05
and in this feedback loop,

00:51:33:05 – 00:51:35:15
and it’s very critical that we

00:51:35:15 – 00:51:37:00
emphasize that.

00:51:37:00 – 00:51:37:16
Theresa saying,

00:51:37:16 – 00:51:39:01
I’d like to add a little bit more.

00:51:39:01 – 00:51:40:01
One thing to think about

00:51:40:01 – 00:51:41:12
is in my past experience,

00:51:41:12 – 00:51:45:13
for many years, living in Mexico,

00:51:45:13 – 00:51:47:00
the deaf population

00:51:47:00 – 00:51:48:20
in the community, in New Mexico,

00:51:48:20 – 00:51:51:22
New Mexico, you know,

00:51:52:19 – 00:51:56:00
they were able to be

00:51:56:00 – 00:51:57:15
mainstreamed into the school

00:51:59:14 – 00:52:00:20
and the dialects are

00:52:00:20 – 00:52:02:22
different in different areas.

00:52:02:22 – 00:52:05:06
I may not recognize that.

00:52:05:06 – 00:52:06:14
So growing up in those schools

00:52:06:14 – 00:52:08:13
where I where the communities were small,

00:52:08:13 – 00:52:10:07
the dialects were diverse,

00:52:10:07 – 00:52:12:21
I may not have that capability

00:52:12:21 – 00:52:15:05
just yet to recognize that.

00:52:15:05 – 00:52:17:17
So we need to make sure that our air

00:52:17:17 – 00:52:20:17
is ready for this specifically,

00:52:20:18 – 00:52:21:14
for example,

00:52:21:14 – 00:52:24:14
with our black deaf signers,

00:52:24:21 – 00:52:26:08
the American sign language

00:52:26:08 – 00:52:27:02
that they use

00:52:27:02 – 00:52:28:17
is so culturally diverse

00:52:28:17 – 00:52:30:12
than American Sign language.

00:52:30:12 – 00:52:32:00
We have to make sure

00:52:32:00 – 00:52:32:19
that we have the right

00:52:32:19 – 00:52:35:10
understanding of interpreting.

00:52:35:10 – 00:52:38:10
And for this approach.

00:52:43:14 – 00:52:44:03
Yes.

00:52:44:03 – 00:52:46:19
And more questions.

00:52:46:19 – 00:52:47:17
Very exciting.

00:52:47:17 – 00:52:49:12
Okay, so another question.

00:52:49:12 – 00:52:52:22
Question number two is related

00:52:52:22 – 00:52:55:22
to the term

00:52:56:06 – 00:52:59:06
sign language

00:53:03:19 – 00:53:05:01
and atomization.

00:53:05:01 – 00:53:08:01

00:53:09:05 – 00:53:09:23
So basically

00:53:09:23 – 00:53:12:23
related to privacy,

00:53:13:03 – 00:53:15:23
protecting individuals and their data.

00:53:18:00 – 00:53:19:14
So from

00:53:19:14 – 00:53:20:21
the research and the report

00:53:20:21 – 00:53:22:15
that you all produced,

00:53:22:15 – 00:53:24:20
can you guys discuss and share a bit

00:53:24:20 – 00:53:27:20
about the data privacy

00:53:28:18 – 00:53:30:06
of confidentiality

00:53:30:06 – 00:53:33:14
and also for automatic interpretation,

00:53:33:23 – 00:53:36:24
would you use an avatar and

00:53:37:10 – 00:53:38:13
what does this look like?

00:53:38:13 – 00:53:39:22
How would you be able to

00:53:39:22 – 00:53:41:06
represent individuals

00:53:41:06 – 00:53:44:06
while also protecting them?

00:53:44:09 – 00:53:44:24
Jeff Here

00:53:44:24 – 00:53:46:03
I would like to take that one,

00:53:46:03 – 00:53:47:18
if that’s all right with the group.

00:53:47:18 – 00:53:50:12
So I prefer to sign

00:53:50:12 – 00:53:51:20
automatic interpreting

00:53:51:20 – 00:53:53:14
for data collection.

00:53:53:14 – 00:53:56:00
The faith is involved in that.

00:53:56:00 – 00:53:58:21
So one aspect of data collection is worth

00:53:58:21 – 00:54:00:05
signing is

00:54:00:05 – 00:54:01:09
they don’t have our faith,

00:54:01:09 – 00:54:02:18
and it’s imperative

00:54:02:18 – 00:54:05:15
to compare that with speech recognition

00:54:05:15 – 00:54:07:06
through automatic speech recognition.

00:54:07:06 – 00:54:08:23
You have voice and intonation,

00:54:08:23 – 00:54:09:15
but with signing,

00:54:09:15 – 00:54:11:06
if you don’t include the faith,

00:54:11:06 – 00:54:12:22
that’s a deep part of the language

00:54:12:22 – 00:54:14:15
that’s missing itself.

00:54:14:15 – 00:54:18:16
So there’s no real easy way to avoid that

00:54:18:16 – 00:54:20:24
with data collection,

00:54:20:24 – 00:54:22:23
because we have to have our faith

00:54:22:23 – 00:54:25:13
for the language foundation and tone

00:54:25:13 – 00:54:27:02
and meaning to be there.

00:54:27:02 – 00:54:28:13
So we have to be very careful

00:54:28:13 – 00:54:30:08
with the data itself.

00:54:30:08 – 00:54:31:24
We have to protect that

00:54:31:24 – 00:54:32:12
to make sure

00:54:32:12 – 00:54:35:03
that people are fully informed.

00:54:35:03 – 00:54:37:10
Now, informed consent,

00:54:37:10 – 00:54:38:22
we’re going to use that in the training,

00:54:38:22 – 00:54:40:02
in AI by AI.

00:54:40:02 – 00:54:41:20
We will be talking about that

00:54:41:20 – 00:54:42:23
facial recognition

00:54:42:23 – 00:54:44:23
and other identifying information.

00:54:44:23 – 00:54:47:07
My background may be there as well.

00:54:47:07 – 00:54:49:22
So we have to really filter out

00:54:49:22 – 00:54:51:04
that information

00:54:51:04 – 00:54:54:03
and see what part of that is protected

00:54:54:03 – 00:54:56:11
and what part of it is stored.

00:54:56:11 – 00:54:58:00
And we also have to think about

00:54:58:00 – 00:55:00:06
what is private.

00:55:00:06 – 00:55:02:21
And with avatars, you know,

00:55:02:21 – 00:55:03:22
it may be possible

00:55:03:22 – 00:55:06:11
to produce signs with avatars,

00:55:06:11 – 00:55:08:03
but it could be a little bit off.

00:55:08:03 – 00:55:11:03
For example, the avatar.

00:55:11:07 – 00:55:13:14
What can we train that avatar

00:55:13:14 – 00:55:14:18
to make the signs?

00:55:14:18 – 00:55:16:24
And can we identify that

00:55:16:24 – 00:55:18:02
and know that that’s

00:55:18:02 – 00:55:20:03
what the person is signing?

00:55:20:03 – 00:55:21:23
That can be a little bit ambiguous

00:55:21:23 – 00:55:23:00
in trying to identify

00:55:23:00 – 00:55:26:15
what the data and the the machine

00:55:26:15 – 00:55:29:22
learning is actually tracking in.

00:55:31:18 – 00:55:34:15
All of that goes back to

00:55:34:15 – 00:55:37:15
the subject of informed consent.

00:55:39:02 – 00:55:40:23
And I’d like to add

00:55:40:23 – 00:55:42:17
to that one more thing.

00:55:42:17 – 00:55:45:18
In terms of informed consent, often

00:55:45:18 – 00:55:47:07
we think, okay,

00:55:47:07 – 00:55:48:20
this is just a one time thing

00:55:48:20 – 00:55:51:08
I’m signing and I’m giving my consent,

00:55:51:08 – 00:55:52:24
but informed consent

00:55:52:24 – 00:55:54:08
really needs to happen

00:55:54:08 – 00:55:56:03
on an ongoing basis.

00:55:56:03 – 00:55:57:17
We need to be reminded

00:55:57:17 – 00:55:58:20
and we need to make sure

00:55:58:20 – 00:56:01:06
that we continue to give that consent

00:56:01:06 – 00:56:04:07
and continue to agree and remind people

00:56:04:13 – 00:56:06:11
that they have permission

00:56:06:11 – 00:56:07:13
to remove themselves

00:56:07:13 – 00:56:09:14
or to take that consent back.

00:56:09:14 – 00:56:09:23

00:56:09:23 – 00:56:10:19
And so they have

00:56:10:19 – 00:56:11:16
the right to be involved.

00:56:11:16 – 00:56:14:16
They had the right to decline.

00:56:19:04 – 00:56:20:09
Okay.

00:56:20:09 – 00:56:21:19
QUESTION

00:56:21:19 – 00:56:23:01
Thank you so much for your question.

00:56:23:01 – 00:56:23:21
So the next question

00:56:23:21 – 00:56:24:21
we’re going to address

00:56:24:21 – 00:56:28:01
is I’d like to add a little bit to

00:56:28:01 – 00:56:29:23
this as well.

00:56:32:08 – 00:56:35:08
The discussion about

00:56:35:23 – 00:56:38:23
I’m sorry, the discussion about how

00:56:40:18 – 00:56:43:24
this task force and this advisory group

00:56:44:05 – 00:56:45:20
has collaborated

00:56:45:20 – 00:56:48:02
with other organizations,

00:56:48:02 – 00:56:50:15
the NAD Gallaudet University,

00:56:50:15 – 00:56:53:18
other educational bodies.

00:56:53:18 – 00:56:55:10
What does that look like?

00:56:55:10 – 00:56:58:10
What does your partnership look like?

00:57:02:05 – 00:57:04:02
And I can answer that question.

00:57:04:02 – 00:57:08:15
And so I think that first of all,

00:57:09:03 – 00:57:09:24
the advisory

00:57:09:24 – 00:57:12:07
council is really a device,

00:57:12:07 – 00:57:13:18
a diverse group.

00:57:13:18 – 00:57:16:14
We welcome people

00:57:16:14 – 00:57:18:15
from various organizations

00:57:18:15 – 00:57:21:06
to make sure that we’re reflecting

00:57:21:06 – 00:57:23:00
various people, like, for example,

00:57:23:00 – 00:57:24:22
from Gallaudet University,

00:57:24:22 – 00:57:28:16
people from Nairobi, people from Nadi.

00:57:29:00 – 00:57:32:00
We have participants from all over, and

00:57:32:06 – 00:57:33:18
I think that’s important.

00:57:33:18 – 00:57:36:01
But also we want to recognize

00:57:36:01 – 00:57:37:06
that we want to continue

00:57:37:06 – 00:57:40:22
to involve individuals like, for example,

00:57:40:22 – 00:57:45:10
management.

00:57:45:10 – 00:57:47:03
Omar That’s another one.

00:57:47:03 – 00:57:51:12
We want to have diverse representation

00:57:51:12 – 00:57:52:21
to be able to continue

00:57:52:21 – 00:57:54:14
to working with everyone,

00:57:54:14 – 00:57:56:22
because there are groups that have been

00:57:56:22 – 00:57:59:15
have not been included

00:57:59:15 – 00:58:01:05
in these types of processes in the past.

00:58:01:05 – 00:58:03:12
So we want to make sure that our research

00:58:03:12 – 00:58:05:06
and our study can continue

00:58:05:06 – 00:58:06:17
that kind of collaboration

00:58:06:17 – 00:58:08:06
that we’ve already established.

00:58:08:06 – 00:58:08:22
Like I said,

00:58:08:22 – 00:58:12:07
whether it be with Gallaudet University,

00:58:12:07 – 00:58:13:09
other organizations, I’m

00:58:13:09 – 00:58:14:04
not sure if anyone else

00:58:14:04 – 00:58:15:20
has something to add to that comment.

00:58:17:23 – 00:58:19:07
I’d like to add to the comment.

00:58:19:07 – 00:58:20:15
So we have individuals

00:58:20:15 – 00:58:23:15
with specific skills and backgrounds.

00:58:23:15 – 00:58:27:08
We have work where, you know, I myself

00:58:27:08 – 00:58:28:19
work at Gallaudet University.

00:58:28:19 – 00:58:32:01
I’m a research focused on bioethics

00:58:32:01 – 00:58:35:12
and I volunteer to participate

00:58:35:12 – 00:58:37:09
as an individual researcher.

00:58:37:09 – 00:58:38:15
I’m not representing

00:58:38:15 – 00:58:39:14
Gallaudet University,

00:58:39:14 – 00:58:40:13
but at the same time,

00:58:40:13 – 00:58:41:24
we do have discussions

00:58:41:24 – 00:58:44:24
with many people in Gallaudet University

00:58:45:09 – 00:58:47:03
throughout different areas,

00:58:47:03 – 00:58:48:23
recognizing the importance of this.

00:58:48:23 – 00:58:50:01
And sometimes we have a person

00:58:50:01 – 00:58:51:08
who’s willing to work

00:58:51:08 – 00:58:53:06
who is working for Gallaudet University,

00:58:53:06 – 00:58:54:10
but they’re not representing

00:58:54:10 – 00:58:55:24
the university itself

00:58:55:24 – 00:58:58:07
in their role in this research.

00:58:58:07 – 00:59:00:14
So in my opinion, my research

00:59:00:14 – 00:59:01:21
and what I’m looking at

00:59:01:21 – 00:59:03:19
is not representing the university

00:59:03:19 – 00:59:05:20
at large itself.

00:59:05:20 – 00:59:07:18
So I hope that clarifies that

00:59:07:18 – 00:59:08:14
to a degree.

00:59:11:14 – 00:59:13:19
This is Anne-Marie and

00:59:13:19 – 00:59:17:09
I think in terms of education

00:59:17:09 – 00:59:20:11
and nonprofit groups and organizations,

00:59:20:11 – 00:59:21:22
as an ADDY

00:59:21:22 – 00:59:23:00
and other organizations

00:59:23:00 – 00:59:24:14
that we’ve had involved,

00:59:24:14 – 00:59:26:15
these different signing

00:59:26:15 – 00:59:29:03
groups are working closely together,

00:59:29:03 – 00:59:32:03
different companies who have technology.

00:59:32:08 – 00:59:33:19
And I’d like to emphasize

00:59:33:19 – 00:59:36:22
that we are all working together

00:59:37:06 – 00:59:40:22
and that the thing is to have

00:59:41:03 – 00:59:42:14
a platform for discussion

00:59:42:14 – 00:59:44:00
for the deaf community.

00:59:44:00 – 00:59:45:08
And I know often

00:59:45:08 – 00:59:46:07
a lot of individuals

00:59:46:07 – 00:59:47:21
are overlooked in our community

00:59:47:21 – 00:59:48:15
are they’re pushed

00:59:48:15 – 00:59:50:15
out of these discussions.

00:59:50:15 – 00:59:51:16
They don’t get the opportunity

00:59:51:16 – 00:59:53:12
to explain their perspective.

00:59:53:12 – 00:59:55:04
There’s education issues.

00:59:55:04 – 00:59:59:20
And so, you know, I think that A.I.

01:00:00:00 – 01:00:02:06
here and we often see people say,

01:00:02:06 – 01:00:03:23
you know, AI is here.

01:00:03:23 – 01:00:05:13
What is that going to look like?

01:00:05:13 – 01:00:06:23
And some people say, no,

01:00:06:23 – 01:00:08:03
we don’t want to see this happen.

01:00:08:03 – 01:00:09:09
We don’t want A.I.

01:00:09:09 – 01:00:11:21
to part of this interpreting process

01:00:11:21 – 01:00:14:11
while we know it’s coming.

01:00:14:11 – 01:00:16:11
And so I think it’s our responsibility

01:00:16:11 – 01:00:18:21
to ensure that this collaborative effort

01:00:18:21 – 01:00:20:17
stays in place

01:00:20:17 – 01:00:21:24
so that we have deaf community

01:00:21:24 – 01:00:23:18
representation and representation

01:00:23:18 – 01:00:26:00
in the development of AI over time,

01:00:26:00 – 01:00:27:16
not just for education,

01:00:27:16 – 01:00:30:16
but also that these all over America,

01:00:30:21 – 01:00:32:04
they have the responsibility

01:00:32:04 – 01:00:34:08
of ensuring that deaf individuals are

01:00:34:08 – 01:00:35:07
they’re at the table

01:00:35:07 – 01:00:36:14
when these discussions happen.

01:00:40:00 – 01:00:42:16
Let’s move on to the next question.

01:00:42:16 – 01:00:42:23
Okay.

01:00:42:23 – 01:00:44:21
So for our next question,

01:00:44:21 – 01:00:47:14
what are we working on related

01:00:47:14 – 01:00:50:20
to the legislation

01:00:50:20 – 01:00:53:20
and for protection of the deaf community?

01:00:54:07 – 01:00:57:06
What is what does that look like?

01:00:57:06 – 01:00:58:18
I could take that.

01:00:58:18 – 01:01:01:19
So in terms of this, we

01:01:02:22 – 01:01:04:11
I can maybe probably clarify

01:01:04:11 – 01:01:05:13
a bit of the explanation

01:01:05:13 – 01:01:07:07
we’ve given already, but

01:01:07:07 – 01:01:08:24
when it comes to structure,

01:01:08:24 – 01:01:11:06
we do have our advisory group.

01:01:11:06 – 01:01:14:06
We also have safe A.I..

01:01:14:06 – 01:01:16:00
It’s a task force.

01:01:16:00 – 01:01:18:11
And that task force represents

01:01:18:11 – 01:01:20:12
so many different languages

01:01:20:12 – 01:01:23:12
and interpreters, providers

01:01:23:21 – 01:01:25:19
and tech people

01:01:25:19 – 01:01:27:21
that are involved in the tech development

01:01:27:21 – 01:01:28:22
of it.

01:01:28:22 – 01:01:31:00
We also have consumers

01:01:31:00 – 01:01:32:21
who would be using the services,

01:01:32:21 – 01:01:33:20
so we have a broad

01:01:33:20 – 01:01:34:20
range of people involved

01:01:34:20 – 01:01:36:05
in the task force.

01:01:36:05 – 01:01:39:24
Now, in terms of the deaf perspective,

01:01:39:24 – 01:01:43:03
we wanted to make sure that

01:01:43:19 – 01:01:44:19
they got involved,

01:01:44:19 – 01:01:46:18
but it was not until a bit later

01:01:46:18 – 01:01:47:12
after task

01:01:47:12 – 01:01:49:00
Force was formed that they got involved.

01:01:49:00 – 01:01:49:14
And so we said,

01:01:49:14 – 01:01:50:08
you know, we want silos,

01:01:50:08 – 01:01:51:00
which include we want

01:01:51:00 – 01:01:52:08
all of this included.

01:01:52:08 – 01:01:55:08
So we established the advisory council

01:01:55:08 – 01:01:58:04
to ensure that AI and say sign language

01:01:58:04 – 01:01:59:23
interpreting was represented

01:01:59:23 – 01:02:01:06
from the deaf perspective.

01:02:01:06 – 01:02:02:08
And of course

01:02:02:08 – 01:02:03:18
there’s diverse organizations,

01:02:03:18 – 01:02:05:15
the academic perspective,

01:02:05:15 – 01:02:07:10
designers, developers, all of that.

01:02:07:10 – 01:02:10:18
They were involved and in the group

01:02:11:02 – 01:02:11:11
so that

01:02:11:11 – 01:02:14:18
they could give their advice to say, I

01:02:15:06 – 01:02:17:03
now say saved by

01:02:17:03 – 01:02:18:02
AI has been working

01:02:18:02 – 01:02:19:08
with the advisory group

01:02:19:08 – 01:02:20:22
and the goal is to continue

01:02:20:22 – 01:02:24:19
to develop the policies and the law.

01:02:24:24 – 01:02:27:21
Suggestions that we have for the group

01:02:27:21 – 01:02:31:19
so that our parts and our recommendations

01:02:31:19 – 01:02:32:12
are included

01:02:32:12 – 01:02:33:20
in their reporting

01:02:33:20 – 01:02:36:08
and in their data collection and surveys.

01:02:36:08 – 01:02:37:07
Right now

01:02:37:07 – 01:02:38:18
there are ten different languages

01:02:38:18 – 01:02:39:23
that are being looked at.

01:02:39:23 – 01:02:41:06
And so in terms of American

01:02:41:06 – 01:02:44:15
Sign Language, it was noted

01:02:44:15 – 01:02:46:04
that there

01:02:46:04 – 01:02:47:23
needed to be another opportunity

01:02:47:23 – 01:02:48:14
to collaborate

01:02:48:14 – 01:02:50:02
more with the deaf community

01:02:50:02 – 01:02:51:05
to send out surveys

01:02:51:05 – 01:02:52:10
in American Sign Language.

01:02:52:10 – 01:02:53:12
So that’s where we are right now.

01:02:53:12 – 01:02:54:13
In the process.

01:02:54:13 – 01:02:56:17
We’re hoping that our dialog

01:02:56:17 – 01:02:57:17
and our discussion

01:02:57:17 – 01:03:00:04
and our collection of information

01:03:00:04 – 01:03:03:17
will become a great contributor

01:03:03:17 – 01:03:05:08
to the bigger picture.

01:03:05:08 – 01:03:06:23
Does anyone have anything to add?

01:03:06:23 – 01:03:08:12
I just wanted to clarify a comment

01:03:08:12 – 01:03:09:14
that I had made earlier.

01:03:15:22 – 01:03:16:19
Okay.

01:03:16:19 – 01:03:19:19
Moving on to the next question

01:03:20:02 – 01:03:23:03
for this report and research.

01:03:23:23 – 01:03:26:06
We are focusing on ASL.

01:03:26:06 – 01:03:28:10
What about other signed languages

01:03:28:10 – 01:03:29:15
in other countries?

01:03:29:15 – 01:03:32:03
Will we be looking at others

01:03:32:03 – 01:03:35:03
in the future?

01:03:36:13 – 01:03:37:22
I can take that question.

01:03:37:22 – 01:03:39:12
It would be great.

01:03:39:12 – 01:03:40:16
It’s a dream

01:03:40:16 – 01:03:43:10
not for ASL to be the only language

01:03:43:10 – 01:03:44:06
that we look at, right?

01:03:44:06 – 01:03:46:04
We want to consider all of this,

01:03:46:04 – 01:03:48:05
but in terms of the task force

01:03:48:05 – 01:03:51:08
and the advisory council where

01:03:51:12 – 01:03:53:01
we’re working with different

01:03:53:01 – 01:03:54:19
American organizations

01:03:54:19 – 01:03:57:20
and we have been mostly focusing on ASL,

01:03:58:01 – 01:04:01:01
but we are seeing more effort right now

01:04:01:01 – 01:04:02:17
in other parts of the world

01:04:02:17 – 01:04:04:05
where they are focusing

01:04:04:05 – 01:04:06:09
on the automatic interpreting

01:04:06:09 – 01:04:08:01
for other languages.

01:04:08:01 – 01:04:11:21
And so I am aware of the fact

01:04:11:21 – 01:04:14:15
that there are some places in Europe

01:04:14:15 – 01:04:15:23
that are focusing on things.

01:04:15:23 – 01:04:17:09
I don’t have the specific names.

01:04:17:09 – 01:04:19:19
They don’t they don’t come to me.

01:04:19:19 – 01:04:20:21
These names are not coming to me

01:04:20:21 – 01:04:21:06
right now.

01:04:21:06 – 01:04:25:06
But yeah, and Tim can talk more about

01:04:26:05 – 01:04:26:12
one of

01:04:26:12 – 01:04:29:12
those organizations.

01:04:31:16 – 01:04:32:16
We do have people involved

01:04:32:16 – 01:04:33:23
in Europe and Canada.

01:04:33:23 – 01:04:35:10
I know that there are many more

01:04:35:10 – 01:04:36:09
all over the world

01:04:36:09 – 01:04:37:09
who are also looking

01:04:37:09 – 01:04:39:05
at the same technology,

01:04:39:05 – 01:04:42:20
but because right now, Safe Eye

01:04:42:21 – 01:04:45:01
Task Force is focusing on

01:04:45:01 – 01:04:47:04
American policy legislation

01:04:47:04 – 01:04:50:12
and all of that current focus is

01:04:50:12 – 01:04:52:20
specifically what’s happening in America

01:04:52:20 – 01:04:53:21
and North America.

01:04:53:21 – 01:04:55:01
But at the same time,

01:04:55:01 – 01:04:58:13
we could have an impact on Canada

01:04:58:22 – 01:04:59:18
some of our research,

01:04:59:18 – 01:05:01:11
could impact Europe as well.

01:05:01:11 – 01:05:03:17
So I think this is with the process,

01:05:03:17 – 01:05:06:17
we will probably

01:05:07:02 – 01:05:09:01
continue to move forward

01:05:09:01 – 01:05:12:01
and see more replication

01:05:12:01 – 01:05:15:01
of our studies or expansion of our focus.

01:05:15:06 – 01:05:17:05
Looking at it on a more global scale

01:05:17:05 – 01:05:19:08
as we move forward.

01:05:19:08 – 01:05:21:24
Theresa Here I’d like to also add as well

01:05:21:24 – 01:05:23:01
that I think it’s imperative

01:05:23:01 – 01:05:24:12
that we recognize people

01:05:24:12 – 01:05:26:17
using sign language here in the U.S.

01:05:26:17 – 01:05:29:02
are not only using ASL.

01:05:29:02 – 01:05:30:19
So it’s important for us to know

01:05:30:19 – 01:05:32:23
that there are many foreign

01:05:32:23 – 01:05:35:16
languages used here in the US

01:05:35:16 – 01:05:38:08
and we’ve feedback from users of sign

01:05:38:08 – 01:05:39:09
language in the U.S.,

01:05:39:09 – 01:05:41:15
not just American Sign language.

01:05:41:15 – 01:05:45:06
We will or we will be reaching out to

01:05:45:07 – 01:05:46:11
other sign languages

01:05:46:11 – 01:05:48:15
and pulling in that data and information

01:05:48:15 – 01:05:50:11
and their experiences as well.

01:05:50:11 – 01:05:51:08
And it’s important

01:05:51:08 – 01:05:54:14
for us to include everyone at the table.

01:05:58:22 – 01:05:59:12
Okay.

01:05:59:12 – 01:06:01:01
Next question.

01:06:01:01 – 01:06:04:01
So this question is related to readiness.

01:06:05:21 – 01:06:07:20
So in terms

01:06:07:20 – 01:06:10:04
of these technological companies

01:06:10:04 – 01:06:11:12
in development

01:06:11:12 – 01:06:13:16
and a lot of these people

01:06:13:16 – 01:06:14:06
who are working

01:06:14:06 – 01:06:14:24
in this

01:06:14:24 – 01:06:16:10
arena and make these decisions

01:06:16:10 – 01:06:18:08
don’t always include deaf people.

01:06:18:08 – 01:06:21:04
So what approach do we think?

01:06:21:04 – 01:06:23:08
Or how can we make sure

01:06:23:08 – 01:06:24:17
that deaf individuals

01:06:24:17 – 01:06:25:07
are involved

01:06:25:07 – 01:06:26:09
in these conversations

01:06:26:09 – 01:06:27:02
and that they’re always

01:06:27:02 – 01:06:31:07
a part of the process for AI by AI,

01:06:32:06 – 01:06:35:06
machine learning and everything are

01:06:37:02 – 01:06:38:12
Emery Here, I’ll take that.

01:06:38:12 – 01:06:39:01
I’d be happy

01:06:39:01 – 01:06:40:18
to make a comment about that.

01:06:40:18 – 01:06:42:18
So with regards to

01:06:42:18 – 01:06:44:13
what we’ve been speaking about,

01:06:44:13 – 01:06:45:15
it’s important

01:06:45:15 – 01:06:49:06
that if all of the impact organizations

01:06:49:10 – 01:06:51:09
that were serving the community,

01:06:51:09 – 01:06:55:01
educational, corporate America,

01:06:55:01 – 01:06:55:24
different areas,

01:06:55:24 – 01:06:57:10
we know that part

01:06:57:10 – 01:06:59:18
of our screening process,

01:06:59:18 – 01:07:01:21
people want to sponsor our organization.

01:07:01:21 – 01:07:04:07
They want to to look into this.

01:07:04:07 – 01:07:05:21
They want to look in the screening

01:07:05:21 – 01:07:10:00
is disability access included?

01:07:10:00 – 01:07:11:01
Is that supported?

01:07:11:01 – 01:07:12:10
Are your hiring deaf

01:07:12:10 – 01:07:14:13
and hard of hearing employees?

01:07:14:13 – 01:07:15:14
There’s so many things

01:07:15:14 – 01:07:17:15
to look at discuss with them.

01:07:17:15 – 01:07:19:10
It’s very creative

01:07:19:10 – 01:07:21:12
in the approach to HIPAA.

01:07:21:12 – 01:07:23:17
And so it’s a big it’s a very hot topic.

01:07:25:01 – 01:07:27:16
Not only are the

01:07:27:16 – 01:07:30:05
the community and organizations at large,

01:07:30:05 – 01:07:31:14
but also the individual.

01:07:31:14 – 01:07:32:17
Again,

01:07:32:17 – 01:07:33:21
what it looks like to us

01:07:33:21 – 01:07:35:13
is we have to really partner

01:07:35:13 – 01:07:38:19
and really push this,

01:07:38:19 – 01:07:41:19
that all of them

01:07:42:05 – 01:07:43:08
it’s important for all

01:07:43:08 – 01:07:45:20
because they’re impacted

01:07:45:20 – 01:07:47:18
by the development and design of this.

01:07:47:18 – 01:07:48:05
It’s important

01:07:48:05 – 01:07:49:24
for all people to be included,

01:07:49:24 – 01:07:52:01
not just hire them for, you know,

01:07:52:01 – 01:07:53:10
a consulting position,

01:07:53:10 – 01:07:55:22
a temporary feedback on this,

01:07:55:22 – 01:07:58:00
but have them involved in the development

01:07:58:00 – 01:08:00:20
and design of this. It’s imperative.

01:08:00:20 – 01:08:03:20
It’s so important.

01:08:06:17 – 01:08:08:10
Okay.

01:08:08:10 – 01:08:10:00
For the next question,

01:08:10:00 – 01:08:12:22
automatic speech recognition.

01:08:12:22 – 01:08:16:19
ESR In comparison to eBay.

01:08:16:24 – 01:08:20:08
AI how comfortable

01:08:20:08 – 01:08:22:15
and how much can you trust these?

01:08:22:15 – 01:08:25:15
Two And can you compare the two?

01:08:25:22 – 01:08:28:02
Jeff Here I’d like to take that.

01:08:28:02 – 01:08:31:07
You know, it’s like comparing apples to

01:08:31:07 – 01:08:32:24
oranges, both of them are fruit,

01:08:32:24 – 01:08:34:15
but they are a little bit different.

01:08:34:15 – 01:08:37:22
So this is a very specific portion

01:08:37:22 – 01:08:39:07
of technology

01:08:39:07 – 01:08:42:05
and it focuses on translating from one

01:08:42:05 – 01:08:45:14
language to another.

01:08:45:14 – 01:08:46:17
So so

01:08:48:00 – 01:08:48:19
in its

01:08:48:19 – 01:08:51:20
form, so spoken speech into written form

01:08:52:02 – 01:08:54:24
Abe AI focuses on interpreting

01:08:54:24 – 01:08:57:20
and that is an automatic on the spot

01:08:57:20 – 01:09:00:17
that moment find production.

01:09:00:17 – 01:09:03:03
So that means we have information

01:09:03:03 – 01:09:07:07
feeding into the process

01:09:07:14 – 01:09:10:13
and speech recognition is just picking up

01:09:10:13 – 01:09:12:08
on the kid’s speech.

01:09:12:08 – 01:09:13:14
The context is different,

01:09:13:14 – 01:09:15:18
the information processing is different,

01:09:15:18 – 01:09:17:15
and most of the equivalent

01:09:17:15 – 01:09:20:14
for ASL and speech recognition

01:09:20:14 – 01:09:23:16
would be ASL are signed recognition

01:09:24:01 – 01:09:27:14
and that technology is part of A.I.

01:09:27:14 – 01:09:28:14
by A.I..

01:09:28:14 – 01:09:29:19
But Abe A.I.

01:09:29:19 – 01:09:31:03
includes a variety

01:09:31:03 – 01:09:32:05
other technologies

01:09:32:05 – 01:09:35:05
to help those components work together.

01:09:35:10 – 01:09:37:22
It picks up on that subject

01:09:37:22 – 01:09:40:22
processing, body language,

01:09:41:05 – 01:09:44:01
situational context, clues

01:09:44:01 – 01:09:45:23
and

01:09:45:23 – 01:09:48:23
pragmatics

01:09:49:04 – 01:09:50:13
are the concepts

01:09:50:13 – 01:09:53:13
of the transcriptions.

01:09:53:24 – 01:09:55:05
And just to add to that,

01:09:55:05 – 01:09:56:07
this is Tim here.

01:09:56:07 – 01:10:00:15
And in terms of speech recognition,

01:10:01:16 – 01:10:02:05
that is

01:10:02:05 – 01:10:05:05
something where we already see

01:10:05:15 – 01:10:06:19
a much more development

01:10:06:19 – 01:10:08:11
because there have been years

01:10:08:11 – 01:10:09:05
of investment

01:10:09:05 – 01:10:11:14
in the development of that technology

01:10:11:14 – 01:10:12:19
while sign language

01:10:12:19 – 01:10:14:24
recognition is behind.

01:10:14:24 – 01:10:18:07
And so that becomes an issue of equity.

01:10:18:14 – 01:10:20:01
And there’s a concern there

01:10:20:01 – 01:10:21:17
that with speech recognition,

01:10:21:17 – 01:10:23:20
because of the time and the investment

01:10:23:20 – 01:10:24:22
that’s already there,

01:10:24:22 – 01:10:26:09
we’ve spoken language,

01:10:26:09 – 01:10:28:15
caring people continue to benefit from,

01:10:28:15 – 01:10:29:10
while deaf

01:10:29:10 – 01:10:31:18
individuals cannot have the same access

01:10:31:18 – 01:10:33:09
or the same benefit.

01:10:33:09 – 01:10:35:00
We go back to, for example,

01:10:35:00 – 01:10:37:00
the invention of the telephone.

01:10:37:00 – 01:10:38:04
Of course, hearing

01:10:38:04 – 01:10:39:17
people were able to use the phone

01:10:39:17 – 01:10:42:02
for many years and enjoy that technology

01:10:42:02 – 01:10:43:06
until 80 years later.

01:10:43:06 – 01:10:44:21
We finally got the video phone

01:10:44:21 – 01:10:46:16
and these other forms of technology

01:10:46:16 – 01:10:46:24
where deaf

01:10:46:24 – 01:10:48:05
people could benefit

01:10:48:05 – 01:10:50:18
from the same kind of experience.

01:10:50:18 – 01:10:53:06
So we always have to look at

01:10:53:06 – 01:10:54:08
these situations

01:10:54:08 – 01:10:56:24
and make sure that there is funding

01:10:56:24 – 01:11:00:12
and that research is being done to try to

01:11:00:19 – 01:11:04:08
ensure that sign language is caught up

01:11:04:12 – 01:11:07:08
to what the hearing community is able

01:11:07:08 – 01:11:07:22
to enjoy.

01:11:11:08 – 01:11:12:23
As A next question is

01:11:12:23 – 01:11:14:06
regarding people

01:11:14:06 – 01:11:18:01
with intellectual disabilities, autism

01:11:18:17 – 01:11:21:06
learning disabilities, language

01:11:21:06 – 01:11:24:06
fluency

01:11:24:08 – 01:11:25:07
and the like.

01:11:25:07 – 01:11:28:18
How does this relate to Abe?

01:11:29:01 – 01:11:32:14
I will their abilities and disabilities

01:11:32:14 – 01:11:35:14
be included in this process?

01:11:36:23 – 01:11:38:24
I can take that, Yeah.

01:11:38:24 – 01:11:41:09
Thank you. This is a great question.

01:11:41:09 – 01:11:42:13
Similar

01:11:42:13 – 01:11:44:12
to what we were talking about earlier.

01:11:44:12 – 01:11:47:22
When it comes to, for example,

01:11:48:16 – 01:11:49:24
the FCC,

01:11:49:24 – 01:11:52:13
the Federal Communications Commission,

01:11:52:13 – 01:11:55:13
and talking about language itself

01:11:55:23 – 01:11:58:23
and speech recognition,

01:12:00:21 – 01:12:03:05
there’s understanding

01:12:03:05 – 01:12:05:05
of people

01:12:05:05 – 01:12:06:23
that have different backgrounds.

01:12:06:23 – 01:12:09:24
Same thing happens with this API

01:12:09:24 – 01:12:11:02
and the approach.

01:12:11:02 – 01:12:12:20
We want to make sure that we resolve

01:12:12:20 – 01:12:13:22
some of these issues.

01:12:13:22 – 01:12:14:24
We need to be able

01:12:14:24 – 01:12:16:17
develop something for them

01:12:16:17 – 01:12:19:20
to make sure that it’s successful.

01:12:20:02 – 01:12:23:02
And so it’s still to be seen.

01:12:23:10 – 01:12:24:20
We are not able

01:12:24:20 – 01:12:26:04
answer that at this time

01:12:26:04 – 01:12:28:04
unless there’s anyone else here involved.

01:12:28:04 – 01:12:30:02
But and from what I understand,

01:12:30:02 – 01:12:31:01
I think it’s still a hot

01:12:31:01 – 01:12:32:10
topic of discussion

01:12:32:10 – 01:12:33:12
and it’s something

01:12:33:12 – 01:12:34:17
that the community of people

01:12:34:17 – 01:12:36:13
is still talking about.

01:12:36:13 – 01:12:38:18
But yeah, great question

01:12:38:18 – 01:12:41:12
and saying, yes, Jeff had mentioned the

01:12:41:12 – 01:12:44:18
data, They’re good data and

01:12:45:23 – 01:12:47:00
good models.

01:12:47:00 – 01:12:50:08
So with that design process,

01:12:50:14 – 01:12:51:17
we have to prepare

01:12:51:17 – 01:12:53:09
for a variety of deaf members

01:12:53:09 – 01:12:54:15
in the community.

01:12:54:15 – 01:12:55:00
You know,

01:12:55:00 – 01:12:55:16
in the beginning

01:12:55:16 – 01:12:57:00
we’ve got to collect data

01:12:57:00 – 01:12:58:23
from the community at large

01:12:58:23 – 01:13:01:23
and make sure that it is

01:13:02:09 – 01:13:05:00
appropriate in moving forward

01:13:05:00 – 01:13:05:16
that it’s going to be

01:13:05:16 – 01:13:07:10
beneficial to the community.

01:13:07:10 – 01:13:08:18
Now, if we don’t include

01:13:08:18 – 01:13:09:20
those in the beginning,

01:13:09:20 – 01:13:12:10
that could be a problem with our models.

01:13:12:10 – 01:13:15:10
They won’t be prepared for that.

01:13:17:10 – 01:13:18:19
This is Teresa,

01:13:18:19 – 01:13:21:17
and I’d like to add a specific example

01:13:21:17 – 01:13:23:07
in terms of writing English.

01:13:23:07 – 01:13:25:01
So we see that I know

01:13:25:01 – 01:13:26:10
most people are probably familiar

01:13:26:10 – 01:13:29:10
with Chat. JPT

01:13:29:24 – 01:13:32:12
Currently you’re able to ask chat.

01:13:32:12 – 01:13:36:20
JPT and to develop

01:13:37:00 – 01:13:40:16
some draft of, for example,

01:13:40:16 – 01:13:44:24
plain language meaning the concept

01:13:44:24 – 01:13:46:19
of using it for people

01:13:46:19 – 01:13:49:16
with intellectual disabilities. So

01:13:51:05 – 01:13:52:07
that reflects the

01:13:52:07 – 01:13:53:23
importance of the inclusion

01:13:53:23 – 01:13:56:08
of people in design.

01:13:56:08 – 01:13:57:17
And also

01:13:57:17 – 01:14:00:05
when we are having these discussions

01:14:00:05 – 01:14:03:01
about how to design technology,

01:14:03:01 – 01:14:04:13
we have to ask the question

01:14:04:13 – 01:14:06:06
who’s going to be involved?

01:14:06:06 – 01:14:08:00
Because sometimes the people who are

01:14:08:00 – 01:14:10:01
there are not the people

01:14:10:01 – 01:14:11:01
that need to be involved.

01:14:11:01 – 01:14:12:05
So we want to make sure

01:14:12:05 – 01:14:13:15
that we recognizing

01:14:13:15 – 01:14:14:10
and not forgetting

01:14:14:10 – 01:14:15:16
about these individuals

01:14:15:16 – 01:14:18:16
and these various communities.

01:14:19:08 – 01:14:20:21
I mean, one thing

01:14:20:21 – 01:14:22:03
that I think about is,

01:14:22:03 – 01:14:23:23
for example, a CDI.

01:14:23:23 – 01:14:24:15
Often

01:14:24:15 – 01:14:25:23
we have seen the use

01:14:25:23 – 01:14:27:21
of a certified deaf interpreter

01:14:27:21 – 01:14:30:05
who comes into the situation

01:14:30:05 – 01:14:32:05
and we look at how that changes

01:14:32:05 – 01:14:34:01
and improves communication.

01:14:34:01 – 01:14:35:09
The experience for the deaf

01:14:35:09 – 01:14:37:04
individual is improved,

01:14:37:04 – 01:14:39:02
and I think we can see that benefit

01:14:39:02 – 01:14:40:02
and understand how it would

01:14:40:02 – 01:14:42:10
apply as well to I by I.

01:14:47:02 – 01:14:47:12
Okay.

01:14:47:12 – 01:14:51:03
Next question is related to

01:14:53:01 – 01:14:55:03
there are three different webinars

01:14:55:03 – 01:14:58:03
that you guys hosted

01:14:58:15 – 01:15:01:08
and during those webinars,

01:15:01:08 – 01:15:04:08
did you have a group

01:15:04:13 – 01:15:07:05
or did you have people

01:15:07:05 – 01:15:10:00
who were not using sign language

01:15:10:00 – 01:15:11:24
involved in the group

01:15:11:24 – 01:15:13:19
that they do not use sign language

01:15:13:19 – 01:15:18:00
at all to communicate, to discuss by I

01:15:23:22 – 01:15:26:12
and this is Emery here.

01:15:26:12 – 01:15:28:09
Another good question

01:15:28:09 – 01:15:32:00
and yeah, we had asked them for

01:15:32:12 – 01:15:34:04
to specifically identify

01:15:34:04 – 01:15:37:14
their level of sign skill and

01:15:38:01 – 01:15:39:06
I know that

01:15:39:06 – 01:15:40:19
the panel was talking about that.

01:15:40:19 – 01:15:43:23
I can’t remember exactly how they said,

01:15:43:23 – 01:15:44:16
but it was

01:15:44:16 – 01:15:45:08
the question

01:15:45:08 – 01:15:47:08
was about their level of signing.

01:15:47:08 – 01:15:50:14
And so the it was an open discussion

01:15:50:14 – 01:15:51:21
and we told everyone that

01:15:51:21 – 01:15:53:16
it was going to be in sign language.

01:15:53:16 – 01:15:55:09
But the percentage of people

01:15:55:09 – 01:15:57:07
who were not fluent in

01:15:57:07 – 01:15:59:10
sign, I’m not totally sure.

01:15:59:10 – 01:16:01:22
I don’t know if anyone has a different

01:16:01:22 – 01:16:03:08
recall, something different and

01:16:05:09 – 01:16:05:19
I don’t

01:16:05:19 – 01:16:06:19
think that we had

01:16:06:19 – 01:16:08:19
collected enough data about that.

01:16:08:19 – 01:16:11:15
I think it was hard for us to evaluate,

01:16:11:15 – 01:16:12:14
let alone

01:16:12:14 – 01:16:14:08
evaluate our own data

01:16:14:08 – 01:16:15:17
that we had collected.

01:16:15:17 – 01:16:18:17
So

01:16:21:08 – 01:16:24:08
and in terms of the captioning

01:16:24:08 – 01:16:26:17
and English based caption,

01:16:26:17 – 01:16:28:16
we were not focused on that.

01:16:28:16 – 01:16:29:05
We were more

01:16:29:05 – 01:16:30:19
focused on the sign language

01:16:30:19 – 01:16:32:10
interpreting aspect.

01:16:32:10 – 01:16:36:16
So we do already have a lot of discussion

01:16:36:21 – 01:16:38:00
happening

01:16:38:00 – 01:16:40:22
to the text based aspect of this,

01:16:40:22 – 01:16:43:13
but sign language, there are some gaps,

01:16:43:13 – 01:16:46:13
so we want to focus on that specifically.

01:16:51:00 – 01:16:54:00
Okay,

01:16:55:18 – 01:16:56:17
please help one moment

01:16:56:17 – 01:16:59:17
while through these questions.

01:17:04:07 – 01:17:04:16
Okay.

01:17:04:16 – 01:17:08:00
This question is regarding

01:17:08:00 – 01:17:11:00
the development of the curriculum

01:17:11:07 – 01:17:13:19
at the university level regarding

01:17:13:19 – 01:17:16:19
different interpreting processes

01:17:19:10 – 01:17:22:04
and

01:17:22:04 – 01:17:24:03
for VRA,

01:17:24:03 – 01:17:26:24
VRA and the like.

01:17:26:24 – 01:17:29:21
Now are we going to be adding a I

01:17:31:02 – 01:17:34:01
and there any feedback

01:17:34:01 – 01:17:36:00
that you could share regarding

01:17:36:00 – 01:17:39:12
curriculum development on this subject?

01:17:47:08 – 01:17:48:17
Okay, that’s a good question.

01:17:48:17 – 01:17:49:07
Again,

01:17:49:07 – 01:17:53:09
I think maybe for a future discussion,

01:17:53:09 – 01:17:54:24
it might be more beneficial.

01:17:54:24 – 01:17:55:14
We haven’t

01:17:55:14 – 01:17:57:08
necessarily gotten to that point yet.

01:17:57:08 – 01:17:59:00
We’ll have to do some training

01:17:59:00 – 01:17:59:19
with the curriculum

01:17:59:19 – 01:18:01:10
and workshops and all of.

01:18:01:10 – 01:18:02:03
But again,

01:18:02:03 – 01:18:02:21
I would say

01:18:02:21 – 01:18:06:12
that it could come up at a symposium

01:18:06:22 – 01:18:09:17
or maybe another opportunity

01:18:09:17 – 01:18:13:20
for discussion could become available

01:18:13:20 – 01:18:16:20
in the future.

01:18:21:17 – 01:18:23:24
We haven’t discussed this as of yet,

01:18:23:24 – 01:18:27:05
but in considering the

01:18:27:05 – 01:18:30:05
ethical foundations of this.

01:18:30:10 – 01:18:33:04
So for interpreting,

01:18:33:04 – 01:18:36:16
we do have the ethical foundation

01:18:36:23 – 01:18:38:00
for the Deaf community.

01:18:38:00 – 01:18:42:02
We do have our ethical expectations

01:18:42:20 – 01:18:44:12
and norms

01:18:44:12 – 01:18:46:05
as well as other aspects of that.

01:18:46:05 – 01:18:48:01
But when we enter into this

01:18:48:01 – 01:18:48:24
with some knowledge,

01:18:48:24 – 01:18:52:13
now we have new technology

01:18:52:18 – 01:18:54:07
and what kind of new questions

01:18:54:07 – 01:18:56:05
are going to arise from this?

01:18:56:05 – 01:18:57:17
And that’s part of what we’re hoping

01:18:57:17 – 01:18:59:24
to have some discussions regarding

01:18:59:24 – 01:19:03:01
that topic in the next symposium

01:19:03:01 – 01:19:03:17
next month.

01:19:07:14 – 01:19:08:04
wow, I forgot.

01:19:08:04 – 01:19:09:18
It’s March already.

01:19:09:18 – 01:19:12:18
It is.

01:19:14:16 – 01:19:15:00
Okay.

01:19:15:00 – 01:19:16:08
Next question.

01:19:16:08 – 01:19:19:01
One person had a comment

01:19:19:01 – 01:19:20:13
and said, a lot is happening

01:19:20:13 – 01:19:20:23
right now

01:19:20:23 – 01:19:24:11
in Europe as it relates to AI by AI.

01:19:24:18 – 01:19:28:04
Are you familiar with that and also how

01:19:28:04 – 01:19:32:16
that relates to D GDP?

01:19:33:14 – 01:19:36:11
R And so that’s

01:19:36:11 – 01:19:39:17
the general data protection

01:19:41:01 – 01:19:44:06
regulation to that law.

01:19:44:15 – 01:19:47:20
It’s a very strict

01:19:48:00 – 01:19:51:00
law on privacy and protection.

01:19:51:09 – 01:19:52:08
So can you guys

01:19:52:08 – 01:19:55:08
discuss in touch a bit on that topic?

01:19:56:11 – 01:19:57:12
I would like to do that.

01:19:57:12 – 01:19:59:14
Thank you for bringing that up.

01:19:59:14 – 01:20:04:16
In general, the data has been

01:20:04:18 – 01:20:07:10
it’s been one of the best

01:20:07:10 – 01:20:09:17
data privacy laws.

01:20:09:17 – 01:20:13:07
Hats off to them to the EU

01:20:13:07 – 01:20:14:10
for developing that.

01:20:14:10 – 01:20:15:09
It’s been wonderful

01:20:15:09 – 01:20:17:17
so I think one of the biggest concepts

01:20:17:17 – 01:20:19:02
or take away from this

01:20:19:02 – 01:20:22:02
is the topic of the right

01:20:22:05 – 01:20:24:02
to be forgotten.

01:20:24:02 – 01:20:25:02
Meaning

01:20:25:02 – 01:20:27:06
we can remind them and say, Hey,

01:20:27:06 – 01:20:29:17
I want you to remove my information

01:20:29:17 – 01:20:32:17
and they have to honor your request.

01:20:32:18 – 01:20:35:00
And that is one of the biggest takeaways

01:20:35:00 – 01:20:37:23
for this foundational concept.

01:20:37:23 – 01:20:40:13
And so another thing to look at as

01:20:40:13 – 01:20:43:08
well is the minors.

01:20:43:08 – 01:20:46:08
The age of the data collection.

01:20:46:10 – 01:20:48:21
Parents have to approve that or

01:20:50:03 – 01:20:51:15
revoke approval.

01:20:51:15 – 01:20:52:14
And so really,

01:20:52:14 – 01:20:54:17
in some of these aspects, the EU

01:20:54:17 – 01:20:55:20
ahead of us

01:20:55:20 – 01:20:59:12
and, you know, the American legislature

01:20:59:12 – 01:21:01:22
does have some protections,

01:21:01:22 – 01:21:03:15
but it’s not as focused.

01:21:03:15 – 01:21:06:07
It’s more focused. Children under age.

01:21:06:07 – 01:21:09:07
I believe it’s 13 or 14,

01:21:10:03 – 01:21:11:15
but the EU is ahead of us

01:21:11:15 – 01:21:13:07
in this respect.

01:21:13:07 – 01:21:16:01
The legislation there

01:21:16:01 – 01:21:17:09
there’s a lot of harm

01:21:17:09 – 01:21:19:10
that can be done with this data.

01:21:19:10 – 01:21:20:21
And so that’s something that really needs

01:21:20:21 – 01:21:22:09
to be fleshed out with anyone else.

01:21:22:09 – 01:21:25:09
Like to add to that,

01:21:30:02 – 01:21:31:05
I could go on for,

01:21:31:05 – 01:21:32:23
for ages about this,

01:21:32:23 – 01:21:34:14
but it’s the biggest thing

01:21:34:14 – 01:21:35:15
since

01:21:35:15 – 01:21:39:09
the organization had been established.

01:21:39:09 – 01:21:44:01
The Data collection and privacy

01:21:44:01 – 01:21:47:20
and who is responsible for making sure

01:21:47:20 – 01:21:50:20
that the data is retained safely,

01:21:51:02 – 01:21:53:12
that it’s not leaked?

01:21:53:12 – 01:21:56:16
And if that data is leaked,

01:21:56:18 – 01:21:59:17
how are they going to inform individuals

01:21:59:17 – 01:22:02:13
whose data has been breached that?

01:22:02:13 – 01:22:04:01
This has occurred.

01:22:04:01 – 01:22:05:21
So that’s part of the process

01:22:05:21 – 01:22:07:14
that should be included

01:22:07:14 – 01:22:09:07
in the transparency

01:22:09:07 – 01:22:12:00
and, you know, maintaining that contact

01:22:12:00 – 01:22:15:11
with individuals on the subject.

01:22:16:16 – 01:22:19:16
Thank you.

01:22:21:17 – 01:22:22:04
Okay.

01:22:22:04 – 01:22:25:04
We still have more questions.

01:22:26:02 – 01:22:27:17
This question is related

01:22:27:17 – 01:22:30:17
to machine learning AML.

01:22:30:20 – 01:22:32:24
So and with sign language

01:22:32:24 – 01:22:35:24
recognition, ASL are

01:22:36:19 – 01:22:38:15
who will be training

01:22:38:15 – 01:22:41:15
and teaching the language data.

01:22:41:22 – 01:22:43:17
Where is it going to come from?

01:22:43:17 – 01:22:46:05
From interpreters, from people?

01:22:46:05 – 01:22:49:05
Where will that come from?

01:22:51:01 – 01:22:52:14
So that is an excellent question.

01:22:52:14 – 01:22:53:03
Again,

01:22:53:03 – 01:22:54:07
that is something that

01:22:54:07 – 01:22:56:20
we can’t really control.

01:22:56:20 – 01:22:59:15
It’s up to the company

01:22:59:15 – 01:23:01:18
who is developing that

01:23:01:18 – 01:23:03:06
and doing the work.

01:23:03:06 – 01:23:06:06
So I know that in academia

01:23:06:08 – 01:23:07:19
there’s a lot of evaluations

01:23:07:19 – 01:23:08:24
on authors,

01:23:08:24 – 01:23:12:11
like, for example Oscar Kilner.

01:23:12:19 – 01:23:15:06

01:23:15:06 – 01:23:19:14
Lopez was an author in the ASL,

01:23:19:14 – 01:23:23:14
are community and used interpreter data.

01:23:23:23 – 01:23:27:24
For example, there was a very well-known

01:23:29:03 – 01:23:31:00
individual from Germany

01:23:31:00 – 01:23:34:03
and they had

01:23:35:05 – 01:23:37:11
a a system

01:23:37:11 – 01:23:38:14
that would do

01:23:38:14 – 01:23:40:03
weather reports and alerts

01:23:40:03 – 01:23:41:06
and it had an interpreter

01:23:41:06 – 01:23:42:13
down in the corner.

01:23:42:13 – 01:23:45:17
They recorded that for many years

01:23:46:07 – 01:23:49:07
and they used that data to train

01:23:49:07 – 01:23:51:11
the machine.

01:23:51:11 – 01:23:55:09
And so there was a very limited context

01:23:55:09 – 01:23:57:23
only using that one interpreter.

01:23:57:23 – 01:24:01:02
So it’s a great idea in theory

01:24:01:02 – 01:24:04:02
in the future to use,

01:24:04:19 – 01:24:06:04
you know, in other situations,

01:24:06:04 – 01:24:08:01
it may impact it in that way.

01:24:08:01 – 01:24:11:01
It’s very hard

01:24:11:24 – 01:24:14:11
because the data requires

01:24:14:11 – 01:24:16:16
retention and storage

01:24:16:16 – 01:24:19:16
and a lot of video data available

01:24:19:16 – 01:24:20:10
on the web out

01:24:20:10 – 01:24:23:04
there is not the best quality.

01:24:23:04 – 01:24:24:03
For example,

01:24:24:03 – 01:24:27:03
if you look at like an ASL, one class

01:24:27:05 – 01:24:31:15
student signing hey or a song or whatever

01:24:31:21 – 01:24:33:07
you’re looking and you’re saying,

01:24:33:07 – 01:24:34:07
Hey, you’re doing a good job,

01:24:34:07 – 01:24:35:11
you’re learning, you’re doing well.

01:24:35:11 – 01:24:36:15
But that’s not the model

01:24:36:15 – 01:24:37:22
that we want to use in

01:24:37:22 – 01:24:39:02
training the machines

01:24:39:02 – 01:24:40:03
for machine learning.

01:24:41:04 – 01:24:44:04
And the issue that arises, you know,

01:24:44:07 – 01:24:47:02
is a standard for A.I.,

01:24:47:02 – 01:24:47:12
you know,

01:24:47:12 – 01:24:48:09
where do we get this

01:24:48:09 – 01:24:49:08
data collection from?

01:24:49:08 – 01:24:50:18
From various,

01:24:50:18 – 01:24:51:02
you know,

01:24:51:02 – 01:24:53:06
and that is a bidirectional approach

01:24:53:06 – 01:24:53:24
to interpreting.

01:24:53:24 – 01:24:55:16
But there’s a big problem with that

01:24:55:16 – 01:24:56:18
as well.

01:24:56:18 – 01:24:58:11
Consent and privacy

01:24:58:11 – 01:25:02:01
and confidentiality are highly regulated.

01:25:02:01 – 01:25:04:10
And so we can’t use VR if

01:25:04:10 – 01:25:06:02
even though that would be the best place

01:25:06:02 – 01:25:07:13
for data collection.

01:25:07:13 – 01:25:10:19
So we have both signed and speech

01:25:11:06 – 01:25:13:04
and we are working on

01:25:13:04 – 01:25:15:04
how the two interact

01:25:15:04 – 01:25:19:05
and that’s what is really good.

01:25:19:05 – 01:25:20:06
But we’ve got to look

01:25:20:06 – 01:25:21:11
at different sources

01:25:21:11 – 01:25:23:18
and the organizations themselves.

01:25:23:18 – 01:25:26:12
It’s not

01:25:26:12 – 01:25:28:01
we’ve we’ve got to choose

01:25:28:01 – 01:25:28:22
what data we use.

01:25:28:22 – 01:25:30:16
It’s a very important to start

01:25:30:16 – 01:25:34:07
thinking about our legal framework

01:25:35:12 – 01:25:36:23
and, try to

01:25:36:23 – 01:25:39:23
encourage that and remind them to use

01:25:40:00 – 01:25:40:07
you know,

01:25:40:07 – 01:25:43:19
we’ve got to make sure that the data set

01:25:44:07 – 01:25:48:11
will be include from different sources.

01:25:48:17 – 01:25:50:13
The accuracy is there,

01:25:50:13 – 01:25:52:09
the variety is there.

01:25:52:09 – 01:25:54:06
And the group models,

01:25:54:06 – 01:25:57:22
for example, may be there

01:25:57:22 – 01:26:00:07
may be some fluent individuals,

01:26:00:07 – 01:26:01:14
deaf individuals there.

01:26:01:14 – 01:26:02:13
And that’s not the best

01:26:02:13 – 01:26:03:23
planning model to have.

01:26:03:23 – 01:26:05:03
I know that I’m not the best

01:26:05:03 – 01:26:06:07
signing model to have.

01:26:06:07 – 01:26:08:02
I’m not perfectly fluent myself.

01:26:08:02 – 01:26:09:18
It’s my native language

01:26:09:18 – 01:26:12:10
and I want to understand me.

01:26:12:10 – 01:26:15:02
And the same applies to other

01:26:15:02 – 01:26:16:14
signing styles

01:26:16:14 – 01:26:18:23
and signers with different abilities.

01:26:18:23 – 01:26:20:08
So there’s so many groups

01:26:20:08 – 01:26:21:13
that need to be included

01:26:21:13 – 01:26:24:13
and represented in this machine learning.

01:26:30:01 – 01:26:31:11
And to add to that,

01:26:31:11 – 01:26:32:23
it can be very challenging,

01:26:32:23 – 01:26:35:23
but also it can become an opportunity.

01:26:36:00 – 01:26:36:20
So for example,

01:26:36:20 – 01:26:38:01
we have a lot of deaf

01:26:38:01 – 01:26:39:14
individuals all over

01:26:39:14 – 01:26:42:04
their stories, their history.

01:26:42:04 – 01:26:43:06
A lot of times

01:26:43:06 – 01:26:44:20
this is not shared

01:26:44:20 – 01:26:47:00
and so we want to collect and and

01:26:47:00 – 01:26:48:03
save that data.

01:26:49:22 – 01:26:50:21
So I think it becomes a

01:26:50:21 – 01:26:51:15
bit of a project

01:26:51:15 – 01:26:54:15
to see what the impact is.

01:26:54:15 – 01:26:55:23
Understanding stories

01:26:55:23 – 01:26:58:24
and being able to record this history.

01:26:58:24 – 01:27:00:12
And also it’s an opportunity

01:27:00:12 – 01:27:01:16
to take a deeper dive

01:27:01:16 – 01:27:05:02
into these generational situations and,

01:27:05:11 – 01:27:06:10
getting information

01:27:06:10 – 01:27:08:21
from all over from diverse groups.

01:27:08:21 – 01:27:11:03
But I think the challenge can be funding,

01:27:11:03 – 01:27:13:13
and it’s also challenging to find people

01:27:13:13 – 01:27:14:21
who are able to go out

01:27:14:21 – 01:27:16:21
and record this information

01:27:16:21 – 01:27:18:02
and a quality way.

01:27:18:02 – 01:27:20:20
And so that’s a project in and of itself.

01:27:20:20 – 01:27:22:24
But we have this opportunity now

01:27:22:24 – 01:27:25:24
to look at different domains

01:27:26:03 – 01:27:27:24
and how to use sign language

01:27:27:24 – 01:27:30:17
and the medical and legal legal realms.

01:27:30:17 – 01:27:31:24
And we can see how

01:27:31:24 – 01:27:35:06
that becomes bigger project.

01:27:36:04 – 01:27:39:04
But again,

01:27:40:09 – 01:27:43:12
I think we need to go into that more

01:27:43:12 – 01:27:44:18
and we need to see where

01:27:44:18 – 01:27:46:01
we can get that funding from,

01:27:46:01 – 01:27:47:20
because that’s one of the challenges.

01:27:53:00 – 01:27:53:15
Okay.

01:27:53:15 – 01:27:55:10
Next question.

01:27:55:10 – 01:28:00:02
We have a question from hang on.

01:28:00:02 – 01:28:03:02
I left a deaf interpreter

01:28:05:02 – 01:28:06:02
and the question

01:28:06:02 – 01:28:10:11
is using AI captioning or work

01:28:10:11 – 01:28:13:23
or for a meeting or something of the like

01:28:14:15 – 01:28:15:20
are one of the many tools

01:28:15:20 – 01:28:17:00
that have been tested

01:28:17:00 – 01:28:18:11
with other countries

01:28:18:11 – 01:28:21:11
and their language in their access.

01:28:21:11 – 01:28:23:15
But it seems to fail

01:28:23:15 – 01:28:26:07
or it’s not as accurate

01:28:26:07 – 01:28:28:09
at capturing everything.

01:28:28:09 – 01:28:31:09
What approach would you use for A.I.

01:28:31:10 – 01:28:35:04
by making sure it’s accurate and

01:28:36:06 – 01:28:39:06
successful?

01:28:44:15 – 01:28:45:05
It’s all about

01:28:45:05 – 01:28:48:05
the data collection.

01:28:48:06 – 01:28:49:07
This is the memory here.

01:28:49:07 – 01:28:52:03
I think. Again, a great question.

01:28:52:03 – 01:28:53:14
And to expand on that,

01:28:53:14 – 01:28:55:17
I think we see the same situation

01:28:55:17 – 01:28:57:11
with live captioning

01:28:57:11 – 01:29:01:00
that is nuanced voice tone.

01:29:01:04 – 01:29:04:04
Sign language is just

01:29:04:04 – 01:29:05:12
we have to see how we’re going

01:29:05:12 – 01:29:07:18
to approach that when it comes to AI,

01:29:08:17 – 01:29:10:11
a AI is already challenging

01:29:10:11 – 01:29:12:09
for American Sign language

01:29:12:09 – 01:29:14:03
because it’s a conceptual

01:29:14:03 – 01:29:16:06
and a visual language.

01:29:16:06 – 01:29:18:12
So to be able to capture that

01:29:18:12 – 01:29:21:02
and replicate it and I you know,

01:29:21:02 – 01:29:22:02
this is a great question

01:29:22:02 – 01:29:24:05
because how can we approach this

01:29:24:05 – 01:29:25:15
to make this happen?

01:29:25:15 – 01:29:27:07
It’s still a hot topic.

01:29:27:07 – 01:29:28:17
There’s a lot of discussion on this

01:29:28:17 – 01:29:31:10
because people are still on this process

01:29:31:10 – 01:29:32:14
of trying to screen

01:29:32:14 – 01:29:35:00
and figure out how this pertains to A.I..

01:29:35:00 – 01:29:37:04
But yeah,

01:29:37:04 – 01:29:38:14
Tim, here I’d like to add,

01:29:38:14 – 01:29:40:13
if you look at different fields,

01:29:40:13 – 01:29:43:13
for example, linguistic studies,

01:29:43:16 – 01:29:45:01
the sign language,

01:29:45:01 – 01:29:48:23
it starts in about 1960,

01:29:49:07 – 01:29:51:23
68 with Stokie and his team.

01:29:51:23 – 01:29:53:23
And so you look at ASL

01:29:53:23 – 01:29:55:06
and you look at that

01:29:55:06 – 01:29:57:05
in how it’s developed over time,

01:29:57:05 – 01:29:58:17
and it’s really only been studied

01:29:58:17 – 01:30:01:02
in depth for 50 to 60 years.

01:30:01:02 – 01:30:03:12
It’s in its infancy at best.

01:30:03:12 – 01:30:05:15
And so there’s so many fine languages

01:30:05:15 – 01:30:06:11
all around the world

01:30:06:11 – 01:30:07:17
that have not been studied

01:30:07:17 – 01:30:09:05
and have not been documented

01:30:09:05 – 01:30:10:13
to that degree.

01:30:10:13 – 01:30:12:24
So we still need some more research.

01:30:12:24 – 01:30:15:24
And where you know,

01:30:16:08 – 01:30:17:12
it’s very important

01:30:17:12 – 01:30:19:13
for us to really look at that

01:30:19:13 – 01:30:20:18
with science languages,

01:30:20:18 – 01:30:22:10
but also with many spoken languages

01:30:22:10 – 01:30:23:07
as well.

01:30:23:07 – 01:30:24:24
There’s, you know,

01:30:24:24 – 01:30:26:21
thousands of languages in the world

01:30:26:21 – 01:30:29:06
and they don’t have a written form

01:30:29:06 – 01:30:31:07
for every one of them.

01:30:31:07 – 01:30:32:22
And so

01:30:32:22 – 01:30:34:10
there’s a lot of minority languages

01:30:34:10 – 01:30:36:18
as well and dialects.

01:30:36:18 – 01:30:40:04
And so those are at risk for extinction

01:30:40:11 – 01:30:41:18
because they’re not written,

01:30:41:18 – 01:30:43:07
they’re not studied,

01:30:43:07 – 01:30:44:24
they’re not documented.

01:30:44:24 – 01:30:46:08
And in a few years,

01:30:46:08 – 01:30:47:13
the most common languages

01:30:47:13 – 01:30:50:02
that will be used are major languages.

01:30:50:02 – 01:30:51:06
In the minority languages

01:30:51:06 – 01:30:52:22
will have dissipated.

01:30:52:22 – 01:30:56:03
And so that is something that is a risk.

01:30:56:03 – 01:30:59:23
And we have to look at we can’t leave

01:30:59:23 – 01:31:00:21
those behind.

01:31:04:20 – 01:31:05:03
I’d like

01:31:05:03 – 01:31:06:15
to add to that comment

01:31:06:15 – 01:31:08:24
when we talk about data.

01:31:08:24 – 01:31:10:16
I made a comment recently about that

01:31:10:16 – 01:31:11:03
I was wrong.

01:31:11:03 – 01:31:13:12
I should have said that more data

01:31:13:12 – 01:31:15:08
is coming from

01:31:15:08 – 01:31:17:18
underrepresented communities

01:31:17:18 – 01:31:19:14
and that we need to identify

01:31:19:14 – 01:31:20:20
those communities

01:31:20:20 – 01:31:23:03
and make them aware

01:31:23:03 – 01:31:24:19
that we’d like them to collaborate

01:31:24:19 – 01:31:25:16
with us

01:31:25:16 – 01:31:26:12
to make sure

01:31:26:12 – 01:31:29:12
that they are represented in the data.

01:31:38:24 – 01:31:39:09
Okay.

01:31:39:09 – 01:31:40:20
Tim recently made a comment

01:31:40:20 – 01:31:42:14
about research,

01:31:42:14 – 01:31:45:17
and I’d like to piggyback

01:31:45:17 – 01:31:46:19
my question off of that.

01:31:46:19 – 01:31:49:19
Are there any publications, books,

01:31:49:24 – 01:31:51:14
resources, articles

01:31:51:14 – 01:31:53:21
related to the issue of A.I.

01:31:53:21 – 01:31:58:01
and the Deaf community in Conflict

01:31:59:17 – 01:32:02:10
and I’m sorry, together

01:32:02:10 – 01:32:05:04
intersection.

01:32:05:04 – 01:32:08:03
Can you spell that again?

01:32:08:03 – 01:32:09:20
I’m sorry.

01:32:09:20 – 01:32:10:16
Intersection.

01:32:10:16 – 01:32:13:16
Intersection.

01:32:13:18 – 01:32:16:13
Okay, so my focus in the research

01:32:16:13 – 01:32:19:19
is more on the technical side of things.

01:32:19:19 – 01:32:22:15
I’m not necessarily too heavy on the

01:32:23:14 – 01:32:24:23
on the other

01:32:24:23 – 01:32:26:20
area with the deaf community

01:32:26:20 – 01:32:28:19
and looking into the socio side of it.

01:32:28:19 – 01:32:29:18
But I do think it’s

01:32:29:18 – 01:32:32:21
very interesting research and

01:32:34:02 – 01:32:36:01
I think it’s a good way to do

01:32:36:01 – 01:32:39:08
some formal search of the literature.

01:32:39:14 – 01:32:40:21
So I would suggest

01:32:40:21 – 01:32:42:07
that if you’re interested in this,

01:32:42:07 – 01:32:43:17
that you look at things

01:32:43:17 – 01:32:45:02
like, for example,

01:32:45:02 – 01:32:48:20
Richard Dana or Danny Bragg

01:32:49:04 – 01:32:51:01
and they both have worked

01:32:51:01 – 01:32:54:02
focus on the ethical aspect

01:32:54:22 – 01:32:59:24
of these and the surrounding topics.

01:32:59:24 – 01:33:01:04
And so it’s a lot,

01:33:01:04 – 01:33:02:08
but I could probably

01:33:02:08 – 01:33:06:13
share a bibliography for you all

01:33:06:13 – 01:33:07:16
so that you could take a look

01:33:07:16 – 01:33:08:15
at those authors

01:33:08:15 – 01:33:11:15
and look more into their work.

01:33:13:21 – 01:33:15:14
I wish we could,

01:33:15:14 – 01:33:15:24
you know,

01:33:15:24 – 01:33:17:09
with the advisory group,

01:33:17:09 – 01:33:20:03
we touch on so many things

01:33:20:03 – 01:33:21:20
and I wish we could keep discussing

01:33:21:20 – 01:33:25:01
how the Germans are.

01:33:25:01 – 01:33:26:23
The shared study and research

01:33:26:23 – 01:33:28:04
really impacts everything.

01:33:30:00 – 01:33:30:19
Good idea.

01:33:30:19 – 01:33:33:19
Good idea.

01:33:40:01 – 01:33:40:16
Okay.

01:33:40:16 – 01:33:43:05
Any more questions?

01:33:43:05 – 01:33:45:07
Okay, I do have

01:33:45:07 – 01:33:47:09
I have more questions here,

01:33:47:09 – 01:33:50:09
so I will

01:33:50:23 – 01:33:53:08
copy exactly from what

01:33:53:08 – 01:33:54:11
I’m seeing here on the question,

01:33:54:11 – 01:33:55:15
this might be a good question

01:33:55:15 – 01:33:57:07
for Jeff to answer.

01:33:57:07 – 01:34:00:08
So, Jeff, we’re wondering

01:34:00:08 – 01:34:02:15
if the deaf community

01:34:02:15 – 01:34:05:15
is open to this system.

01:34:06:05 – 01:34:09:05
So for each individual user,

01:34:11:19 – 01:34:14:06
will they all be trained

01:34:14:06 – 01:34:17:19
and taught in their language of ASL?

01:34:17:19 – 01:34:18:14
For example,

01:34:18:14 – 01:34:20:07
if there’s a website,

01:34:20:07 – 01:34:23:07
well, that have American Sign language

01:34:23:09 – 01:34:26:09
and will it be copied and saved,

01:34:26:18 – 01:34:30:07
and will there be anything like

01:34:30:07 – 01:34:31:11
if this is saved,

01:34:31:11 – 01:34:32:16
will personal information

01:34:32:16 – 01:34:34:14
be saved in a server?

01:34:34:14 – 01:34:37:04
How will they ensure confidentiality?

01:34:37:04 – 01:34:37:16
While deaf

01:34:37:16 – 01:34:41:12
individuals have personal rights

01:34:41:12 – 01:34:43:06
to say yes, I’m

01:34:43:06 – 01:34:44:19
okay with releasing my information

01:34:44:19 – 01:34:46:14
at this company, How will that work?

01:34:48:14 – 01:34:50:10
Yeah, that’s a great question.

01:34:50:10 – 01:34:53:04
I think the formal term for

01:34:53:04 – 01:34:56:04
that is called fine tuning.

01:34:57:13 – 01:35:00:08
So with fine tuning you can go in

01:35:00:08 – 01:35:03:13
and make sure that the model is a fit

01:35:03:13 – 01:35:05:12
for your personal style,

01:35:05:12 – 01:35:08:11
your personal terms of choice.

01:35:08:11 – 01:35:11:15
So we can do that currently with English.

01:35:11:15 – 01:35:13:07
When it comes to Elm,

01:35:13:07 – 01:35:15:08
large language models like for example,

01:35:15:08 – 01:35:18:19
Chat, GPT, you can go in and personalize

01:35:18:19 – 01:35:20:01
and bind tune

01:35:20:01 – 01:35:21:23
the data that comes out of it.

01:35:21:23 – 01:35:23:15
And so I would imagine

01:35:23:15 – 01:35:24:09
that the same thing

01:35:24:09 – 01:35:25:09
will eventually happen

01:35:25:09 – 01:35:28:18
when it comes to AI by AI and also

01:35:28:18 – 01:35:31:18
in terms of sharing with others.

01:35:31:23 – 01:35:34:19
I don’t say I don’t see any reason why

01:35:34:19 – 01:35:36:01
it would take a long time

01:35:36:01 – 01:35:36:14
to be able

01:35:36:14 – 01:35:38:07
to for people

01:35:38:07 – 01:35:40:01
to give their informed consent.

01:35:40:01 – 01:35:41:15
So it is your data

01:35:41:15 – 01:35:42:20
and you’ll be able to do

01:35:42:20 – 01:35:44:22
whatever you want to do with it.

01:35:44:22 – 01:35:46:23
That would be the idea.

01:35:46:23 – 01:35:48:02
I don’t know if anyone has

01:35:48:02 – 01:35:49:08
anything else to add to that

01:35:51:14 – 01:35:52:16
Emery here, I

01:35:52:16 – 01:35:54:05
would like to add a comment.

01:35:54:05 – 01:35:55:19
I think that

01:35:55:19 – 01:35:58:19
the community feels empowered

01:35:59:14 – 01:36:02:11
to have these options available.

01:36:02:11 – 01:36:05:00
So I think whether or not

01:36:05:00 – 01:36:07:02
everyone agrees or not,

01:36:07:02 – 01:36:08:18
the idea of having options

01:36:08:18 – 01:36:11:10
available to personalize their data

01:36:11:10 – 01:36:12:23
I think would be optimal

01:36:12:23 – 01:36:15:23
and well received in general

01:36:17:16 – 01:36:19:01
to say, okay, you know,

01:36:19:01 – 01:36:20:15
I see that this is happening,

01:36:20:15 – 01:36:22:10
but I’m just one person.

01:36:22:10 – 01:36:25:01
But that would be my my guess.

01:36:25:01 – 01:36:26:17
This is Tim.

01:36:26:17 – 01:36:28:22
I remember back in what was it, 2001

01:36:28:22 – 01:36:30:11
when I was in college,

01:36:30:11 – 01:36:33:02
we had speech recognition software

01:36:33:02 – 01:36:35:05
and it was called Dragon

01:36:35:05 – 01:36:37:04
Natural Speaking.

01:36:37:04 – 01:36:41:04
And so that required a lot of training,

01:36:41:12 – 01:36:44:24
a lot of feeding into the technology to

01:36:44:24 – 01:36:46:00
develop it.

01:36:46:00 – 01:36:47:22
But there was some control

01:36:47:22 – 01:36:49:18
of sound and clarity.

01:36:49:18 – 01:36:51:24
But I think today the technology

01:36:51:24 – 01:36:54:01
seems to have gotten even better,

01:36:54:01 – 01:36:57:06
and so there’s less of a curve there.

01:36:57:06 – 01:36:59:09
But I think it would be possible

01:36:59:09 – 01:37:01:16
for bias to still be involved.

01:37:01:16 – 01:37:03:09
And so, of course, we see

01:37:03:09 – 01:37:04:16
that it’s quite standard English,

01:37:04:16 – 01:37:06:05
especially in particular tool

01:37:06:05 – 01:37:07:03
I was just discussing.

01:37:07:03 – 01:37:10:08
So when it comes to dialect and accents

01:37:10:08 – 01:37:11:07
and that kind of thing,

01:37:11:07 – 01:37:12:13
for my understanding,

01:37:12:13 – 01:37:14:17
the bigger concern is coming from

01:37:14:17 – 01:37:15:20
the greater community

01:37:15:20 – 01:37:16:14
based on what we see

01:37:16:14 – 01:37:18:03
on speech recognition.

01:37:18:03 – 01:37:20:02
So we can only imagine

01:37:20:02 – 01:37:20:21
what it might look like

01:37:20:21 – 01:37:21:19
with sign language.

01:37:21:19 – 01:37:23:08
So with over 20 years

01:37:23:08 – 01:37:25:11
invested in these types of tools

01:37:25:11 – 01:37:27:10
and and seeing how that goes,

01:37:27:10 – 01:37:28:04
I think that

01:37:28:04 – 01:37:29:23
we need to take into consideration

01:37:29:23 – 01:37:32:04
that we would need better technology

01:37:32:04 – 01:37:35:04
and we would need, you know, that,

01:37:35:14 – 01:37:37:13
for example, leapfrog technology

01:37:37:13 – 01:37:40:13
where we would come in and

01:37:40:20 – 01:37:41:24
be able to say,

01:37:41:24 – 01:37:43:04
we’ve already seen these things

01:37:43:04 – 01:37:45:05
develop, we’ve seen how this has gone.

01:37:45:05 – 01:37:46:22
So maybe we wouldn’t

01:37:46:22 – 01:37:48:23
need as much time to catch up.

01:37:48:23 – 01:37:49:24
I would hope,

01:37:49:24 – 01:37:50:22
like I mentioned, that

01:37:50:22 – 01:37:53:22
that dragon technology’s like 20 years in

01:37:53:22 – 01:37:57:22
so I think that I’m not necessarily

01:37:57:22 – 01:38:00:12
into the technical aspect as much as,

01:38:00:12 – 01:38:01:02
but that’s just

01:38:01:02 – 01:38:02:17
based on my experience time.

01:38:02:17 – 01:38:04:06
I predict it could go.

01:38:07:12 – 01:38:08:24
Holly says the person that asked

01:38:08:24 – 01:38:10:08
that question added a comment

01:38:10:08 – 01:38:12:11
and said, Yes,

01:38:12:11 – 01:38:14:09
I know exactly what you’re talking about.

01:38:14:09 – 01:38:15:09
The dragon, naturally

01:38:15:09 – 01:38:16:10
speaking technology.

01:38:16:10 – 01:38:18:20
I’m familiar with that.

01:38:18:20 – 01:38:20:14
Okay.

01:38:20:14 – 01:38:23:14
Another question says,

01:38:23:23 – 01:38:26:23
I am a deaf leader in my community.

01:38:28:13 – 01:38:32:19
How can we build more of a

01:38:33:04 – 01:38:34:17
bond,

01:38:34:17 – 01:38:37:18
a stronger bond with the interpreters?

01:38:38:03 – 01:38:40:14
Because many of them are afraid

01:38:40:14 – 01:38:43:14
to reach out to the deaf?

01:38:45:01 – 01:38:47:08
And said General question.

01:38:47:08 – 01:38:49:10
Emery says, Can you repeat the question,

01:38:49:10 – 01:38:52:10
please?

01:38:53:08 – 01:38:54:13
It’s a general question.

01:38:54:13 – 01:38:55:16
Holly says,

01:38:55:16 – 01:38:58:16
I am a deaf leader in my community

01:38:59:10 – 01:39:02:06
and I want to know how we as the deaf

01:39:02:06 – 01:39:05:07
community, can build stronger bonds

01:39:06:19 – 01:39:09:18
with interpreters.

01:39:09:18 – 01:39:12:17
Many interpreters afraid,

01:39:12:17 – 01:39:15:17
and they don’t reach out to us and from

01:39:15:17 – 01:39:16:11
the deaf community,

01:39:19:01 – 01:39:21:14
Tim says.

01:39:21:14 – 01:39:24:08
I think that

01:39:24:08 – 01:39:26:18
it’s all comes back to trust

01:39:26:18 – 01:39:28:10
trust issues.

01:39:28:10 – 01:39:29:23
So from my understanding,

01:39:29:23 – 01:39:32:06
many deaf people are afraid

01:39:32:06 – 01:39:34:03
and resistant to technology

01:39:34:03 – 01:39:36:17
because they’re afraid that

01:39:36:17 – 01:39:38:03
this technology

01:39:38:03 – 01:39:39:10
would take away their ability

01:39:39:10 – 01:39:40:20
to have an informed choice

01:39:40:20 – 01:39:42:00
to make decisions.

01:39:42:00 – 01:39:44:03
Same things with same thing with video

01:39:44:03 – 01:39:45:05
remote interpreting.

01:39:45:05 – 01:39:47:04
A lot of deaf people didn’t want that

01:39:47:04 – 01:39:48:00
because,

01:39:48:00 – 01:39:48:05
you know,

01:39:48:05 – 01:39:49:23
they have to fight the system

01:39:49:23 – 01:39:51:22
to get an in-person interpreter.

01:39:51:22 – 01:39:53:11
So based on that experience

01:39:53:11 – 01:39:54:16
and, those challenges,

01:39:54:16 – 01:39:57:16
it’s caused a lot of issues with trust.

01:39:57:19 – 01:40:00:01
And

01:40:00:01 – 01:40:00:16
I know

01:40:00:16 – 01:40:03:03
especially when it comes to health care,

01:40:03:03 – 01:40:04:12
it’s already challenging

01:40:04:12 – 01:40:05:14
to make appointments

01:40:05:14 – 01:40:06:07
to come

01:40:06:07 – 01:40:09:07
in, to have access and all of that.

01:40:09:10 – 01:40:12:20
So it can be very exhausting and cause

01:40:12:20 – 01:40:13:15
deaf individuals

01:40:13:15 – 01:40:14:04
to feel like

01:40:14:04 – 01:40:15:08
they don’t want to go to the doctor

01:40:15:08 – 01:40:16:09
because they don’t want to deal

01:40:16:09 – 01:40:17:14
with all of that.

01:40:17:14 – 01:40:19:17
Now, when it comes to interpreters,

01:40:19:17 – 01:40:23:10
of course, there are some things

01:40:23:10 – 01:40:24:00
to consider,

01:40:24:00 – 01:40:25:21
like the code of ethics, code of coverage

01:40:25:21 – 01:40:26:18
reality.

01:40:26:18 – 01:40:28:13
And I think deaf people may feel,

01:40:28:13 – 01:40:29:09
you know what,

01:40:29:09 – 01:40:32:09
I would prefer to have a I, because

01:40:32:23 – 01:40:35:12
in this case, there’s no baggage.

01:40:35:12 – 01:40:36:15
I don’t have to deal with

01:40:36:15 – 01:40:39:15
the human aspect of trust issues.

01:40:39:16 – 01:40:42:11
I know that with AI

01:40:42:11 – 01:40:45:08
and that kind of collaboration and dialog

01:40:45:08 – 01:40:47:12
also, it can lead to the discussion

01:40:47:12 – 01:40:48:18
of what the meaning

01:40:48:18 – 01:40:50:12
of trust is, what confidentiality

01:40:50:12 – 01:40:51:14
should look like,

01:40:51:14 – 01:40:53:24
the code of ethics, how that applies,

01:40:53:24 – 01:40:55:04
and just to make sure

01:40:55:04 – 01:40:58:02
that if our expectations of AI

01:40:58:02 – 01:41:00:02
are high and

01:41:01:14 – 01:41:02:15
we need to know

01:41:02:15 – 01:41:03:09
those accurate

01:41:03:09 – 01:41:04:20
or what could we expect something

01:41:04:20 – 01:41:06:08
that we’ve had with the human experience.

01:41:06:08 – 01:41:09:08
So that’s another discussion to have.

01:41:10:15 – 01:41:12:03
Theresa I don’t know if you want to add

01:41:12:03 – 01:41:13:01
to that.

01:41:13:01 – 01:41:15:02
Theresa says Yes, I’m just thinking.

01:41:15:02 – 01:41:18:23
I think that maybe the first step

01:41:19:02 – 01:41:21:18
would be to start these discussions

01:41:21:18 – 01:41:24:15
and to have some time in smaller

01:41:24:15 – 01:41:25:16
communities

01:41:25:16 – 01:41:28:08
where everyone knows each other, right?

01:41:28:08 – 01:41:29:11
We are all familiar

01:41:29:11 – 01:41:30:23
with those types of situations

01:41:30:23 – 01:41:32:07
where the interpreters and all the deaf

01:41:32:07 – 01:41:33:04
people and the deaf people,

01:41:33:04 – 01:41:34:07
not the interpreters,

01:41:34:07 – 01:41:37:22
but maybe we would start with some kind

01:41:37:22 – 01:41:39:12
of, let’s say, for example,

01:41:39:12 – 01:41:43:07
have your local deaf

01:41:43:07 – 01:41:47:17
community groups and local or I.D.

01:41:47:19 – 01:41:49:10
or interpreting organizations

01:41:49:10 – 01:41:50:20
come together.

01:41:50:20 – 01:41:51:13
So, for example,

01:41:51:13 – 01:41:52:13
maybe they come together

01:41:52:13 – 01:41:53:18
and watch this video

01:41:53:18 – 01:41:55:17
and then they host a discussion

01:41:55:17 – 01:41:57:02
and ask questions.

01:41:57:02 – 01:41:59:22
But I’m just thinking, how can we start

01:41:59:22 – 01:42:01:17
to develop this discussion

01:42:01:17 – 01:42:03:02
and this dialog?

01:42:03:02 – 01:42:05:05
Because without the dialog,

01:42:05:05 – 01:42:07:08
there’s so many misunderstandings.

01:42:07:08 – 01:42:08:19
And I think that

01:42:08:19 – 01:42:10:19
with the dialog, misunderstandings

01:42:10:19 – 01:42:12:08
will still happen as well,

01:42:12:08 – 01:42:14:06
but is definitely an opportunity

01:42:14:06 – 01:42:14:23
to have

01:42:14:23 – 01:42:16:00
more understanding

01:42:16:00 – 01:42:17:03
and more of an opportunity

01:42:17:03 – 01:42:18:15
to listen to each other

01:42:18:15 – 01:42:20:20
and figure out how we can discuss

01:42:20:20 – 01:42:23:09
this together is huge.

01:42:23:09 – 01:42:24:13
It’s coming

01:42:24:13 – 01:42:26:08
and we need to be sure

01:42:26:08 – 01:42:28:18
that we know how to respond to this.

01:42:28:18 – 01:42:30:08
We need to start with

01:42:30:08 – 01:42:32:24
the grassroots community

01:42:32:24 – 01:42:34:08
and go from there.

01:42:34:08 – 01:42:35:22
So the grassroots

01:42:35:22 – 01:42:36:16
community is the heart

01:42:36:16 – 01:42:37:16
of our deaf community.

01:42:43:04 – 01:42:44:04
Thank you, Anne Marie.

01:42:44:04 – 01:42:44:24
And Jeff, would you

01:42:44:24 – 01:42:46:13
do you have anything you want to add?

01:42:46:13 – 01:42:49:13
Just saying I agree wholeheartedly.

01:42:50:04 – 01:42:51:20
Emery Here,

01:42:51:20 – 01:42:55:18
this one topic itself is just it.

01:42:56:03 – 01:42:57:09
It’s huge.

01:42:57:09 – 01:42:59:09
There’s no way to describe it.

01:42:59:09 – 01:43:00:05
Otherwise,

01:43:00:05 – 01:43:01:21
there are so many things

01:43:01:21 – 01:43:04:17
to look at the process itself,

01:43:04:17 – 01:43:08:02
the trust in the process, the transport

01:43:08:03 – 01:43:11:03
tenancy, the data collection, all of it

01:43:11:06 – 01:43:12:17
together.

01:43:12:17 – 01:43:13:17
But you know,

01:43:13:17 – 01:43:16:06
and how everything moves in tandem.

01:43:16:06 – 01:43:17:06
It’s an opportunity

01:43:17:06 – 01:43:20:06
to always create a safe space

01:43:20:23 – 01:43:23:06
for the community to come together

01:43:23:06 – 01:43:24:17
and to discuss.

01:43:24:17 – 01:43:25:19
And it’s important

01:43:25:19 – 01:43:26:08
for us

01:43:26:08 – 01:43:29:08
to look at that process as a whole.

01:43:31:13 – 01:43:33:04
Very good discussion.

01:43:33:04 – 01:43:35:11
Very good discussion,

01:43:35:11 – 01:43:35:24
Ali, saying,

01:43:35:24 – 01:43:38:24
okay, next question is about the research

01:43:38:24 – 01:43:42:14
process from last October.

01:43:42:14 – 01:43:45:14
You had three different webinar sessions,

01:43:45:18 – 01:43:48:15
the deaf participants in those sessions.

01:43:48:15 – 01:43:49:24
What were their backgrounds,

01:43:49:24 – 01:43:52:19
where were they from geographically?

01:43:52:19 – 01:43:54:07
And

01:43:55:22 – 01:43:58:22
demographically?

01:43:59:03 – 01:44:00:20
Just saying, go ahead.

01:44:00:20 – 01:44:01:07
Emery

01:44:01:07 – 01:44:03:21
So the backgrounds were very diverse.

01:44:03:21 – 01:44:05:18
We had some that were deaf interpreters,

01:44:05:18 – 01:44:07:13
deaf consumers,

01:44:07:13 – 01:44:11:05
deaf individuals who are professionals

01:44:11:10 – 01:44:14:08
working in the field of education,

01:44:14:08 – 01:44:16:07
working in the field of interpreting

01:44:16:07 – 01:44:17:21
very different varieties

01:44:17:21 – 01:44:20:21
of backgrounds and

01:44:21:12 – 01:44:23:04
areas as well.

01:44:23:04 – 01:44:24:22
Jeff Did you want to add some more?

01:44:24:22 – 01:44:25:20
Jeff Yes,

01:44:25:20 – 01:44:26:22
I think the next step

01:44:26:22 – 01:44:28:22
is to decide how we expand

01:44:28:22 – 01:44:30:11
and how we grow our audience

01:44:30:11 – 01:44:32:05
from those webinars

01:44:32:05 – 01:44:34:00
and to really include

01:44:34:00 – 01:44:35:21
even more of the community at large,

01:44:35:21 – 01:44:38:21
to have all those perspectives as well

01:44:40:06 – 01:44:41:09
in the saying yes,

01:44:41:09 – 01:44:42:23
the webinars were really

01:44:42:23 – 01:44:47:15
our first step into this

01:44:47:15 – 01:44:49:19
realm of testing out this,

01:44:49:19 – 01:44:50:21
looking at different things,

01:44:50:21 – 01:44:52:02
and this symposium

01:44:52:02 – 01:44:53:04
will just continue

01:44:53:04 – 01:44:55:21
to be a springboard into the future

01:44:55:21 – 01:44:57:17
and as long as we have a

01:44:57:17 – 01:44:59:13
I will be having these discussions

01:45:00:24 – 01:45:01:16
in raising.

01:45:01:16 – 01:45:02:11
I’d like to add

01:45:02:11 – 01:45:06:05
also that all of the individuals,

01:45:06:13 – 01:45:08:03
the participants here today,

01:45:08:03 – 01:45:09:16
you guys are critical

01:45:09:16 – 01:45:11:16
for this process as well.

01:45:11:16 – 01:45:13:01
We are not finished now

01:45:13:01 – 01:45:14:15
that this report is published.

01:45:14:15 – 01:45:16:05
This is just the beginning.

01:45:16:05 – 01:45:18:12
US and even all of your questions

01:45:18:12 – 01:45:20:04
have really spurred our thoughts

01:45:20:04 – 01:45:22:04
into how we move this forward

01:45:22:04 – 01:45:23:04
and what the next steps

01:45:23:04 – 01:45:24:06
are in this process.

01:45:24:06 – 01:45:27:06
We’ve just scratched the surface.

01:45:34:04 – 01:45:35:21
Okay, I’ll help you think.

01:45:35:21 – 01:45:36:23
We have navigated

01:45:36:23 – 01:45:38:06
through all of the questions

01:45:38:06 – 01:45:39:22
that we have for the day.

01:45:39:22 – 01:45:45:02
We do have one more, but it is

01:45:45:07 – 01:45:46:17
with this

01:45:46:17 – 01:45:49:17
webinar, complete this report, complete

01:45:49:21 – 01:45:52:01
the sharing of the recording

01:45:52:01 – 01:45:53:10
When will that be done?

01:45:53:10 – 01:45:57:08
The Deaf Safe

01:45:57:16 – 01:46:01:10
Advisory website, the report,

01:46:01:16 – 01:46:03:05
if that will be shared,

01:46:03:05 – 01:46:06:09
and how to register

01:46:06:09 – 01:46:09:09
for the Brown University Symposium.

01:46:09:16 – 01:46:10:14
Several people have asked

01:46:10:14 – 01:46:12:06
for that information as well.

01:46:14:18 – 01:46:17:11
Tim here, so I will answer that.

01:46:17:11 – 01:46:21:10
The report we have both the safe

01:46:21:18 – 01:46:25:19
I their report as well as ours,

01:46:25:19 – 01:46:27:00
and we’ve been working

01:46:27:00 – 01:46:29:08
with Katha research.

01:46:29:08 – 01:46:31:10
They have spent so much time

01:46:31:10 – 01:46:32:21
going through this project.

01:46:32:21 – 01:46:36:04
The surveys, volunteers, so much work

01:46:36:04 – 01:46:37:13
has gone into that report

01:46:37:13 – 01:46:38:04
and that will

01:46:38:04 – 01:46:40:00
and will be published as well.

01:46:40:00 – 01:46:41:15
We had a tech

01:46:41:15 – 01:46:43:12
a presentation scheduled yesterday,

01:46:43:12 – 01:46:45:11
but there was technological problems

01:46:45:11 – 01:46:47:00
and so we’re going to be rescheduling

01:46:47:00 – 01:46:47:24
that for Wednesday morning,

01:46:47:24 – 01:46:50:24
I believe, at 11 Eastern time.

01:46:51:00 – 01:46:54:02
And so with that being said,

01:46:54:02 – 01:46:55:13
I really encourage you all

01:46:55:13 – 01:46:57:12
to watch that presentation.

01:46:57:12 – 01:46:59:08
It’s about languages in general,

01:46:59:08 – 01:47:00:14
not just sign languages,

01:47:00:14 – 01:47:01:23
it’s languages in general.

01:47:01:23 – 01:47:03:17
And so that will be made

01:47:03:17 – 01:47:05:10
available to the public as well.

01:47:05:10 – 01:47:06:06
This one,

01:47:06:06 – 01:47:09:17
I believe we have some editing to do

01:47:09:23 – 01:47:12:10
and we will make it available here soon.

01:47:12:10 – 01:47:13:10
The presentation

01:47:13:10 – 01:47:14:01
next Wednesday

01:47:14:01 – 01:47:16:07
will also be available online.

01:47:16:07 – 01:47:18:22
Our website does have links

01:47:18:22 – 01:47:21:22
to Safe A.I.

01:47:22:01 – 01:47:23:11
Advisory Group,

01:47:23:11 – 01:47:25:04
and then we have our own Death

01:47:25:04 – 01:47:27:04
Advisory Group page

01:47:27:04 – 01:47:29:02
and the links to those.

01:47:29:02 – 01:47:30:08
I will thin them out

01:47:30:08 – 01:47:32:07
and make it available there

01:47:32:07 – 01:47:34:16
also the symposium

01:47:34:16 – 01:47:37:16
we are currently working on that platform

01:47:37:20 – 01:47:38:22
and the save

01:47:38:22 – 01:47:40:12
the date is just announced today.

01:47:40:12 – 01:47:42:16
This is our first announcement for that

01:47:42:16 – 01:47:44:19
I will send out a more formal

01:47:44:19 – 01:47:45:14
save the date

01:47:45:14 – 01:47:47:22
with information on registration

01:47:47:22 – 01:47:49:02
and the like.

01:47:49:02 – 01:47:50:05
It will be sent out.

01:47:50:05 – 01:47:52:08
That is a currently in process.

01:47:52:08 – 01:47:53:19
So Look forward to that.

01:47:55:06 – 01:47:55:21
I do.

01:47:55:21 – 01:47:56:20
Any of the other

01:47:56:20 – 01:47:58:06
advisory council members

01:47:58:06 – 01:47:58:21
have something

01:47:58:21 – 01:48:01:10
that they would like to add.

01:48:01:10 – 01:48:02:05
Theresa Thing

01:48:02:05 – 01:48:03:23
I just want to thank all of you

01:48:03:23 – 01:48:06:12
and Mary saying, yes, I agree. Thank you.

01:48:06:12 – 01:48:08:09
Thank you for your interest

01:48:08:09 – 01:48:09:04
in this topic.

01:48:09:04 – 01:48:10:04
Thank you for coming

01:48:10:04 – 01:48:12:03
and listening to our presentation.

01:48:12:03 – 01:48:13:15
We Appreciate it.

01:48:13:15 – 01:48:15:09
We we think that it’s wonderful

01:48:15:09 – 01:48:16:19
that all of you were involved

01:48:16:19 – 01:48:18:22
and we really appreciate everyone

01:48:18:22 – 01:48:20:12
that was involved in the study

01:48:20:12 – 01:48:21:04
just said yes.

01:48:21:04 – 01:48:24:04
Thank you so much.

01:48:24:11 – 01:48:25:17
Thank you, Holly, as well

01:48:25:17 – 01:48:26:21
for your time today.

01:48:26:21 – 01:48:29:06
We appreciate you joining us

01:48:29:06 – 01:48:32:06
in this presentation.

01:48:32:07 – 01:48:35:07
Thank you very much,

01:48:37:06 – 01:48:37:12
Holly.

01:48:37:12 – 01:48:37:19
Same.

01:48:37:19 – 01:48:39:09
I want to make sure that we have a clear

01:48:39:09 – 01:48:40:13
answer about

01:48:40:13 – 01:48:40:22

01:48:40:22 – 01:48:43:10
if this recording of the webinar

01:48:43:10 – 01:48:44:22
will be broadcast

01:48:44:22 – 01:48:46:09
and then

01:48:46:09 – 01:48:47:17
we will be sending up

01:48:47:17 – 01:48:49:06
a follow up email as well.

01:48:49:06 – 01:48:49:19
For everyone

01:48:49:19 – 01:48:50:16
who registered

01:48:50:16 – 01:48:53:16
with the website information

01:48:53:16 – 01:48:56:24
and Brown University Symposium

01:48:58:07 – 01:48:59:23
information

01:48:59:23 – 01:49:03:02
and what else was there.

01:49:03:02 – 01:49:03:24
I think that’s it.

01:49:03:24 – 01:49:06:24
So we will correct?

01:49:07:15 – 01:49:08:21
Yes, simply yes,

01:49:08:21 – 01:49:09:24
we will all just say yes,

01:49:09:24 – 01:49:12:24
it will be all available.

01:49:15:22 – 01:49:18:15
The webinar also that we have,

01:49:18:15 – 01:49:19:23
we have the video

01:49:19:23 – 01:49:21:03
recordings of the webinars,

01:49:21:03 – 01:49:22:13
if you’d like to see those as well.

01:49:22:13 – 01:49:25:13
We those available.

01:49:29:23 – 01:49:31:24
All right,

01:49:31:24 – 01:49:33:18
Tim thing, I believe this concludes

01:49:33:18 – 01:49:36:18
our meeting.

01:49:38:24 – 01:49:40:10
Just saying Thank you, everyone.

01:49:40:10 – 01:49:42:06
We appreciate your time.

01:49:42:06 – 01:49:43:18
Thank you so much. Bye bye.

Are Deaf Communities Ready for AI Interpreting?

3 Videos
Use our hashtag #DeafSafeAI when sharing on social media!

The Advisory Group on AI and Sign Language Interpreting is supported by the following organizations:

Liaison Team Members

Eileen Forestal, PhD

Portrait of Star Grieser

Star Grieser

Chief Executive Officer at Registry of Interpreters for the Deaf

Star is a Member of the Stakeholders Assembly of the SAFE AI Task Force and the CEO of RID since July of 2021. Star graduated from the Rochester Institute of Technology with a B.S. in Professional and Technical Communication, and McDaniel College with a Master’s in Deaf Education (2001) and been active in advocacy and has worked among the Deaf and interpreting communities, be it medical, mental health care, Deaf education, legislative advocacy, interpreter education, etc., before becoming the Director of Testing for CASLI in 2017, and the CEO of RID 2021. She currently holds a RID certification as a CDI and is also ICE-CCP.

Portrait of Steph Kent

Stephanie Jo Kent, PhD

Community Interpreter, American Sign Language and English at Learning Lab for Resiliency®

Stephanie Jo Kent, CI, PhD, is a Member of the Stakeholders Assembly of the SAFE AI Task Force. Steph has a dual career in professional sign language interpreting and academic-activism for the social good. The combination of experience inside “the ivory tower” while also interpreting “in the real world” informs her/their unique facilitation of action research as group-level growth and change, especially how plurilingual communication can leverage diverse lived experiences into collaborations for co-creating a humane and exhilarating future. 

Portrait of Timothy Riker

Timothy Riker

Senior Lecturer in American Sign Language at Brown University

Tim Riker is a Member of the Advisory Group (AG) on AI and Sign Language Interpreting. He is a Senior Lecturer in American Sign Language at Brown University and research co-investigator with the DeafYES! Center for Deaf Empowerment and Recovery. Through his Deaf-led team’s community-engaged research, Riker has collaborated to develop linguistically and sociopolitically correct methods of inquiry when conducting Deaf qualitative research. The research team leveraged technology so they could analyze sign language data while reducing bias found in traditional methods of inquiry.

Erin Sanders-Sigmon

Advisory Group Members

Portrait of Teresa Blankmeyer Burke

Teresa Blankmeyer Burke

Professor of Philosophy and Bioethicist at Gallaudet University

Teresa Blankmeyer Burke is Professor of Philosophy and Bioethics at Gallaudet University. Her published research areas are bioethics, including the ethics of genetic technology; ethics of signed language interpreting; Deaf philosophy; philosophy of disability; and the ethics of emerging technologies, including AI. She has provided her expertise to a variety of entities, including the United Nations, World Federation of the Deaf, National Council on Disability (USA), the Canadian Association of Sign Language Interpreters, the Registry of Interpreters for the Deaf (USA) and is co-Editor in Chief of the Journal of Philosophy of Disability.

Portrait of Ryan Commerson

Ryan Commerson

Product Strategist at Sorenson

An experienced corporate leader in marketing and product strategy, Ryan Commerson resides in Colorado Springs, Colorado with a partner of 20 years and a perfect mess of a dog. During his free time, he plays in the mountains and trains for ultras and triathlons.

Portrait of Ryan Hait-Campbell

Ryan Hait-Campbell

Creative Strategist at GoSign.AI

Ryan Hait-Campell is a seasoned creative strategist with a decade of experience, recognized for innovative contributions to emerging technologies. His accolades include being named one of Time Magazine’s 25 Best Inventions of 2014 among others.

Currently, Ryan operates as a Program Lead at www.convorelay.com where he guides new product initiatives aligning with global expansion strategies. On the side, Ryan is one of the owners of www.gosign.ai where they focus on pioneering AI solutions for deaf and hearing individuals, breaking accessibility barriers worldwide.

You can learn more about Ryan at his website www.ryanhc.com

Portrait of Travis Dougherty

Travis Dougherty

Chair at National Association of State Relay Administration

Travis Dougherty is a dedicated advocate for communication accessibility, serving as the Chair of the National Association of State Relay Administration and the Maryland Relay Manager. His work focuses on enhancing communication services for the deaf and hard of hearing community. Additionally, Travis is an AI & Sign Language Recognition Consultant, where his expertise bridges technology with accessibility. As a serial entrepreneur, he leverages his knowledge and passion to innovate and advocate for more inclusive communication solutions, making significant strides in the intersection of technology and accessibility.

Portrait of Cody Francisco

M. Cody Francisco

Director, Deaf and Hard-of-Hearing at MasterWord Services, Inc.

M. Cody Francisco serves as the Director of Interpreting Services: Deaf and Hard of Hearing at MasterWord in Houston, Texas. He completed his undergraduate studies at Oklahoma State University. Furthermore, Cody earned his second Bachelor’s in Social Work & Master’s in Multidisciplinary Studies with a specialization in Deaf Studies and Counseling from the Rochester Institute of Technology. With an extensive 18-year career, Cody has held leadership positions in various capacities, including roles at a prominent Video Relay Service provider, Interpreting Agencies, and Interpreter Training Programs. He is also a Certified Deaf Interpreter

Portrait of Abraham Glasser

Abraham Glasser

Assistant Professor at Gallaudet University

Dr. Abraham Glasser is a Deaf at birth, native ASL signer from Rochester, NY. He attended RIT from 2015-2018 for his B.S. in Computer Science (mathematics minor), and 2019-2022 for his PhD in Computing and Information Sciences. His dissertation was about ASL interaction with personal assistant devices (e.g. Google Assistant and Alexa), and he has over 30 peer-reviewed academic publications.

He is currently an Assistant Professor at Gallaudet University, where he is committed to growing the new M.S. in Accessible Human-Centered Computing program, as well as continuing Human-Computer Interaction and accessible technology research.

Portrait of Holly Jackson

Holly Jackson, MEd, BEI

NAOBI-Atlanta, Inc.

Holly was an IT Consultant for 10 years before switching her career to interpreting. She was a staff interpreter at Georgia State University for 8 years but now teaches in the ASL and Deaf Studies programs. She has worked in VRS and community settings for over 10 years and presents annually to K-12 schools for Career Day. She has a B.S. in Computer Science from Clark Atlanta University, an ITP Certificate from Georgia Perimeter College, an M.Ed. in ASL-English Interpretation from University of North Florida, and has trained/worked with Deaf-Blind individuals at Helen Keller National Center in New York. She is BEI-certified, ASLPI 4.

Portrait of Nikolas Kelly

Nikolas Kelly

Co-Founder & Chief Product Officer at Sign-Speak, Inc.

Nikolas Kelly is a Missouri native and is Deaf. ASL is his first language. Nikolas Kelly is Sign-Speak’s Co-founder and Chief Product Officer where he shapes the product with the Deaf and Hard of Hearing community. Nikolas obtained a BS in Supply Chain Management. He studied for an MBA with a concentration in Data Analytics from the National Technical Institute of the Deaf (NTID) and the Rochester Institute of Technology (RIT). He started Sign-Speak with Nicholas Wilkins and Yami Payano to solve an unmet market demand and remove communication hurdles that he and others in the community face.

Portrait of David Kerr

David Kerr

Executive Director at Canadian Association of Sign Language Interpreters

David has been an activist in the Deaf community and a leader within various Deaf organizations since the 1980s. With over 25 years of experience in senior management, David has worked in interpreting services as well as the delivery of social and health services. He held senior management positions including Executive Director of Durham Deaf Services, Regional Director with the Canadian Hearing Society, Deputy Director of DCARA, and is currently the Executive Director of the Canadian Association of Sign Language Interpreters (CASLI).

Portrait of AnnMarie Killian

AnnMarie Killian

CEO at TDIforAccess (TDI)

AnnMarie Killian is a trailblazing leader and advocate with over 20 years of experience
championing accessibility and inclusion in information communication technology. (ICT) As the Chief Executive Officer of TDIforAccess (TDI), AnnMarie’s remarkable career has been marked by her unwavering commitment to advancing communication access for the Deaf, DeafBlind and hard of hearing community. She is deeply involved in shaping and influencing public policy through her engagement with the FCC Disability Advisory Council (DAC).

Portrait of Silvia Muturi

Silvia Muturi

Silvia is a social entrepreneur passionate about the Deaf community and partners with the public and private sectors to champion Deaf rights. With over 5 years of experience in providing research, data collection, and analysis, she has helped design programs on inclusion and diversity for organizations working in marginalized communities.

Ms. Muturi has published articles on best practices regarding the Deaf in the justice, social media, and transport sectors. She lives with her teenage son and loves nothing better than to dance, listen to music, read a good book, and travel.

Portrait of Jeff Shaul

Jeff Shaul

Co-Founder at GoSign.AI

I am from Cincinnati, Ohio. I am interested in developing novel approaches to data farming for accessibility applications. Along with Ryan Hait-Campbell and Calvin Young, we cofounded GoSign.AI, a company dedicated to collecting data of the sign languages of the world. Currently, there is great disparity in the robustness of AI tools designed for the hearing and those for sign language users. We aim to help fix that.

Portrait of Neal Tucker

Neal Tucker

Director of Government Affairs, Public Policy & Advocacy at Registry of Interpreters for the Deaf (RID)

Neal has spent over a decade of his career working in areas intersecting with disability rights and government affairs at local, state, and federal levels. His passion for influencing and implementing positive change is the driving force behind his professional pursuits with the approach of, “Leave the world a better place than you found it.” He is honored to work for RID knowing that he and his team have a direct impact on the lives of the diverse Deaf and ASL using communities.