Legislature(2023 - 2024)GRUENBERG 120
03/22/2024 01:00 PM House JUDICIARY
Note: the audio
and video
recordings are distinct records and are obtained from different sources. As such there may be key differences between the two. The audio recordings are captured by our records offices as the official record of the meeting and will have more accurate timestamps. Use the icons to switch between them.
| Audio | Topic |
|---|---|
| Start | |
| Confirmation Hearing(s): Alaska State Commission for Human Rights | |
| HB358 | |
| Adjourn |
* first hearing in first committee of referral
+ teleconferenced
= bill was previously heard/scheduled
+ teleconferenced
= bill was previously heard/scheduled
| += | HB 358 | TELECONFERENCED | |
| += | HB 107 | TELECONFERENCED | |
| + | TELECONFERENCED | ||
| + | TELECONFERENCED |
ALASKA STATE LEGISLATURE
HOUSE JUDICIARY STANDING COMMITTEE
March 22, 2024
1:16 p.m.
MEMBERS PRESENT
Representative Sarah Vance, Chair
Representative Ben Carpenter
Representative Craig Johnson
Representative Jesse Sumner
Representative Andrew Gray
Representative Cliff Groh
MEMBERS ABSENT
Representative Jamie Allard, Vice Chair
COMMITTEE CALENDAR
CONFIRMATION HEARING(S)
Alaska State Commission for Human Rights
Rebecca Carrillo - Juneau
- CONFIRMATION(S) ADVANCED
HOUSE BILL NO. 358
"An Act relating to use of artificial intelligence to create or
alter a representation of the voice or likeness of an
individual."
- HEARD & HELD
HOUSE BILL NO. 107
"An Act relating to criminal law definitions."
- BILL HEARING CANCELED
PREVIOUS COMMITTEE ACTION
BILL: HB 358
SHORT TITLE: PROHIBIT AI-ALTERED REPRESENTATIONS
SPONSOR(s): REPRESENTATIVE(s) CRONK
02/20/24 (H) READ THE FIRST TIME - REFERRALS
02/20/24 (H) JUD
03/13/24 (H) JUD AT 1:00 PM GRUENBERG 120
03/13/24 (H) Heard & Held
03/13/24 (H) MINUTE(JUD)
03/15/24 (H) JUD AT 1:00 PM GRUENBERG 120
03/15/24 (H) Heard & Held
03/15/24 (H) MINUTE(JUD)
03/20/24 (H) JUD AT 1:00 PM GRUENBERG 120
03/20/24 (H) <Bill Hearing Canceled>
03/22/24 (H) JUD AT 1:00 PM GRUENBERG 120
WITNESS REGISTER
REBECCA CARRILLO, Appointee
Alaska State Commission for Human Rights
Juneau, Alaska
POSITION STATEMENT: Testified as appointee to the Alaska State
Commission for Human Rights.
BOB BALLINGER, Staff
Representative Sarah Vance
Alaska State Legislature
Juneau, Alaska
POSITION STATEMENT: Gave an explanation of changes in the
proposed committee substitute (CS) for HB 358, Version U, on
behalf of Representative Vance.
DAVE STANCLIF, Staff
Representative Mike Cronk
Alaska State Legislature
Juneau, Alaska
POSITION STATEMENT: Answered questions during the hearing on
CSHB 358, Version U, on behalf of Representative Cronk, prime
sponsor.
IAN WALSH, Attorney
Legislative Legal Services
Legislative Affairs Agency
Juneau, Alaska
POSITION STATEMENT: Answered questions during the hearing on
CSHB 358, Version U.
ACTION NARRATIVE
1:16:23 PM
CHAIR VANCE called the House Judiciary Standing Committee
meeting to order at 1:16 p.m. Representatives Carpenter, C.
Johnson, and Vance were present at the call to order.
Representatives Sumner, Groh, and Gray arrived as the meeting
was in progress.
^CONFIRMATION HEARING(S): Alaska State Commission for Human
Rights
CONFIRMATION HEARING(S): Alaska State Commission for Human
Rights
1:17:00 PM
CHAIR VANCE announced that the first order of business would be
a confirmation hearing for Rebecca Carrillo to the Alaska State
Commission for Human Rights (ASCHR).
1:17:24 PM
REBECCA CARRILLO, Appointee, Alaska State Commission for Human
Rights, shared that this was her second appointment to ASCHR.
She said she would like to continue the work she began three
years ago on the commission, which was a continuation of the
work she conducted as a 32-year employee of the state on the
Alaska Supplemental Food Program for Women and Children. She
stated that ASCHR dealt with issues that are important to her as
an individual and a minority. She added that the commission
resolves issues, such as discrimination and inequality, in
manner that is meaningful, fair, and equitable.
1:20:05 PM
REPRESENTATIVE CARPENTER asked Ms. Carrillo to share one thing
that she's proud of and one thing that she looks forward to
achieving.
MS. CARRILLO stated that she is proud of the commission's
conversations about relevant issues, such as improving education
for children with visual disabilities and advocating for the
elderly. She added that as a commission, ASCHR looks forward to
facilitating fairness and improving the quality of life for
these particular demographics.
REPRESENTATIVE CARPENTER inquired about the most recent issues
that ASCHR had been working on.
MS. CARRILLO stated that ASCHR had recently moved offices to
improve staff and client safety. In addition, the commission
was instituting automated services and advertising to better
serve the community.
1:26:45 PM
REPRESENTATIVE CARPENTER asked Ms. Carrillo what strengths she
would bring to the commission.
MS. CARRILLO explained that as a minority, she had lived in two
different worlds: one in the Philippines, a third world country
with many challenges, and Juneau, Alaska. She said she had
navigated different cultures and experienced discrimination in
many forms.
CHAIR VANCE opened public testimony on the appointee. After
ascertaining that no one wished to testify, she closed public
testimony.
1:30:56 PM
CHAIR VANCE moved to advance the confirmation of Rebecca
Carrillo to the Alaska State Commission for Human Rights to the
joint session of the House and Senate for consideration. He
reminded members that signing the reports regarding appointment
to boards and commissions in no way reflects individual members'
approval or disapproval of the appointees, and that the
nominations are merely forwarded to the full legislature for
confirmation or rejection. There being no objection, the
confirmation was advanced.
1:31:18 PM
The committee took an at-ease from 1:31 p.m. to 1:33 p.m.
HB 358-PROHIBIT AI-ALTERED REPRESENTATIONS
1:33:11 PM
CHAIR VANCE announced that the final order of business would be
HOUSE BILL NO. 358, "An Act relating to use of artificial
intelligence to create or alter a representation of the voice or
likeness of an individual."
1:33:36 PM
REPRESENTATIVE CARPENTER moved to adopt the proposed committee
substitute (CS) for HB 358, Version 33-LS1272\U, Walsh, 3/21/24,
as the working document.
REPRESENTATIVE GROH objected for the purpose of discussion.
1:33:57 PM
REPRESENTATIVE GROH expressed concern that Version U had been
distributed on short notice.
1:34:26 PM
The committee took an at-ease from 3:34 p.m. to 1:36 p.m.
1:36:00 PM
REPRESENTATIVE GROH removed his objection.
REPRESENTATIVE CARPENTER objected for the purpose of discussion.
1:37:09 PM
BOB BALLINGER, Staff, Representative Sarah Vance, Alaska State
Legislature, gave an explanation of changes in the proposed
committee substitute (CS) for HB 358, Version U, on behalf of
Representative Vance. Section 1 addresses civil liability for
defamation based on deepfakes and allows for a claim of
defamation per se. Section 2 addresses child pornography that
has been manipulated, created, or modified using artificial
intelligence (AI) to appear to depict a child under 18 years of
age. Section 3 defines "artificial intelligence" as an
automated system that uses data input, human-defined objectives
and machine learning, natural language processing, or other
computational processing techniques of similar or greater
complexity to make a decision or facilitate human decision
making. Section 4 addresses deepfakes in electioneering
communications and allows for the following disclosure statement
when included in the communication to be used as a defense:
"This (image/video/audio) has been manipulated." Section 4 also
defines "deepfake" as any visual or audio media that is created,
altered, or otherwise manipulated by artificial intelligence,
and defines "information content provider" and "interactive
computer service" with the meanings given in 47 U.S.C. 230.
Section 5 indicates that the law would not be retroactive.
1:47:17 PM
REPRESENTATIVE SUMNER expressed concern about the phrase "human
defined objectives" in the proposed definition of AI and
suggested striking it from the bill.
CHAIR VANCE pointed out that the definition includes the words
"machine learning" and asked whether that would assuage
Representative Sumner's concern.
REPRESENTATIVE SUMNER stated that machines can be trained by
other AI programs on AI-generated data, thereby completely
removing humans from the process. He maintained his belief that
the definition would not be hindered by removing the phrase
"human defined objectives."
MR. BALLINGER opined that the definition should include the
phrase "human defined objectives" for the purpose of the
legislation, which seeks to hold an individual liable, either
criminally or civilly. Nonetheless, he agreed that if the goal
was to generally define "artificial intelligence," it would make
sense to remove it.
1:50:11 PM
REPRESENTATIVE CARPENTER removed his objection. There being no
further objection, Version U was adopted as the working
document.
1:50:43 PM
REPRESENTATIVE CARPENTER considered a hypothetical scenario and
asked whether it would fall under the definition of deepfake.
MR. BALLINGER said he did not know the answer. He explained
that it would be hard to argue that the scenario posed by
Representative Carpenter portrays a fundamentally different
understanding or impression from the altered original version.
REPRESENTATIVE CARPENTER asked whether superimposing a speech
that was made by a legislator on the House or Senate floor onto
the same legislator in a different location would be considered
a deepfake.
MR. BALLINGER suspected that it probably wouldn't rise to the
level of deepfake. He noted that the video would have to be
used in electioneering communication before considering whether
it was a deepfake.
1:55:03 PM
REPRESENTATIVE GRAY considered a scenario in which the meaning
of a video is altered to convey the opposite messaging and then
used on a political stage. He asked whether that would be
considered a deepfake.
MR. BALLINGER explained that if AI is used to convey a
fundamentally different message, it would fall under the
definition of deepfake.
REPRESENTATIVE GRAY asked whether the bill would be "hurt" by
including this scenario in the language.
MR. BALLINGER stated that it would be a policy call.
1:56:56 PM
DAVE STANCLIF, Staff, Representative Mike Cronk, Alaska State
Legislature, on behalf of Representative Cronk, prime sponsor of
CSHB 358, Version U, advised that the bill should remain
narrowly focused on deepfakes.
REPRESENTATIVE GRAY questioned the legality of altering a
person's words or image without the use of AI in political
advertising.
1:58:38 PM
IAN WALSH, Attorney, Legislative Legal Services, Legislative
Affairs Agency (LAA), shared his understanding that the creation
of falsely edited media would likely fall under defamation. He
offered to follow up with the requested information as it
pertains to election law. He noted that currently, the bill
applies to AI that modifies something to be fundamentally
different than it appears.
1:59:30 PM
REPRESENTATIVE C. JOHNSON shared his understanding that per
Version U, the disclaimer [stating "This (image/video/audio) has
been manipulated"] must be the same text size as the headline.
He expressed concern that [the size] is not workable.
MR. BALLINGER noted that this provision would only apply when a
deepfake is used in electioneering communications.
REPRESENTATIVE C. JOHNSON pointed out that not all deepfakes are
necessarily bad and gave the example of Photoshopping a dog into
a photo to induce a loving feeling.
MR. BALLINGER contended that inserting a dog would not
fundamentally change the meaning.
2:04:26 PM
REPRESENTATIVE GROH asked whether "electioneering" is defined
elsewhere in statute.
MR. BALLINGER answered yes, "in two parts," indicating that both
"electioneering" and "communication" are defined separately in
statute.
REPRESENTATIVE GROH asked how many people would need to see the
deepfake for it to become actionable.
MR. BALLINGER indicated that it would depend on whether the
material was generated with the intent to alter a person's vote.
He added that if all elements were met and damages were
demonstrated, then the individual could be held liable.
2:07:07 PM
REPRESENTATIVE GROH asked whether the disclosure in a falsified
image should be at least as big as the size of the headline.
MR. BALLINGER said it would be a policy call.
REPRESENTATIVE GROH highlighted the definition of "artificial
intelligence" on page 3, lines 19-22 of Version U. He
referenced Photoshop, ChatGPT, and Grammarly and asked what
would fit under the proposed definition of AI.
MR. BALLINGER answered yes, all of those applications are
considered AI.
2:09:47 PM
REPRESENTATIVE GROH asked whether there is a distinction between
child pornography that manipulates the likeness of a real child
versus that which is entirely generated by AI.
MR. BALLINGER shared his understanding that AI generated child
pornography [that does not depict a real child] is not
prosecutable under current laws.
REPRESENTATIVE GROH questioned whether the intent of the
legislation would be to ask the jury whether the material
depicts an identifiable, actual child.
MR. BALLINGER explained that the jury would need to consider
whether a reasonable person could think that the material
depicts an actual child.
REPRESENTATIVE GROH said he understood the dual purposes of the
legislation, which both involve complicated issues. He stressed
the importance of thinking through the implications of the bill.
2:15:17 PM
REPRESENTATIVE SUMNER suggested that the creation of deepfake
pornography based on the likeness of an adult should be, at a
minimum, defamation per se.
MR. BALLINGER clarified that it would be considered defamation
per se because damages are assumed.
2:16:46 PM
REPRESENTATIVE GRAY asked whether a campaign photo with
artificially whitened teeth would be considered a deepfake and
would require the inclusion of a disclosure statement.
MR. BALLINGER said that would not be considered a deepfake
because it would not [convey a fundamentally different
understanding].
2:18:42 PM
REPRESENTATIVE GRAY asked what would happen if the language on
page 2, line 5, "and the depiction is recognizable as an
identifiable actual child" was deleted.
MR. BALLINGER remarked, "It probably, almost definitely, would
be considered unconstitutional" until it reaches the Supreme
Court.
2:22:01 PM
REPRESENTATIVE SUMNER recalled a previous election in which his
political opponent Photoshopped him into a Planned Parenthood t-
shirt. He asked whether that would be considered defamation per
se under the proposed legislation.
MR. BALLINGER responded, "probably not," because the standard
for political officials is much higher. However, if Version U
were to pass, he suspected that the photo could be considered a
deepfake if it were made to look legitimate.
2:23:32 PM
REPRESENTATIVE CARPENTER shared a hypothetical example and asked
whether it would be considered a deepfake in addition to
defamation.
MR. BALLINGER reiterated that it would depend on whether [the
manipulated material] conveyed a fundamentally different
meaning.
2:25:47 PM
REPRESENTATIVE CARPENTER sought to confirm that if the material
included a disclosure statement, it would not be considered a
deepfake.
MR. BALLINGER confirmed that if the disclosure were used, there
would be "no harm no foul."
REPRESENTATIVE CARPENTER pointed out that there are years of
quoted text available to AI, which could be used to create a
speech in video form to manipulate the context and portray the
candidate in a certain light. He sought to confirm that this
conduct would not be covered by the bill.
MR. BALLINGER maintained that if the quote was used with the
intent to convey a completely different message, that would be
considered [a deepfake].
2:28:03 PM
MR. STANCLIFF asked the committee to identify the most
significant constitutional issues with the bill.
MR. BALLINGER shared his belief that the biggest constitutional
issue surrounds AI generated child pornography because, under
current jurisprudence, it is victimless.
CHAIR VANCE emphasized the importance of ensuring that any
sweeping changes pertaining to AI would not unintentionally
impact current electioneering practices. Furthermore, she
opined that there needs to be a disclaimer on the use of
deepfakes for consumer protection.
2:33:50 PM
REPRESENTATIVE CARPENTER suggested that the committee take a
step back from child pornography [that is generated
artificially] in an effort to realize that it may not involve
the filming of humans.
CHAIR VANCE agreed that much of this discussion bleeds over into
First Amendment rights. She shared her belief that [the bill]
should not push the envelope with regard to constitutionality in
an effort to protect Alaskans' rights. She added that the use
of AI would be an ongoing conversation because it's moving at a
much faster pace than the legislature.
[CSHB 358, Version U, was held over.]
2:37:13 PM
CHAIR VANCE issued closing remarks on future business.
2:37:23 PM
ADJOURNMENT
There being no further business before the committee, the House
Judiciary Standing Committee meeting was adjourned at 2:37 p.m.
| Document Name | Date/Time | Subjects |
|---|---|---|
| Rebecca Carrillo Human Rights App_Redacted.pdf |
HJUD 3/22/2024 1:00:00 PM |
|
| Rebecca Carrillo Human Rights Resume_Redacted.pdf |
HJUD 3/22/2024 1:00:00 PM |
|
| HB 358 - Proposed CS v.U.pdf |
HJUD 3/22/2024 1:00:00 PM HJUD 3/25/2024 1:00:00 PM |
HB 358 |