Please sign and share the petition 'Tighten regulation on taking, making and faking explicit images' at Change.org initiated by Helen Mort to the w:Law Commission (England and Wales) to properly update UK laws against synthetic filth. Only name and email required to support, no nationality requirement.
Current and possible laws and their application
Laws and their application
Law on synthetic filth in Virginia 2019
Since July 1 2019 w:Virginia w:has criminalized the sale and dissemination of unauthorized synthetic pornography, but not the manufacture., as section § 18.2-386.2 titled 'Unlawful dissemination or sale of images of another; penalty.' became part of the w:Code of Virginia.
Code of Virginia (TOC) » Title 18.2. Crimes and Offenses Generally » Chapter 8. Crimes Involving Morals and Decency » Article 5. Obscenity and Related Offenses » Section § 18.2-386.2. Unlawful dissemination or sale of images of another; penalty
A. Any w:person who, with the w:intent to w:coerce, w:harass, or w:intimidate, w:maliciously w:disseminates or w:sells any videographic or still image created by any means whatsoever that w:depicts another person who is totally w:nude, or in a state of undress so as to expose the w:genitals, pubic area, w:buttocks, or female w:breast, where such person knows or has reason to know that he is not w:licensed or w:authorized to disseminate or sell such w:videographic or w:still image is w:guilty of a Class 1 w:misdemeanor.
- For purposes of this subsection, "another person" includes a person whose image was used in creating, adapting, or modifying a videographic or still image with the intent to depict an actual person and who is recognizable as an actual person by the person's w:face, w:likeness, or other distinguishing characteristic.
B. If a person uses w:services of an w:Internet service provider, an electronic mail service provider, or any other information service, system, or access software provider that provides or enables computer access by multiple users to a computer server in committing acts prohibited under this section, such provider shall not be held responsible for violating this section for content provided by another person.
C. Venue for a prosecution under this section may lie in the w:jurisdiction where the unlawful act occurs or where any videographic or still image created by any means whatsoever is produced, reproduced, found, stored, received, or possessed in violation of this section.
The identical bills were House Bill 2678 presented by w:Delegate w:Marcus Simon to the w:Virginia House of Delegates on January 14 2019 and three day later an identical Senate bill 1736 was introduced to the w:Senate of Virginia by Senator w:Adam Ebbin.
Law on synthetic filth in Texas 2019
On September 1 2019 w:Texas Senate bill SB 751 - Relating to the creation of a criminal offense for fabricating a deceptive video with intent to influence the outcome of an election w:amendments to the election code came into effect in the w:Law of Texas, giving w:candidates in w:elections a 30-day protection period to the elections during which making and distributing digital look-alikes or synthetic fakes of the candidates is an offense. The law text defines the subject of the law as "a video, created with the intent to deceive, that appears to depict a real person performing an action that did not occur in reality" SB 751 was introduced to the Senate by w:Bryan Hughes (politician).
The text of S.B. No. 751 is as follows
AN ACT relating to the creation of a criminal offense for fabricating a deceptive video with intent to influence the outcome of an election.
BE IT ENACTED BY THE LEGISLATURE OF THE STATE OF TEXAS:
- (d) A person commits an offense if the person, with intent to injure a candidate or influence the result of an election:
- creates a deep fake video; and
- causes the deep fake video to be published or distributed within 30 days of an election.
- (e) In this section, "deep fake video" means a video, created with the intent to deceive, that appears to depict a real person performing an action that did not occur in reality.
SECTION 2. This Act takes effect September 1, 2019.
Law on synthetic filth in California 2020
January 1 2020  the w:California w:US state law "AB-602 Depiction of individual using digital or electronic technology: sexually explicit material: cause of action." came into effect in the civil code of the w:California Codes banning the manufacturing and w:digital distribution of synthetic pornography without the w:consent of the people depicted. AB-602 provides victims of synthetic pornography with w:injunctive relief and poses legal threats of w:statutory and w:punitive damages on w:criminals making or distributing synthetic pornography without consent. The bill AB-602 was signed into law by California w:Governor w:Gavin Newsom on October 3 2019 and was authored by w:California State Assemblymember w:Marc Berman and an identical Senate bill was coauthored by w:California Senator w:Connie Leyva. AB602 at trackbill.com
Introduction by Assemblymember Marc Berman:
AB 602, Berman. Depiction of individual using digital or electronic technology: sexually explicit material: cause of action.
Existing law creates a private w:right of action against a person who intentionally distributes a photograph or recorded image of another that exposes the intimate body parts of that person or of a person engaged in a sexual act without the person’s consent if specified conditions are met.
This bill would provide that a depicted individual, as defined, has a w:cause of action against a person who either
- (1) creates and intentionally discloses sexually explicit material if the person knows or reasonably should have known the depicted individual did not w:consent to its creation or disclosure or
- (2) who intentionally discloses sexually explicit material that the person did not create if the person knows the depicted individual did not consent to its creation.
The bill would specify exceptions to those provisions, including if the material is a matter of legitimate public concern or a work of political or newsworthy value.
The bill would authorize a prevailing w:plaintiff who suffers harm to seek w:injunctive relief and recover reasonable w:attorney’s fees and costs as well as specified monetary w:damages, including statutory and w:punitive damages.
The law is as follows:
SECTION 1. Section 1708.86 is added to the Civil Code of California, to read:
1708.86. (a) For purposes of this section:
- (1) “Altered depiction” means a performance that was actually performed by the depicted individual but was subsequently altered to be in violation of this section.
- (2) “Authorized Representative” means an attorney, talent agent, or personal manager authorized to represent a depicted individual if the depicted individual is represented.
- (3) (A) “Consent” means an agreement written in plain language signed knowingly and voluntarily by the depicted individual that includes a general description of the sexually explicit material and the audiovisual work in which it will be incorporated.
- (3) (B) A depicted individual may rescind consent by delivering written notice within three business days from the date consent was given to the person in whose favor consent was made, unless one of the following requirements is satisfied:
- (i) The depicted individual is given at least 72 hours to review the terms of the agreement before signing it.
- (ii) The depicted individual’s authorized representative provides written approval of the signed agreement.
- (4) “Depicted individual” means an individual who appears, as a result of digitization, to be giving a performance they did not actually perform or to be performing in an altered depiction.
- (5) “Despicable conduct” means conduct that is so vile, base, or contemptible that it would be looked down on and despised by a reasonable person.
- (6) “Digitization” means to realistically depict any of the following:
- (A) The nude body parts of another human being as the nude body parts of the depicted individual.
- (B) Computer-generated nude body parts as the nude body parts of the depicted individual.
- (C) The depicted individual engaging in sexual conduct in which the depicted individual did not engage.
- (7) “Disclose” means to publish, make available, or distribute to the public.
- (8) “Individual” means a natural person.
- (9) “Malice” means that the defendant acted with intent to cause harm to the plaintiff or despicable conduct that was done with a willful and knowing disregard of the rights of the plaintiff. A person acts with knowing disregard within the meaning of this paragraph when they are aware of the probable harmful consequences of their conduct and deliberately fail to avoid those consequences.
- (10) “Nude” means visible genitals, pubic area, anus, or a female’s postpubescent nipple or areola.
- (11) “Person” means a human being or legal entity.
- (12) “Plaintiff” includes cross-plaintiff.
- (13) “Sexual conduct” means any of the following:
- (A) Masturbation.
- (B) Sexual intercourse, including genital, oral, or anal, whether between persons regardless of sex or gender or between humans and animals.
- (C) Sexual penetration of the vagina or rectum by, or with, an object.
- (D) The transfer of semen by means of sexual conduct from the penis directly onto the depicted individual as a result of ejaculation.
- (E) Sadomasochistic abuse involving the depicted individual.
(14) “Sexually explicit material” means any portion of an audiovisual work that shows the depicted individual performing in the nude or appearing to engage in, or being subjected to, sexual conduct.
(b) A depicted individual has a cause of action against a person who does either of the following:
- (1) Creates and intentionally discloses sexually explicit material and the person knows or reasonably should have known the depicted individual in that material did not consent to its creation or disclosure.
- (2) Intentionally discloses sexually explicit material that the person did not create and the person knows the depicted individual in that material did not consent to the creation of the sexually explicit material.
(c) (1) A person is not liable under this section in either of the following circumstances:
- (A) The person discloses the sexually explicit material in the course of any of the following:
- (i) Reporting unlawful activity.
- (ii) Exercising the person’s law enforcement duties.
- (iii) Hearings, trials, or other legal proceedings.
- (B) The material is any of the following:
- (i) A matter of legitimate public concern.
- (ii) A work of political or newsworthy value or similar work.
- (iii) Commentary, criticism, or disclosure that is otherwise protected by the California Constitution or the United States Constitution.
- (2) For purposes of this subdivision, sexually explicit material is not of newsworthy value solely because the depicted individual is a public figure.
(d) It shall not be a defense to an action under this section that there is a disclaimer included in the sexually explicit material that communicates that the inclusion of the depicted individual in the sexually explicit material was unauthorized or that the depicted individual did not participate in the creation or development of the material.
(e) (1) A prevailing plaintiff who suffers harm as a result of the violation of subdivision (b) may recover any of the following:
- (A) An amount equal to the monetary gain made by the defendant from the creation, development, or disclosure of the sexually explicit material.
- (B) One of the following:
- (i) Economic and noneconomic damages proximately caused by the disclosure of the sexually explicit material, including damages for emotional distress.
- (ii) Upon request of the plaintiff at any time before the final judgment is rendered, the plaintiff may instead recover an award of statutory damages for all unauthorized acts involved in the action, with respect to any one work, as follows:
- (I) A sum of not less than one thousand five hundred dollars ($1,500) but not more than thirty thousand dollars ($30,000).
- (II) If the unlawful act was committed with malice, the award of statutory damages may be increased to a maximum of one hundred fifty thousand dollars ($150,000).
- (C) Punitive damages.
- (D) Reasonable attorney’s fees and costs.
- (E) Any other available relief, including injunctive relief.
(2) The remedies provided by this section are cumulative and shall not be construed as restricting a remedy that is available under any other law.
(f) An action under this section shall be commenced no later than three years from the date the unauthorized creation, development, or disclosure was discovered or should have been discovered with the exercise of reasonable diligence.
(g) The provisions of this section are severable. If any provision of this section or its application is held invalid, that invalidity shall not affect other provisions.
Law on synthetic filth in China 2020
On January 1 2020 Chinese law requiring that synthetically faked footage should bear a clear notice about its fakeness came into effect. Failure to comply could be considered a w:crime the w:Cyberspace Administration of China (cac.gov.cn) stated on its website. China announced this new law in November 2019. The Chinese government seems to be reserving the right to prosecute both users and w:online video platforms failing to abide by the rules. 
Bills in the works
Law on synthetic filth in New York 20??
In the 2021-2022 w:New York State Senate regular sessions, on 2021-01-14 Senator w:James Skoufis (official website) sponsored and Senators w:Brian Benjamin (official website) and w:Todd Kaminsky (official website) of the New York State Senate co-sponsored New York Senate bill S1641 to add section § 250.70 UNLAWFUL ELECTRONIC TRANSMISSION OF SEXUALLY EXPLICIT VISUAL MATERIAL to the Article 250 of the penal law. On 2021-03-19 an identical New York Assembly bill A6517 - Establishes the crime of unlawful electronic transmission of sexually explicit visual material was introduced to the w:New York State Assembly by Assembly Member w:Aileen Gunther (official website).[1st seen in 1]
- Title of bill: An act to amend the penal law, in relation to the creation of the criminal offense of unlawful electronic transmission of sexually explicit visual material
- Purpose: The purpose of this bill is to make it unlawful to send sexually explicit material through electronic means unless the material is sent at the request of, or with the express consent of the recipient.
- Summary of provisions: Adds a new section 250.70 to the penal law making it unlawful to knowingly transmit by electronic means visual material that depicts any person engaging in sexual conduct or with a person's intimate parts exposed unless the material is sent at the request of, or with the express consent of the recipient.
- Justification: Currently under New York State law, indecent exposure in person is a crime, but it is not unlawful to send sexually explicit photos to nonconsenting adult recipients through electronic transmission. With the growing modem age of online dating, many individuals are receiving sexually explicit visual content without their consent from strangers. No person should be forced to view sexually explicit material without their consent.
The bill offers a clear deterrent to those considering sending unsolicited sexual pics and similar inappropriate conduct, and protects the unwilling recipients who currently have no legal recourse for such abuses.
What is illegal in the real world must be illegal in the digital world, and this legislation is a first step in the right direction in adding that accountability.
- Legislative history:
- Senate - 2020 - S5949 Referred to Codes
- Assembly - 2020 - A7801 Referred to Codes
- Fiscal implications: Minimal
- Effective date: This act shall take effect on the first of November next succeeding the date on which it shall have become a law.
The text of the bill is, as of 2021-03-24, as follows:
- "Section 1. The penal law is amended by adding a new section 250.70 to read as follows:
- § 250.70 UNLAWFUL ELECTRONIC TRANSMISSION OF SEXUALLY EXPLICIT VISUAL MATERIAL.
- A person is guilty of unlawful electronic transmission of sexually explicit visual material if a person knowingly transmits by electronic means visual material that depicts any person engaging in sexual conduct or with a person's intimate parts exposed or depicts the covered genitals of a male person that are in a discernibly turgid state and such visual material is not sent at the request of or with the express consent of the recipient. For purposes of this section the term "intimate parts" means the naked genitals, pubic area, anus, or female postpubescent nipple of the person and the term "sexual conduct" shall have the same meaning as defined in section 130.00 (Sex offenses; definitions of terms) of this chapter. Unlawful electronic transmission of sexually explicit visual material is a class a misdemeanor.
- § 2. This act shall take effect on the first of November next succeeding the date on which it shall have become a law."
NY Assembly Member w:Aileen Gunther (official website) presented an identical New York Assembly bill A6517 - Establishes the crime of unlawful electronic transmission of sexually explicit visual material to the w:New York State Assembly on 2021-03-19.
Law on synthetic filth in the UK 20??
The UK law does not seem very up-to-date on the issue of synthetic filth.
The independent w:Law Commission (England and Wales) is currently reviewing the law as it applies to taking, making and sharing intimate images without consent. The outcome of the consultation is due to be published later in 2021.
"In 2019, law expert Dr Aislinn O’Connell told w:The Independent that our current laws on image sharing are piecemeal and not fit for purpose. In October 2018 The w:Women and Equalities Committee called on the UK Government to introduce new legislation on image-based sexual abuse in order to criminalise ALL non-consensual creation and distribution of intimate sexual images." This call is for similar laws as California put in place on January 1 2020.
The petition 'Tighten regulation on taking, making and faking explicit images' at Change.org by w:Helen Mort aims to petition the UK govt for proper legislation against synthetic filth. See the mediatheque for a video by Helen Mort on her ordeal of becoming the victim of covert disinformation attacks.
Law proposals against synthetic filth by Juho Kunsola
- Audience: Developed with suitability for national, supranational and UN treaty levels.
- Writing context:
- Written from context of inclusion to criminal codes.
- I'm a Finn so this has been worded to fit in the Chapter 24 of the Criminal Code of Finland (in Finnish at finlex.fi) titled "Offences against privacy, public peace and personal reputation"
- Access the English translations of the Finnish Criminal Code at finlex.fi or go straight to the latest .pdf from 2016. Chapter 24 starts on page 107.
- History: This version is an evolution of a Finnish language original written in 2016.
Existing law in Chapter 24. of the Finnish Criminal Code - "Offences against privacy, public peace and personal reputation" seems to be ineffective against many synthetic human-like fake attack and seems it could be used to frame victims for crimes with digital sound-alikes.
The portions affected by or affecting the synthetic filth situation in bold font:
- Section 1 - Invasion of domestic premises (879/2013)
- Section 1(a) - Harassing communications (879/2013)
- Section 2 - Aggravated invasion of domestic premises (531/2000)
- Section 3 - Invasion of public premises (585/2005)
- Section 4 - Aggravated invasion of public premises (531/2000)
- Section 5 - Eavesdropping (531/2000)
- Section 6 - Illicit observation (531/2000)
- Section 7 - Preparation of eavesdropping or illicit observation (531/2000)
- Section 8 - Dissemination of information violating personal privacy (879/2013)
- Section 8(a) - Aggravated dissemination of information violating personal privacy (879/2013)
- Section 9 - Defamation (879/2013)
- Section 10 - Aggravated defamation (879/2013)
- Section 11 - Definition (531/2000)
- Section 12 - Right to bring charges (879/2013)
- Section 13 - Corporate criminal liability (511/2011)
Law proposal to ban visual synthetic filth
§1 Models of human appearance
A model of human appearance means
- A realistic 3D model
- A 7D bidirectional reflectance distribution function model
- A direct-to-2D capable w:machine learning model
- Or a model made with any technology whatsoever, that looks deceivingly like the target person.
§2 Producing synthetic pornography
Making projections, still or videographic, where targets are portrayed in a nude or in a sexual situation from models of human appearance defined in §1 without express consent of the targets is illegal.
§3 Distributing synthetic pornography
Distributing, making available, public display, purchase, sale, yielding, import and export of non-authorized synthetic pornography defined in §2 are punishable.[footnote 1]
§4 Aggravated producing and distributing synthetic pornography
If the media described in §2 or §3 is made or distributed with the intent to frame for a crime or for blackmail, the crime should be judged as aggravated.
The original idea I had was to ban both the raw materials i.e. the models to make the visual synthetic filth and also the end product weaponized synthetic pornography, but then in July 2019 it appeared to me that Adequate Porn Watcher AI (concept) could really help in this age of industrial disinformation if it were built, trained and operational. Banning modeling of human appearance was in conflict with the revised plan.
It is safe to assume that collecting permissions to model each pornographic recording is not plausible, so an interesting question is that can we ban covert modeling from non-pornographic pictures, while still retaining the ability to model all porn found on the Internet.
In case we want to pursue banning modeling people's appearance from non-pornographic images/videos without explicit permission be pursued it must be formulated so that this does not make Adequate Porn Watcher AI (concept) illegal / impossible. This would seem to lead to a weird situation where modeling a human from non-pornographic media would be illegal, but modeling from pornography legal.
Motivation: The current situation where the criminals can freely trade and grow their libraries of stolen voices is unwise.
Acquiring such a model of a human's voice, that deceptively resembles some dead or living person's voice and the possession, purchase, sale, yielding, import and export without the express consent of the target are punishable.
Producing and making available media from covert voice models defined in §1 is punishable.
If the produced media is for a purpose to
- frame a human target or targets for crimes
- to attempt extortion or
- to defame the target,
the crime should be judged as aggravated.
Resources and reporting on law
AI and law in general
Reviews and regulation From the w:Library of Congress:
- 'Regulation of Artificial Intelligence' at loc.gov
- 'Regulation of Artificial Intelligence: Comparative Summary' at loc.gov
- 'Regulation of Artificial Intelligence: International and Regional Approaches' (loc.gov)
- 'Regulation of Artificial Intelligence: The Americas and the Caribbean' (loc.gov)
- 'Regulation of Artificial Intelligence: East/South Asia and the Pacific' (loc.gov)
- 'Regulation of Artificial Intelligence: Europe and Central Asia' loc.gov
- 'Regulation of Artificial Intelligence: Middle East and North Africa' (loc.gov)
- 'Regulation of Artificial Intelligence: Sub-Saharan Africa' (loc.gov)
w:Gibson Dunn & Crutcher (gibsondunn.com) publishes a quarterly legal update on 'Artificial Intelligence and Autonomous Systems'. Gibson Dunn & Crutcher is a global w:law firm, founded in Los Angeles in 1890.
- 'Artificial Intelligence and Autonomous Systems Legal Update' Quarter 4 2018 at Gibson & Dunn
- 'Artificial Intelligence and Autonomous Systems Legal Update Quarter 1 2019'
- 'Artificial Intelligence and Autonomous Systems Legal Update Quarter 2 2019'
- 'Artificial Intelligence and Autonomous Systems Legal Update Quarter 3 2019'
- 'Artificial Intelligence and Autonomous Systems Legal Update Quarter 4 2019'
- 'The ethics of artificial intelligence: Issues and initiatives' (.pdf) at europarl.europa.eu, a March 2020 study by the w:European Parliamentary Research Service Starting from page 37 the .pdf lists organizations in the field.
Synthetic filth in the law and media
- '"The New Weapon of Choice": Law's Current Inability to Properly Address Deepfake Pornography' at scholarship.law.vanderbilt.edu, October 2020 Notes by Anne Pechenik Gieseke published in the The w:Vanderbilt Law Review, the flagship w:academic journal of w:Vanderbilt University Law School.
- 'Deepfakes and Synthetic Media in the Financial System: Assessing Threat Scenarios' at carnegieendowment.org, a 2020-07-08 assessment identifies some types of criminalities that can be made using synthetic human-like fakes.
- 'Don’t Believe Your Eyes (or Ears): The Weaponization of Artificial Intelligence, Machine Learning, and Deepfakes' at ssri.duke.edu, an October 2019 news article by Joe Littell, published by the Social Science Research Institute at the w:Duke University
- 'Deepfakes: False Pornography Is Here and the Law Cannot Protect You' at scholarship.law.duke.edu, published in 2019 in the Duke Law Journal, a student-run law review.
The countries that have unfortunately banned full face veil
“There are currently 16 nations that have banned the burqa (not to be confused with the hijab), including w:Tunisia, w:Austria, w:Denmark, w:France, w:Belgium, w:Tajikistan, w:Latvia, w:Bulgaria, w:Cameroon, w:Chad, w:Congo-Brazzaville, w:Gabon, w:Netherlands, w:China, w:Morocco, and w:Switzerland.”
Taking into consideration these times of industrial disinformation, it is vicious and uncivilized to have laws banning wearing a the full face veil in public.
Quotes on the current laws and their application
“If no-one who wants to hurt you knows what you look like, so how could someone malevolent make a covert digital look-alike of you?”
- People who are found in possession of this synthetic pornography should probably not be penalized, but rather advised to get some help.
1st seen in
- First seen in the suggestions for similar bills for Bills similar to CA AB602 by trackbill.com.
- "New state laws go into effect July 1".
- "§ 18.2-386.2. Unlawful dissemination or sale of images of another; penalty". w:Virginia. Retrieved 2021-01-23.
"Relating to the creation of a criminal offense for fabricating a deceptive video with intent to influence the outcome of an election". w:Texas. 2019-06-14. Retrieved 2021-01-23.
In this section, "deep fake video" means a video, created with the intent to deceive, that appears to depict a real person performing an action that did not occur in reality
- Johnson, R.J. (2019-12-30). "Here Are the New California Laws Going Into Effect in 2020". KFI. iHeartMedia. Retrieved 2021-01-23.
- "AB 602 - California Assembly Bill 2019-2020 Regular Session - Depiction of individual using digital or electronic technology: sexually explicit material: cause of action". openstates.org. openstates.org. Retrieved 2021-03-24.
- Mihalcik, Carrie (2019-10-04). "California laws seek to crack down on deepfakes in politics and porn". w:cnet.com. w:CNET. Retrieved 2021-01-23.
- Berman, Marc; Leyva, Connie (2019), "AB-602 Depiction of individual using digital or electronic technology: sexually explicit material: cause of action.", w:California
- "China seeks to root out fake news and deepfakes with new online content rules". w:Reuters.com. w:Reuters. 2019-11-29. Retrieved 2021-01-23.
- Statt, Nick (2019-11-29). "China makes it a criminal offense to publish deepfakes or fake news without disclosure". w:The Verge. Retrieved 2021-01-23.
Royle, Sara (2021-01-05). "'Deepfake porn images still give me nightmares'". w:BBC Online. w:BBC. Retrieved 2021-01-31.
She alerted the police to the images but was told that no action could be taken. Dr Aislinn O'Connell, a lecturer in law at Royal Holloway University of London, explained that Helen's case fell outside the current law.
Mort, Helen (2020). "Change.org petition: 'Tighten regulation on taking, making and faking explicit images'". w:Change.org. w:Change.org. Retrieved 2021-01-31.
Unlike other forms of revenge porn, creating pictures or videos like this is not yet illegal in the UK, though it is in some places in the US. The police were unable to help me.
- "Tunisian PM bans wearing of niqab in public institutions". Reuters. 5 July 2019. Retrieved 2021-03-13.
- "A European government has banned Islamic face veils despite them being worn by just three women". 21 April 2016. Retrieved 2021-03-13.
- Bulgaria the latest European country to ban the burqa and [niqab in public places, Smh.com.au: accessed 5 December 2016.
- Halasz, Stephanie; McKenzie, Sheena (27 June 2018). "The Netherlands introduces burqa ban in some public spaces" (27 June 2018). CNN. CNN. Retrieved 2021-03-13.
- Phillips, Tom (13 January 2015). "China bans burqa in capital of Muslim region of Xinjiang". The Telegraph (13 January 2015). The Telegraph. Retrieved 2021-03-13.