QUOTE OF THE DAY: "One harrowing potential eventuality: Fake video and audio may become so convincing that it can’t be distinguished from real recordings, rendering audio and video evidence inadmissible in court."
PASSAGE OF THE DAY: "A program called Face2Face, developed at Stanford, films one person speaking, then manipulates that person’s image to resemble someone else’s. Throw in voice manipulation technology, and you can literally make anyone say anything — or at least seem to. The technology isn’t quite there; Princess Leia was a little wooden, if you looked carefully. But it’s closer than you might think."
COMMENTARY: "Our Hackable Political Future, by reporters Henry J. Farrell and Rick Perlsteinfeb, published by The New York Times on February 4, 2018. (Henry J. Farrell is a professor of political science and international affairs at the George Washington University. Rick Perlstein is the author, most recently, of “The Invisible Bridge: The Fall of Nixon and the Rise of Reagan.”)
GIST: "Imagine
it is the spring of 2019. A bottom-feeding website, perhaps tied to
Russia, “surfaces” video of a sex scene starring an 18-year-old Kirsten
Gillibrand. It is soon debunked as a fake, the product of a
user-friendly video application that employs generative adversarial
network technology to convincingly swap out one face for another. It
is the summer of 2019, and the story, predictably, has stuck around —
part talk-show joke, part right-wing talking point. “It’s news,”
political journalists say in their own defense. “People are talking
about it. How can we not?” Then
it is fall. The junior senator from New York State announces her
campaign for the presidency. At a diner in New Hampshire, one “low
information” voter asks another: “Kirsten What’s-her-name? She’s running
for president? Didn’t she have something to do with pornography?” Welcome
to the shape of things to come. In 2016 Gareth Edwards, the director of
the Star Wars film “Rogue One,” was able to create a scene featuring a
young Princess Leia by manipulating images of Carrie Fisher as she
looked in 1977. Mr. Edwards had the best hardware and software a $200
million Hollywood budget could buy. Less than two years later, images of
similar quality can be created with software available for free
download on Reddit. That was how a faked video supposedly of the actress
Emma Watson in a shower with another woman ended up on the website
Celeb Jihad. Programs
like these have many legitimate applications. They can help
computer-security experts probe for weaknesses in their defenses and
help self-driving cars learn how to navigate unusual weather conditions.
But as the novelist William Gibson once said, “The street finds its own
uses for things.” So do rogue political actors. The implications for
democracy are eye-opening. The
conservative political activist James O’Keefe has created a cottage
industry manipulating political perceptions by editing footage in
misleading ways. In 2018, low-tech editing like Mr. O’Keefe’s is already
an anachronism: Imagine what even less scrupulous activists could do
with the power to create “video” framing real people for things they’ve
never actually done. One harrowing potential eventuality: Fake video and
audio may become so convincing that it can’t be distinguished from real
recordings, rendering audio and video evidence inadmissible in court. A
program called Face2Face, developed at Stanford, films one person
speaking, then manipulates that person’s image to resemble someone
else’s. Throw in voice manipulation technology, and you can literally
make anyone say anything — or at least seem to. The
technology isn’t quite there; Princess Leia was a little wooden, if you
looked carefully. But it’s closer than you might think. And even when
fake video isn’t perfect, it can convince people who want to be
convinced, especially when it reinforces offensive gender or racial
stereotypes. Another
harrowing potential is the ability to trick the algorithms behind
self-driving cars to not recognize traffic signs. Computer scientists
have shown that nearly invisible changes to a stop sign can fool
algorithms into thinking it says yield instead. Imagine if one of these
cars contained a dissident challenging a dictator. In
2007, Barack Obama’s political opponents insisted that footage existed
of Michelle Obama ranting against “whitey.” In the future, they may not
have to worry about whether it actually existed. If someone called their
bluff, they may simply be able to invent it, using data from stock
photos and pre-existing footage. The
next step would be one we are already familiar with: the exploitation
of the algorithms used by social media sites like Twitter and Facebook
to spread stories virally to those most inclined to show interest in
them, even if those stories are fake. It
might be impossible to stop the advance of this kind of technology. But
the relevant algorithms here aren’t only the ones that run on computer
hardware. They are also the ones that undergird our too easily hacked
media system, where garbage acquires the perfumed scent of legitimacy
with all too much ease."
The entire story can be found at:https://www.nytimes.com/2018/02/04/opinion/hacking-politics-future.html
PUBLISHER'S NOTE: I am monitoring this case/issue. Keep your eye on the Charles Smith Blog for reports on developments. The Toronto Star, my previous employer for more than twenty incredible years, has put considerable effort into exposing the harm caused by Dr. Charles Smith and his protectors - and into pushing for reform of Ontario's forensic pediatric pathology system. The Star has a "topic" section which focuses on recent stories related to Dr. Charles Smith. It can be found at: http://www.thestar.com/topic/c