Ryan Calo is an assistant professor at the University of Washington School of Law and a former research director at CIS. A nationally recognized expert in law and emerging technology, Ryan's work has appeared in the New York Times, the Wall Street Journal, NPR, Wired Magazine, and other news outlets. Ryan serves on several advisory committees, including the Electronic Frontier Foundation, the Electronic Privacy Information Center, and the Future of Privacy Forum. He co-chairs the American Bar Association Committee on Robotics and Artificial Intelligence and serves on the program committee of National Robotics Week.
The website Wikileaks recently published hundreds of thousands of confidential State Department cables. These communications apparently reveal the details of conversations with, and personal impressions and assessments of, foreign leaders and diplomats. Many fear that the leak will undermine international relations in profound and unknowable ways. One of the unintended consequence of the leak, however, may be to strengthen the case for a national consumer privacy law. Read more » about Wikileaks: Lessons For Consumer Privacy
UPDATE: As told to Jules Polonetsky over at The Future of Privacy Forum, Capital One was engaging in "totally random" rate changes that were not related to browser type. On the other hand, according to the Wall Street Journal, Capital One was at one point using [x+1] data to calibrate what credit card offers to show.
The other day, I suggested that the facts of the Clementi suicide may perfectly illustrate why no actual transfer of information is necessary for someone to suffer a severe subjective privacy harm. (Thanks to TechDirt and PogoWasRight for the write ups.)
Just now I learned about an allegation against Capital One that the company offered someone a different lending rate on the basis of what browser he used (Chrome vs. Firefox). A similar allegation was made against Amazon, which apparently used cookies for a time to calibrate the price of DVDs.
Here you have a clear objective privacy harm: your information (browser type) is being used adversely in a tangible and unexpected way. It matters not at all whether a human being sees the information or whether a company knows "who you are." Neither personally identifying information, nor the revelation of information to a person, is necessary for there to be a privacy harm. Read more » about Browser Snobbery As Objective Privacy Harm (UPDATE)
Ann Bartow once criticized Daniel Solove for not providing enough “dead bodies” in his discussion of privacy. I tend to disagree that such proof is necessary. But privacy has seen a dead body recently—that of Rutgers University student Tyler Clementi.
The narrative around Clementi’s tragic suicide continues to shift. The press originally reported that Clementi killed himself after his roommate invited the entire campus to view footage of Clementi having sex with another man. The Associated Press is now reporting that, according to the roommate’s defense attorney, no one but he and his friend ever saw the video.
The question of whether the defendants recorded or broadcast the web cam is highly relevant to whether there has been a privacy violation. Yet it is hardly relevant at all to the question of whether there has been a privacy harm. Read more » about Clementi And The Nature Of Privacy Harm
I don’t know that generativity is a theory, strictly speaking. It’s more of a quality. (Specifically, five qualities.) The attendant theory, as I read it, is that technology exhibits these particular, highly desirable qualities as a function of specific incentives. These incentives are themselves susceptible to various forces—including, it turns out, consumer demand and citizen fear.
The law is in a position to influence this dynamic. Thus, for instance, Comcast might have a business incentive to slow down peer-to-peer traffic and only refrain due to FCC policy. Or, as Barbara van Schewick demonstrates inter alia in Internet Architecture and Innovation, a potential investor may lack the incentive to fund a start up if there is a risk that the product will be blocked.
Similarly, online platforms like Facebook or Yahoo! might not facilitate communication to the same degree in the absence of Section 230 immunity for fear that they will be held responsible for the thousand flowers they let bloom. I agree with Eric Goldman’s recent essay in this regard: it is no coincidence that the big Internet players generally hail from these United States. Read more » about Will Robots Be 'Generative'?
Prohibition wasn’t working. President Hoover assembled the Wickersham Commission to investigate why. The Commission concluded that despite an historic enforcement effort—including the police abuses that made the Wickersham Commission famous—the government could not stop everyone from drinking. Many people, especially in certain city neighborhoods, simply would not comply. The Commission did not recommend repeal at this time, but by 1931 it was just around the corner.
Five years later an American doctor working in a chemical plant made a startling discovery. Several workers began complaining that alcohol was making them sick, causing most to stop drinking it entirely—“involuntary abstainers,” as the doctor, E.E. Williams, later put it. It turns out they were in contact with a chemical called disulfiram used in the production of rubber. Disulfiram is well-tolerated and water-soluble. Today, it is marketed as the popular anti-alcoholism drug Antabuse.
Were disulfiram discovered just a few years earlier, would federal law enforcement have dumped it into key parts of the Chicago or Los Angeles water supply to stamp out drinking for good? Probably not. It simply would not have occurred to them. No one was regulating by architecture then. To dramatize this point: when New York City decided twenty years later to end a string of garbage can thefts by bolting the cans to the sidewalk, the decision made the front page of the New York Times. The headline read: “City Bolts Trash Baskets To Walks To End Long Wave Of Thefts.”
In an important but less discussed chapter in The Future of the Internet, Jonathan Zittrain explores our growing taste and capacity for “perfect enforcement."
Readers are likely familiar with the cyberlaw mantra that “code is law.” What’s striking is that since Lawrence Lessig published Code in 1999, relatively little has been written about the dangers of regulation by architecture, particularly outside of the context of intellectual property. Many legal scholars—Neil Katyal, Elizabeth Joh, Edward Cheng—have instead argued for more regulation by architecture on the basis that it is less discriminatory or more effective. Read more » about (Im)Perfect Enforcement
I am proud to say that I helped found the Robot Block Party in Silicon Valley. Now in its fifth year, the event brings together industry, academia, and the hobbyist community to demo robots in celebration of National Robotics Week. We held the first one in Paul Brest Hall at Stanford Law School. The second, third, and fourth Robot Block Parties took place nearby at the Volkswagen Automotive Innovation Lab (where Stanford University develops driverless cars). Each event drew at least a thousand visitors. Read more » about Even (Some) Law Firms Think Robots Are The Next Big Thing
"In his paper entitled Digital Market Manipulation, Professor Ryan Calo suggests that a market mediated by technology creates a “sea change in the way companies use data to persuade”. He argues that some digital advertising could even constitute market manipulation. " Read more » about Are business owners who don’t invest in social media negligent?
"It’s quite clear: for most people, the link between government surveillance and freedom is more plainly understood by cars, rather than personal computers. As more and more objects become connected to the Internet these questions will grow in importance.And cars in particular might become, as Ryan Calo puts it in a 2011 article on drones, “a privacy catalyst”; an object giving us an opportunity to drag our privacy laws into the 21st century; an object that restores our mental model of what a privacy violation is." Read more » about Self-Driving Cars Will Turn Surveillance Woes Into a Mainstream Worry
English translation: "The National Commission for Data Protection ( CNIL ) does not sound the tocsin against civilian drones, but it intends to encourage the "vigilance" face that the American university Ryan Calo, expert on protection of privacy, says be "technological and cold incarnation of observation ' . "Without seeking to dramatize it seems essential to anticipate , " believes Edward Geffroy, the general secretary of the CNIL has committed a "forward thinking" with actors sector." Read more » about UAVs in civil
"“Criminal law is going to be looking for a guilty mind, a particular mental state — should this person have known better?” Mr. Calo said.
“If you’re not driving the car, it’s going to be difficult.”" “It’s the one headline, ‘machine kills child,’ rather than the 30,000 obituaries we have every year from humans killed on the roads,” said Bryant Walker Smith, a fellow at Stanford University’s Center for Automotive Research." Read more » about When Driverless Cars Break the Law
""I worry that profitable organizations will use big data to disadvantage consumers," Calo adds. "Intimate knowledge of consumers coupled with being able design every aspect of interaction--that’s going to create incentives to exploit."" Read more » about 3 Ways Big Data Is Going To Be Used Against You In The Future
For more information and to RSVP visit The New America Foundation's website. Webcast also available.
CIS Affiliate Scholar Ryan Calo part of panel titled "Delivery Drones and Robot Babysitters". Read more » about Can We Imagine Our Way to a Better Future?
Roundtable with experts Professor Ronald C. Arkin, Professor Ryan Calo, Dr. Kate Darling, Professor Illah Nourbakhsh, and Professor Noel Sharkey
Moderated by Professor Jennifer Urban
Friday, July 11, 3:30 pm
Boalt Hall Goldberg Room
Robots are quickly moving out of controlled environments into public spaces and homes, and researchers are developing artificial intelligence systems that will allow robots to make decisions autonomously. How should society plan for this transition? Read more » about Our Robot Future: The Moral, Ethical, and Legal Challenges of Ubiquitous Robotic Systems
Humans and Machines — Drones, Phones, and Robotic Friends: Where is Emergent Technology Taking Us? On June 27 at 8:30 p.m. with speakers Mary “Missy” Cummings, Ryan Calo, Ken Goldberg and moderator David Kirkpatrick.
As the landscape of high tech is increasingly modernized through applications of robotics from operating theaters to rescue missions, smarter phones that manage our lives, and flying technologies that put cameras (and weapons) in the air (if not everywhere), how will the balance of law, ethics, and relationships between humans and machines change us? Read more » about Drones, Phones, and Robotic Friends: Where is Emergent Technology Taking Us?
2013 PRIVACY PAPERS FOR POLICY MAKERS
The Future of Privacy Forum
Co-chairs Jules Polonetsky and Christopher Wolf
in conjunction with Congresswoman Sheila Jackson Lee invite you to
“Privacy Papers for Policy Makers”
A discussion of leading privacy research Read more » about Privacy Papers for Policy Makers
CIS Affiliate Scholars Peter Asaro, Ryan Calo and Woodrow Hartzog are listed as participants for We Robot 2014. Robotics is becoming a transformative technology. We Robot 2014 builds on existing scholarship exploring the role of robotics to examine how the increasing sophistication of robots and their widespread deployment everywhere from the home, to hospitals, to public spaces, and even to the battlefield disrupts existing legal regimes or requires rethinking of various policy issues. If you are on the front lines of robot theory, design, or development, we hope to see you. Read more » about We Robot 2014
On April 10, 2013, Stanford's Center for Law and the Biosciences welcomed CIS Affiliate Scholar Ryan Calo to campus for a discussion on law and emerging technology, with an emphasis on spyware for your brain. Read more » about The Center for Law and the Biosciences presents Ryan Calo
Hearing before the Senate Committee on the Judiciary on “The Future of Drones in America: Law Enforcement and Privacy Considerations” Read more » about The Future of Drones in America: Law Enforcement and Privacy Considerations
CIS Affiliate Scholar Ryan Calo interviews Neal Stephenson, author of Readme. Topics include privacy, virtual economics and security. Beth Cantrell, Greg Lastowka, and Tadayoshi Kohno also included in panel interview. This event was hosted by the University of Washington Law School. Read more » about Open Book Club: A Conversation With Neal Stephenson
It is not hard to imagine why robots raise privacy concerns. Practically by definition, robots are equipped with the ability to sense, process, and record the world around them. Robots can go places humans cannot go, see things humans cannot see. Robots are, first and foremost, a human instrument. And after industrial manufacturing, the principal use to which we’ve put that instrument has been surveillance. This talk explores the various ways that robots implicate privacy and why, absent conscientious legal and design interventions, we may never realize the potential of this transformative technology. Read more » about Robots, Privacy & Society- Cal Poly