Facebook 鈥渓ikes鈥 can tell a lot about a person. Maybe even enough to fuel a voter-manipulation effort like the one a Trump-affiliated data-mining firm stands accused of 鈥 and which Facebook may have enabled.
The social network is under fire after The New York Times and The Guardian newspaper reported that former Trump campaign consultant Cambridge Analytica used data, including user likes, inappropriately obtained from roughly 50 million Facebook users to try to influence elections.
Monday was a wild roller coaster ride for Facebook, whose shares plunged 7 per cent in its worst one-day decline since 2014. Officials in the EU and the U.S. sought answers, while Britain鈥檚 information commissioner said she will seek a warrant to access Cambridge Analytica鈥檚 servers because the British firm had been 鈥渦nco-operative鈥 in her investigation. The first casualty of that investigation was an audit of Cambridge that Facebook had announced earlier in the day; the company said it 鈥渟tood down鈥 that effort at the request of British officials.
Adding to the turmoil, the New York Times reported that Facebook security chief Alex Stamos will step down by August following clashes over how aggressively Facebook should address its role in spreading disinformation. In a tweet , Stamos said he鈥檚 still fully engaged at Facebook but that his role has changed.
Related:
It would have been quieter had Facebook likes not turned out to be so revealing. Researchers in a 2013 study found that likes on hobbies, interests and other attributes can predict personal attributes such as sexual orientation and political affiliation. Computers analyze such data to look for patterns that might not be obvious, such as a link between a preference for curly fries and higher intelligence.
Chris Wylie, a Cambridge co-founder who left in 2014, said the firm used such techniques to learn about individuals and create an information cocoon to change their perceptions. In doing so, he said, the firm 鈥渢ook fake news to the next level.鈥
鈥淭his is based on an idea called 鈥榠nformational dominance,鈥 which is the idea that if you can capture every channel of information around a person and then inject content around them, you can change their perception of what鈥檚 actually happening,鈥 Wylie said Monday on NBC鈥檚 鈥淭oday.鈥 It鈥檚 not yet clear exactly how the firm might have attempted to do that.
Late Friday, Facebook said Cambridge improperly obtained information from 270,000 people who downloaded an app described as a personality test. Those people agreed to share data with the app for research 鈥 not for political targeting. And the data included who their Facebook friends were and what they liked 鈥 even though those friends hadn鈥檛 downloaded the app or given explicit consent.
Related:
Cambridge got limited information on the friends, but machines can use detailed answers from smaller groups to make good inferences on the rest, said Kenneth Sanford of the data science company Dataiku.
Cambridge was backed by the conservative billionaire Richard Mercer, and at one point employed Stephen Bannon 鈥 later President Donald Trump鈥檚 campaign chairman and White House adviser 鈥 as a vice-president. The Trump campaign paid Cambridge roughly $6 million according to federal election records, although officials have more recently played down that work.
The type of data mining reportedly used by Cambridge Analytica is fairly common, but is typically used to sell diapers and other products. Netflix, for instance, provides individualized recommendations based on how a person鈥檚 viewing behaviours fit with what other customers watch.
But that common technique can take on an ominous cast if it鈥檚 connected to possible elections meddling, said Robert Ricci, a marketing director at Blue Fountain Media.
Wylie said Cambridge Analytica aimed to 鈥渆xplore mental vulnerabilities of people.鈥 He said the firm 鈥渨orks on creating a web of disinformation online so people start going down the rabbit hole of clicking on blogs, websites etc. that make them think things are happening that may not be.鈥
Wylie told 鈥淭oday鈥 that while political ads are also targeted at specific voters, the Cambridge effort aimed to make sure people wouldn鈥檛 know they were getting messages aimed at influencing their views.
The Trump campaign has denied using Cambridge鈥檚 data. The firm itself denies wrongdoing, and says it didn鈥檛 retain any of the data pulled from Facebook and didn鈥檛 use it in its 2016 campaign work.
Yet Cambridge boasted of its work after another client, Texas Republican Sen. Ted Cruz, won the Iowa caucus in 2016.
Cambridge helped differentiate Cruz from similarly minded Republican rivals by identifying automated red light cameras as an issue of importance to residents upset with government intrusion. Potential voters living near the red light cameras were sent direct messages saying Cruz was against their use.
Even on mainstay issues such as gun rights, Cambridge CEO Alexander Nix said at the time, the firm used personality types to tailor its messages. For voters who care about tradition, it could push the importance of making sure grandfathers can offer family shooting lessons. For someone identified as introverted, a pitch might have described keeping guns for protection against crime.
It鈥檚 possible that Cambridge tapped other data sources, including what Cruz鈥檚 campaign app collected. Nix said during the Cruz campaign that it had five or six sources of data on each voter.
Facebook declined to provide officials for interview and didn鈥檛 immediately respond to requests for information beyond its statements Friday and Monday. Cambridge also didn鈥檛 immediately respond to emailed questions.
Facebook makes it easy for advertisers to target users based on nuanced information about them. Facebook鈥檚 mapping of the 鈥渟ocial graph鈥 鈥 essentially the web of people鈥檚 real-life connections 鈥 is also invaluable for marketers.
For example, researchers can look at people鈥檚 clusters of friends and get good insight as to who is important and influential, said Jonathan Albright, research director at the Tow Center for Digital Journalism at Columbia University. People who bridge different friend networks, for example, can have more influence when they post something, making them prime for targeting.
Two-thirds of Americans get at least some of their news on social media, according on Pew Research Center. While people don鈥檛 exist in a Facebook-only vacuum, it is possible that bogus information users saw on the site could later be reinforced by the 鈥渞abbit hole鈥 of clicks and conspiracy sites on the broader internet, as Wylie described.
Related
___
AP technology reporter Ryan Nakashima contributed to this report from Menlo Park, California.
Barbara Ortutay And Anick Jesdanun, The Associated Press
Like us on and follow us on .