Google as God?
Opportunities and Risks of the Information Age
By Dirk Helbing
"You're already a walking sensor platform… You
are aware of the fact that somebody can know where you are at all times because
you carry a mobile device, even if that mobile device is turned off. You
know this, I hope? Yes? Well, you should… Since
you can't connect dots you don't have, it drives us into a mode of, we
fundamentally try to collect everything and hang on to it forever… It is really
very nearly within our grasp to be able to compute
on all human generated information."
If God did not exist – people would invent one! The development of human civilization requires mechanisms
promoting cooperation and social order. One of these mechanisms is based on the
idea that everything we do is seen and judged by God – bad deeds will be
punished, while good ones will be rewarded. The Information Age has now fueled the dream that God-like omniscience
and omnipotence can be created by man.
For many decades, the processing power of computer chips has
increased exponentially – a process known as "Moore's Law". Storage
capacity is growing even faster. We are now entering a phase of the
"Internet of Things", where computer chips and measurement sensors
will soon be scattered everywhere producing huge amounts of data ("Big
Data"). It's not just cell phones, computers and factories that are more and
more connected, but our coffee machines, fridges, shoes and clothes, among
others.
Gold Rush for the 21st
Century Oil
This huge amount of data, including credit card transactions,
communication with friends and colleagues, mobility data and more is already
celebrated as the “Oil of the 21st Century”. The gold rush to exploit this
valuable resource is just starting. But the more data are generated, stored and
interpreted, the more will it be possible for companies and secret services to
know us better than our friends and families do. For example, the company
"Recorded Future" – apparently a joint initiative between Google and the
CIA – seems to investigate people’s social networks and mobility profiles.
Furthermore, credit card companies analyze "consumers’ genes" – the
factors that determine our consumer behavior.
Our individual motivations are analyzed in order to understand our
decisions and influence our behavior through personalized search,
individualized advertisements, and recommendations or decisions of our Facebook
friends. But how many of these “friends” are trustable, how many of them are
paid to influence us, and how many are software robots?
Humans Controlled by
Computers?
Today, computers autonomously perform the majority of financial
transactions. They decide how much we have to pay for our loans or insurances,
based on our behavioral data and on those of our friends, neighbors and
colleagues. People are increasingly discriminated by obscure "machine
learning" algorithms, which are neither transparent nor have to meet
particular quality standards. People classified as dangerous are now eliminated
by drones, without a chance to prove their innocence, while some countries are
discussing robots rights. Soon, Google will drive our cars. And in ten years,
supercomputers will exceed the performance of human brains.
Is Privacy Still Needed?
What will the role of privacy be in such an information society? Some
companies are already trying to turn privacy into a marketable commodity. This
is done by first taking away our privacy and then selling it back to us. The
company Acxiom, for example, is said to sell detailed data about more than 500
million people. Would it be possible to know beforehand whether the data will be
used for good or bad? Many will pay to have their personal data removed from
the Internet and commercial databases. And where data removal is not possible,
fake identities and mobility profiles will be offered for sale, to obfuscate
our traces.
Information Overload
“Big Data” do not necessarily mean that we'll see the world more
accurately. Rather, we will have to pay for “digital eyewear” that allows us to
keep an overview in the data deluge. Those not willing to pay (possibly also
with personal data) will be blinded by an information overload. Already today,
we cannot assess the quality of the answers we get online. As the way in which
the underlying data are processed remains hidden to the user, it is hard to know
how much we are being manipulated by web services and social media. But given the huge economic potential, it is
pretty clear that manipulation is happening.
The Knowledge-Is-Power
Society
The statement "knowledge is power" seems to imply that
"omniscience is omnipotence" – a tempting idea indeed. Therefore, who
collects all the data in the world, such as the National Security Agency (NSA)
in the United States, might hope to become almighty, especially if equipped
with suitable manipulation tools. By knowing everything about us, one can always
find an Achille’s heel. Even CIA director General David Petraeus was not immune
to this risk. He became the victim of a love affair irrelevant to his duty.
The developments outlined above are not fantasy - they are already
taking place behind the scenes or are just around the corner. Yet, our society
and legal system are not well prepared for this.
Each 1 dollar bill may
be read to suggest that omniscience and omnipotence (see the “God’s eye”) and
the belief in God (“In God We Trust”) should be the basis of a new world order
(“Novus Ordo Seclorum”). Source: Wikimedia Common, see
A New World Order Based on
Information?
Some people may see information and technologies as new tools to
create social order. Why should one object to a computer or government or
company taking decisions for us, as long as they are in our interest? But who
would decide how to use these tools? Can the concept of a ‘caring state’ or a ‘benevolent
dictator’ really work? In other words, can supercomputers enabled by Big Data
take the decisions that are best for us?
This has always failed in the past, and it will be also
unsuccessful in the future. Not only do many systems fail under asymmetric
information (if some stakeholders are very well informed and others very badly).
The performance of all computers in the world will also never be sufficient to
optimize our economy and society in real time. Supercomputers cannot even
optimize the traffic lights of a big city in real time. This is because the
computational effort explodes with the size and complexity of the system. Just
a very simple society could be optimized top down, but who would want to live
in it?
Privacy and Socio-Diversity
Need Protection
The aforementioned "omniscient almighty society" cannot
work. If we all did what a super-intelligent institution thinks is right – it
would be as if children always did what parents are asking for, never becoming
adolescents. Then they would never take autonomous decisions, and find their
own way. Privacy is a necessary ingredient for the development of individual responsibility
and for society. It should not to be understood as a concession to the
citizens.
“Private” and “public” are two sides of the same coin, each of which
cannot exist without the other. People can only adjust to the thousands of normative
public expectations every day, if there is a private, protected space where
they can be free and relax. Privacy is an invention that reduces mutual
interference to a degree that allows us to "live and let live". If we
knew what others think, we would have far more conflicts.
The importance of unobserved opinion formation is demonstrated by
the crucial role of anonymous votes in democracies. Would we only adjust ourselves
to expectations of others, many new ideas would not emerge or spread. Social
diversity would decrease, and thus the ability of our society to adapt.
Innovation requires the protection of minorities and new ideas. It is an engine
of our economy. Social diversity also promotes happiness, social well-being, and
the ability of our society to recover from shocks ("resilience").
Social diversity must be protected just as much as biodiversity.
Today, however, the Internet recommends us opinions about books, music, movies
and even about friends and partners. This undermines the principle of the
"wisdom of crowds" and collective intelligence. Why should a company
decide what is good for us? Why can’t we determine the recommendation
algorithms ourselves? Why don’t we get access to relevant data?
An Alternative Vision of
the Information Age
Also in an increasingly unstable world, surveillance, combined
with the manipulation or suppression of undesired behaviors, is not a
sustainable solution. But is there an alternative to the omniscient almighty
state that matches our ethical values? An alternative that can create cultural
and economic prosperity? Yes, indeed!
Our society and economy are currently undergoing a fundamental
transformation. Global networking creates increasing complexity and instability
that cannot be properly managed by planning, optimization and top-down control.
A flexible adaptation to local needs works better for complex, variable systems.
This means that managing complexity requires a stronger bottom-up component.
In the economy and the organization of the Internet, decentralized
self-organization principles have always played a big role. Now they have also
spread to intelligent energy networks ("smart grids") and traffic
control. One day, societal decision-making and economic production processes
will also be run in a more participatory way to better manage the increase in complexity.
It seems the natural course of history. A growing desire of citizens to
participate in social, political and economic affairs is already found in many parts
of the world.
The Democratic,
Participatory Market Society
In connection with a participatory economy, one often speaks of
“prosumers”, i.e. co-producing consumers. Advanced collaboration platforms will
allow anyone to set up projects with others to create their own products, for
example with 3D printers. Thus, classical companies and political parties and
institutions might increasingly be replaced by project-based initiatives – a
form of organization that I would like to call “democratic, participatory
market society”.
To ensure that the participatory market society will work well and
create jobs on a large scale, the right decisions will have to be taken. For
example, it seems essential that the information systems of the future will be
open, transparent and participatory. This requires us to create a participatory
information and innovation ecosystem, i.e. to make large amounts of data accessible
to everyone.
The Benefit of Opening Data
to All
The great advantage of information is that it is (re)producible in
a cheap and almost unlimited way, so that the eternal struggle for limited
resources might be overcome. It is important that we take advantage of this and
open the door to an age of creativity rather than limiting access to
information, thereby creating artificial scarcity again. Today, many companies
collect data, but lack access to other important data. The situation is as if
everyone owned a few words of our language, but had to pay for the use of all the
other words. It is pretty clear that, under such conditions, we could not fully
capitalize on our communicative potentials.
To overcome this dissatisfactory data exchange situation and achieve
"digital literacy", one could decide to open up data for all.
Remember that most countries have also decided to turn the privilege of reading
and writing into a public good by establishing public schools. It is well known
that this step has boosted the development of modern societies. Similarly, "Open
Data" could boost the development of the information society, but the
producers of data must be adequately compensated.
A New Paradigm to Manage Complexity
Access to data is essential for the successful management of
complex dynamical systems, as it requires three elements: (i) proper systems design to support predictability
and controllability, (ii) probabilistic short-term forecasts of the system
dynamics, which need plenty of reliable real-time data, and (iii) suitable adaptive
mechanisms ("feedback loops") that support the desired system
behavior.
Managing complexity should build on the natural tendency of
complex dynamical systems to self-organize. To enable self-organization, it is
crucial to find the right institutional settings and suitable ‘rules of the
game”, while avoiding too much top down control. Then, complex systems can
essentially regulate themselves.
One must be aware, however, that complex systems often behave in
counterintuitive ways. Hence, it is easy to choose the wrong rules, thereby ending
up with suboptimal results, unwanted side effects, or unstable system behaviors
that can lead to man-made disasters. The financial system, which went out of
control, might serve as a warning. These problems have traditionally been
managed by top-down regulation, which is usually inefficient and expensive.
Loss of Control due to a Wrong
Way of Thinking
Whether a system can be adequately managed or is self-organizing
in the way we want is a matter of systems design. If the system is designed in
the wrong way, then it will get out of control sooner or later, even if all
actors involved are highly trained, well equipped and highly motivated to do
the right things. “Phantom traffic jams” and crowd disasters are examples of unwanted
situations that occur despite all efforts to prevent them from happening. Likewise,
financial crises, conflicts and wars can be unintended consequences of systemic
instabilities. Even today, we are still not immune to them.
Therefore, we need a much better understanding of our
techno-socio-economic-ecological systems and their interdependencies. Appropriate
institutions and rules for our highly networked world must still be found. The
information age will revolutionize our economy and society in a dramatic way.
If we do not pay sufficient attention to these developments, we will suffer the
fate of a car driving too fast on a foggy day.
Decisions Needed to Use
Opportunities and Avoid Risks
To meet the challenges of the 21st century and benefit from its
great opportunities, a Global Systems Science needs to be established in order to
fill the current knowledge gaps. It aims to generate new insights allowing politics,
economy and society to take better informed, more successful decisions. This
could help us to use the chances of the information age and minimize its risks.
We must be aware that everything is possible – ranging from a Big Brother
society to a participatory economy and society. The choice is ours!
About the author
Dirk Helbing is Professor of Sociology,
in particular of Modeling and Simulation at ETH Zurich, and member of its
Computer Science Department. He earned a PhD in physics and was Managing
Director of the Institute of Transport & Economics at Dresden University of
Technology in Germany. He is internationally known for his work on pedestrian
crowds, vehicle traffic, and agent-based models of social systems. Furthermore,
he is coordinating the FuturICT Initiative (http://www.futurict.eu), which focuses on the
understanding of techno-socio-economic systems, using Big Data. His work is
documented by hundreds of scientific articles, dozens of keynote talks, and
media reports in all major languages. Helbing is elected member of the World
Economic Forum’s Global Agenda Council on Complex Systems and of the German
Academy of Sciences “Leopoldina”. He is also chairman of the Physics of
Socio-Economic Systems Division of the German Physical Society and co-founder
of ETH Zurich’s Risk Center.
[1] See: http://www.businessinsider.com/cia-presentation-on-big-data-2013-3?op=1 and http://gigaom.com/2013/03/20/even-the-cia-is-struggling-to-deal-with-the-volume-of-real-time-social-data/2/. For similar recent FBI priorities see http://www.slate.com/blogs/future_tense/2013/03/26/andrew_weissmann_fbi_wants_real_time_gmail_dropbox_spying_power.html
Dirk Helbing:
ReplyDeleteRe: Inquiry about working with FuturICT on certain critical issues I do not see being addressed.
In the course of a Linked-In discussion that tried to articulate a 'Systems Thinking' response to UN Secretary General BAn Ki-Moon's call for 'revolutionary thinking and action to ensure an economic model for survival' at the WEF meeting in Davos 2011, I tried to sketch some ideas both for the kind of global discourse framework that will be needed to develop meaningful and feasible strategies to tackle the problems we face, but also for some specific problems for which even the outline of solutions are not sufficiently well discussed. One of them has to do with the problems you address in your FuturICT work and your postings on the issue of 'Big Data' and the implications for personal privacy. Briefly, I have come to the conclusion that one of the key things that should be developed is a system of 'sanctions' -- provisions to ensure that laws, agreements, treaties etc. are adhered to -- that do not depend on the currently predominant arrangement of detection, investigation of such violations and the efforts to apprehend violators by 'enforcement' agencies, but 'sanctions' that are triggered by the very act or attempt at violation. Instead of having to rely on increasingly comprehensive and intrusive 'big data' systems, as the current direction implies, such a system would limit the activity information to some 'civic credit' account (for example) that is kept on a person's individual data device. These credits in turn entitle the bearer to various 'rights' (such as the right to drive vehicles, or make financial transactions) and the sanctions would consist of reduction or complete removal of the respective rights. Technology is already being developed that could make such systems possible, but I see the overall policies directed at the big data/big brother applications, which are justifiably resisted and feared by many.
I have drafted a report on the discussion mentioned above; the group running the LinkedIn forum (Systems Thinking World) is currently considering its possible publication in its 'SystemsThinking World Journal and has posted it for review and comments on Googledocs : https://docs.google.com/document/d/1JJGsP7V799VImq52dsN4Dmx7qqwjgmVW8lYLZUkn8EE/edit?pli=1
I see this and similar issues as critical to the efforts of FuturICT, and would be interested in working with appropriate groups within FuturICT to pursue them, but I am not sure how to get these ideas on the agenda.
If you share my sentiment on the importance of these problems, I can send you (and other interested parties) more detail on my work on some of these issues, and would appreciate your thoughts on these ideas.
Sincerely
Thorbjoern Mann