Home ยป Articles ยป How tech billionaires’ visions of human nature shape our world

How tech billionaires’ visions of human nature shape our world

President Trump shakes Peter Thiel’s hand during a December 2016 meeting with technology executives at Trump Tower in New York. (Drew Angerer/Getty Images)

PROLOGUE

Peter Andreas Thiel is a German-American billionaire entrepreneur and venture capitalist. He is a co-founder of PayPalPalantir TechnologiesFounders Fund, and was the first investor in Facebook were he sits on the company board.[2][3] Thiel spoke at the Republican National Convention in 2016, where he criticized the state of the economy and described Mr. Trump as a man set to rebuild America. One of the few tech leaders to publicly back Mr. Trump, he also donated $1.25 million to the Trump campaign and related groups.

The Guardian: Why Silicon Valley billionaires are prepping for the apocalypse in New Zealand. How an extreme libertarian tract predicting the collapse of liberal democracies โ€“ written by Jacob Rees-Moggโ€™s father โ€“ inspired the likes of Peter Thiel to buy up property across the Pacific

Peter Tiel stood in front of screen displaying computer code.

Peter Thiel considers the code of human nature.
Heisenberg Media/Flickr, CC BY-SA

Simon McCarthy-Jones, Trinity College Dublin

In the 20th century, politiciansโ€™ views of human nature shaped societies. But now, creators of new technologies increasingly drive societal change. Their view of human nature may shape the 21st century. We must know what technologists see in humanityโ€™s heart.

The economist Thomas Sowell proposed two visions of human nature. The utopian vision sees people as naturally good. The world corrupts us, but the wise can perfect us.

The tragic vision sees us as inherently flawed. Our sickness is selfishness. We cannot be trusted with power over others. There are no perfect solutions, only imperfect trade-offs.

Science supports the tragic vision. So does history. The French, Russian and Chinese revolutions were utopian visions. They paved their paths to paradise with 50 million dead.

The USAโ€™s founding fathers held the tragic vision. They created checks and balances to constrain political leadersโ€™ worst impulses.

Technologistsโ€™ visions

Yet when Americans founded online social networks, the tragic vision was forgotten. Founders were trusted to juggle their self-interest and the public interest when designing these networks and gaining vast data troves.

Users, companies and countries were trusted not to abuse their new social-networked power. Mobs were not constrained. This led to abuse and manipulation.

Belatedly, social networks have adopted tragic visions. Facebook now acknowledges regulation is needed to get the best from social media.

Tech billionaire Elon Musk dabbles in both the tragic and utopian visions. He thinks โ€œmost people are actually pretty goodโ€. But he supports market, not government control, wants competition to keep us honest, and sees evil in individuals.

Muskโ€™s tragic vision propels us to Mars in case short-sighted selfishness destroys Earth. Yet his utopian vision assumes people on Mars could be entrusted with the direct democracy that Americaโ€™s founding fathers feared. His utopian vision also assumes giving us tools to think better wonโ€™t simply enhance our Machiavellianism.

Bill Gates leans to the tragic and tries to create a better world within humanityโ€™s constraints. Gates recognises our self-interest and supports market-based rewards to help us behave better. Yet he believes โ€œcreative capitalismโ€ can tie self-interest to our inbuilt desire to help others, benefiting all.

A different tragic vision lies in the writings of Peter Thiel. This billionaire tech investor was influenced by philosophers Leo Strauss and Carl Schmitt. Both believed evil, in the form of a drive for dominance, is part of our nature.

Thiel dismisses the โ€œEnlightenment view of the natural goodness of humanityโ€. Instead, he approvingly cites the view that humans are โ€œpotentially evil or at least dangerous beingsโ€.

The consequences of seeing evil

The German philosopher Friedrich Nietzsche warned that those who fight monsters must beware of becoming monsters themselves. He was right.

People who believe in evil are more likely to demonise, dehumanise, and punish wrongdoers. They are more likely to support violence before and after anotherโ€™s transgression. They feel that redemptive violence can eradicate evil and save the world. Americans who believe in evil are more likely to support torture, killing terrorists and Americaโ€™s possession of nuclear weapons.

Technologists who see evil risk creating coercive solutions. Those who believe in evil are less likely to think deeply about why people act as they do. They are also less likely to see how situations influence peopleโ€™s actions.

Two years after 9/11, Peter Thiel founded Palantir. This company creates software to analyse big data sets, helping businesses fight fraud and the US government combat crime.

Thiel is a Republican-supporting libertarian. Yet, he appointed a Democrat-supporting neo-Marxist, Alex Karp, as Palantirโ€™s CEO. Beneath their differences lies a shared belief in the inherent dangerousness of humans. Karpโ€™s PhD thesis argued that we have a fundamental aggressive drive towards death and destruction.

Just as believing in evil is associated with supporting pre-emptive aggression, Palantir doesnโ€™t just wait for people to commit crimes. It has patented a โ€œcrime risk forecasting systemโ€ to predict crimes and has trialled predictive policing. This has raised concerns.

Karpโ€™s tragic vision acknowledges that Palantir needs constraints. He stresses the judiciary must put โ€œchecks and balances on the implementationโ€ of Palantirโ€™s technology. He says the use of Palantirโ€™s software should be โ€œdecided by society in an open debateโ€, rather than by Silicon Valley engineers.

Yet, Thiel cites philosopher Leo Straussโ€™ suggestion that America partly owes her greatness โ€œto her occasional deviationโ€ from principles of freedom and justice. Strauss recommended hiding such deviations under a veil.

Thiel introduces the Straussian argument that only โ€œthe secret coordination of the worldโ€™s intelligence servicesโ€ can support a US-led international peace. This recalls Colonel Jessop in the film, A Few Good Men, who felt he should deal with dangerous truths in darkness.

Can we handle the truth?

Seeing evil after 9/11 led technologists and governments to overreach in their surveillance. This included using the formerly secret XKEYSCORE computer system used by the US National Security Agency to colllect peopleโ€™s internet data, which is linked to Palantir. The American people rejected this approach and democratic processes increased oversight and limited surveillance.

Facing the abyss

Tragic visions pose risks. Freedom may be unnecessarily and coercively limited. External roots of violence, like scarcity and exclusion, may be overlooked. Yet if technology creates economic growth it will address many external causes of conflict.

Utopian visions ignore the dangers within. Technology that only changes the world is insufficient to save us from our selfishness and, as I argue in a forthcoming book, our spite.

Technology must change the world working within the constraints of human nature. Crucially, as Karp notes, democratic institutions, not technologists, must ultimately decide societyโ€™s shape. Technologyโ€™s outputs must be democracyโ€™s inputs.

This may involve us acknowledging hard truths about our nature. But what if society does not wish to face these? Those who cannot handle truth make others fear to speak it.

Straussian technologists, who believe but dare not speak dangerous truths, may feel compelled to protect society in undemocratic darkness. They overstep, yet are encouraged to by those who see more harm in speech than its suppression.

The ancient Greeks had a name for someone with the courage to tell truths that could put them in danger – the parrhesiast. But the parrhesiast needed a listener who promised to not to react with anger. This parrhesiastic contract allowed dangerous truth-telling.

We have shredded this contract. We must renew it. Armed with the truth, the Greeks felt they could take care of themselves and others. Armed with both truth and technology we can move closer to fulfilling this promise.The Conversation

Simon McCarthy-Jones, Associate Professor in Clinical Psychology and Neuropsychology, Trinity College Dublin

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Enter your email address below to subscribe to our newsletter
Subscribe
Subscribe To Bergensia's Newsletter
Be the first to get latest updates and exclusive content straight to your email inbox.
Subscribe
Give it a try, you can unsubscribe anytime.
Send this to a friend