Toggle light / dark theme

A Future of Fewer Words? Five Trends Shaping the Future of Language
By Lawrence Baines
Published in 2012 in THE FUTURIST 46(2), 42–47.

Summary: Natural selection is as much a phenomenon in human language as it is in natural ecosystems. An ongoing “survival of the fittest” may lead to continuing expansion of image-based communications and the extinction of more than half the world’s languages by this century’s end.

Just after I moved to Oklahoma three years ago, I was invited to a meeting of the state’s Department of Education to discuss Native American languages. I learned that, of the 37 or so Native American languages represented in the state, 22 are already extinct. The last speakers of the Delaware and Mesquakie tongues had recently died; several other languages had only one or two speakers left.

Vanishing languages are not unique to Oklahoma. K. David Harrison, author of When Languages Die (Oxford University Press, 2008), estimates that, of the 6,900 or so languages spoken on the planet, more than half are likely to become extinct over the next century. Today, 95% of people speak one of just 400 languages. The other 6,509 languages are unevenly distributed among the remaining 5%. Hundreds of languages, most with only a few speakers still living, are teetering on oblivion at this very moment.

Why are the world’s languages disappearing? Like living organisms, languages morph over time in response to continuous evolutionary pressures. Any language is in serious trouble if it is spoken by few people or is confined to a remote geographic area. Many of the languages in northeastern Asia, for example, are in isolated, inhospitable regions where low birthrates and high morbidity rates have been facts of life for hundreds of years.

Geography and Distribution of Languages and Speakers

Geographic isolation is a problem that Oklahoma’s dying Native American languages have in common. For example, speakers of Ottawa, of which there may be only three still living Oklahoma, live in the northeastern part of the state, a location that draws few tourists and little business. If the remaining speakers of Ottawa are still alive, there is a good chance that they are over age 70 and rarely travel outside of the community. Anyone who would want to learn the Ottawa language would have to journey down dirt roads and knock on some unfamiliar doors to find out where these speakers live. Once you arrived on their doorstep, they still might not talk to you, especially if you are not a member of the tribe.

In New Guinea, a country that hosts a cauldron of language diversity, villagers on one side of a mountain often speak a completely different language from villagers who may live less than a kilometer away on the other side. If travel to a geographic location is difficult or interactions with speakers of other languages is restricted, then a language has no way to flourish. Like a plant that receives no pollination, a language without some kind of interaction eventually dies.

A second factor contributing to a language’s health is its social desirability. In some parts of the United States, children of first-generation immigrants often grow up in English-speaking neighborhoods, go to English-speaking schools, and come to think of English as the language of acceptance and power. Some of my Texas friends whose parents emigrated from Mexico do not know how to speak, read, or write in Spanish. One friend told me that his parents actually forbade him from speaking Spanish when he was growing up because they considered mastery of English to be essential for success in America.

According to the Global Language Monitor (www.languagemonitor.com, 2011), almost 2 billion people around the globe speak English as either a first or second language, making it the most widely spoken language in the history of the world. The closest runner-up is Mandarin Chinese, with roughly 1 billion speakers, the majority located in or around China. Spanish is the third most widely spoken language, with 500 million speakers, while speakers of Hindi and Arabic come in at fourth and fifth respectively, with between 450 million and 490 million speakers.

French was the most popular language in the world in 1800, but today, Spanish speakers outnumber French speakers worldwide by more than a 2:1 margin. English speakers outnumber French speakers by 10:1.

As the table below shows, 10 languages constitute a combined 82% of all content and traffic on the Internet. Six of them—English, Chinese, Spanish, Arabic, French, and Russian—also happen to be the six “official” languages of the United Nations. The ubiquity of these languages on the Internet, and in international relations and commerce, assures their advance for the foreseeable future.

Languages Represented on the Internet, 2011 est.

Language

Number of Users

Percent of Total Users

Percent Increase, 2000–2011

English

565 million

27%

301%

Chinese

510 million

24

1,478

Spanish

165 million

8

807

Japanese

99 million

5

110

Portuguese

83 million

4

990

German

75 million

4

174

Arabic

65 million

3

2,501

French

60 million

3

398

Russian

60 million

3

1,825

Korean

39 million

2

107

Source: Internet World Stats, www.internetworldstats.com/stats7.htm

Trend 1: Images Are Subverting Words

Not only is the world using fewer languages on a daily basis, but it is also using fewer words. Consider the rich vocabulary and complex sentence constructions in extemporaneous arguments of politicians in earlier centuries against the slick, simplistic sound bites of contemporary times. No politician today speaks like Thomas Jefferson, whose 1801 inaugural address began with the following two sentences: “Called upon to undertake the duties of the first Executive office of our country, I avail myself of the presence of that portion of my fellow citizens which is here assembled to express my grateful thanks for the favor with which they have been pleased to look towards me, to declare a sincere consciousness that the task is above my talents, and that I approach it with those anxious and awful presentiments which the greatness of the charge, and the weakness of my powers so justly inspire. A rising nation, spread over a wide and fruitful land, traversing all the seas with the rich productions of their industry, engaged in commerce with nations who feel power and forget right, advancing rapidly to destinies beyond the reach of mortal eye; when I contemplate these transcendent objects, and see the honour, the happiness, and the hopes of this beloved country committed to the issue and the auspices of this day, I shrink from the contemplation & humble myself before the magnitude of the undertaking.”

During the nineteenth century, Abraham Lincoln and Stephen Douglas debated for hours in open, public arenas around the country. An examination of their spontaneous verbal sparring reveals dexterous vocabulary and complex thought processes delivered in speeches loaded with clarity and wit. In contrast, during the 2004 presidential campaign, George W. Bush’s campaign logo for the election was a simple W alongside an American flag, an essentially wordless communiqué.

The move from language to image is perhaps most apparent in advertisements, which increasingly emphasize sound and image to the exclusion of language. A winner of the 2010 CLIO award for the best commercial of the year was Volkswagen, whose commercial featured a series of rapid close-ups of a man and woman intimately dancing to rap music, followed in the last few seconds by a picture of a car and just two words: “Tough. Beautiful.”

To help encourage communication among tribes who have been long-time rivals, organizations working in Tanzania, where 129 “official languages” exist, have turned to images, not words, to try to get the tribes to communicate with one another. As John Wesley Young reports in Totalitarian Language (University of Virginia, 1992), translators in these organizations have found that trying to find a common language was cumbersome and fraught with unexpected problems, such as the “loaded connotations” of words like comrade and enemy. To de-escalate tensions, translators try to establish communications using only images, which require no intermediary translation and are not as encumbered by pejoration.

Trend 2: The Written Word Is Losing Authority

In the Bible, John 1:1 begins, “In the beginning was the Word, and the Word was with God, and the Word was God.” In Isaiah 48:13, God says, “By my word, I have founded the earth.”

In Christianity, as in most religions, holy words are assumed to have potency well beyond human comprehension, and the mere utterance of a holy word is assumed to have mystical power. J. K. Rowling borrowed this aspect of religious texts in writing the Harry Potter series of books, where her characters are often too fearful to even mention “him that need not be named” (Voldemort).

To get a sense of the power of words in earlier times, it is instructive to read the literate stirrings of a sixteenth-century Italian peasant named Domenico Scandella and his attempt to understand the Bible on his own terms. Scandella’s interpretation of the world as a ball of cheese infested with worms (angels) was considered blasphemous by the priests of the local Catholic Diocese, and he was thrown in prison several times over the course of his life, eventually dying there.

The Church assumed that Scandella’s linguistic interpretation could influence other parishioners in nefarious ways, so it silenced him. Today, thousands of political dissidents around the world are imprisoned on the same principle—that a few well-chosen words have the potency to change society.

The power of words is also substantiated by endless volumes of legal documents. In most countries, an agreement between individuals may be binding only if it is in writing and features the signatures of all involved. In courts of law, the presence of written documentation trumps oral agreements.

With the proliferation of electronic documents, clicking “I ACCEPT” has become equivalent to a written signature. Software programs downloaded from the Internet whose long, legal agreements momentarily flash upon a computer screen, are, in actuality, legally binding documents. In this manner, “proof of click” is replacing the multi-page, hand-signed documents in the legal system.

At first blush, the popularity of texting might be construed as a sort of affirmation for writing. Upon closer inspection, text messages and e-mails have more in common with oral language than written language. Text messages are usually spontaneous, one-shot efforts, written with little to no revision, often in response to a previous communication. They may include pauses (communicated through additional spaces or …), facial expressions (communicated through emoticons such as ;D for a wink and a smile), simple vocabulary, and recursive, sometimes incoherent construction, all of which are characteristics of oral language. Not surprisingly, text messages are generated by a device originally designed for speaking—a telephone.

Some texts are tweets, which are limited to 140 characters. Obviously, a 140-character limit restricts both linguistic complexity and sentence length. Few tweeters are likely to become the next William Faulkner, who commonly used more than 140 words (not characters) in a single sentence.

The cell phone has become a ubiquitous, all-purpose communications tool. However, its small keyboard and tiny screen limit the complexity, type, and length of written messages. Because no sane person wants to read streams of six-point font on a three-inch video screen, phones today are built with menus of images up to the presentation point of the messages themselves.

Trend 3: Changing Environment for Words

Most public libraries around the world are transforming from institutions focused on archives and research to centers for information and entertainment. The old conception of the library, with its mammoth, unabridged dictionary, ordered sets of reference books, and collections of bounded materials, has become a relic. Now most libraries feature large open spaces with Wi-Fi access, plenty of computer terminals, and as many film DVDs and audio CDs as can be purchased on a dwindling budget. Most libraries today spend more on non-print media than on books and magazines. In my local, college-town library, the computer stations always have a line of patrons waiting to log on, and the DVD aisles are packed with browsers, while the book stacks are relatively deserted.

Libraries are simply responding to changes in human behavior. In 1996, Americans spent more time reading than using the Internet. The following year, time spent on the Internet eclipsed reading, and the gap between reading and Internet usage has been expanding every year since. On a typical weekend, when individuals can choose how to fill their time, they read for about five minutes and they watch television, socialize, text, click around the Internet, and play video games for about five hours. In other words, the ratio of time spent with electronic media to time spent reading has ballooned to 60:1.

Research by the Kaiser Foundation found that adolescents, who are particularly heavy users of electronic media, pack a total of 10 hours and 45 minutes of media content into seven and a half hours of media interactions per day. In his book Everything Bad Is Good for You (Riverhead, 2005), Steven Johnson observes that electronic media have been shown to enhance student decision-making processes, to improve hand–eye coordination, and to promote collaborative thinking. However, most electronic media do not build vocabulary, enhance reading comprehension, or improve the quality of writing.

Television shows, even critically acclaimed series, are notoriously simplistic in their use of language. Script analyses of popular television shows such as South Park, 24, CSI, American Idol, and Friday Night Lights, all reveal a preponderance of monosyllabic words and short sentences.

Language simplification is apparent in cinema, as well. Film scripts from Avatar, Planet of the Apes, Transformers, Lord of the Rings, and Star Wars are written at a second- or third-grade readability level. The basic unit of communication for film is the image, with music and special effects playing significant, supplementary roles. Words serve only as minor support.

The move toward grander spectacle through computer-generated images moves film even more toward the visual and farther away from the linguistic. The complete dialogue for the first Terminator film, which served as a harbinger for a new era of special effects, is just 3,850 words—about as long as this magazine article.

Trend 4: Effects of Neural Darwinism

Nobel Prize winning neuroscientist Gerald Edelman postulated that the brain constantly undergoes a “survival of the fittest” process, in which cells respond to environmental stimuli and, in turn, battle for dominance. Thus, avid readers build parts of their brains that are associated with reading while those parts of the brain associated with other tasks, such as hand–eye coordination (exercised during the playing of video games, for example), stabilize or atrophy. This “neural Darwinism,” the constant fight for dominance in the brain, is evident even in very young children. The stimuli that newborns choose to pay attention to will strengthen the circuits and synapses in the brain related to the stimuli. If certain parts of the brain are not stimulated, that part of the brain will not develop.

More than 20 years ago, neuroscientist Marian Diamond noted that enriched environments increase the size of the cortex at any age. Incredibly, detectable increases in cortical development become apparent after only four days.

The hypotheses of Edelman and Diamond have been confirmed in non-laboratory settings by sociologists Betty Hart and Todd Risley, who did studies of language use among parents and children in professional and welfare homes. The sociologists observed that, by age 3, children in professional homes had twice the vocabularies of children in welfare homes.

To find out why, they recorded oral exchanges between parent and children in both environments and found that professional parents averaged 487 utterances per hour with their children with a ratio of positive to negative comments of 6:1 (six positive comments for every one negative comment). On the other hand, in welfare homes, parents only averaged 178 utterances per hour with a ratio of positive to negative comments of 1:2 (one positive comment for every two negative comments).

By age 3, the average IQ of children of professional parents was 117; the average IQ of children of welfare parents was 79. Thus, much of the achievement gap may be attributable to impoverished environments in the early years.

In a more recent Carnegie Mellon University study, psychologists Timothy A. Keller and Marcel Adam Just took PET images (Positron Emission Tomography) of the brains of children who were poor readers, and then offered the children 100 hours of intensive “reading therapy” designed to improve reading effectiveness. Upon the conclusion of the 100 hours of therapy, the students showed significant improvements in their ability to read. When photographs were taken of the children’s brains after the 100 hours of therapy, the physical structure of their brains had changed to look more like the brains of avid readers.

As the world recedes from the written word and becomes inundated with multisensory stimuli (images, sound, touch, taste, and smell), the part of the human brain associated with language will regress. While there are benefits to becoming more visually astute and more aurally discriminating, the areas of the brain associated with language are also associated with critical thinking and analysis. So, as the corpus of language shrinks, the human capacity for complex thinking may shrink with it.

Trend 5: Translating Machines

Imagine a hand-held device that can translate simple phrases into any of several foreign languages. You type a phrase in your native language and the machine instantly translates and pronounces the desired phrase in the target language. Actually, such a machine already exists and may be purchased for about $50.

While today’s machine translators are not perfect, they are surprisingly functional. Most rely on translation algorithms that depend upon the most commonly occurring words in a language. Plain-language-in and plain-language-out enhances the probability that a word is contained in the database of the device and is understandable by the listener. An erudite translation could result in misunderstanding and confusion. That is, the machine is programmed explicitly for the most common words and phrases.

The inevitable proliferation and technological improvements of translating devices will mean more plainspeak, more monosyllabic words, and fewer polysyllabic words. As world commerce continues to expand and the need to communicate in several languages becomes a standard expectation, the emphasis will be on functionality—a few, useful words and durable phrases. Again, the universe of words seems destined to shrink.

The Rise and Fall of Languages, and What Comes Next

Foreign-language courses in K-12 schools and colleges used to focused upon culture as much as language. Students would study the government, religion, history, customs, foods, and etiquette of a country as much as its language. Today, foreign-language teaching is moving away from cultural awareness and toward language as a transaction. If you own a factory in Norway and you want to export your products to Vietnam, it would be in your best interest to become competent in Vietnamese as quickly as you can. What level of competency do you need to achieve your goals? How long will it take to get there? When the enterprise in Vietnam dries up, then the urgency to learn Vietnamese ceases to exist. While the loss of cultural transmission is lamentable, the focus on functionality is understandable, especially in light of widening international trade.

The reverberations of the shift from words as the dominant mode of communication to image-based media are becoming apparent. As we click more and write less, the retreat of polysyllabic words, particularly words with complex or subtle meanings, seems inevitable. The rich vocabulary in books occurs in the exposition, not the dialogue. When a book is adapted for film, a video game, or a television series, the exposition is translated into images, so the more complex language never reaches the ears of the audience. Media associated with the print world (books, magazines, newspapers) are the repositories of sophisticated language, so as individuals read less, they will have less exposure to sophisticated language.

Losing polysyllabic words will mean a corresponding loss of eloquence and precision. Today, many of the most widely read texts emanate from blogs and social networking sites, such as Facebook. Authors of these sites may be non-readers who have little knowledge of effective writing and may have never developed an ear for language. Over the next century, a rise in “tone deaf” writing seems certain.

Finally, more and more languages will disappear from the face of the planet, and world languages will coalesce into pidgin dialects as communication among cultures continues to accelerate. There will be an ongoing “survival of the fittest” battle among languages. If a language is not needed for commerce, identity, or communication, then it will shrink and possibly die.

The French novelist Gustave Flaubert once wrote, “Human language is like a cracked kettle on which we beat out tunes for bears to dance to.” As images replace words, they will foster faster comprehension, enable easier communication, support stronger retention, and stimulate new ways of thinking. The possible consequences of the contraction of written communication are difficult to discern, but ready or not, the age of the image is upon us.

I have just watched this video by Global Futures 2045.

This is my list of things I disagree with:

It starts with scary words about how every crisis comes faster and faster. However this is untrue. Many countries have been running deficits for decades. The financial crisis is no surprise. The reason the US has such high energy costs goes back to government decisions made in the 1970s. And many things that used to be crises no longer happen, like the Black Plague. We have big problems, but we’ve also got many resources we’ve built up over the centuries to help. Much of the challenges we face are political and social, not technical.

We will never fall into a new Dark Ages. The biggest problem is that we aren’t advancing as fast as we could and many are still starving, sick, etc. However, it has always been this way. The 20th century was very brutal! But we are advancing and it is mostly known threats like WMDs which could cause a disaster. In the main, the world is getting safer every day as we better understand it.

We aren’t going to build a new human. It is more like a Renaissance. Those who lost limbs will get increasingly better robotic ones, but they will still be humans. The best reason to build a robotic arm is to attach it to a human.

The video had a collectivist and authoritarian perspective when it said:

“The world’s community and leaders should encourage mankind instead of wasting resources on solving momentary problems.”

This sentence needs to be deconstructed:

1. Government acts via force. Government’s job is to maintain civil order, so having it also out there “encouraging” everyone to never waste resources is creepy. Do you want your policeman to also be your nanny? Here is a quote from C.S. Lewis:

“Of all tyrannies, a tyranny sincerely exercised for the good of its victims may be the most oppressive. It would be better to live under robber barons than under omnipotent moral busybodies. The robber baron’s cruelty may sometimes sleep, his cupidity may at some point be satiated; but those who torment us for our own good will torment us without end for they do so with the approval of their own conscience.”

2. It is wrong to think government is the solution to our problems. Most of the problems that exist today like the Greek Debt Crisis, and the US housing crisis were caused by governments trying to do too much.

3. There is no such thing as the world’s leaders. There is the UN, which doesn’t act in a humanitarian crisis until after everyone is dead. In any case, we don’t need the governments to act. We built Wikipedia.

4. “Managing resources” is codeword for socialism. If their goal is to help with the development of new technologies, then the task of managing existing resources is totally unrelated. If your job is to build robots, then your job is not also to worry about whether the water and air are dirty. Any scientist who talks about managing resources is actually a politician. Here is a quote from Frederic Hayek:

“The curious task of economics is to demonstrate to men how little they really know about what they imagine they can design. Before the obvious economic failure of Eastern European socialism, it was widely thought that a centrally planned economy would deliver not only “social justice” but also a more efficient use of economic resources. This notion appears eminently sensible at first glance. But it proves to overlook the fact that the totality of resources that one could employ in such a plan is simply not knowable to anybody, and therefore can hardly be centrally controlled.”

5. We should let individuals decide what to spend their resources on. People don’t only invest in momentary things. People build houses. In fact, if you are looking for an excuse to drink, being poor because you live in a country with 70% taxes is a good one.

The idea of tasking government to finding the solutions and to do all futuristic research and new products to shove down our throats is wrong and dangerous. We want individuals, and collections of them (corporations) to do it because they will best put it to use in ways that actually improve our lives. Everything is voluntary which encourages good customer relationships. The money will be funded towards the products people actually care about, instead of what some mastermind bureaucrat thinks we should spend money on. There are many historical examples of how government doesn’t innovate as well as the private sector: the French telephone system, Cuba, expensive corn-based ethanol, the International Space Station, healthcare. The free market is imperfect but it leads to fastest technological and social progress for the reasons Frederic Hayek has explained. A lot of government research today is wasted because it never gets put to use commercially. There are many things that can be done to make the private sector more vibrant. There are many ways government can do a better job, and all that evidence should be a warning to not use governments to endorse programs with the goal of social justice. NASA has done great things, but it was only because it existed in a modern society that it was possible.

They come up with a nice list of things that humanity can do, but they haven’t listed that the one of the most important first steps is more Linux. We aren’t going to get cool and smart robots, etc. without a lot of good free software first.

The video says:

“What we need is not just another technological revolution, but a new civilization paradigm, we need philosophy and ideology, new ethics, new culture, new psychology.”

It minimizes the technology aspect when this is the hard work by disparate scientists that will bring us the most benefits.

It is true that we need to refine our understandings of many things, but we are not starting over, just evolving. Anyone who thinks we need to start over doesn’t realize what we’ve already built and all the smart people who’ve come before. The basis of good morals from thousands of years ago still apply. It will just be extended to deal with new situations, like cloning. The general rules of math, science, and biology will remain. In many cases, we are going back to the past. The Linux and free software movement is simply returning computer software to the hundreds of years-old tradition of science. Sometimes the idea has already been discovered, but it isn’t widely used yet. It is a social problem, not a technical one.

The repeated use of the word “new”, etc. makes this video like propaganda. Cults try to get people to reset their perspective into a new world, and convince them that only they have the answers. This video comes off as a sales pitch with them as the solution to our problems, ignoring that it will take millions. Their lists of technologies are random. Some of these problems we could have solved years ago, and some we can’t solve for decades, and they mix both examples. It seems they do no know what is coming next given how disorganized they are. They also pick multiple words that are related and so are repeating themselves. Repetition is used to create an emotional impact, another trait of propaganda.

The thing about innovation and the future is that it is surprising. Many futurists get things wrong. If these guys really had the answers, they’d have invented it and made money on it. And compared to some of the tasks, we are like cavemen.

Technology evolves in a stepwise fashion, and so looking at it as some clear end results on some day in the future is wrong.

For another example: the video makes it sound like going beyond Earth and then beyond the Solar System is a two-step process when in fact it is many steps, and the journey is the reward. If they were that smart, they’d endorse the space elevator which is the only cheap way to get out there, and we can do it in 10 years.

The video suggests that humanity doesn’t have a masterplan, when I just explained that you couldn’t make one.

It also suggests that individuals are afraid of change, when in fact, that is a trait characteristic of governments as well. The government class has known for decades that Social Security is going bankrupt, but they’d rather criticize anyone who wants to reform it rather than fix the underlying problem. This video is again trying to urge collectivism with its criticism of the “mistakes” people make. The video is very arrogant at how it looks down at “the masses.” This is another common characteristic of collectivism.

Here is the first description of their contribution:

“We integrate the latest discoveries and developments from the sciences: physics, energetics, aeronautics, bio-engineering, nanotechnology, neurology, cybernetics, cognitive science.”

That sentence is laughable because it is an impossible task. To understand all of the latest advances would involve talking with millions of scientists. If they are doing all this integration work, what have they produced? They want everyone to join up today, work to be specified later.

The challenge for nuclear power is not the science, it is the lawyers who outlawed new ones in 1970s, and basically have halted all advancements in building safer and better ones. Some of these challenges are mostly political, not scientific. We need to get engineers in corporations like GE, supervised by governments, building safer and cleaner nuclear power.

If you wanted to create all of what they offer, you’d have to hire a million different people. If you were building the pyramids, you could get by with most of your workers having one skill, the ability to move heavy things around. However, the topics they list are so big and complicated, I don’t think you could build an organization that could understand it all, let alone build it.

They mention freedom and speak in egalitarian terms, but this is contradicted by their earlier words. In their world, we will all be happy worker bees, working “optimally” for their collective. Beware of masterminds offering to efficiently manage your resources.

I support discussion and debate. I am all for think-tanks and other institutions that hire scientists. However, those that lobby government to act on their behalf are scary. I don’t want every scientist lobbying the government to institute their pet plan, no matter how good it sounds. They will get so overwhelmed that they won’t be able to do their actual job. The rules of the US Federal government are very limited and generally revolve around an army and a currency. Social welfare is supposed to be handled by the states.

Some of their ideas cannot be turned into laws by the US Congress because they don’t have this authority — the States do. Obamacare is likely to be ruled unconstitutional, and their ideas are potentially much more intrusive towards individual liberty. It would require a Constitutional Amendment, which would never pass and we don’t need.

They offer a social network where scientists can plug in and figure out what they need to do. This could also be considered an actual concrete example of something they are working on. However, there are already social networks where people are advancing the future. SourceForge.net is the biggest community of programmers. There is also Github.com with 1,000,000 projects. Sage has a community advancing the state of mathematics.

If they want to create their own new community solving some aspect, that is great, especially if they have money. But the idea that they are going to make it all happen is impossible. And it will never replace all the other great communities that already exist. Even science happens on Facebook, when people chat about their work.

If they want to add value, they need to specialize. Perhaps they come up with millions of dollars and they can do research in specific areas. However, their fundamental research would very likely get used in ways they never imagined by other people. The more fundamental, the more no one team can possibly take advantage of all aspects of the discovery.

They say there is some research lab they’ve got working on cybernetics. However they don’t demonstrate any results. I don’t imagine they can be that much ahead of the rest of the world who provides them the technology they use to do their work. Imagine a competitor to Henry Ford. Could he really build a car much better given the available technology at the time? My response to anyone who has claims of some advancements is: turn it into a demo or useful product and sell it. All this video offer as evidence here is CGI, which any artist can make.

I support the idea of flying cars. First we need driverless cars and cheaper energy. Unless they are a car or airplane company, I don’t see what this organization will have to do with that task. I have nothing against futuristic videos, but they don’t make clear what is their involvement and instances of ambiguity should be noted.

They are wrong when they say we won’t understand consciousness till 2030 because we already understand it at some level today. Neural networks have been around for decades. IBM’s Jeopardy-playing Watson was a good recent example. However, it is proprietary so not much will come of that particular example. Fortunately, Watson was built on lots of free software, and the community will get there. Google is very proprietary with their AI work. Wolfram Alpha is also proprietary. Etc. We’ve got enough the technical people for an amazing world if we can just get them to work together in free software and Python.

The video’s last sentence suggests that spiritual self-development is the new possibility. But people can work on that today. And again, enlightenment is not a destination but a journey.

We are a generation away from immortality unless things greatly change. I think about LibreOffice, cars that drive themselves and the space elevator, but faster progress in biology is also possible as well if people will follow the free software model. The Microsoft-style proprietary development model has infected many fields.

I spend most of my time thinking about software, and occasionally I come across issues that are relevant to futurists. I wrote my book about the future of software in OpenOffice, and needed many of its features. It might not be the only writing / spreadsheet / diagramming / presentation, etc. tool in your toolbox, but it is a worthy one. OpenDocument Format (ODF) is the best open standard for these sorts of scenarios and LibreOffice is currently the premier tool to handle that format. I suspect many of the readers of Lifeboat have a variant installed, but don’t know much of the details of what is going on.

The OpenOffice situation has been a mess for many years. Sun didn’t foster a community of developers around their work. In fact, they didn’t listen to the community when it told them what to do. So about 18 months ago, after Oracle purchased Sun and made the situation worse, the LibreOffice fork was created with most of the best outside developers. LibreOffice quickly became the version embraced by the Linux community as many of the outside developers were funded by the Linux distros themselves. After realizing their mess and watching LibreOffice take off within the free software community, Oracle decided to fire all their engineers (50) and hand the trademark and a copy of the code over to IBM / Apache.

Now it would be natural to imagine that this should be handed over to LibreOffice, and have all interested parties join up with this effort. But that is not what is happening. There are employees out there whose job it is to help Linux, but they are actually hurting it. You can read more details on a Linux blog article I wrote here. I also post this message as a reminder about how working together efficiently is critical to have faster progress on complicated things.

It’s been a while since anyone contributed a post on space exploration here on the Lifeboat blogs, so I thought I’d contribute a few thoughts on the subject of potential hazards to interstellar travel in the future — if indeed humanity ever attempts to explore that far in space.

It is only recently that the Voyager probes provided us with some idea of the nature of the boundary of our solar system with what is commonly referred to as the local fluff, The Local Interstellar Cloud, through which we have been travelling for the past 100,000 years or so, and which we will continue to travel through for another 10,000 or 20,000 years yet. The cloud has a temperate of about 6000°C — albeit very tenuous.

We are protected by the effects of the local fluff by the solar wind and the sun’s magnetic field, the front between the two just beyond the termination shock where the solar wind slows to subsonic velocities. Here, in the heliosheath, the solar wind becomes turbulent by its interaction with the interstellar medium, and keeping the interstellar medium at bay from the inners of the solar system, the region currently under study by the Voyager 1 and Voyager 2 space probes. It has been hypothesised that there may be a hydrogen wall further out between the bow shock and the heliopause composed of ISM interacting with the edge of the heliosphere, another obstacle to consider with interstellar travel.

The short end of the stick is that what many consider ‘open space’ to traverse once we get beyond the Kuiper belt may in fact be many more mission-threatening obstacles to traverse to reach beyond our solar system. Opinions welcome. I am not an expert on this.

Russia’s hastily convened international conference in St. Petersburg next month is being billed as a last-ditch effort at superpower cooperation in defense of Earth against dangers from space.

But it cannot be overlooked that this conference comes in response to the highly controversial NATO anti-ballistic missile deployments in Eastern Europe. These seriously destabilizing, nuclear defenses are pretexted as a defense against a non-nuclear Iran. In reality, the western moves of anti-missile systems into Poland and Romania create a de facto nuclear first-strike capability for NATO, and they vacate a series of Anti-Ballistic Missile Treaties with the Russians that go back forty years.

Deeply distrustful of these new US and NATO nuclear first-strike capabilities, the Russians announced they will not attend NATO’s planned deterrence summit in Chicago this month. Instead, they are testing Western intentions with a proposal for cooperative project for near-space mapping, surveillance, and defense against Earth-crossing asteroids and other dangerous space objects.

The Russians have invited NATO members as well as forward-thinking space powers to a conference in June in Petrograd. The agenda: Planetary defense against incursions by objects from space. It would be a way of making cooperative plowshares from the space technologies of hair-trigger nuclear terror (2 minutes warning, or less, in the case of the Eastern European ABMs).

It’s an offer the US and other space powers should accept.

High energy experiments like the LHC at the nuclear research centre CERN are extreme energy consumers (needing the power of a nuclear plant). Their construction is extremely costly (presently 7 Billion Euros) and practical benefits are not in sight. The experiments eventually pose existential risks and these risks have not been properly investigated.

It is not the first time that CERN announces record energies and news around April 1 – apparently hoping that some critique and concerns about the risks could be misinterpreted as an April joke. Additionally CERN regularly starts up the LHC at Easter celebrations and just before week ends, when news offices are empty and people prefer to have peaceful days with their friends and families.

CERN has just announced new records in collision energies at the LHC. And instead of conducting a neutral risk assessment, the nuclear research centre plans costly upgrades of its Big Bang machine. Facing an LHC upgrade in 2013 for up to CHF 1 Billion and the perspective of a Mega-LHC in 2022: How long will it take until risk researchers are finally integrated in a neutral safety assessment?

There are countless evidences for the necessity of an external and multidisciplinary safety assessment of the LHC. According to a pre-study in risk research, CERN fits less than a fifth of the criteria for a modern risk assessment (see the press release below). It is not acceptable that the clueless member states point at the operator CERN itself, while this regards its self-set security measures as sufficient, in spite of critique from risk researchers, continuous debates and the publication of further papers pointing at concrete dangers and even existential risks (black holes, strangelets) eventually arising from the experiments sooner or later. Presently science has to admit that the risk is disputed and basically unknown.

It will not be possible to keep up this ostrich policy much longer. Especially facing the planned upgrades of the LHC, CERN will be confronted with increasing critique from scientific and civil side that the most powerful particle collider has yet not been challenged in a neutral and multidisciplinary safety assessment. CERN has yet not answered to pragmatic proposals for such a process that also should constructively involve critics and CERN. Also further legal steps from different sides are possible.

The member states that are financing the CERN budget, the UN or private funds are addressed to provide resources to finally initiate a neutral and multidisciplinary risk assessment.

German version of this article published in Oekonews: http://www.oekonews.at/index.php?mdoc_id=1069458

Related LHC-Critique press release and open letter to CERN:

https://lifeboat.com/blog/2012/02/lhc-critique-press-release…hc-upgrade

Typical physicist’s April joke on stable black holes at the LHC (April 1 2012, German): http://www.scienceblogs.de/hier-wohnen-drachen/2012/04/stabi…rzeugt.php

Latest publications of studies demonstrating risks arising from the LHC experiment:

Prof Otto E. Rössler: http://www.academicjournals.org/AJMCSR/PDF/pdf2012/Feb/9%20Feb/Rossler.pdf

Thomas Kerwick B.Tech. M.Eng. Ph.D.: http://www.vixra.org/abs/1203.0055

Brief summary of the basic problem by LHC-Kritik (still valid since Sep. 2008): http://lhc-concern.info/wp-content/uploads/2008/12/lhc-kriti…ry-908.pdf

Detailed summary of the scientific LHC risk discussion by LHC-Kritik and ConCERNed International: http://lhc-concern.info/wp-content/uploads/2010/03/critical-…ed-int.pdf

We wish you happy Easter and hope for your support of our pragmatic proposals to urgently increase safety in these new fields of nuclear physics.

LHC Critique / LHC Kritik — Network for Safety at nuclear and sub-nuclear high energy Experiments.

www.LHC-concern.info

[email protected]

Tel.: +43 650 629 627 5

New Facebook group: http://www.facebook.com/groups/LHC.Critique/

GatgetBridge is currently just a concept. It might start its life as a discussion forum, later turn into a network or an organisation and hopefully inspire a range of similar activities.

We will soon be able to use technology to make ourselves more intelligent, feel happier or change what motivates us. When the use of such technologies is banned, the nations or individuals who manage to cheat will soon lord it over their more obedient but unfortunately much dimmer fellows. When these technologies are made freely available, a few terrorists and psychopaths will use them to cause major disasters. Societies will have to find ways to spread these mind enhancement treatments quickly among the majority of their citizens, while keeping them from the few who are likely to cause harm. After a few enhancement cycles, the most capable members of such societies will all be “trustworthy” and use their skills to stabilise the system (see “All In The Mind”).

But how can we manage the transition period, the time in which these technologies are powerful enough to be abused but no social structures are yet in place to handle them? It might help to use these technologies for entertainment purposes, so that many people learn about their risks and societies can adapt (see “Should we build a trustworthiness tester for fun”). But ideally, a large, critical and well-connected group of technology users should be part of the development from the start and remain involved in every step.

To do that, these users would have to spend large amounts of money and dedicate considerable manpower. Fortunately, the basic spending and working patterns are in place: People already use a considerable part of their income to buy consumer devices such as mobile phones, tablet computers and PCs and increasingly also accessories such as blood glucose meters, EEG recorders and many others; they also spend a considerable part of their time to get familiar with these devices. Manufacturers and software developers are keen to turn any promising technology into a product and over time this will surely include most mind measuring and mind enhancement technologies. But for some critical technologies this time might be too long. GadgetBridge is there to shorten it as follows:

- GadgetBridge spreads its philosophy — that mind-enhancing technologies are only dangerous when they are allowed to develop in isolation — that spreading these technologies makes a freer world more likely — and that playing with innovative consumer gadgets is therefore not just fun but also serves a good cause.

- Contributors make suggestions for new consumer devices based on the latest brain research and their personal experiences. Many people have innovative ideas but few are in a position to exploit them. Contributors rather donate their ideas that see them wither away or claimed by somebody else.

- All ideas are immediately published and offered free of charge to anyone who wants to use them. Companies select and implement the best options. Users buy their products and gain hands-on experience with the latest mind measurement and mind enhancement technologies. When risks become obvious, concerned users and governments look for ways to cope with them before they get out of hand.

- Once GadgetBridge produces results, it might attract funding from the companies that have benefited or hope to benefit from its services. GadgetBridge might then organise competitions, commission feasibility studies or develop a structure that provides modest rewards to successful contributors.

Your feedback is needed! Please be honest rather than polite: Could GadgetBridge make a difference?

- CERN’s annual meeting to fix LHC schedules in Chamonix: Increasing energies. No external and multi-disciplinary risk assessment so far. Future plans targeting at costly LHC upgrade in 2013 and Mega-LHC in 2022.

- COMMUNICATION to CERN – For a neutral and multi-disciplinary risk assessment before any LHC upgrade

According to CERN’s Chamonix workshop (Feb. 6–10 2012) and a press release from today: In 2012 the collision energies of the world’s biggest particle collider LHC should be increased from 3.5 to 4 TeV per beam and the luminosity is planned to be increased by a factor of 3. This means much more particle collisions at higher energies.

CERN plans to shut down the LHC in 2013 for about 20 months to do a very costly upgrade (for CHF 1 Billion?) to run the LHC at double the present energies (7 TeV per beam) afterwards.

Future plans: A High-Luminosity LHC (HL-LHC) is planned, “tentatively scheduled to start operating around 2022” — with a beam energy increased from 7 to 16.5 TeV(!):
http://cdsweb.cern.ch/journal/CERNBulletin/2012/06/News%20Articles/1423292?ln=en

One might really ask where this should lead to – sooner or later – without the risks being properly investigated. Many critics from different fields are severely alarmed.

For comparison: The AMS 2 experiment for directly measuring cosmic rays in the atmosphere operates on a scale around 1.5 TeV. Very high energetic cosmic rays have only been measured indirectly (their impulse). Sort, velocity, mass and origin of these particles are unknown. In any way, the number of collisions under the extreme and unprecedented artificial conditions at the LHC is of astronomical magnitudes higher than anywhere else in the nearer cosmos.

There were many talks on machine safety at the Chamonix meeting. The safety of humans and environment obviously were not an official topic. That’s why critics turned to CERN in an open letter:

———————————————————–
Communication on LHC Safety directed to CERN

For a neutral and multidisciplinary risk assessment to be done before any LHC upgrade

—————————-
Communiqué to CERN
—————————-

Dear management and scientists at CERN,

Astronomer and Leonardo-publisher Roger Malina recently emphasized that the main problem in research is that “curiosity is not neutral”. And he concluded: “There are certain problems where we cannot cloister the scientific activity in the scientific world, and I think we really need to break the model. I wish CERN, when they had been discussing the risks, had done that in an open societal context, and not just within the CERN context.”

Video of Roger Malina’s presentation at Ars Electronica, following prominent philosopher and leading constructivist Humberto Maturana’s remarkable lecture on science and “certainy”: https://www.youtube.com/watch?v=DOZS2qJrVkU

In the eyes of many critics a number of questions related to LHC safety are not ruled out and some of them have concrete and severe concerns. Also the comparability of the cosmic ray argument is challenged.

Australian risk researcher and ethicist Mark Leggett concludes in a paper that CERN meets less than a fifth of the criteria of a modern risk assessment:
http://lhc-concern.info/wp-content/uploads/2009/09/leggett_r…_1__09.pdf

Without getting into details of the LHC safety discussion – this article in the well-recognized Physics arXiv Blog (MIT’s Technology Review) states: “Black Holes, Safety, and the LHC Upgrade — If the LHC is to be upgraded, safety should be a central part of the plans.”

Similar to pragmatic critics, the author claims in his closing remarks: “What’s needed, of course, is for the safety of the LHC to be investigated by an independent team of scientists with a strong background in risk analysis but with no professional or financial links to CERN.”
http://www.technologyreview.com/blog/arxiv/27319/

The renowned Institute for Technology Assessment and Systems Analysis (ITAS) in Karlsruhe and other risk researchers have already signalized interest in cooperation. We think, in such a process, naturally also CERN and critics should be constructively involved.

Please act in favour of such a neutral and multi-disciplinary assessment, maybe already following the present Chamonix meeting. Even if you feel sure that there are no reasons for any concerns, this must be in your interest, while also being of scientific and public concern.

In the name of many others:
[…]
————————–
LHC-Kritik / LHC-Critique
www.LHC-concern.info

Direct link to this Communication to CERN:
http://lhc-concern.info/?page_id=139
Also published in “oekonews”: http://www.oekonews.at/index.php?mdoc_id=1067776

CERN press release from Feb 13 2012:
http://press.web.cern.ch/press/PressReleases/Releases2012/PR01.12E.html

“Badly designed to understand the Universe — CERN’s LHC in critical Reflection by great Philosopher H. Maturana and Astrophysicist R. Malina”:
https://lifeboat.com/blog/2012/02/badly-designed-to-understa…t-r-malina

“LHC-Kritik/LHC-Critique – Network for Safety at experimental sub-nuclear Reactors”, is a platform articulating the risks related to particle colliders and experimental high energy physics. LHC-Critique has conducted a number of detailed papers demonstrating the insufficiency of the present safety measures under well understandable perspectives and has still got a law suit pending at the European Court of Human Rights.

More info at LHC-Kritik / LHC-Critique:
www.LHC-concern.info
[email protected]
+43 650 629 627 5

Famous Chilean philosopher Humberto Maturana describes “certainty” in science as subjective emotional opinion and astonishes the physicists’ prominence. French astronomer and “Leonardo” publisher Roger Malina hopes that the LHC safety issue would be discussed in a broader social context and not only in the closer scientific framework of CERN.

(Article published in “oekonews”: http://oekonews.at/index.php?mdoc_id=1067777 )

The latest renowned “Ars Electronica Festival” in Linz (Austria) was dedicated in part to an uncritical worship of the gigantic particle accelerator LHC (Large Hadron Collider) at the European Nuclear Research Center CERN located at the Franco-Swiss border. CERN in turn promoted an art prize with the idea to “cooperate closely” with the arts. This time the objections were of a philosophical nature – and they had what it takes.

In a thought provoking presentation Maturana addressed the limits of our knowledge and the intersubjective foundations of what we call “objective” and “reality.” His talk was spiked with excellent remarks and witty asides that contributed much to the accessibility of these fundamental philosophical problems: “Be realistic, be objective!” Maturana pointed out, simply means that we want others to adopt our point of view. The great constructivist and founder of the concept of autopoiesis clearly distinguished his approach from a solipsistic position.

Given Ars Electronica’s spotlight on CERN and its experimental sub-nuclear research reactor, Maturana’s explanations were especially important, which to the assembled CERN celebrities may have come in a mixture of an unpleasant surprise and a lack of relation to them.

During the question-and-answer period, Markus Goritschnig asked Maturana whether it wasn’t problematic that CERN is basically controlling itself and discarding a number of existential risks discussed related to the LHC — including hypothetical but mathematically demonstrable risks also raised — and later downplayed — by physicists like Nobel Prize winner Frank Wilczek, and whether he thought it necessary to integrate in the LHC safety assessment process other sciences aside from physics such as risk search. In response Maturana replied (in the video from about 1:17): “We human beings can always reflect on what we are doing and choose. And choose to do it or not to do it. And so the question is, how are we scientists reflecting upon what we do? Are we taking seriously our responsibility of what we do? […] We are always in the danger of thinking that, ‘Oh, I have the truth’, I mean — in a culture of truth, in a culture of certainty — because truth and certainty are not as we think — I mean certainty is an emotion. ‘I am certain that something is the case’ means: ‘I do not know’. […] We cannot pretend to impose anything on others; we have to create domains of interrogativity.”

Disregarding these reflections, Sergio Bertolucci (CERN) found the peer review system among the physicists’ community a sufficient scholarly control. He refuted all the disputed risks with the “cosmic ray argument,” arguing that much more energetic collisions are naturally taking place in the atmosphere without any adverse effect. This safety argument by CERN on the LHC, however, can also be criticized under different perspectives, for example: Very high energetic collisions could be measured only indirectly — and the collision frequency under the unprecedented artificial and extreme conditions at the LHC is of astronomical magnitudes higher than in the Earth’s atmosphere and anywhere else in the nearer cosmos.

The second presentation of the “Origin” Symposium III was held by Roger Malina, an astrophysicist and the editor of “Leonardo” (MIT Press), a leading academic journal for the arts, sciences and technology.

Malina opened with a disturbing fact: “95% of the universe is of an unknown nature, dark matter and dark energy. We sort of know how it behaves. But we don’t have a clue of what it is. It does not emit light, it does not reflect light. As an astronomer this is a little bit humbling. We have been looking at the sky for millions of years trying to explain what is going on. And after all of that and all those instruments, we understand only 3% of it. A really humbling thought. […] We are the decoration in the universe. […] And so the conclusion that I’d like to draw is that: We are really badly designed to understand the universe.”

The main problem in research is: “curiosity is not neutral.” When astrophysics reaches its limits, cooperation between arts and science may indeed be fruitful for various reasons and could perhaps lead to better science in the end. In a later communication Roger Malina confirmed that the same can be demonstrated for the relation between natural sciences and humanities or social sciences.

However, the astronomer emphasized that an “art-science collaboration can lead to better science in some cases. It also leads to different science, because by embedding science in the larger society, I think the answer was wrong this morning about scientists peer-reviewing themselves. I think society needs to peer-review itself and to do that you need to embed science differently in society at large, and that means cultural embedding and appropriation. Helga Nowotny at the European Research Council calls this ‘socially robust science’. The fact that CERN did not lead to a black hole that ended the world was not due to peer-review by scientists. It was not due to that process.”

One of Malina’s main arguments focused on differences in “the ethics of curiosity”. The best ethics in (natural) science include notions like: intellectual honesty, integrity, organized scepticism, dis-interestedness, impersonality, universality. “Those are the believe systems of most scientists. And there is a fundamental flaw to that. And Humberto this morning really expanded on some of that. The problem is: Curiosity is embodied. You cannot make it into a neutral ideal of scientific curiosity. And here I got a quote of Humberto’s colleague Varela: “All knowledge is conditioned by the structure of the knower.”

In conclusion, a better co-operation of various sciences and skills is urgently necessary, because: “Artists asks questions that scientists would not normally ask. Finally, why we want more art-science interaction is because we don’t have a choice. There are certain problems in our society today that are so tough we need to change our culture to resolve them. Climate change: we’ve got to couple the science and technology to the way we live. That’s a cultural problem, and we need artists working on that with the scientists every day of the next decade, the next century, if we survive it.

Then Roger Malina directly turned to the LHC safety discussion and articulated an open contradiction to the safety assurance pointed out before: He would generally hope for a much more open process concerning the LHC safety debate, rather than discussing this only in a narrow field of particle physics, concrete: “There are certain problems where we cannot cloister the scientific activity in the scientific world, and I think we really need to break the model. I wish CERN, when they had been discussing the risks, had done that in an open societal context, and not just within the CERN context.”

Presently CERN is holding its annual meeting in Chamonix to fix LHC’s 2012 schedules in order to increase luminosity by a factor of four for maybe finally finding the Higgs Boson – against a 100-Dollar bet of Stephen Hawking who is convinced of Micro Black Holes being observed instead, immediately decaying by hypothetical “Hawking Radiation” — with God Particle’s blessing. Then it would be himself gaining the Nobel Prize Hawking pointed out. Quite ironically, at Ars Electronica official T-Shirts were sold with the “typical signature” of a micro black hole decaying at the LHC – by a totally hypothetical process involving a bunch of unproven assumptions.

In 2013 CERN plans to adapt the LHC due to construction failures for up to CHF 1 Billion to run the “Big Bang Machine” at double the present energies. A neutral and multi-disciplinary risk assessment is still lacking, while a couple of scientists insist that their theories pointing at even global risks have not been invalidated. CERN’s last safety assurance comparing natural cosmic rays hitting the Earth with the LHC experiment is only valid under rather narrow viewpoints. The relatively young analyses of high energetic cosmic rays are based on indirect measurements and calculations. Sort, velocity, mass and origin of these particles are unknown. But, taking the relations for granted and calculating with the “assuring” figures given by CERN PR, within ten years of operation, the LHC under extreme and unprecedented artificial circumstances would produce as many high energetic particle collisions as occur in about 100.000 years in the entire atmosphere of the Earth. Just to illustrate the energetic potential of the gigantic facility: One LHC-beam, thinner than a hair, consisting of billions of protons, has got the power of an aircraft carrier moving at 12 knots.

This article in the Physics arXiv Blog (MIT’s Technology Review) reads: “Black Holes, Safety, and the LHC Upgrade — If the LHC is to be upgraded, safety should be a central part of the plans.”, closing with the claim: “What’s needed, of course, is for the safety of the LHC to be investigated by an independent team of scientists with a strong background in risk analysis but with no professional or financial links to CERN.”
http://www.technologyreview.com/blog/arxiv/27319/

Australian ethicist and risk researcher Mark Leggett concluded in a paper that CERN’s LSAG safety report on the LHC meets less than a fifth of the criteria of a modern risk assessment. There but for the grace of a goddamn particle? Probably not. Before pushing the LHC to its limits, CERN must be challenged by a really neutral, external and multi-disciplinary risk assessment.

Video recordings of the “Origin III” symposium at Ars Electronica:
Presentation Humberto Maturana:


Presentation Roger Malina:

“Origin” Symposia at Ars Electronica:
http://www.aec.at/origin/category/conferences/

Communication on LHC Safety directed to CERN
Feb 10 2012
For a neutral and multidisciplinary risk assessment to be done before any LHC upgrade
http://lhc-concern.info/?page_id=139

More info, links and transcripts of lectures at “LHC-Critique — Network for Safety at experimental sub-nuclear Reactors”:

www.LHC-concern.info

As we all know, Venus’s atmosphere & temperature makes it too hostile for colonization: 450°C temperatures and an average surface pressure almost 100 times that of Earth. Both problems are due to the size of its atmosphere — massive — and 95% of which is CO2.

The general consensus is that Venus was more like that of the Earth several billion years ago, with liquid water on the surface, but a runaway greenhouse effect may have been caused by the evaporation of the surface water and subsequent rise of greenhouse gases.

It poses not just a harsh warning of the prospects of global warming on Earth, but also a case study for how to counter such effects — reversing the runaway greenhouse effect.

I have wondered if anyone has given serious thought to chemical processes which could be set in motion on Venus to extract the carbon dioxide from the atmosphere. The most common gas in the Universe is of course hydrogen, and if sufficient quantities could be introduced to the Venusian atmosphere, with the appropriate catalysts, could the carbon dioxide in the atmosphere be eventually reversed back into solid carbon compounds, water vapor and oxygen? The effect of this would of course not only bring down the temperature, but return the surface pressure, with 95% of its atmosphere removed, to one more similar to that of Earth. Perhaps in adding other aerosols the temperatures could be reduced further and avoid a re-runaway effect.

I’d like to hear others thoughts on this. It would be a long term project — but would perhaps make our closest planet our most habitable one in the future — one we could turn into a habitat that would be very accessible, with ample oxygen, water and mineral resources… The study of such a process would also greatly benefit Earth in the event that theorized runaway greenhouse effects start to occur on our own planet, the strategies learned could save it. Other issues to address regarding Venus: lack of magnetic field and its slow rotation would have to be considered, though hardly off-putting, and 150ppm sulfur dioxide in the atmosphere would need to be cleansed — surely not insurmountable.