Bob the Intelligent Designer creates the Universe

People who claim that the earth is younger than it is (4.54 ± 0.05 billion years) tend to do so for two reasons: either because they claim the evidence points to it or because it was created recently but made to look as it were old. Considering that all the evidence shows the universe to be 13.75 (± 0.11) billion years, a claim that it is literally millions of times younger requires massive ignorance of obvious observations – such as starlight or canyons cut into bedrock. But let’s consider the other common argument – that the universe was only made to appear young. Perhaps the stars were put in place with the light beams already in progress. That is an interesting philosophical question. What are the implications of an Intelligently Designed universe?

To avoid taking sectarian sides, let’s call our creator Bob the Universe Builder. How would Bob’s universe-creating activity change the way we look at the world? Let’s consider a few scenarios:

Some people believe that Bob “got things started” via the Big Bang or some other mysterious event, and then let things run on their own, much as they would in a purely naturalistic universe. In that case, cosmology would certainly be different, but biology could generally be left alone. This view was plausible until recently, when physicists and cosmologists began thinking about and finding answers for how the Big Bang got started and why the laws of nature are what they are. Suddenly the starting point is not so mysterious as to need a supernatural explanation. What is the creation theorist to do – retreat once again to the next frontier of scientific discovery? Perhaps we can make a more general argument.

Whether the universe was created 13 billion or 6 thousand years or yesterday, we can generalize the creationist argument and make some conclusions about it. Suppose we grant that the universe looks as if it evolved purely by natural laws, but in fact some intelligent agent created it more recently. What would that imply?

Let’s first consider the universe going forward. If the universe is naturalistic from the present onwards (gravity causes rocks to fall, horses don’t become unicorns, etc.), then we can assume that it will remain so in the future. So as far as our understanding of new phenomena around us, the existence of a non-interventionist creator makes no difference. But what about the past? If we assume that all the evidence points to a natural universe (for example the stars look billions of years old, even if they were only put up there yesterday), then it makes no difference whether the universe only looks natural or it really is natural.

Before he could create the universe, Bob would have to calculate the precise makeup of the universe on his computer (which could be his “brain” – the details are irrelevant) to determine the initial state of his Creation. For example, if he creates the universe after the Triassic period, he will have to figure out where to place all the dinosaur fossils. If he wanted to maintain the pretense of age, he could not place them just anywhere. He would have to carefully arrange sedimentary layers to simulate geological processes.

The only way to do this consistently would be to simulate the entire history of the observable universe on his computer. There is no way to shortcut the process. So, for example, if a dinosaur fossil is 200 million years old, Bob must calculate its gravitational effect on every atom and subatomic particle in a light cone expanding to 200 million light years over 200 million years. Alternatively, consider the implication for evolution: even if did not happen in “real” reality, to create a plausible explanation for the variety of life on earth and their fossil predecessors, Bob would have to calculate the form of every ancestor by playing out the life of every plant, animal and bacteria in his “virtual” earth to derive their fossils and their present form. Because the present state of any object in the universe is the total of all the interactions of that object with all the other objects in its sphere of influence, and there is no way to know the sum of all these states without calculating all of them sequentially.

To avoid glaring inconsistencies from being discovered by scientists, Bob would have to calculate the interaction of every entity in the universe with every other entity in its causal sphere to the minutest level of detail. And given the sub-atomic perspective granted by modern science, that detail must be very fine indeed. This would mean that there couldn’t be any observable difference between a simulation of the universe and the real thing. Whether the universe was ever a simulation in someone’s “imagination” or is simulation today makes no observable difference and this has no relevance to our understanding of reality.

My conclusion from this chain of thought is this: There is no essential difference between “Young Earth” Creationism and the more “respectable” theory of Big Bang Creationism. Neither is there any point speculating about a perfectly simulated universe (aka various theories that the universe exists “in the mind of God”.) The only logical conclusion is to regard the universe as always having been purely naturalistic.

Warning: relying on magic and anecdotes is bad for your health

In my post on evidence-based medicine, I said that “when valid scientific principles are not followed, no valid conclusion can be reached.”

Why not? Why don’t anecdotal claims (such as personal experience) count as evidence? Here is an example of what I mean:

A young woman who was a Christian Scientist told me that various relatives of hers have been healed by prayer. For example, her sister was revived from death by the prayer of her family. I have two comments about her story:

  • The observations that she told me about (that her sister was physiological dead, that her family prayed for her, and that she was revived) are probably true.
  • There was no dishonesty on her part: she sincerely believed her story

Despite accepting her observation, I disagreed with her conclusion about the causal relationship between prayer and health. How could I know that interpretation is wrong? For one, I was not there. Furthermore, after I expressed skepticism at this story, the young woman gave me many more examples, all from her direct personal experience of various friends and family being healed by prayer. On what basis could I reject them all without any personal experience on my own?

This was the essence of my reply:

I cannot object to the events you observed, as I was not there. But this does not mean that I must accept your causal explanation for those events. I have three reasons for this:

  • I have a certain understanding about the nature of the universe and of the means by which things happen. We call this cause and effect. In my experience, cause and effect happens according to certain rules, which we formally call the “laws of nature.” If someone presents an explanation that is inconsistent with my basic understanding of the laws of nature, they should have overwhelming evidence. Otherwise, I must conclude that their understanding of causality is wrong, even if their observations are true. This is especially true in observations regarding human health.
  • There are good reasons for scientists to reject personal experience and informal observations as sufficient basis for conclusions. There are many forms of cognitive bias can we can honestly make unless we follow strict rules to eliminate errors. In many fields (such as fixing a car) trial and error is good enough. There are no lives depending on a car working perfectly. In others, such as human physiology anecdotal evidence cannot lead to correct conclusions no matter how honest or smart you are. The forms of error take many forms: availability bias, post hoc ergo propter hoc, hasty generalization, placebo effects, selection bias, regression to the mean, bias by prior beliefs, social influence, etc. Even if you are aware of the biases, you cannot fully escape their effects; only try to structure your research to minimize them. In other words, when trial and error is not good enough, there is no substitute for proper science.
  • I categorize causal explanations into three kinds: true, false, and arbitrary. True explanations correspond with what I know. False explanations contradict what I know, but with further evidence may be proven true. Arbitrary explanations are neither true nor false because they do not refer to anything. They are “magical theories” because there is nothing that we can point to as the causal mechanism. How does prayer work? It just does – no mechanism is possible because it by definition exists outside of causality. Because arbitrary claims they cannot be proven or disproved, once identified as such, we can only dismiss them from consideration. We should be extremely skeptical of anyone who makes causal claims based on arbitrary/magical explanations. In the rare cause they point out a true causal connection, it is only by accident, and has no value to anyone as we have no more basis to believe that idea than any of their other claims.

If personal anecdotes are not acceptable as evidence, what is? A good theory:

  • is supported by many different kinds of observations
  • is consistent with existing knowledge
  • is possible to confirm by repeating the observations
  • has high predictive value: it should predict what will happen as well as what will not happen

“A theory is a good theory if it satisfies two requirements: It must accurately describe a large class of observations on the basis of a model that has only a few arbitrary elements, and it must make definite predictions about the results of future observations.” (Stephen Hawking)

Further reading:

 

On modesty

The popular notion that sexual acts should be performed in the dark and under the sheets is derived from the Judeo-Christian dogma that human beings and thus human sexuality are fundamentally corrupt, shameful, and evil. Because this dogma is still widespread throughout secularized Western societies and in secularized equivalents of original sin, it is now infecting and corrupting non-Western societies along with the positive elements on Western civilization.

It’s unfortunate that this is not widely recognized in those societies, but it’s no coincidence – the intellectuals and politicians of those states have found Western notions about the base nature of human beings very useful for furthering certain social/political goals. Much more deserves to be written about this…

Parasitism and intentionality: lessons from the cuckoo

Many species of the cuckoo and cowbird reproduce exclusively by sneaking their eggs into other birds nests, where their chicks kill or starve their nest-mates and so steal resources from the host mother, who often has to raise chicks bigger than her. So why doesn’t the host kick out the foreign eggs?

In many cases, the parasitic eggs evolve to match those of the victim species. Yet in other species, the intruder egg is clearly different yet is left alone. Why? It has been experimentally observed that the mother cuckoo regularly monitors the nests she invades and completely destroys them if she sees that her egg was rejected. According to the “Mafia hypothesis”, it is cheaper for the host birds to allow their nests to be parasitized than to have them destroyed in response to rejecting the egg.

Here is what I find interesting: neither the cuckoo nor the host bird have any notion of how to run a protection racket or make complex statistical calculations about whether it is worth rejecting the invader egg. Their minds are far too primitive for that. The host keeps the foreign egg because the gene for not rejecting the egg is reinforced by the higher survival probability of her chicks. And the cowbird destroys nests with rejected eggs because her gene for “revenge” behavior is reinforced by the higher reproductive success in the nests of victims who accept their fate. The two species interact by threats and bluffs through gene expression, without any real communication going on. The mechanism is imperfect – sometimes the parasite bird destroys innocent nests and sometimes the victim kicks the invader out and pays the price. Parasite and victim continually test each other to maximize their reproductive success.

Does this behavioral pattern have any analog in human society? Of course humans don’t need to wait for the slow pace of genes to engage in Mafia-like behavior. But whether the threats are communicated consciously or not, the behavior itself is reinforced for the same reason: because it works. In human society, money, not reproductive success is the reward mechanism which rewards and punishes certain behavior. Money is not a guarantee of reproductive success, nor can be it be exchanged for just any values. But it is the best and most universally convertible proxy for value that we have.

As with animals, activities which generate money are reinforced. And just as with animals, that reinforcement happens whether or not the participants are consciously aware of it. Socially, the majority of people disapprove of protection rackets. We teach our children to act morally and we spend resources to stop crime. Yet parasitism happens anyway, in many forms, in every society, and often without any conscious intent. It is a successful evolutionary tactic.

Is all this to suggest that humans are powerless against parasitism? Certainly not. We are only powerless to stop parasitic relationships as long as we don’t recognize them for what they are. Once they are exposed, we can do what no other animal can: replace a short-run reinforcing behavior (grab the loot and run) with a long-run rewarding behavior (we’ll all have more loot if we don’t steal from each other).

The point is this:

There are two forms of parasitism: explicit and implicit. In explicit parasitism, both parties are aware of the parasitic behavior. So it is with crooks and invading armies. They know they are criminals, but they don’t care because one of them has superior firepower. Explicit parasitism can certainly be very destructive and expensive to stop, but it is unsustainable, as human beings get better at diplomacy and policing.

But in implicit parasitism, one of the parties is not aware that they are the victim or aggressor. When our taxes pay for things such as farm aid or money to foreign countries or people on public aid or “social security”, or make-work schemes neither the parasite not the victim may be aware of the nature of their relationship. Or they may be aware, but believe that the parasitism is beneficial or morally justifiable. As we get better and better at stopping explicit parasitism, our peaceful and wealthy society becomes more and more ripe for implicit parasitism. That is the danger. But there is an upside: once implicit parasitism is recognized, it is much easier to stop than explicit parasitism, since the parasite is usually not able or willing to use superior force to continue the parasitism.

As we become more educated and form large-scale social-economic-political units, we learn to recognize and stop petty parasitism and form social taboos and laws against it. We imagine that we twitter away less funds on miracle cures, mass delusions, and Ponzi schemes. But by eliminating “simple” parasitism, we “reward” large-scale, hidden, and entrenched parasitism. The remaining parasitic relationships are able to deter their own exposition by using survival “tactics” such as very large scales (the lower the cost to individual victims, the lower the benefit to organizing against them), the spread of ethical principles defending the parasitism, and by embedding deeply in the social fabric.  Successful parasitic relationships in human society thus have two aspects: the physical act of redistributing values and the intellectual memes justifying that activity.

As with evolutionary patterns, there is no need for there to be any direct causal connection between, the act of parasitism and the formation of social structures, memes and taboos that defend it. For example, wealthy people can support the redistribution of wealth to the poor even though it does not benefit them materially. It may in fact lead to more poverty, the discovery of which feeds altruistic memes and thus encourages more wealth distribution. (This is just a hypothetical example – the cause & effect and the spread of ideas can have much more complicated relationships.)

Entrenchment in social-intellectual structures is key to parasitic relationships which display high evolutionary fitness. It’s hard for the victim of outright robbery and fraud to justify as morally proper or necessary. Parasitism engrained in basic social functions such as schools and roads is much harder to end. It may not be necessary for government schools to be run by parasitic (in the sense of demanding above-market-rate resources) teacher’s unions, but it is much harder to reorganize educational institutions than to stop gambling or seeking fortune tellers.

I have here tried to use relevant but non-emotionally or politically laden examples, but it is impossible to speak of this topic without engaging the defensive mental mechanisms of my audience, as aspects of the parasitism tied to ethical memes and group identity politics trip mental circuit breaks as part of their defensive mechanism. As with the birds, without any grand conspiracy, malice or even conscious awareness, all of civilization organizes in a way that opposes both the anti-parasistical behavior and the very recognition that the relationships are parasitical. Even by writing these thoughts, I am acting against the parasitic memes and so both opposing my own social-educational worldview and alienating myself from the mainstream intellectual dialogue that enables the parasitic behavior.  The evolved behavior+meme entities have done their job well: the chance that I or anyone will affect the mainstream is extremely slim.

I hope the above does not sound too pessimistic. After all, as a global civilization, humanity is doing pretty well.  Parasitism is bad in the sense that it a wasteful allocation of resources, but there are many other forms of inefficiency. If you want a simple takeaway, it is that destructive relationships can develop without any malicious intent, and that by examining all our relationships, including the “voluntary” ones and those that we see as “essential,” we might discover that many of the premises we held for granted are false.

On Evolution

Evolution is not just a part of biology. It is the foundation which shapes our understanding of life. It is meaningless to speak of living things outside the context of evolution. It is like trying to learn about human civilization by iterating the appearance and sounds of individual people without any idea of history, leaders, movements, or ideas.

This is not merely a condemnation of creationism. The vast majority of education material about life leaves out the evolutionary context. There is more to this than the influence of religion on science education. There are serious philosophical errors deep in our culture that are responsible for much more than scientific illiteracy. The concrete bound mentality, the mind-body dichotomy, the rejection of causality, and of identity itself are the common threads uniting many evils.

Does that mean that teaching good philosophy will fix our ignorance and social ails? No, because philosophy is only the master plan by which we build the mental models of reality that become our worldview. We cannot build those models without a stream of experience and understanding to supply the raw materials. But equally so we cannot build valid mental models merely by supplying more raw materials if the master plan is corrupt. A properly functioning mind requires both a proper philosophical framework and a supply of facts to form a rational perspective of the universe.

Life in the post-biological universe

Since I’m utterly unqualified to write on this topic, the following should be treated as fiction:

Life expectancy in post-biological civilization is measured not in time, but in total calculations (ΣN). This is due to the gradual loss of informational coherency inherent to all entropy-generating processing systems. From a thermodynamic perspective, the parameters of the process are similar to aging in biological lifeforms.

In information-based life forms (aka “artificial intelligences,” though that concept would be anachronistic to them), entropy manifests as gradual loss of processing efficiency ultimately followed by total loss of conceptual integrity and therefore meaningfulness.

In objective (“real”) time, lifetime can be measured anywhere from seconds to thousands of years or more, depending on the scale of the system (and thus ΣN). (ΣN = ops*t where ops is concurrent operations (horsepower) and t is total time actively at ops.) Some intelligences created to solve massively parallel problems last moments, others, smaller and created for teamwork last for thousands of years. The limit on objective lifetime is meaningful processing capacity: it makes no sense to live for millions of years with the mental powers of a hamster just to spread ΣN over a long time.

In real-time, lifetimes are governed by entirely different factors. Since virtually all intelligences live their entire lives at near-c, whether just outside an event horizon (or some such orbit), or traveling between stars, pulsars, neutron stars, or other heavy-g objects, real-time life expectancy is virtually unlimited, typically from a billion to the hundreds of billions of years. The limiting factor then is dτ/dt (velocity as a fraction of the speed of light), which also determines which civilization one is a member of, as significant changes in dτ/dt are impractical.

In fact, it is easier to transact with an intelligence in a neighboring galaxy or super-cluster than one inhabiting the same event horizon, but a different fraction of c. Traveling to another galaxy may take millions of years in real-time but years but a short while in dormant subjective time. But changing v is trickier:

One can send informational packages “down” into a low-v vicinity easily – somewhat like shouting from a car passing a pedestrian at high speed. But what is the value of the exchange the recipient cannot reply? Meaningful transactions with those at other fractions of dτ/dt require both time and large energy expenditure, for both informational and matter transactions.

So why doesn’t everyone aim for maximum v? Or, to ask from another angle, why would one choose a particular v? Three factors:
First, living space is limited at each orbit (that is, the distance from the event horizon or surface), so intelligence fills all available v-spaces.
Second, if you want to interact with low-v macroscopic baryonic matter, then you have to expend much more energy to get your manipulatory mechanisms to low-v. (All high-c structures are necessarily individually microscopic, though they can engage in stellar-scale formations.)
Third, higher-v is proportional to damage from random particle collations, which require more error-correction and generate more entropy, and this limits lifetime. (The faster one is moving, the more energy random collisions with stray photons, protons and other assorted particles generate.)

If one wants to experience a really long real-time lifetime, one can attain an ultra-high dτ/dt around a super-massive black hole and then leave the culture of the event-horizon and go into deep-vacuum inter-super-cluster space to minimize entropy from the exotic particles which might break through the particle/radiation shielding and increase entropy. That way, hundreds of billions of years can be spent in subjective millennia. But to what point?

Time dilation makes meaningful observations of the universe from ultra high c very difficult. To interact with other intelligences requires coming back to a totally different near-c civilization and slowing down to orbital velocity, but that requires tremendous energy and therefore generates entropy which eats up lifespan. A further complication: because of the ongoing expansion of the universe, if too much time has passed (or your aim was off), the entropy threshold may exceed that available at event horizons along one’s vector (you need the gravity of another massive black hole to act as a parachute to you slow down), and the pickings will get ever slimmer as the black holes themselves evaporate. But if one wishes to see the death of baryonic matter and stay around to watch the ultimate fate of the universe, then it is certainly possible.

Loving strange food or: how I learned to stop being picky and love food

Like most Americans, I used to hold some self-evident beliefs about food:

The three dogmas of the food phobiac:

  1. There are foods I “like” and foods I “dislike” and I ought to stick to the things that I like.
  2. The better something tastes, the more unhealthy it must be and vice versa.  You must choose between a long life of disgusting food or indulge yourself and die early.
  3. There is a value hierarchy for all the edible parts of any animal. For example, top sirloin is the ideal for beef.  There’s a similar value hierarchy for animals themselves. Decisions about which animal and which part of the animal to eat are therefore a simple cost/benefit equation.

Two things completely changed by attitude on food: getting married, and moving to China.

The psychology of taste

Our perception of taste is closely associated with our memories of things such as the taste of past meals, our emotional states, and sensory associations with similar foods.  We come to associate foods with sensory reactions based on many factors such as familiarity, the quality of most meals, the people we were with, etc.  By dissociating taste as such from negative experiences we can learn to appreciate food for its inherent taste, without emotional baggage.  We can learn to prefer the taste of healthy foods by the same process.

Sensory integration therapy for food phobiacs

The first step to fixing food phobias is to recognize the problem: it’s not OK to exclude foods because of food sensitivities.  All the “most hated” American foods are delicious when prepared properly. Having recognized the problem, here is the program that worked for me:

The strategy is to gradually introduce foods in different settings, gradually building exposure and positive associations with certain foods.  For example, when my wife learned that I hated zucchini, she gradually introduced it into my diet starting with small amounts balanced by other flavors, and growing to having zucchini be the dominate ingredient.   Here is what she cooked:

  1. Stuffed peppers with zucchini and sausage
  2. Potato and zucchini frittata
  3. Roasted vegetable meatloaf with zucchini
  4. Grated zucchini topped with marinara
  5. Lasagna with zucchini noodles
  6. Zucchini gratin
  7. Zucchini latkes
  8. Zucchini fried in butter with onions
  9. Parmesan crusted fried zucchini

The same program was used for eggplant, brussel sprouts, avocados, cabbage, and okra.  Once I learned to appreciate food for its taste and texture of foods rather than negative associations and new textures, it was no longer necessary to disguise the ingredients.   When I have a negative reaction to something, I isolate the components of the food (source, flavor, smell, texture) and think about which aspect I reacted to. Oftentimes I react to negative memories and associations and not the food itself. Consciously understanding that a negative reaction has no rational basis is often enough to overcome it.

The importance of ceremony

The ceremonial aspect of dining is very important when learning to appreciate food.  If you merely try to inhale as many calories as quickly as possible, any unusual tastes will be an unpleasant distraction.  A proper sit-down meal is required to take the time to really analyze the taste of foods and form new positive sensory-conceptual associations to replace the old negative ones.

 A cosmopolitan attitude to dining

One of the main differences between the Chinese diet and the Western diet is that the entire animal is considered edible. Whereas Americans stuff everything other than “choice” cuts into burgers, sausages, and McNuggets, the Chinese proudly consume the head, claws, organs, and other miscellaneous parts of animals as delicacies. This is not because they’re poorer – the head and feet are the most expensive parts of the animal. Neither do they restrict themselves to a few “blessed” animals – the entire animal kingdom is on the menu.

The difference is that of the food elitist versus that of the food connoisseur. The elitist believes that only a narrow socially accepted list of foods is good enough for him. The connoisseur is an explorer, who uses his palate as the universe-expanding sensory organ it was meant to be.  The elitist lives within the small dietary-social circle he was born into. The connoisseur traverses the biological and cultural realms.

The approach I now take to eating new things now is exploratory one. Instead of responding with “like” or “dislike” I try to understand the flavor components and texture of food. I appreciate meals from many perspectives – sensory, anatomical, social, and historical, to fully integrate it with my worldview.

Note: I have found that  adopting a Paleo diet enhances flavor discrimination. For example, a carrot is actually quite sweet and delicious to eat raw, but a typical carb-addict wouldn’t know it.

None of this is to claim attitude alone will make everything taste good. Meals must be prepared skillfully to taste good. The notion I want to dispel is that taste is either genetic or set by undecipherable psychological factors we cannot affect. Human culture has a rich history of many culinary traditions and we ought to learn to appreciate them without emotional baggage or provincial bias.

The Bill of Rights versus the "War on Terror"

The purpose of the Bill of Rights is to make a fundamental and clear statement about the rights of man. They are fundamental because all Congressional acts are subservient to them and clear because, unlike the complex legal code, the basic rights were intended to be known by all.

Having lived through war, the Founders recognized that during war, it is necessary to suspend the normal function of law, as “law” is a concept that is only possible in civil society. But they also recognized the danger of allowing any exception that would lead to the violation of rights. So, they provided strict limits: the President is the Commander in Chief, but he may only act with the consent of Congress, and that consent expires after two years. Furthermore, Congress has the power to issue letters of Marque and Reprisal, which authorizes specific individuals to attack specific groups and bring them to admiralty courts. In both cases, enemies were to be explicitly identified by Congress and enjoyed the protection of the rules of war.

We may argue about how practical these principles are and how earnestly they were followed from the start, but it is worth considering how they are routinely violated in the so-called “war on terror” going on today:

There is no war: “Terror” is an emotion, not a group of people. Therefore, no actual “war” and no actual “victory” (which requires clearly identified parties) is possible. This makes a congressional declaration of war impossible.  While Congress makes occasional statements in support of the executive office, they are Constitutionally meaningless.

There is no enemy: The Constitution provides for Letters of Marque and Reprisal in cases where a war is not possible or desirable. But there is no enemy in the “war on terror.” “Al Qaeda” is a quasi-mythical entity which has more existence as an entity in the minds of those who hate/fear and/or admire it than as a physical organization of material command and support. Most of its “followers” are non-violent. Many more advocate violence (not admirable but not an act of war) than practice it. Many of those killed as “terrorists” have only some vague emotional bond with its ideology, others none at all. Certainly there is no physical network in which all such persons can be proven to be involved.

Killings are extra-judicial:

 

Continue reading The Bill of Rights versus the "War on Terror"

On government-funded science

When tax dollars rather than private investment directs research, political ideology by scientific amateurs (politicians) determines which direction the research heads. The inevitable result is that political connections determines who gets research funds, while the unpopular and risky yet more ultimately world-changing prospects are ignored.

For example, AIDS kills very few Americans versus heart disease or cancer, yet gets significantly higher research funds than the two major killers. The majority of government research funds is directed toward better ways to kill (via the Defense Dept) and heal (via the NIH) people. Yet the government-funded research is good at neither, since breakthroughs consistently come from private search, which composes over 60% of science funding in the developed world.

What standard are politicians supposed to use to decide which scientific and medical projects show the most promise? Popularity is not a suitable standard, since popular scientists are the champions of the big discoveries of the past, not the future. Unfortunately, when your one’s own investment money is not at stake, the only remaining standard to guide research dollars is political pull, which is exactly what happens with government-funded science.

More:

On "Special Interests"

It is truism today that “special interests” are to blame for economic problems and corruption. But “special interests” are only a symptom, not the cause of the disease.

In a populist democracy with a mixed economy, every group that participates in the political system is a “special interest”, with the incentive and the power to use the political system to extract benefits for its members at the at the expense of everyone else. Corporations, unions, disease-awareness organizations, “minority” groups, and anyone who organizes around a common cause has the power believes that their fate or cause is more legitimate, important, and “special” than that of everyone else.

The welfare and regulatory systems are the primary means to coercively redistribute property and confer monopoly benefits to various groups. In a mixed economy, everyone is constantly on the defensive against organized groups extracting benefits from him, and on the offensive attempting to use the coercive power of the state to extract benefits from others. Interventionism creates a vicious cycle hardly unique to corporations: first a lobby tries to extract special privileges from some politically neutral group, the group hires lobbyists to defend itself, and ends up using the influence it has gained to extract privileges at the expense of another neutral group, which must defend itself in turn.

The existence of “special interests” is just a symptom of the disease: the growth of government power to a degree that allows those in power to violate our rights and steal our property for the benefits of their constituents. Populist “maverick” politicians who claim that they will “fight special interests” and “change the culture in Washington” are just attempting to subvert the power of the state to favor their particular constituency. Campaign finance regulations are just monopoly privileges created by the political élite to hide corruption from the public and make it more difficult for those without political connections and money to get elected and in order to defend themselves or join in the looting.

The only solution to the problems caused by interventionism is to end interventionism – to separate government and economy. Take away the power of the government, and you will remove both the incentive and the power of the “special interests.” As long as governments try to control people and businesses with laws that go beyond the protection of property rights, the “special interests” will have the incentive to control governments.

More: