Wednesday, April 07, 2010

Many Species of Animal Law

Today I went to hear Bruce Wagman speak on the subject of "Many Species of Animal Law" at ASU's Sandra Day O'Connor College of Law.  Wagman, an attorney with Schiff Hardin who is also an outside litigator for the Animal Legal Defense Fund, has litigated cases involving animals for 18 years, written a case book on animal law, and teaches animal law courses at several law schools as an adjunct faculty member.  He was introduced by ASU Law School Dean Paul Berman and Arizona Court of Appeals Judge Pat Norris.

Wagman began by defining "animal law" as any law where the status of an animal matters--psychological, biological, welfare, etc. status of the animal, as opposed to its value as property.  He suggested that animal law attorneys "may be the only lawyers on earth whose clients are all innocent."

He divided his talk up into multiple "species" of animal law.

Species 1: Companion Animal Issues

He said this makes up the majority of his cases, and includes injuries by or to animals, including veterinary malpractice.  The challenge is to get courts to recognize that animals are not merely property, since historically companion animals have been viewed as property with low or even zero market value.  In cases where an animal is injured or killed, the market value doesn't recognize the interests of the animal or other kinds of value that companion animals give.  Under the American Law Institute's Restatements of the Law, however, there is a notion of "special property" (or "peculiar property" in California's statutes) which allows quantification of other kinds of worth to an animal owner, for instance if the animal is a therapy dog.  There are no emotional stress damages available.

Other sorts of companion animal cases include custody disputes, which often occur as a result of one partner just trying to inflict distress on another rather than having actual interest in the animal.  Wagman said that courts are beginning to take a better look at the interests of the animal in such cases, and be willing to appoint a guardian ad litem, as occurred in the Michael Vick case and in another case in Tennessee where there was a dispute over custody of a dog between a dead man's girlfriend and parents.

There are dangerous dog issues, where an attorney may be fighting against the classification of a dog as a dangerous or vicious animal, or against its euthanasia--what he called "capital cases" for animals.  In three counties surrounding San Francisco, what happens in the case of a dog biting another dog that requires stitches varies dramatically.  In one county, the dog gets a period of probation.  In another, the dog gets labeled as a dangerous or vicious dog, which requires the owner to meet various conditions of housing the dog, having a certain height of fence, carry additional insurance, and so forth.  And in Santa Clara County, the dog gets euthanized.  He pointed out that that county's statute has an exemption for "mitigating circumstances" which he's successfully used to prevent dogs from being euthanized.

Finally, there are wills and trusts--he said he doesn't do that sort of work, but that 48 states now have mechanisms for having trusts for animals.

He said he considers companion animals to be a sort of "gateway animal" for getting recognition of animals in the law, and noted that we tend to be "speciesists" who would feel very different about snakes vs. Labrador Retrievers.  [IMO, this is rational to the extent that animals differ in cognitive capacities, and I note that at no point did he discuss litigating on behalf of cockroaches against pest control companies.]

Species 2: Farm animal issues--legislation and litigation.

His second species of animal law was about animals killed for food--about 10 billion per year in the United States.  He said the goal here is not to stop the killing, but just to improve the living conditions of animals before they're killed for food.  This is problematic, however, because the animal cruelty statutes are criminal rather than civil (with an exception in North Carolina that will be discussed with regard to Species 3 of animal law), and that the criminal law for animal cruelty excludes farm animals in 35 states.  He discussed a few of the more abusive methods of animal treatment in factory farming--calf crates, in which calves are placed for about the first 60 days of life, gestation crates for pigs (outlawed in Arizona since 2006, as well as illegal in Florida, Oregon, Colorado, and California), and battery cages for chickens.

He also discussed downer animals--animals which are either so seriously injured or ill that they are unable to move, which the meat industry wants to drag in that condition to slaughter.  Wagman raised the concern that such animals, if sick, could potentially spread illness to humans, and listed a bunch of diseases that could potentially so spread, with BSE (mad cow) at the top of the list along with avian flu.  Of these, only BSE has been documented to spread to humans, and the industry position is that there should be no restrictions on downer pigs unless and until a human actually gets sick.  The state of California passed a law that said that all downer animals must be euthanized on the spot; the meat industry sued and overturned the statute in federal district court, but the 9th Circuit just reversed it last week (National Meat Association v. Brown).

Species 3: Animal hoarding--private ownership, breeders, and the sanctuary that is not

Wagman said that there have been 250,000 documented cases of animal hoarding, and that they are difficult cases to work with in multiple ways.  He said he believes such cases involve mental illness, but while the APA has a diagnosis for "hoarding" behavior, it excludes animal hoarding which is considered to be different.  How many animals constitutes hoarding?  He said he likes to say "more than eight," because he has eight animals at home.  Hoarders characteristics include possessing more animals than they can care for, having a sense of being persecuted, and living in deplorable conditions.

He discussed two cases that he litigated, ALDF v. Barbara & Robert Woodley, and ALDF v. Janie Conyers, which involved over 500 animals between them.  The former case, in North Carolina, was able to use North Carolina statute 19a, which allows a civil cause of action for animal cruelty.  Wagman had some horrifying photos from the Woodley case.  They had hundreds of dogs in their home living in their own feces, where ammonia levels were 20 times the USDA maximum allowed in a pig facility.  These ammonia levels caused blindness in the dogs, as well as chemical burns to bare skin that contacted the floor, such as dogs' scrotums.  Multiple dogs were kept in wooden boxes with lids on them, and never let out.  Mrs. Woodley's favorite dog, Buddy, not only had his eyes burned to blindness from ammonia, but the bone in the dog's jaw deteriorated from malnutrition.  Local officials had known of Woodley's problem for 20 years, but considered themselves powerless to do anything about it, since the scale of the problem was so large--the local shelter had only eight kennels, while the Woodleys had about 450 animals.  The ALDF had to coordinate a massive effort to manage the rescue of the animals through their case.

Conyers was an AKC poodle breeder who had 106 poodles living in their own feces.

Wagman said that animal psychological suffering is difficult to show, but it can be done; demonstrating physiological suffering is easier, with objective criteria like the ammonia levels and physical injuries to animals.

There is no law against hoarding (except in Hawaii), just the criminal abuse statutes (and civil in NC).  In the hoarding cases the abuse is typically neglect rather than active abuse.

Species 4: Exotic animal ownership

Wagman has handled about 10 chimpanzee cases.  One was a case involving a couple in West Covina, California who had a chimp named Moe for 35 years that bit two people.  He argued for a guardian ad litem to determine what was in the best interests of the chimp, and arranged to get Jane Goodall and Roger Fouts for that role.  The court looked upon it favorably, but the couple came to an out-of-court settlement.

He also briefly discussed the Stamford, Connecticut case of Travis, the 200-pound chimpanzee who attacked a woman that was in the news last year.

He argued that there should be a legislative fix to ban exotic animal ownership completely--they're wild animals.  [A complete ban seems to me too much--there should be exceptions for research, conservation, breeding programs for endangered species, and so forth.  And shouldn't it be possible to domesticate other wild animals?]  Connecticut has taken the step of banning chimp ownership.

Species 5: Shelter practices - euthanasia, veterinary care, adequate food, water, and sanitation, and hold periods

Animal shelters have an overwhelming job, said Wagman.  The County of Los Angeles, which he sued, operates seven shelters which handle tens of thousands of animals per year.  California law says that all animals must get veterinary care and be held for five days, and allowing animal suffering without treatment is not permissible.  The shelters' own records showed that they weren't meeting that standard for thousands of animals, but they're now working to meet them and having their activity monitored for compliance.  A similar set of cases occurred in Kentucky, when the state transferred all shelter responsibility to the counties.  Although the standards of care were minimal, they weren't meeting it, and there were nutrition, veterinary care, and euthanasia issues.  Upon getting notice, they quickly took action to remedy.

In Georgia, by contrast, there is a statute that prohibits the use of gas chambers for euthanization at shelters, but the Commissioner of Agriculture sent out letters to the shelters asking that they purchase gas chambers for euthanization.  Gas chambers apparently have very ugly results in some cases, such as with unhealthy dogs.  A lawsuit against the state of Georgia for its failure to comply with its own statute resulted an an injunction, which they then immediately violated by sending out more letters asking for gas chamber purchases.  After obtaining a contempt ruling from the court, they finally got compliance.

Species 6: Entertainment

Wagman called this category both the most obvious and the most hidden.  The use of animals in entertainment is obvious, but what is not obvious is what goes on behind the scenes, the knowledge of which drains the fun out of the entertainment.

Circuses, zoos, film and TV ads, animal fighting, public appearances, racing and rodeos, and hunting and fishing are all cases of animals used for entertainment.  Wagman first discussed elephants in circuses, commenting on a recent Ringling Brothers case which was tossed out on an issue of standing.  The case involved the use of bullhooks for elephant training, which injures the animals.  The defense didn't deny use of bullhooks, but claimed that they only use them as "guides."

Elephant treatment in zoos is also problematic, since standing around on hard surfaces causes painful arthritis.  In the wild, elephants are awake 21 hours a day and may move 35 miles per day.

Wagman discussed dog fighting, and said that the Michael Vick case was a wakeup call for America to the reality of dog fighting, which exists in every state and most major cities.

He argued that the use of great apes in film and television should be banned, because of how the training process works.  He said that while trainers claim to use only positive reinforcement training, an undercover person who volunteered for a year and a half with trainer Sid Yost found otherwise.  A young chimpanzee is immediately treated to beating and punching to get them to comply.  Their performance lifetime is about 3-5 years, after which they become to strong to conrol, and end up in private homes, in research, or in zoos, often all alone in barren cases.  Wagman pointed out that the common use of a "smiling" chimpanzee is actually a fear grimace.  He does lots of work for sanctuaries, of which there are nine in the U.S. for chimpanzees (including chimpsanctuarynw.org).

Regarding hunting, he distinguished traditional hunting from canned hunting and Internet hunting.  Hunting is protected in most states, including in many state constitutions.  Canned hunting ranches, where animals are fed by hand by humans before they are flushed out into open areas to be shot, are not considered to be hunting by most traditional hunters.  [But is considered hunting by our former Vice President, Dick Cheney.]  Internet hunting, where a rifle can be fired at live animals over the Internet, has been banned in 30 states.

He mentioned mountain lion hunting in the Black Hills of South Dakota, where mountain lions have become fairly scarce.  A lawsuit was filed to try to stop the hunting on grounds of near-extinction of the animals, but the injunction was denied on the grounds that there were unlikely to be any mountain lions even found and killed.  Two mountain lions were killed shortly thereafter in fairly quick succession, and even though there was a law that prohibited killing female mountain lions with cubs, the second one killed had a cub, and there was no prosecution.

Some Adidas shoes are made with kangaroo skin, and the state of California has banned the importation of kangaroo skin, which Adidas ignored.  Adidas was sued as a result, and they lost at the California Supreme Court--but they responded by persuading the legislature to repeal the ban rather than changing their practices.

Species 7: Species and breed-specific legislation and ADA breedism case.

A variety of dog breeds have been considered at various times and places to be "bad dogs" that create a special danger.  After WWII, it was German Shepherds and Dobermans.  All cases to stop such breed-specific legislation have failed, because the "rational relation" standard is met by only a single case of harm.  A case in progress right now in Concord, California involves Theresa Huerta, a woman suing under the Americans with Disabilities Act to keep her pit bull therapy dog from being euthanized.

Wagman concluded by saying that his overall objective is to keep the public and the courts focused on the real issue, which is ending blatant cases of animal abuse.  Animal law is a growing field, and there's an annual animal law conference in Portland that's now in its fifth year.

Tuesday, April 06, 2010

First two stray dogs of 2010

I caught these two male dogs in the front yard this afternoon--they wandered in while the gate was open, and I closed it to catch them.  No collars, no tags, and the pit mix was unneutered (didn't check the Spitz mix or whatever he is).  At first they were very skittish, but after they finally approached me, both wanted my constant attention.  They were both quickly picked up by the Maricopa County pound--I'm sure they'll get taken to the east side.

As I was closing the gate to catch these guys, I heard a car honk its horn and a dog yelp, and looked up to see the car drive away as a man, woman, and dog stood on the sidewalk, the dog limping.  I asked the man if the dog had just been hit, and if it was his dog, and he answered yes to both.  They walked off, the dog limping (and off leash, with no collar or tags).

Please, if you own animals, be a responsible pet owner.

Against "coloring book" history of science

It's a bad misconception about evolution that it proceeds in a linear progression of one successfully evolving species after another displacing its immediate ancestors.  Such a conception of human history is equally mistaken, and is often criticized with terms such as "Whiggish history" or "determinism" with a variety of adjectives (technological, social, cultural, historical).  That includes the history of science, where the first version we often hear is one that has been rationally reconstructed by looking back at the successes and putting them into a linear narrative.  Oh, there are usually a few errors thrown in, but they're usually fit into the linear narrative as challenges that are overcome by the improvement of theories.

The reality is a lot messier, and getting into the details makes it clear that not only is a Whiggish history of science mistaken, but that science doesn't proceed through the algorithmic application of "the scientific method," and in fact that there is no such thing as "the scientific method."  Rather, there is a diverse set of methods that are themselves evolving in various ways, and sometimes not only do methods which are fully endorsed as rational and scientific produce erroneous results, sometimes methods which have no such endorsement and are even demonstrably irrational fortuitously produce correct results.  For example, Johannes Kepler was a neo-pythagorean number mystic who correctly produced his second law of planetary motion by taking an incorrect version of the law based on his intuitions and deriving the correct version from it by way of a mathematical argument that contained an error.  Although he fortuitously got the right answer and receives credit for devising it, he was not justified in believing it to be true on the basis of his erroneous proof.  With his first law, by contrast, he followed an almost perfectly textbook version of the hypothetico-deductive model of scientific method of formulating hypotheses and testing them against Tycho Brahe's data.

The history of the scientific revolution includes numerous instances of new developments occurring piecemeal, with many prior erroneous notions being retained.  Copernicus retained not only perfectly circular orbits and celestial spheres, but still needed to add epicycles to get his theory any where close to the predictive accuracy of the Ptolemaic models in use.  Galileo insisted on retaining perfect circles and insisting that circular motion was natural motion, refusing to consider Kepler's elliptical orbits.  There seems to be a good case for "path dependence" in science.  Even the most revolutionary changes are actually building on bits and pieces that have come before--and sometimes rediscovering work that had already been done before, like Galileo's derivation of the uniform acceleration of falling bodies that had already been done by Nicole Oresme and the Oxford calculators.  And the social and cultural environment--not just the scientific history--has an effect on what kinds of hypotheses are considered and accepted.

This conservativity of scientific change is a double-edged sword.  On the one hand, it suggests that we're not likely to see claims that purport to radically overthrow existing theory (that "everything we know is wrong") succeed--even if they happen to be correct.  And given that there are many more ways to go wrong than to go right, such radical revisions are very likely not to be correct.  Even where new theories are correct in some of their more radical claims (e.g., like Copernicus' heliocentric model, or Wegener's continental drift), it often requires other pieces to fall into place before they become accepted (and before it becomes rational to accept them).  On the other hand, this also means that we're likely to be blinded to new possibilities by what we already accept that seems to work well enough, even though it may be an inaccurate description of the world that is merely predictively successful.  "Consensus science" at any given time probably includes lots of claims that aren't true.

My inference from this is that we need both visionaries and skeptics, and a division of cognitive labor that's largely conservative, but with tolerance for diversity and a few radicals generating the crazy hypotheses that may turn out to be true.  The critique of evidence-based medicine made by Kimball Atwood and Steven Novella--that it fails to consider prior plausibility of hypotheses to be tested--is a good one that recognizes the unlikelihood of radical hypotheses to be correct, and thus that huge amounts of money shouldn't be spent to generate and test them.  (Their point is actually stronger than that, since most of the "radical hypotheses" in question are not really radical or novel, but are based on already discredited views of how the world works.)  But that critique shouldn't be taken to exclude anyone from engaging in the generation and test of hypotheses that don't appear to have a plausible mechanism, because there is ample precedent for new phenomena being discovered before the mechanisms that explain them.

I think there's a tendency among skeptics to talk about science as though it's a unified discipline, with a singular methodology, that makes continuous progress, and where the consensus at any moment is the most appropriate thing to believe.  The history of science suggests, on the other hand, that it's composed of multiple disciplines, with multiple methods, that proceeds in fits and starts, that has dead-ends, that sometimes rediscovers correct-but-ignored past discoveries, and is both fallible and influenced by cultural context.  At any given time, some theories are not only well-established but unified well with others across disciplines, while others don't fit comfortably well with others, or may be idealized models that have predictive efficacy but seem unlikely to be accurate descriptions of reality in their details.  To insist on an overly rationalistic and ahistorical model is not just out-of-date history and philosophy of science, it's a "coloring book" oversimplification.  While that may be useful for introducing ideas about science to children, it's not something we should continue to hold to as adults.

Friday, April 02, 2010

Scientific autonomy, objectivity, and the value-free ideal

It has been argued by many that science, politics, and religion are distinct subjects that should be kept separate, in at least one direction if not both.  Stephen Jay Gould argued that science and religion have non-overlapping areas of authority (NOMA, or non-overlapping magisteria), with the former concerned about how questions and the latter with why questions, and that conflicts between them won’t occur if they stick to their own domain.  Between science and politics, most have little problem with science informing politics, but a big problem with political manipulation of science.  Failure to properly maintain the boundaries leads to junk science, politicized science, scientism, science wars, and other objectionable consequences.

Heather E. Douglas, in Science, Policy, and the Value-Free Ideal argues that notions of scientific autonomy and a scientific ideal of being isolated from questions of value (political or otherwise) are mistaken, and that this idea of science without regard to value questions (apart from epistemic virtues) is itself a contributing factor to such consequences.  She attributes blame for this value-free ideal of science to post-1940 philosophy of science, though the idea of scientific autonomy appears to me to have roots much further back, including in Galileo’s “Letter to Castelli” and "Letter to the Grand Duchess Christina" and John Tyndall’s 1874 Belfast Address, which were more concerned to argue that religion should not intrude into the domain of science rather than the reverse.  (As I noted in a previous post about Galileo, he did not carve out complete autonomy for natural philosophy from theology, only for those things which can be demonstrated or proven, which he argued that scripture could not contradict--and where it apparently does, scripture must be interpreted allegorically.)

Douglas describes a “topography of values” in the categories of cognitive, ethical, and social values, and distinguishes direct and indirect roles for them.  Within the “cognitive” category go values pertaining to our ability to understand evidence, such as simplicity, parsimony, fruitfulness, coherence, generality, and explanatory power, but excluding truth-linked epistemic virtues such as internal consistency and predictive competency or adequacy, which she identifies not as values but as minimal negative conditions that theories must necessarily meet.  Ethical values and social values are overlapping categories, the former concerned with what’s good or right and the latter with what a particular society values, such as “justice, privacy, freedom, social stability, or innovation” (Douglas, p. 92).  Her distinction between a direct and indirect role is that the former means that values can act directly as reasons for decisions, versus indirectly as a factor in decision-making where evidence is uncertain.

Douglas argues that values can legitimately play a direct role in certain phases of science, such as problem selection, selection of methodology, and in the policy-making arena, but should be restricted to an indirect role in phases such as data collection and analysis and drawing conclusions from evidence.  She identifies some exceptions, however--problem selection and method selection can’t legitimately be guided by values in a way that undermines the science by forcing a pre-determined conclusion (e.g., by selecting a method that is guaranteed to be misleading), and a direct role for ethical values can surface in later stages by discovering that research is causing harm.

Her picture of science is one where values cannot directly intrude between the collection of data and the inference of the facts from that data, but the space between evidence and fact claims is somewhat more complex than she describes.  There is the inference by a scientist of a fact from the evidence, the communication of that fact to other scientists, the publication of that fact in the scientific literature, and its communication to the general public and policy makers.  All but the first of these are not purely epistemic, but are also forms of conduct.  It seems to me that there is, in fact, a potential direct role for ethical values, at the very least, for each such type of conduct, in particular circumstances, which could merit withholding of the fact claim.  For example, a scientist in Nazi Germany could behave ethically by withholding information about how to build an atomic bomb.

Douglas argues that the motivation for the value-free ideal is as a mechanism for preserving scientific objectivity; she therefore gives an account of objectivity that comports with her account of science with values.  She identifies seven types of objectivity that are relevant in three different domains (plus one she rejects), all of which have to do with a shared ground for trust.  First, within the domain of human interactions with the world, are “manipulable objectivity,” or the ability to repeatably and reliably make interventions in nature that give the same result, and “convergent objectivity,” or having supporting evidence for a conclusion from multiple independent lines of evidence.  Second, in the realm of individual thought processes, she identifies “detached objectivity”--a scientific disinterest, freedom from bias, and eschewing the use of values in place of evidence.  There’s also “value-free objectivity,” the notion behind the value-free ideal, which she rejects.  And there’s “value-neutral objectivity,” or leaving personal views aside in, e.g., conducting a review of the literature in a field and identifying possible sets of explanations, or taking a "centrist" or "balanced" view of potentially relevant values.  Finally, in the domain of social processes, Douglas identifies “procedural objectivity,” where use of the same procedures produces the same results regardless of who engages in the procedure, and “intersubjectivity” in two senses--“concordant objectivity,” agreement in judgments between different people, and “interactive objectivity,” agreement as the result of argument and deliberation.

Douglas writes clearly and concisely, and makes a strong case for the significance of values within science as well as in its application to public policy.  Though she limits her discussion to natural science (and focuses on scientific discovery rather than fields of science that involve the production of new materials, an area where more direct use of values is likely appropriate), her account could likely be extended with the introduction of a bit more complexity.  While I don’t think she has identified all or even the primary causes of the “science wars,” which she discusses at the beginning of her book, I think her account is more useful in adjudicating the “sound science”/“junk science” debate that she also discusses, as well as identifying a number of ways in which science isn’t and shouldn’t be autonomous from other areas of society.

[A slightly different version of the above was written as a comment for my Human and Social Dimensions of Science and Technology core seminar. Thanks to Judd A. for his comments.]

Thursday, April 01, 2010

Galileo on the relation between science and religion

Galileo’s view of natural philosophy (science) is that it is the study of the book of nature,” “written in mathematical language” (Finocchiaro 2008, p. 183), as contrasted with theology, the study of the book of Holy Scripture and revelation.  Galileo endorses the idea that theology is the “queen” of the “subordinate sciences” (Finocchiaro 2008, p. 124), by which he means not that theology trumps science in any and all matters.  He distinguishes two senses of theology being “preeminent and worthy of the title of queen”: (1) That “whatever is taught in all the other sciences is found explained and demonstrated in it [theology] by means of more excellent methods and of more sublime principles,” [Note added 12/14/2012: which he rejects] and (2) That theology deals with the most important issues, “the loftiest divine contemplations” about “the gaining of eternal bliss,” but “does not come down to the lower and humbler speculations of the inferior sciences ... it does not bother with them inasmuch as they are irrelevant to salvation” [Note added 12/14/2012: which he affirms] (quotations from Finocchiaro 2008, pp. 124-125).  Where Holy Scripture makes reference to facts about nature, they may be open to allegorical interpretation rather than literal interpretation, unless their literal truth is somehow necessary to the account of “the gaining of eternal bliss.”

Galileo further distinguishes two types of claims about science:  (1) “propositions about nature which are truly demonstrated” and (2) “others which are simply taught” (Finocchiaro 2008, p. 126).  The role of the theologian with regard to the former category is “to show that they are not contrary to Holy Scripture,” e.g., by providing an interpretation of Holy Scripture compatible with the proposition; with regard to the latter, if it contradicts Holy Scripture, it must be considered false and demonstrations of the same sought (Finocchiaro 2008, p. 126).  Presumably, if in the course of attempting to demonstrate that a proposition in the second category is false, it is instead demonstrated to be true, it then must be considered to be part of the former category.  Galileo’s discussion allows that theological condemnation of a physical proposition may be acceptable if it is shown not to be conclusively demonstrated (Finnochiaro 2008, p. 126), rather than a more stringent standard that it must be conclusively demonstrated to be false, which, given his own lack of conclusive evidence for heliocentrism, could be considered a loophole allowing him to be hoist with his own petard.

Galileo also distinguishes between what is apparent to experts vs. the layman (Finnochiaro 2008, p. 131), denying that popular consensus is a measure of truth, but regarding that this distinction is what lies behind claims made in Holy Scripture about physical propositions that are not literally true.  With regard to the theological expertise of the Church Fathers, their consensus on a physical proposition is not sufficient to make it an article of faith unless such consensus is upon “conclusions which the Fathers discussed and inspected with great diligence and debated on both sides of the issue and for which they then all agreed to reject one side and hold the other” (Finnochiaro 2008, p. 133).  Or, in a contemporary (for Galileo) context, the theologians of the day could have a comparably weighted position on claims about nature if they “first hear the experiments, observations, reasons, and demonstrations of philosophers and astronomers on both sides of the question, and then they would be able to determine with certainty whatever divine inspiration will communicate to them” (Finnochiaro 2008, p. 135).

Galileo’s conception of science that leads him to take this position appears to be drawn from what Peter Dear (1990, p. 664), drawing upon Thomas Kuhn (1977), calls “the quantitative, ‘classical’ mathematical sciences” or the “mixed mathematical sciences,” identifying this as a predominantly Catholic conception of science, as contrasted with experimental science developed in Protestant England.  The former conception is one in which laws of nature can be recognized through idealized thought experiments based on limited (or no) actual observations, but demonstrated conclusively by means of rational argument.  This seems to be the general mode of Galileo’s work.  Dear argues that this notion of natural law allows for a conception of the “ordinary course of nature” which can be violated by an observed miraculous event, which comports with a Catholic view that miracles continue to occur in the world.

By contrast, the experimentalist views of Francis Bacon and Robert Boyle involve inductively inferring natural laws on the basis of observations, in which case observing something to occur makes it part of nature that must be accounted for in the generalized law--a view under which a miracle seems to be ruled out at the outset, which was not a problem for Protestants who considered the “age of miracles” to be over (Dear 1990, pp. 682-683).  Dear argues that for the British experimentalists, authentication of an experimental result was in some ways like the authentication of a miracle for the Catholics--requiring appropriately trustworthy observations--but that instead of verifying a violation of the “ordinary course of nature,” it verified what the “ordinary course of nature” itself was (Dear 1990, p. 680).  Where the Catholics like Galileo and Pascal derived conclusions about particulars from universal laws recognized by observation, reasoning, and mathematical demonstration, the Protestants like Bacon and Boyle constructed universal laws by inductive generalization from observations of particulars, and were notably critical of failing to perform a sufficient number of experiments before coming to conclusions (McMullin 1990, p. 821), and put forth standards for hypotheses and experimental method (McMullin 1990, p. 823; Shapin & Schaffer 1985, pp. 25ff & pp. 56-59).  The English experimentalist tradition, arising at a time of political and religious confusion after the English Civil War and the collapse of the English state church, was perhaps an attempt to establish an independent authority for science.  By the 19th century, there were explicit (and successful) attempts to separate science from religious authority and create a professionalized class of scientists (e.g., as Gieryn 1983, pp. 784-787 writes about John Tyndall).

The English experimentalists followed the medieval scholastics (Pasnau, forthcoming) in adopting a notion of “moral certainty” for “the highest degree of probabilistic assurance” for conclusions adopted from experiments (Shapin 1994, pp. 208-209).  This falls short of the Aristotelian conception of knowledge, yet is stronger than mere opinion.  They also placed importance on public demonstration in front of appropriately knowledgeable witnesses--with both the credibility of experimenter and witness being relevant to the credibility of the result.  Where on Galileo’s conception expertise appears to be primarily a function of possessing rational faculties and knowledge, on the experimentalist account there is importance to skill in application of method and to the moral trustworthiness of the participants as a factor in vouching for the observational results.  In the Galilean approach, trustworthiness appears to be less relevant as a consequence of actual observation being less relevant--though Galileo does, from time to time, make remarks about observations refuting Aristotle, e.g., in “Two New Sciences” where he criticizes Aristotle’s claims about falling bodies (Finnochiaro 2008, pp. 301, 303).

The classic Aristotelian picture of science is similar to the Galilean approach, in that observation and data collection is done for the purpose of recognizing first principles and deriving demonstrations by reason from those first principles.  What constitutes knowledge is what can be known conclusively from such first principles and what is derived by necessary connection from them; whatever doesn’t meet that standard is mere opinion (Posterior Analytics, Book I, Ch. 33; McKeon 1941, p. 156).  The Aristotelian picture doesn’t include any particular deference to theology; any discipline could could potentially yield knowledge so long as there were recognizable first principles. The role of observation isn’t to come up with fallible inductive generalizations, but to recognize identifiable universal and necessary features from their particular instantiations (Lennox 2006).  This discussion is all about theoretical knowledge (episteme) rather than practical knowledge (tekne), the latter of which is about contingent facts about everyday things that can change.  Richard Parry (2007) points out an apparent tension in Aristotle between knowledge of mathematics and knowledge of the natural world on account of his statement that “the minute accuracy of mathematics is not to be demanded in all cases, but only in the case of things which have no matter.  Hence its method is not that of natural science; for presumably the whole of nature has matter” (Metaphysics, Book II, Ch. 3, McKeon 1941, p. 715).

The Galilean picture differs from the Aristotelian in its greater use of mathematics (geometry)--McMullin writes that Galileo had “a mathematicism ... more radical than Plato’s” (1990, pp. 822-823) and by its inclusion of the second book, that of revelation and Holy Scripture, as a source of knowledge.  But while the second book is one which can trump mere opinion--anything that isn’t conclusively demonstrated and thus fails to meet Aristotle’s understanding of knowledge--it must be held compatible with anything that does meet those standards.

References
  • Peter Dear (1990) “Miracles, Experiments, and the Ordinary Course of Nature,” ISIS 81:663-683.
  • Maurice A. Finocchiaro, editor/translator (2008) The Essential Galileo.  Indianapolis: Hackett Publishing Company.
  • Thomas Gieryn (1983) “Boundary Work and the Demarcation of Science from Non-Science: Strains and Interests in Professional Ideologies of Scientists,” American Sociological Review 48(6, December):781-795.
  • Thomas Kuhn (1957) The Copernican Revolution: Planetary Astronomy in the Development of Western Thought.  Cambridge, Mass.: Harvard University Press.
  • Thomas Kuhn (1977) The Essential Tension.  Chicago: The University of Chicago Press.
    Lennox, James (2006) “Aristotle’s Biology,” Stanford Encyclopedia of Philosophy, online at http://plato.stanford.edu/entries/aristotle-biology/, accessed March 18, 2010.
  • Richard McKeon (1941) The Basic Works of Aristotle. New York: Random House.
  • Ernan McMullin (1990) “The Development of Philosophy of Science 1600-1900,” in Olby et al. (1990), pp. 816-837.
  • R.C. Olby, G.N. Cantor, J.R.R. Christie, and M.J.S. Hodge (1990) Companion to the History of Science.  London: Routledge.
  • Parry, Richard (2007) “Episteme and Techne,” Stanford Encyclopedia of Philosophy, online at http://plato.stanford.edu/entries/episteme-techne/, accessed March 18, 2010.
  • Robert Pasnau (forthcoming) “Medieval Social Epistemology: Scienta for Mere Mortals,” Episteme, forthcoming special issue on history of social epistemology.  Online at http://philpapers.org/rec/PASMSE, accessed March 18, 2010. 
  • Steven Shapin and Simon Schaffer (1985) Leviathan and the Air Pump: Hobbes, Boyle, and the Experimental Life.  Princeton, N.J.: Princeton University Press.
  • Steven Shapin (1994) A Social History of Truth: Civility and Science in Seventeenth-Century England. Chicago: The University of Chicago Press.
[The above is slightly modified from one of my answers on a midterm exam.  My professor observed that another consideration on the difference between Catholic and Protestant natural philosophers is that theological voluntarism, more prevalent among Protestants, can suggest that laws of nature are opaque to human beings except through inductive experience.  NOTE ADDED 13 April 2010: After reading a couple of chapters of Margaret Osler's Divine Will and the Mechanical Philosophy: Gassendi and Descartes on Contingency and Necessity in the Created World (2005, Cambridge University Press), I'd add Pierre Gassendi to the experimentalist/inductivist side of the ledger, despite his being a Catholic--he was a theological voluntarist.]

Thursday, March 11, 2010

Representation, realism, and relativism

The popular view of the “science wars” of the 1990s is that it involved scientists and philosophers criticizing social scientists for making and accepting absurd claims as a result of an extreme relativistic view about scientific knowledge. Such absurd claims included claims like “the natural world in no way constrains what is believed to be,” “the natural world has a small or nonexistent role in the construction of scientific knowledge,” and “the natural world must be treated as though it did not affect our perception of it” (all due to Harry Collins, quoted in Yves Gingras’ scathingly critical review of his book (PDF), Gravity’s Shadow: The Search for Gravitational Waves). Another example was Bruno Latour’s claim that it was impossible for Ramses II to have died of tuberculosis because the tuberculosis bacillus was not discovered until 1882. This critical popular view is right as far as it goes--those claims are absurd--but the popular view of science also tends toward an overly rationalistic and naively realistic conception of scientific knowledge that fails to account for social factors that influence science as actually practiced by scientists and scientific institutions. The natural world and our social context both play a role in the production of scientific knowledge.

Mark B. Brown’s Science in Democracy: Expertise, Institutions, and Representation tries to steer a middle course between extremes, but periodically veers too far in the relativist direction. Early on, in a brief discussion of the idea of scientific representations corresponding to reality, he writes (p. 6): “Emphasizing the practical dimensions of science need not impugn the truth of scientific representations, as critics of science studies often assume ...” But he almost immediately seems to retract this when he writes that “science is not a mirror of nature” (p. 7) and, in one of several unreferenced and unargued-for claims appealing to science studies that occur in the book, that “constructivist science studies does undermine the standard image of science as an objective mirror of nature” (p. 16). Perhaps he merely means that scientific representations are imperfect and fallible, for he does periodically make further attempts to steer a middle course, such as when he quotes Latour: “Either they went on being relativists even about the settled parts of science--which made them look ridiculous; or they continued being realists even about the warm uncertain parts--and they made fools of themselves” (p. 183). It’s surely reasonable to take an instrumentalist approach to scientific theories that aren’t well established, are somewhat isolated from the rest of our knowledge, or are highly theoretical, but also to take a realist approach to theories that are well established with evidence from multiple domains and have remained stable while being regularly put to the test. The evidence that we have today for a heliocentric solar system, for common ancestry of species, and for the position and basic functions of organs in the human body is of such strength that it is unlikely that we will see that knowledge completely overthrown in a future scientific revolution. But Brown favorably quotes Latour: “Even the shape of humans, our very body, is composed to a great extent of sociotechnical negotiations and artifacts.” (p. 171) Our bodies are not “composed” of “sociotechnical negotiations and artifacts”--this is either a mistaken use of the word “composed” (instead of perhaps “the consequence of”) or a use-mention error (referring to “our very body” instead of our idea of our body).

In Ch. 6, in a section titled “Realism and Relativism” that begins with a reference to the “science wars,” he follows the pragmatist philosopher John Dewey in order to “help resolve some of the misunderstandings and disagreements among today’s science warriors” such as that “STS scholars seem to endorse a radical form of relativism, according to which scientific accounts of reality are no more true than those of witchcraft, astrology, or common sense” (p. 156). Given that Brown has already followed Dewey’s understanding of scientific practice as continuous with common sense (pp.151-152), it’s somewhat odd to see it listed with witchcraft and astrology in that list--though perhaps in this context it’s not meant as the sort of critical common sense Dewey described, but more like folk theories that are undermined or refuted by science.

Brown seems to endorse Dewey’s view that “reality is the world encountered through successful intervention” and favorably quotes philosopher Ian Hacking that “We shall count as real what we can use to intervene in the world to affect something else, or what the world can use to affect us” (pp. 156-157), but he subsequently drops the second half of Hacking’s statement when he writes “If science is understood in terms of the capacity to direct change, knowing cannot be conceived on the model of observation.” Such an understanding may capture experimental sciences, but not observational or historical sciences, an objection Brown attributes to Bertrand Russell, who “pointed out in his review of Dewey’s Logic that knowledge of a star could not be said to affect the star” (p. 158). Brown, however, follows Latour and maintains that “the work of representation ... always transforms what it represents” (p. 177). Brown defends this by engaging in a use-mention error, the failure to properly distinguish between the use of an expression and talking about the expression, when he writes that stars as objects of knowledge are newly created objects (p. 158, more below). Such an error is extremely easy to make when talking about social facts, where representations are themselves partly constitutive of the facts, such as in talk about knowledge or language.

Brown writes that “People today experience the star as known, differently than before ... The star as an object of knowledge is thus indeed a new object” (p. 158). But this is unnecessary given the second half of Hacking’s statement, since we can observe and measure stars--they have impact upon us. Brown does then talk about impact on us, but only by the representation, not the represented: “...this new object causes existential changes in the knower. With the advent of the star as a known object, people actually experience it differently. This knowledge should supplement and not displace whatever aesthetic or religious experiences people continue to have of the star, thus making their experiences richer and more fulfilling” (p. 158). There may certainly be augmented experience with additional knowledge, which may not change the perceptual component of the experience, but I wonder what the Brown’s basis is for the normative claim that religious experiences in particular shouldn’t be displaced--if those religious experiences are based on claims that have been falsified, such as an Aristotelian conception of the universe, then why shouldn’t they be displaced? But perhaps here I’m making the use-mention error, and Brown doesn’t mean that religious interpretations shouldn’t be displaced, only experiences that are labeled as “religious” shouldn’t be displaced.

A few other quibbles:

Brown writes that “all thought relies on language” (p. 56). If this is the case, then nonhuman animals that have no language cannot have thoughts. (My commenter suggested that all sentient beings have language, and even included plants in that category. I think the proposal that sentience requires language is at least plausible, though I wouldn’t put many nonhuman animals or any plants into that category--perhaps chimps, whales, and dolphins. Some sorts of “language” extend beyond that category, such as the dance of honeybees that seems to code distance and direction information, but I interpreted Brown’s claim to refer to human language with syntax, semantics, generative capacity, etc., and to mean that one can’t have non-linguistic thoughts in the form of, say, pictorial imagery, without language. I.e., that even such thoughts require a “language of thought,” to use Jerry Fodor’s expression.)

Brown endorses Harry Collins’ idea of the “experimenter’s regress,” without noting that his evidence for the existence of such a phenomenon is disputed (Allan Franklin, “How to Avoid the Experimenters’ Regress,” Studies in History and Philosophy of Science 25(3, 1994): 463-491). (Franklin also discusses this in the entry on "Experiment in Physics" at the Stanford Encyclopedia of Philosophy.)

Brown contrasts Harry Collins and Robert Evans with Hobbes on the nature of expertise: The former see “expertise as a ‘real and substantive’ attribute of individuals” while “For Hobbes, in contrast, what matters is whether the claims of reason are accepted by the relevant audience.” (p. 116). Brown sides with Hobbes, but this is to make a similar mistake to that Richard Rorty made when claiming that truth is what you can get away with, which is false by its own definition--since philosophers didn’t let him get away with it. This definition doesn’t allow for the existence of a successful fake expert or con artist, but we know that such persons exist from examples that have been exposed. Under this definition, such persons were experts until they were unmasked.

Brown’s application of Hobbes’ views on political representation to nature is less problematic when he discusses the political representation of environmental interests (pp. 128-131) than when he discusses scientific representations of nature (pp. 131-132). The whole discussion might have been clearer had it taken account of John Searle’s account of social facts (in The Construction of Social Reality).

Brown writes that “Just as recent work in science studies has shown that science is not made scientifically ...” (p. 140), without argument or reference.

He apparently endorses a version of Dewey’s distinction between public and private actions with private being “those interactions that do not affect anyone beyond those engaged in the interaction; interactions that have consequences beyond those so engaged he calls public” (p. 141). This distinction is probably not tenable since the indirect consequences of even actions that we’d consider private can ultimately affect others, such as a decision to have or not to have children.

On p. 159, Brown attributes the origin of the concept of evolution to “theories of culture, such as those of Vico and Comte” rather than Darwin, but neither of them had theories of evolution by natural selection comparable to Darwin’s innovation; concepts of evolutionary change go back at least to the pre-Socratic philosophers like the Epicureans and Stoics. (Darwin didn't invent natural selection, either, but he was the first to put all the pieces together and recognize that evolution by natural selection could serve a productive as well as a conservative role.)

[A slightly different version of the above was written as a comment for my Human and Social Dimensions of Science and Technology core seminar. Thanks to Brenda T. for her comments. It should be noted that the above really doesn't address the main arguments of the book, which are about the meaning of political representation and representation in science, and an argument about proper democratic representation in science policy.]

Wednesday, February 24, 2010

Science as performance

The success of science in the public sphere is determined not just by the quality of research but by the ability to persuade. Stephen Hilgartner’s Science on Stage: Expert Advice as Public Drama uses a theatrical metaphor, drawing on the work of Erving Goffman, to shed light on and explain the outcomes associated with three successive reports on diet and nutrition issued by the National Academies of Science, one of which was widely criticized by scientists, one of which was criticized by food industry groups, and one of which was never published. They differed in “backstage” features such as how they coordinated their work and what sources they drew upon, in “onstage” features such as the composition of experts on their committees and how they communicated their results, and how they responded to criticism.

The kinds of features and techniques that Hilgartner identifies as used to enhance perceptions of credibility--features of rhetoric and performance--are the sorts of features relied upon by con artists. If there is no way to distinguish such features as used by con artists from those used by genuine practitioners, if all purported experts are on equal footing and only the on-stage performances are visible, then we have a bit of a problem. All purported experts of comparable performing ability are on equal footing, and we may as well flip coins to distinguish between them. But part of a performance includes the propositional content of the performance--the arguments and evidence deployed--and these are evaluated not just on aesthetic grounds but with respect to logical coherence and compatibility with what the audience already knows. Further, the performance itself includes an interaction with the audience that strains the stage metaphor. Hilgartner describes this as members of the audience themselves taking the stage, yet audience members in his metaphor also interact with each other, individually and in groups, through complex webs of social relationships.

The problem of expert-layman interaction is that the layman in most cases lacks the interactional expertise to even be able to communicate about the details of the evidence supporting a scientific position, and must rely upon other markers of credibility which may be rhetorical flourishes. This is the problem of Plato’s “Charmides,” in which Socrates asserts that only a genuine doctor can distinguish a sufficiently persuasive quack from a genuine doctor. A similar position is endorsed by philosopher John Hardwig, in his paper “Epistemic Dependence,” (PDF) and by law professor Scott Brewer in “Scientific Expert Testimony and Intellectual Due Process,” which points out that the problem faces judges and juries. There are some features which enable successful distinctions between genuine and fake experts in at least the more extreme circumstances--examination of track records, credentials, evaluations by other experts or meta-experts (e.g., experts in methods used across multiple domains, such as logic and mathematics). Brewer enumerates four strategies of nonexperts in evaluating expert claims: (1) “substantive second-guessing,” (2) “using general canons of rational evidentiary support,” (3) “evaluating demeanor,” and (4) “evaluating credentials.” Of these, only (3) is an examination of the merely surface appearances of the performance (which is not to say that it can’t be a reliable, though fallible, mechanism). But when the evaluation is directed not at distinguishing genuine expert from fake, but conflicting claims between two genuine experts, the nonexpert may be stuck in a situation where none of these is effective and only time (if anything) will tell--but in some domains, such as the legal arena, a decision may need to be reached much more quickly than a resolution might become available.

One novel suggestion for institutionalizing a form of expertise that fits into Hilgartner’s metaphor is philosopher Don Ihde’s proposal of “science critics”, in which individuals with at least interactional expertise within the domain they criticize serve a role similar to art and literary critics in evaluating a performance, including its content and not just its rhetorical flourishes.

[A slightly different version of the above was written as a comment for my Human and Social Dimensions of Science and Technology core seminar. The Hardwig and Brewer articles are both reprinted in Evan Selinger and Robert P. Crease, editors, The Philosophy of Expertise. NY: Columbia University Press, 2006, along with an excellent paper I didn't mention above, Alvin I. Goldman's "Experts: Which Ones Should You Trust?" (PDF). The term "interactional expertise" comes from Harry M. Collins and Robert Evans, "The Third Wave of Science Studies: Studies of Expertise and Experience," also reprinted in the Selinger & Crease volume; a case study of such expertise is in Steven Epstein's Impure Science: AIDS, Activism, and the Politics of Knowledge, Berkeley: University of California Press, 1996. Thanks to Tim K. for his comments on the above.]

Monday, February 22, 2010

Is knowledge drowning in a flood of information?

There have long been worries that the mass media are producing a “dumbing down” of American political culture, reducing political understanding to sound bites and spin. The Internet has been blamed for information overload, and, like MTV in prior decades, for a reduction in attention span as the text-based web became the multimedia web, and cell phones have become a more common tool for its use. Similar worries have been expressed about public understanding of science. Nicholas Carr has asked the question, “Is Google Making Us Stupid?”

Yaron Ezrahi’s “Science and the political imagination in contemporary democracies” (a chapter in Sheila Jasanoff's States of Knowledge: The Co-Production of Science and Social Order) argues that the post-Enlightenment synthesis of scientific knowledge and politics in democratic societies is in decline, on the basis of a transition of public discourse into easily consumed, bite-sized chunks of vividly depicted information that he calls “outformation.” Where, prior to the Enlightenment, authority had more of a religious basis and the ideal for knowledge was “wisdom”--which Ezrahi sees as a mix of the “cognitive, moral, social, philosophical, and practical” which is privileged, unteachable, and a matter of faith, the Enlightenment brought systematized, scientific knowledge to the fore. Such knowledge was formalized, objective, universal, impersonal, and teachable--with effort. When that scientific knowledge is made more widely usable, “stripped of its theoretical, formal, logical and mathematical layers” into a “think knowledge” that is context-dependent and localized, it becomes “information.” And finally, when information is further stripped of its context and design for use for a particular purpose, yet augmented with “rich and frequently intense” representations that include “cognitive, emotional, aesthetic, and other dimensions of experience,” it becomes “outformation.”

According to Ezrahi, such “outformations” mix references to objective and subjective reality, and they become “shared references in the context of public discourse and action.” They are taken to be legitimated and authoritative despite lacking any necessary grounding in “observations, experiments, and logic.” He describes this shift as a shift from a high-cost political reality to a low-cost political reality, where “cost” is a measure of the recipient’s ability to consume it rather than the consequences to the polity of its consumption and use as the basis for political participation. This shift, he says, “reflects the diminished propensity of contemporary publics to invest personal or group resources in understanding and shaping politics and the management of public affairs.”

But, I wonder, is this another case of reflecting on “good old days” that never existed? While new media have made new forms of communication possible, was there really a time when the general public was fully invested in “understanding and shaping politics” and not responding to simplifications and slogans? And is it really the case, as Ezrahi argues, that while information can be processed and reconstructed into knowledge, the same is not possible for outformations? Some of us do still read books, and for us, Google may not be “making us stupid,” but rather providing a supplement that allows us to quickly search a vast web of interconnected bits of information that can be assembled into knowledge, inspired by a piece of “outformation.”

[A slightly different version of the above was written as a comment on Ezrahi's article for my Human and Social Dimensions of Science and Technology core seminar. Although I wrote about new media, it is apparent that Ezrahi was writing primarily about television and radio, where "outformation" seems to be more prevalent than information. Thanks to Judd A. for his comments on the above.]

UPDATE (April 19, 2010): Part of the above is translated into Italian, with commentary from Ugo Bardi of the University of Florence, at his blog.

Saturday, February 20, 2010

Seeing like a slime mold

Land reforms instituted in Vietnam under French rule, in India under the British, and in rural czarist Russia introduced simplified rights of ownership and standardized measurements of size and shape that were primarily for the benefit of the state, e.g., for tax purposes. James C. Scott’s Seeing as a State: How Certain Schemes to Improve the Human Condition Have Failed gives these and numerous other examples of ways in which standardization and simplification have been used by the state to make legible and control resources (and people) within its borders. He recounts cases of how the imposition of such standardization often fails or at least has unintended negative consequences, such as his example of German scientific forestry’s introduction of a monoculture of Norway spruce or Scotch pine designed to maximize lumber production, but which led to die-offs a century later. (The monoculture problem of reduced resilience/increased vulnerability is one which has been recognized in an information security context, as well, e.g., in Dan Geer et al.'s paper on Microsoft monoculture that got him fired from @stake and his more recent work.)

Scott’s examples of state-imposed uniformity should not, however, be misconstrued to infer that any case of uniformity is state-imposed, or that such regularities, even if state-imposed, don't have underlying natural constraints. Formalized institutions of property registration and title have appeared in the crevices between states, for example in the squatter community of Kowloon Walled City that existed from 1947-1993 on a piece of the Kowloon peninsula that was claimed by both China and Britain, yet governed by neither. While the institutions of Kowloon Walled City may have been patterned after those familiar to its residents from the outside world, they were internally imposed rather than by a state.

Patterns of highway network design present another apparent counterexample. Scott discusses the design of highways around Paris as being designed by the state to intentionally route traffic through Paris, as well as to allow for military and law enforcement activity within the city in order to put down insurrections. But motorway patterns in the UK appear to have a more organic structure, as a recent experiment with slime molds oddly confirmed. Two researchers at the University of West of England constructed a map of the UK out of agar, putting clumps of oat flakes at the locations of the nine most populous cities. They then introduced a slime mold colony to the mix, and in many cases it extruded tendrils to feed on the oat flakes creating patterns which aligned with the existing motorway design, with some variations. A similar experiment with a map of cities around Tokyo duplicated the Tokyo railway network, slime-mold style. The similarity between transportation networks and evolved biological systems for transporting blood and sap may simply be because they are efficient and resilient solutions.

These examples, while not refuting Scott’s point about frequent failures in top-down imposition of order, suggest that it may be possible for states to achieve success in certain projects by facilitating bottom-up development of ordered structures. The state often imposes an order that has already been developed via some other means--e.g., electrical standards were set up by industry bodies before being codified, IETF standards for IP which don't have the force of law yet are globally implemented. In other cases, states may ratify an emerging order by, e.g., preempting a diversity of state rules with a set that have been demonstrated to be successful, though that runs the risk of turning into a case like Scott describes, if there are local reasons for the diversity.

[A slightly different version of the above was written as a comment on the first two chapters of Scott's book for my Human and Social Dimensions of Science and Technology core seminar. I've ordered a copy of the book since I found the first two chapters to be both lucidly written and extremely interesting. Thanks to Gretchen G. for her comments that I've used to improve (I hope) the above.]

UPDATE (April 25, 2010): Nature 407:470 features "Intelligence: Maze-solving by an amoeboid organism."

Rom Houben not communicating; blogger suppresses the evidence

It has now been demonstrated, as no surprise to skeptics, that Rom Houben was not communicating via facilitated communication, a discredited method by which facilitators have typed for autistic children. A proper test was conducted by Dr. Steven Laureys with the help of the Belgian Skeptics, and it was found that the communications were coming from the facilitator, not from Houben.

A blogger who was a vociferous critic of James Randi and Arthur Caplan for pointing out that facilitated communication is a bogus technique and who had attempted to use Houben's case to argue that Terri Schiavo also may have been conscious is not only unwilling to admit he was wrong, but is deleting comments that point to the results of this new test. I had posted a comment along the lines of "Dr. Laureys performed additional tests with Houben and the facilitator and found that, in fact, the communications were coming from the facilitator, not Houben" with a link to the Neurologica blog; this blogger called that "spam" (on the basis of my posting a similar comment on another blog, perhaps) and "highly misleading" (on the basis of nothing).

As I've said all along, this doesn't mean that Houben isn't "locked in" and conscious, but facilitated communication provides no evidence that he is.

(Previously, previously.)