Blood

Christopher Lawrence. Scientific Thought: In Context. Editor: K Lee Lerner & Brenda Wilmoth Lerner. Volume 1. Detroit: Gale, 2009.

Introduction

For thousands of years people put the seat of individuality in the heart; in the modern West we locate it in the brain. By contrast, blood has long been designated the place where group identity resides. For ancient peoples, and many modern ones, blood is the fluid that binds together family, clan, order, race, and even nation. It is said to be “thicker than water,” and noble blood is proverbially blue. Blood is deeply connected to breeding and lineage. Breeders of racehorses talk of bloodlines as a geneticist would speak about DNA.

Blood is prominent in the earliest written records. The Sumerians (c.3000 BC) developed a pictogram for it, and the Egyptian Ebers papyrus (c.1600 BC) described the belief that the heart turned food into blood. Ancient peoples seemed to have debated whether the heart or blood—or both—were the seat of life (no one gave much thought to the brain). The Hebrews seemed to have voted for blood and prohibited eating or drinking of it. A passage in Genesis that forbids eating blood is today interpreted by Jehovah’s Witnesses as a prohibition against receiving a blood transfusion.

Historical Background and Scientific Foundations

Body fluids—humors—had far more importance in ancient medical writings than solid body organs. This is hardly surprising: They seem a natural measure of health. Every day we take in fluids and pass urine and sweat. If we are sick, the most obvious signs are disturbances of the fluids: In the summer we may have diarrhea, and in the winter we might cough up runny phlegm. The works associated with the Greek physician Hippocrates (c.460-c.370 BC), although clearly from many different authors and sects, use humors extensively to explain health and disease.

The main humors appearing in the Hippocratic texts—blood, phlegm, black bile, and yellow bile—were codified into a theory of medicine by Galen of Pergamum (AD c.129-c.216), a Greek physician who practiced in Rome. Humoral theory served medicine for 1,400 years, both in Islam and the West. Humors were associated with seasons, age, sex, constitution, and many other things, and blood predominated in young men and in the spring. When Tennyson wrote “In the spring a young man’s fancy lightly turns to thoughts of love,” he was merely expressing an ancient sentiment associated with rising sap or a surge of blood. Bleeding or bloodletting (also known as venesection or phlebotomy) is an ancient practice. It was used in the West until relatively recently and is still widespread in some non-Western cultures.

Ancient medical ideas of the humors (particularly blood) began to be challenged during the scientific revolution of the seventeenth century. The fluid so obviously essential to life became a major object of speculation and experiment, especially after English physician William Harvey (1578-1657) published Exercitatio Anatomica de Motu Cordis et Sanguinis in Animalibus (An anatomical exercise concerning the motion of the heart and blood in animals) in 1628. Interest was also fueled by the growth of chemistry and the new corpuscular (atomistic) philosophy.

In Oxford and London a circle developed around the Earl of Cork, Robert Boyle (1627-1691), an enthusiast for mechanical and chemical inquiries. These investigators used the newly invented air pump to study the functions of blood. The circle included physicians Robert Hooke (1635-1703), curator of experiments at the newly established Royal Society, and Richard Lower (1631-1691). Together they conducted experiments on dogs, exposing their lungs and artificially ventilating them with bellows. Lower, a passionate supporter of the idea of circulation, reported that blood changed color from dark blue to bright red as it passed through the lungs, not in the heart as was previously believed. A younger collaborator of Lower, chemist and physiologist John Mayow (1640-1679), on the basis of experiments and chemical theory, postulated that “elastic” particles in the air were necessary to support combustion and life and were taken into the blood during respiration.

The most remarkable and daring exertions of this band of virtuosi were those involving blood transfusion. Their experiments confirmed that the scientific revolution was constituted as much by a quest for practical results as it was by disinterested inquiry. Their work was provoked by the studies and writings of another member of the group, Sir Christopher Wren (1632-1723), architect of St. Paul’s cathedral. Wren had injected various fluids into the veins of animals and observed the effects. In 1665 Lower used hollow quills to transfuse blood from one dog to another without apparent adverse effects. In 1667 he transfused blood from a sheep to a mad clergyman, Arthur Coga, testing the theory that blood from a “gentle” animal like a lamb might quiet a troubled and disturbed spirit. The patient reported himself “very well.” However, a similar but highly controversial experiment in France, after which the patient died (whether or not from transfusion was and is contested), resulted in a ban on further trials until the nineteenth century.

After the failed blood transfusion experiment in the seventeenth century, there were no further recorded attempts until 1817, when a London obstetrician, James Blondell (c.1790-c.1877), sought a treatment for the often-fatal hemorrhage that could occur following childbirth. He began to experiment with transfusion between dogs, but insisted that only human blood should be used on human patients. He attempted transfusion on ten very sick women. Two were so ill that nothing availed, but five of the remaining eight recovered (of course we cannot be sure this was owing to transfusion). After this, numerous doctors, notably surgeons, tried transfusion with mixed success.

Blood’s Physical Properties

Of importance for future studies of blood was the rise of gas chemistry. From the middle of the seventeenth century onward, common air was analyzed, separating it into a number of constituent gases. Other gases were generated experimentally. At the forefront of this work were the English dissenting minister Joseph Priestley (1733-1804) and the wealthy French chemist Antoine Lavoisier (1743-1794). By the end of the century Lavoisier’s innovative chemical nomenclature had swept all before it and his newly named “oxygen” was made an integral constituent of the investigation of blood. Chemists quickly saw in the study of gases a fast lane to understanding the blood’s respiratory functions.

The nineteenth century saw the beginning of a two-pronged scientific study of blood that has continued ever since. Both inquiries were underpinned by new laboratory sciences. Side by side, physicochemical analysis and biological investigation broke blood into myriad constituents and numbered its countless functions. Nonetheless, scientists failed to demystify it.

Blood is unlike other tissues because of its fluid, mediating role between the environment and the interior of the organism. Ultimately all studies of blood seek to comprehend how this mediation is effected. In the nineteenth century, the physicochemical approach built on the foundation of eighteenth-century gas chemistry and centered on blood’s role in respiration. Much of this work was done in Germany, where physiologists readily embraced the knowledge and technologies of the new professional disciplines of physics and chemistry. In 1838 the iron-containing compound hematin was isolated from blood. Over the next two decades hemoglobin was crystallized, and in 1864 Ernst Felix Hoppe-Seyler (1825-1895), a physiological chemist in Germany, demonstrated that hemoglobin could be split into hematin and a protein.

By the beginning of the twentieth century the hybrid discipline biochemistry had been created, and it soon became the premier science of the blood’s physicochemical properties. In many ways biochemists still approach blood in the same way as their nineteenth-century predecessors. Rather than being limited to respiration, however, modern studies investigate blood’s role in keeping the internal environment of animals constant in the face of an ever-changing external world. This capacity to regulate the internal environment was named homeostasis by the American neurologist Walter Bradford Cannon (1871-1945), building on the studies of the French physiologist Claude Bernard (1813-1878).

In one way or another, the huge array of studies encompassed by biochemistry bear on the blood’s mediating role. These studies include investigation of the blood’s electrical, optical, and thermal properties, iron metabolism, respiratory heme pigments, biosynthetic pathways, enzymes, metabolite transport, maintenance of acid-base balance, and so on. In the biochemical laboratory of a modern hospital, a vast range of substances can now be measured to indicate not only the condition of the blood itself but the state of every organ in the body. The “blood test” has the sort of status in medicine today that bloodletting did 200 years ago. Forensic investigation of blood’s physical properties has also played an important part in decoding the “blood spatter” produced by criminal violence.

Cell theory was first proposed by microscopists in Germany in the 1840s; with it, red blood cells (erythrocytes) and many varieties of white cells were described. (Microscopic red discs had been seen in the blood since the seventeenth century, but without a unifying theory their nature was open to all sorts of speculation). Leukemia (an excess of white cells in the blood) was one of the first cellular diseases to be described. An important account of the disorder was given by Rudolph Virchow (1821-1902), a German pathologist who, in 1856, also named the condition.

It was, however, modern germ theory, created in the 1880s (again largely in Germany), that brought blood to the fore as a biological substance. Hardly had that theory been accepted than the problems of resistance and immunity to bacterial attack raised their heads. Working in Germany, bacteriologists Emil von Behring (1854-1917) and Shibasaburo Kitasato (1853-1931) injected various laboratory animals (mice, rabbits, guinea pigs) with nonlethal doses of the toxin produced by the tetanus bacillus. They found that blood serum from these animals could protect other animals of the same species from otherwise lethal doses of the toxin. Some animals, under the right circumstances, could produce substances that appeared in their blood (antitoxins) that could neutralize the most vicious bacterial poisons. In 1891-1892, using this principle, von Behring successfully treated children suffering from diphtheria with antiserum produced in horses.

Not long after this, bacteria and their products were cultivated, attenuated, and made into prophylactic vaccines—agents that are injected to protect people from diseases such as typhoid and tetanus. In this case, as with curative sera, crucial transformations seemed to have taken place in the blood. Scientists noticed too that injection with horse sera sometimes produced very nasty (anaphylactic) reactions. Later, disorders such as hay fever and asthma were seen to produce similar responses. What was to become the modern massive discipline of immunology was being created in the early twentieth century, and its territory was the blood.

At the beginning of the century two theories of long-term immunity and its immediate response to bacterial invasion and toxin production were propounded. A humoral theory placed resistance in the serum. The most famous supporter of this view was German medical scientist Paul Ehrlich (1854-1915). The cellular theory put forward by the Russian zoologist Élie Metchnikoff (1845-1916) proposed that a special sort of white cell in the blood could ingest alien material. These cells, he said, were phagocytes (from the Greek phagein, “to eat”). The theories were resolved by the view that both sorts of immunity existed and, indeed, complemented one another.

There was yet another piece of laboratory work in the early twentieth century that contributed to the explosion of confusing observations about blood. It was a study that turned out to have major practical consequences. In 1900, in Vienna, Austrian immunologist Karl Landsteiner (1868-1943) mixed blood cells and sera from different individuals in his laboratory. In some cases the cells clumped together (agglutination), in others they did not. Landsteiner and others then divided human blood into four groups, A, B, AB, and O, on the basis of antigens contained in the red cells.

In 1940 Landsteiner discovered a subdivision of these groups, the rhesus or Rh factor. This is a specific agglutinating substance, a protein found in rhesus monkey blood and that of most humans—at least 85% of the world’s population. Those with the protein were labeled Rh positive; those lacking it were called Rh negative. Giving Rh positive blood to an Rh negative individual—whether a laboratory animal or a human being—produced severe agglutinating reactions.

The knowledge of blood groups would prove vital to making transfusions work. In addition, discoveries of blood’s other physical properties lead to knowledge critically important for the use of blood transfusion on a fairly wide scale. One major problem with transfusion was resolved at the beginning of World War I (1914-1918) when it was discovered that sodium citrate maintained the blood’s fluidity outside the body. These discoveries facilitated the use of transfusions in a variety of settings.

In 1937 Cook County Hospital in Chicago established a blood bank where citrated, refrigerated blood was stored for up to ten days. The onset of World War II (1939-1945) saw the creation of many such banks and the mobilization of donors on a massive scale. Most donors then were volunteers; this remains so today. In 1940 a program for enrolling volunteers was begun in New York; by the end of the war 13 million units of plasma, which has a much longer shelf life than blood, had been shipped. Once the hostilities ceased, surgeons back in civilian life regarded blood banks as a medical essential and no longer a luxury.

Modern Cultural Connections

Since then, collection and storage have changed considerably. Preservation has been improved, thanks especially to the plastic bags launched in 1950. New anticoagulants have been introduced, and different components, such as platelets, can now be separated and stored. Freezing red blood cells can extend their shelf life to ten years. In 1947 the American Association of Blood Banks (now the AABB) was formed, and today AABB facilities collect and store nearly all of the nation’s donated blood.

The post-Second World War era saw the emergence of unimagined problems associated with blood transfusion, notably the transmission of infectious diseases, including syphilis, hepatitis B and C, and HIV (human immunodeficiency virus). Many countries now screen donated blood intensely to detect such pathogens, but in areas without these standards—rural China, for example—large-scale outbreaks of HIV have occurred.

Although the study of blood has been a medical specialty since the 1920s—hematology—it has become a subject of universal significance in the life sciences. Unlike an organ that performs only one or two functions, blood is unique. It is the river through which all the body’s activities flow and is therefore of interest to scientists in every discipline. In this sense the huge amount of twenty-first-century research devoted to blood is simply a massive scale increase of the sort of work carried out 100 years ago when blood was assigned the place it still has in biological systems.

Immunology is perhaps the most prominent modern science centered on blood, notably because of immunological problems associated with organ transplants. Molecular biologists have found a gold mine for research in the blood’s complex proteins. The reticulocyte (immature red blood cell) has been a model for protein synthesis. The structure of hemoglobin has proved a source of endless fascination and research. But other scientists have been drawn to study blood as well. Geneticists work on the inheritance of blood groups and the genetics of conditions such as high blood pressure and hemophilia; the latter also attracts molecular biologists and hematologists who are experts in clotting mechanisms. In the field, students of evolution and human populations use blood groups as markers.

Blood continues to have significant racial dimensions. Despite laboratory analysis and reduction to objective elements, ancient ideas of its associations with life and race persist in both popular and scientific culture. Belief in vampires—reanimated corpses that live on human blood—can be traced deep into history. Bram Stoker’s (1847-1912) 1897 novel Dracula hit a resonant chord with the Victorian public; versions of it have been in wide circulation ever since.

In popular and serious scientific thought after World War I (1914-1919), blood was deeply associated with race, a powerful, emotive term that pervaded eugenic rhetoric. In 1924 Robert Allen, a Democratic congressman from West Virginia declared: “The primary reason for the restriction of the alien stream … is the necessity for purifying and keeping pure the blood of America.” The Nazis consistently identified racial purity with “blood” (literally and metaphorically) of Germanic people of so-called pure Ayran (Nordic) descent. Any other “blood” was regarded as polluting.

Today the study of sickle-cell anemia brings together blood, genetics, and race. The disease was first described in 1910 when a Chicago physician, James B. Herrick (1861-1954), described sickle-shaped red corpuscles in an African-American patient’s blood. It was soon recognized as a hereditary disorder and thought to be carried by a dominant gene, raising fear among some whites that interracial marriage would further disseminate the condition. It was cited as a disorder of “Negro blood,” a phrase then in everyday use in biological, social and public health literature. Conceptions of the disease as carried by a dominant gene chimed perfectly with the rediscovery of Mendelian genetics and eugenic sentiments. After World War II, however, sickle-cell anemia was found to be a recessive disorder manifested in molecular abnormality and thus could not be spread by a single individual.

Although this reevaluation was based on the use of a new medical technology—electrophoresis—it has been suggested that this reconstruction of its identity was as much sociopolitical as it was clinical and technological. Molecular biologists are a powerful community. They can inform social policy and influence spending on health. This new view of sickle-cell disease signaled the decline of the doctor as “race detective” and the rise of the “molecular engineer.”