Just before his appointment as head of the US National Institutes of Health (NIH), Francis Collins, the most prominent medical geneticist of our time, had his own genome scanned for disease susceptibility genes. He had decided, so he said, that the technology of personalised genomics was finally mature enough to yield meaningful results. Indeed, the outcome of his scan inspired The Language of Life, his recent book which urges every individual to do the same and secure their place on the personalised genomics bandwagon.
So, what knowledge did Collins’s scan produce? His results can be summarised very briefly. For North American males the probability of developing type 2 diabetes is 23%. Collins’s own risk was estimated at 29% and he highlighted this as the outstanding finding. For all other common diseases, however, including stroke, cancer, heart disease, and dementia, Collins’s likelihood of contracting them was average.
Predicting disease probability to within a percentage point might seem like a major scientific achievement. From the perspective of a professional geneticist, however, there is an obvious problem with these results. The hoped-for outcome is to detect genes that cause personal risk to deviate from the average. Otherwise, a genetic scan or even a whole genome sequence is showing nothing that wasn’t already known. The real story, therefore, of Collins’s personal genome scan is not its success, but rather its failure to reveal meaningful information about his long-term medical prospects. Moreover, Collins’s genome is unlikely to be an aberration. Contrary to expectations, the latest genetic research indicates that almost everyone’s genome will be similarly unrevealing.
We must assume that, as a geneticist as well as head of NIH, Francis Collins is more aware of this than anyone, but if so, he wrote The Language of Life not out of raw enthusiasm but because the genetics revolution (and not just personalised genomics) is in big trouble. He knows it is going to need all the boosters it can get.
What has changed scientifically in the last three years is the accumulating inability of a new whole-genome scanning technique (called Genome-Wide Association studies; GWAs) to find important genes for disease in human populations1. In study after study, applying GWAs to every common (non-infectious) physical disease and mental disorder, the results have been remarkably consistent: only genes with very minor effects have been uncovered (summarised in Manolio et al 2009; Dermitzakis and Clark 2009). In other words, the genetic variation so confidently expected by medical geneticists to explain common diseases, cannot be found.
There are, nevertheless, certain exceptions to this blanket statement. One group are the single gene, mostly rare, genetic disorders whose discovery predated GWA studies2. These include cystic fibrosis, sickle cell anaemia and Huntington’s disease. A second class of exceptions are a handful of genetic contributors to common diseases and whose discovery also predated GWAs. They are few enough to list individually: a fairly common single gene variant for Alzheimer’s disease, and the two breast cancer genes BRCA 1 and 2 (Miki et al. 1994; Reiman et al. 1996). Lastly, GWA studies themselves have identified five genes each with a significant role in the common degenerative eye disease called age-related macular degeneration (AMD). With these exceptions duly noted, however, we can reiterate that according to the best available data, genetic predispositions (i.e. causes) have a negligible role in heart disease, cancer3, stroke, autoimmune diseases, obesity, autism, Parkinson’s disease, depression, schizophrenia and many other common mental and physical illnesses that are the major killers in Western countries4.
For anyone who has read about ‘genes for’ nearly every disease and the deluge of medical advances predicted to follow these discoveries, the negative results of the GWA studies will likely come as a surprise. They may even appear to contradict everything we know about the role of genes in disease. This disbelief is in fact the prevailing view of medical geneticists. They do not dispute the GWA results themselves but are now assuming that genes predisposing to common diseases must somehow have been missed by the GWA methodology. There is a big problem, however, in that geneticists have been unable to agree on where this ‘dark matter of DNA’ might be hiding.
If, instead of invoking missing genes, we take the GWA studies at face value, then apart from the exceptions noted above, genetic predispositions as significant factors in the prevalence of common diseases are refuted. If true, this would be a discovery of truly enormous significance. Medical progress will have to do without genetics providing “a complete transformation in therapeutic medicine” (Francis Collins, White House Press Release, June 26, 2000). Secondly, as with Francis Collins’ ftest, genetic testing will never predict an individual’s personal risk of common diseases. And most important of all, if the enormous death toll from common Western diseases cannot be attributed to genetic predispositions it must predominantly originate in our wider environment. In other words, diet, lifestyle and chemical exposures, to name a few of the possibilities.
The question, therefore, of whether medical geneticists are acting reasonably in proposing some hitherto unexpected genetic hiding place, or are simply grasping at straws, is a hugely significant one. And there is more than one problem with the medical geneticists’ position. Firstly, as lack of agreement implies, they have been unable to hypothesise a genetic hiding place that is both plausible and large enough to conceal the necessary human genetic variation for disease. Furthermore, for most common diseases there exists plentiful evidence that environment, and not genes, can satisfactorily explain their existence. Finally, the oddity of denying the significance of results they have spent many billions of dollars generating can be explained by realising that a shortage of genes for disease means an impending oversupply of medical geneticists.
You will not, however, gather this from the popular or even scientific media, or even the science journals themselves. No-one so far has been prepared to point out the weaknesses in the medical geneticist’s position. The closest up to now is from science journalist Nicholas Wade in the New York Times who has suggested that genetic researchers have “gone back to square one.” Even this is a massive understatement, however. Human genetic research is not merely at an impasse, it would seem to have excluded inherited DNA, its central subject, as a major explanation of most diseases.
The failure to find major ‘disease genes’
Advances in medical genetics have historically centered on the search for genetic variants conferring susceptibility to rare diseases. Such genes are most easily detected when their effects are very strong (in genetics this is called highly penetrant), or a gene variant is present in unusually inbred human populations such as Icelanders or Ashkenazi Jews. This strategy, based on traditional genetics, has uncovered genes for cystic fibrosis, Huntington’s disease, the breast cancer susceptibility genes BRCA 1 and 2, and many others. Important though these discoveries have been, these defective genetic variants are relatively rare, meaning they do not account for disease in most people2. To find the genes expected to perform analogous roles in more common diseases, different genetic tools were needed, ones that were more statistical in nature.
The technique of genome wide association (GWA) was not merely the latest hot thing in genetics. It was in many ways the logical extension of the human genome sequencing project. The original project sequenced just one genome but, genetically speaking, we are all different. These differences are, for many geneticists, the real interest of human DNA. Many thousands of minor genetic differences between individuals have now been catalogued and medical geneticists wanted to use this seemingly random variation to tag disease genes. Using these minor DNA differences to screen large human populations, GWA studies were going to identify the precise location of the gene variants associated with susceptibility to common disorders and diseases.
To date, more than 700 separate GWA studies have been completed, covering about 80 different diseases. Every common disease, including dozens of cancers, heart disease, stroke, diabetes, mental illnesses, autism, and others, has had one or more GWA study associated with it (Hindorff et al. 2009). At a combined cost of billions of dollars, it was expected at last to reveal the genes behind human illness. And, once identified, these gene variants would become the launchpad for the personalised genomic revolution.
But it didn’t work out that way. Only for one disease, AMD, have geneticists found any of the major-effect genes they expected and, of the remaining diseases, only for type 2 diabetes does the genetic contribution of the genes with minor effects come anywhere close to being of any public health significance (Dermitzakis and Clark 2009; Manolio et al. 2009). In the case of AMD, the five genes determine approximately half the predicted genetic risk (Maller et al. 2006). Apart from these, GWA studies have found little genetic variation for disease. The few conclusive examples in which genes have a significant predisposing influence on a common disease remain the gene variant associated with Alzheimer’s disease and the breast cancer genes BRCA1 and 2, all of which were discovered well before the GWA era (Miki et al. 1994 and Reiman et al. 1996).
Though they have not found what their designers hoped they would, the results of the GWA studies of common diseases do support two distinct conclusions, both with far-reaching implications. First, apart from the exceptions noted, the genetic contribution to major diseases is small, accounting at most for around 5 or 10% of all disease cases (Manolio et al. 2009). Secondly, and equally important, this genetic contribution is distributed among large numbers of genes, each with only a minute effect (Hindorff et al. 2009). For example, the human population contains at least 40 distinct genes associated with type I diabetes (Barrett et al. 2009). Prostate cancer is associated with 27 genes (Ioannidis et al. 2010); and Crohn’s disease with 32 (Barrett et al. 2008).
The implications for understanding how each person’s health is affected by their genetic inheritance are remarkable. For each disease, even if a person was born with every known ‘bad’ (or ‘good’) genetic variant, which is statistically highly unlikely, their probability of contracting the disease would still only be minimally altered from the average.
DNA is not the language of life or death
This dearth of disease-causing genes is without question a scientific discovery of tremendous significance. It is comparable in stature to the discovery of vaccination, of antibiotics, or of the nature of infectious diseases, because it tells us that most disease, most of the time, is essentially environmental in origin.
But such significance leaves a puzzle. Huge quantities of newspaper space has been devoted to genes, or even to hints of genes for various diseases5. By rights then, reports of the GWA results should have filled the front pages of every world newspaper for a week. So, why has this coverage not occurred?
It is possible to conceive of excuses for lack of coverage: refutation is inherently less interesting, and the GWA results have been reported piecemeal, but the more likely reason is the disturbing implications for medical geneticists who are its discoverers. The GWA studies were not envisaged as a test of the hypothesis: do genes cause common diseases? Rather, they were expected merely to straightaway identify the guilty genes that everyone “knew” were there. By apparently refuting the entire concept of genes for common diseases, the GWA studies raise fundamental questions about money spent, hopes raised, and judgments made by medical researchers.
In the first place, the GWA results raise what are probably insurmountable questions for the prospective ‘genetic revolution’ in healthcare. What use will personalised DNA testing (or sequencing) be if genes cannot predict disease for the vast majority of people? Are genes with only extremely minor effects going to be of value as drug targets? How hard is it going to be to untangle their roles in disease when they have hardly any measurable effect? Should we still suppose that pouring more resources into human genetic research is going to rescue industry’s faltering drug development pipelines? All of a sudden, the future of medicine, especially in the specialities dealing with degenerative diseases and mental illness, looks very different and a lot less promising. We no longer have a ‘complete transformation’ to look forward to, only a continuation of the incremental improvements and setbacks that have characterised medicine for the last fifty years.
Shoring up the good ship medical genetics
In a rare public sign of the struggle to come to terms with this genetically impoverished world-view, the authors of a brief review in Science magazine, Andrew Clark of Cornell University and Emmanouil Dermitzakis of the University of Geneva Medical School, Switzerland have been alone in stating the case even partly straightforwardly. According to them, the GWA studies tell us that “the magnitude of genetic effects is uniformly very small” and therefore “common variants provide little help in predicting risk” (Dermitzakis and Clark 2009). Consequently, the likelihood that personalised genomics will ever predict the occurrence of common diseases is “bleak”. This aim, they believe, will have to be abandoned altogether.
The first conclusion to be drawn from these quotes is that such directness implies that if the GWA findings are not finding their way to the front page the reason is not ambiguity in the results themselves. From a scientific perspective the GWA results, though negative, are robust and clear.
Most human geneticists view the GWA results somewhat differently, however. An invited workshop, convened by Collins and others, discussed the then-accumulating results in February 2009. The most visible outcome of this workshop was a lengthy review published in Nature and titled: “Finding the Missing Heritability of Complex Diseases.” (Manolio et al. 2009).
For a review paper that does not lay out any new concepts or directions, 27 senior scientists as coauthors might be considered overkill. “Finding the Missing Heritability”, however, should be understood not so much as a scientific contribution but as an effort to conceal the gaping hole in the science of medical genetics.
In their Science article, which was published almost simultaneously, Dermitzakis and Clark paused only briefly to consider whether so many genes could have been overlooked. Apparently, they thought it an unlikely possibility. Manolio et al., however, frame this as the central issue. According to them, since heritability measurements suggest that genes for disease must exist, they must be hiding under some as-yet-unturned genetic rock. They list several possible hiding places: there may be very many genes with exceedingly small effects; genes for disease may be highly represented by rare variants with large effects; disease genes may have complex genetic architectures; or they may exist as gene Copy Number Variants (CNVs). Since Manolio et al. presented their list, the scientific literature has seen further suggestions for where disease genes might be hiding. These include in mitochondrial DNA, epigenetics and in statistical anomalies (e.g. Eichler et al. 2010; Petronis 2010).
A problem for all these hypotheses, however, is that anyone wishing to take them seriously needs to consider one important question. How likely is it that a quantity of genetic variation that could only be called enormous (i.e. more than 90-95% of that for 80 human diseases) is all hiding in what until now had been considered genetically unlikely places? In other words, they all require the science of genetics to be turned on its head. For epigenetics, for example, there is scant evidence that important traits can be inherited through acquired modifications of DNA. Similarly, if rare variants with strong effects keep appearing in the population and causing major illnesses, why is there no evidence for this phenomenon, since it must have been occurring in the past? With unanswered questions such as these, it is unsurprising that none of the mooted explanations has attracted any kind of consensus among geneticists and in fact the CNV explanation is already looking highly unlikely (Conrad et al. 2010; The Wellcome Trust Case Control Consortium 2010). As the first of these two papers summarised “we conclude that, for complex traits, the heritability void left by genome-wide association studies will not be accounted for by CNVs” (Conrad et al. 2010).
Now, it is not impossible that human diseases follow unique genetic rules, but the apparently overlooked possibility is that the GWA studies are indicating a simple truth: that genes are not important causes of major diseases.
As stated so far, the case against the importance of genes for disease seems strong. However, the ‘missing heritability’ argument is based on numerous predictions of a large genetic contribution to human diseases that are derived from heritability measurements. These heritability estimates are obtained from the study of identical and non-identical twins. A crucial question becomes, therefore, are these estimates truly reliable?
How robust is the historical evidence for genetic causation?
A perennial feature of research into human health has always been the mountain of evidence that environment is overwhelmingly important in disease. People who migrate acquire the spectrum of diseases of their adopted country. Populations who take up Western habits, or move to cities with Western lifestyles, acquire Western diseases, and so on (e.g. Campbell and Campbell 2008). These data are hard to refute, not least because they are so simple, but geneticists, when discussing them, invariably wheel out their own version of incontrovertible evidence: twin studies of the heritability of complex diseases. When Francis Collins talks about ‘missing heritability’ it is to studies such as these that he is referring. They provide the basic evidence for genetic influences on human disease.
A classic example of this contradiction is myopia. A large body of evidence suggests that myopia is an environment-induced disorder caused by some combination of night lighting, close reading, lack of distance viewing and diet (e.g. Quinn et al. 1999). Moreover, under the influence of Westernisation, genetically unchanged populations, for example, are known to have switched in a single generation from close to 0% to a prevalence of myopia of over 80% (Morgan 2003). And myopia is only one of many examples of diseases with very strong evidence for its environmental origin. In 2009, for example, researchers demonstrated that very moderate improvements in lifestyle could reduce an individual’s probability of contracting type 2 diabetes by 89% (Mozzafarian et al. 2009). The subjects of this study just had to smoke less than the average, keep trim, exercise moderately and not eat too much fat.
In stark contrast, twin studies (which compare the extent of similarity exhibited by identical and non-identical twins) estimate that myopia is a disease with a heritability (called h2) of about 0.8 (out of a possible 1.0), indicating that for myopia genetic causes dominate environmental ones. These findings are clearly incompatible with the available epidemiological data on myopia and no satisfactory resolution to them has ever been proposed (e.g. Rose et al. 2002; Morgan 2003). This contradiction, between the results of twin studies and the results of epidemiological and clinical research, is repeated for almost every human disease.
A meaningful resolution to these contradictions is, nevertheless, necessary. Since it is unlikely that the many observations identifying environment as a dominant disease-causing factor are all incorrect, the parsimonious solution to the conundrum, even before the GWA studies were reported, was to propose that heritability studies of twins are inherently mistaken or misinterpreted.
Studies of human twins estimate heritability (h2) by calculating disease incidence in monozygotic (genetically identical) twins versus dizygotic (fraternal) twins (who share 50% of their DNA). If monozygotic twin pairs share disorders more frequently than do dizygotic twins, it is presumed that a genetic factor must be involved. A problem arises, however, when the number resulting from this calculation is considered to be an estimate of the relative contribution of genes and environment over the whole population (and environment) from which the twins were selected. This is because the measurements are done in a series of pairwise comparisons, meaning that only the variation within each twin pair is actually being measured. Consequently, the method implicitly defines as environment only the difference within each twin pair. Since each twin pair normally shares location, parenting styles, food, schooling, etc., much of the environmental variability that exists between individuals in the wider population is de facto excluded from the analysis. In other words, heritability (h2), when calculated this way, fails to adequately incorporate environmental variation and inflates the relative importance of genes.
Heritability studies of humans are classic experiments that have been conducted many times and they have strong defenders among modern geneticists (e.g. Visscher et al. 2008). Nevertheless, criticisms such as those above are not novel. They are a specific example of the general problem, formulated by Richard Lewontin (of Harvard University), that the contributions of genes to a trait normally depend on the particular environment. And further, that susceptibility to environment depends on genes. In consequence, there can be no universal constant (such as h2) that defines their relationship to one another (Lewontin, Rose and Kamin 1984; Lewontin 1993). Lewontin is not alone among geneticists in his dismissal of heritability as it is used in human genetics. Martin Bobrow of Cambridge University, for example, has called human heritability “a poisonous concept” and “almost uninterpretable”6.
If one accepts either that h2 is consistently inflated, or that it is essentially meaningless, even “poisonous”, then the only current evidence supporting genetic susceptibility as a major cause of disease disappears. “The Missing Heritability of Complex Diseases”, DNAs’ so-called ‘dark matter’, becomes simply an artefact arising from overinterpretation of twin studies.
A mutually convenient untruth
Genetic determinist ideas, especially in the form of explanations for health and disease, are powerful forces in our society (Lewontin 1993). Their pervasive influence, however, requires some explanation because the purely scientific evidence for genetic causation has always been weak, since it depended heavily on disputed heritability studies. To understand the significance a repudiation of inherited DNA as a disease explanation has, it is first necessary to understand the role genetic determinism plays in consolidating the social order.
Politicians like genetic determinism as a theory of disease because it substantially reduces their responsibility for people’s ill-health. By shifting blame towards individuals and their genetic ‘predispositions’ it greatly dilutes the pressure they may feel to regulate, ban, or tax harmful products and contaminants, courses of action that typically offend their business constituents. For a politician, therefore, spending tax dollars on medical genetics is an easy and even popular decision.
Corporations like genetic determinism, again because it shifts blame. The Salt Institute website, for example, currently maintains that diseases linked to salt reflect the existence of a small number of highly predisposed individuals. This assertion, sandwiched (on the website) between other questions about salt and health, is clearly intended to undermine efforts to restrict salt in the diet. For the same reason, the tobacco industry has for many years encouraged research into the genetics of nicotine addiction (Gundle et al. 2010). This same reasoning, that disease is the fault of the victim’s genes, also protects corporate defendants from after-the-fact liability. If lung cancer patients, for example, suffer from even the possibility of a genetic predisposition, suing tobacco companies is very much harder than it would be otherwise (Tokuhata and Lilienfeld, 1963). There is evidence, too, that genetic determinism influences decisions well before the full facts are known. At least sometimes, it can even encourage the vendor knowingly to place on the market products with harmful effects (Gundle et al. 2010).
Medical researchers are also partial to genetic determinism. They have noticed that whenever they focus on genetic causation, they can raise research dollars with relative ease. The last fifteen years, coinciding with the rise of medical genetics, have seen unprecedented sums of money directed at medical research. At the same time, research on pollution, nutrition and epidemiology has not benefited in any comparable way. It is hard not to conclude that this funding disparity is strongly influenced by the fit of genetics to the needs of businesses and politicians. In the words of Homer Simpson, “It takes two to lie, Marge. One to lie and one to listen”.
Recognising their value, these groups have tended to elevate genetic explanations for disease to the status of unquestioned scientific facts, thus making their dominance of official discussions of health and disease seem natural and logical. This same mindset is accurately reflected in the media where even strong environmental links to disease often receive little attention, while speculative genetic associations can be front page news. It is astonishing to think that all this has occurred in spite of the reality that genes for common diseases were essentially hypothetical entities.
Mutually convenient or not, by the criteria normally applied in science, the hypothesis that genes are significant causes of common diseases stands refuted. The history of scientific refutation, however, is that adherents of established theories construct ever more elaborate or unlikely explanations to fend off their critics (Ziman 2000). The invocation of genetic ‘dark matter’ and the search for ‘hiding’ genetic variation shows that the process of special pleading is already well underway (e.g. Manolio et al. 2009; Eichler et al. 2010). Implausible though the suggested hiding places seem, it is nevertheless going to be difficult to rule them all out in the near future. Consequently, those geneticists wishing to do so will have the opportunity to obfuscate for some while yet.
Needed: A declaration of dependence
In societies, including our own, much of the social fabric is arranged around our conception of the ‘proper’ place of death and disease. Confidence in the genetic paradigm has led us to explain non-infectious disease as primarily a natural manifestation of genetic predispositions and thus a normal outcome of aging. This normalisation of diseases has obscured the contrary evidence that these same diseases can be all but absent in other cultures and often were rare in historical times. With the GWA results confirming the epidemiological studies, however, we are confronted with the necessity of constructing a new narrative. To be consistent with the facts, this new narrative must incorporate Western diseases not as unavoidable, but as indicators of human fragility in the face of industrialisation and modern life.
That we are so vulnerable to our social and physical surroundings, is an uncompromising message. But to the very best of our scientific knowledge it is the truth. Fortunately, it is a truth that offers hope. If we can change our environment for the worse, we can also change it for the better. And if a magic medical cure-all pill is not going to materialise after all, it may be that it wasn’t needed in the first place.
Change for better health can occur in part through individual effort. The new understanding implies that we are not fated to develop any of the common diseases and that the efforts we make to eat well and live a healthy life will be amply rewarded. We should not be surprised if specific lifestyle changes can reverse decades of disease progression (Esselstyn et al. 1995). Or that Seventh Day Adventists, who are non-smoking, non-drinking vegetarians, live on average to 88, eight years beyond the average American’s life expectancy (Fraser and Shavlik 2001). These examples suggest what can be achieved with relatively modest lifestyle changes. By focusing more exclusively on health-related lifestyle modifications than even Seventh Day Adventists do, we could probably extend our life expectancy still further. Exactly how much further is now a much more interesting question than we previously thought.
For most people, life expectancy is only truly of value if it is accompanied by life quality. We should expect, however, any future diminution of the burden of degenerative diseases from lifestyle modification to both extend life expectancy and enhance life quality7. If so, it might make the most common end-of-life experience very different from the actual prospect facing most Westerners for whom old age is commonly a process of ever more aggressive medical intervention culminating in a hospital room attached to drips and electrodes.
While individual effort has a place, many positive lifestyle and social changes require the cooperation of the state. Nevertheless, most governments cooperate far more, for example, with their food industries than with those who wish to eat a healthy diet. The laying to rest of genetic determinism for disease, however, provides an opportunity to shift this cynical political calculus. It raises the stakes by confronting policy-makers as never before with the fact that they have every opportunity, through promoting food labeling, taxing junk food, or funding unbiased research, to help their electorates make enormously positive lifestyle choices. And, when their constituents realise that current policies are robbing every one of them of perhaps whole decades of healthy living, these citizens might start to apply the necessary political pressure.
Addendum: Following publication of ‘The Great DNA Data Deficit’ various readers have contacted us with relevant publications and books of which we were unaware. These important contributions to the issue of whether genes might cause disease extend or otherwise support the discussion considerably. They are listed below in chronological order. Our sincere thanks to readers for sending these in:
(1) ‘Genes for’ disease is shorthand for genetic variants predisposing the carrier to disease.
(2) The definition of a genetically rare disease is usually that it affects fewer than 1 in 1,000 people. Approximately 6,000 rare diseases have been identified in humans.
(3) The famous alleles BRCA 1 and 2 are important in some families and populations but otherwise are fairly rare.
(4) According to the World Health Organisation, heart disease causes 17.1% of all deaths worldwide. Cancer causes 15% of all deaths. Stroke causes 10% of all deaths. WHO factsheet.
(5) The explanation of the contradiction between the GWA studies and the newspaper reports is that much of this coverage was hype. Almost without exception these newspaper reports covered discoveries whose significance could be questioned. Typically, they concerned unsubstantiated results, or the genes were for very minor diseases or the medical and genetic implications of the discovery were substantially overplayed.
(6) Why geneticists disagree about heritability has a historical context that usefully illuminates this issue. Once upon a time the term heritability was used differently. When Sewall Wright, one of the founders of genetics, developed the concept of heritability, he titled a key paper “The Relative Importance of Heredity and Environment in Determining the Piebald Pattern of Guinea Pigs” (Wright, 1920). He used this title even though all animals in the study were kept in identical conditions. Clearly, therefore, he wasn’t defining ‘environment’ as we now do. Instead, in that paper he explicitly defined environment as “the irregularities of development due to the intangible sorts of causes to which the word chance is applied”. All of the subsequent questions (like Lewontin’s) surrounding the validity of twin studies have arisen precisely because Wright’s method, which defined heritability in opposition to chance variations in development, was extended to populations of humans living in variable and varying environments.
(7) In popular speech, aging and degeneration are often conflated, leading sometimes to a rejection of health advice as simply life-span extension. However, aging, by definition, is simply the passage of time and research shows that typically, extended life expectancy is correlated with improved health, when age is taken into account (Fraser and Shavlik 2001). In case one is tempted to confuse aging and disease, it may be helpful to think of children. For them, aging is a process of becoming stronger.
Barrett J. et al. (2008) Genome-wide association defines more than 30 different susceptibility loci for Crohn’s disease. Nature Genet. 40: 955-962. Barrett J. et al. (2009) Genome-wide association study and meta-analysis find that over 40 different loci affect risk of type 1 diabetes. Nature Genet. 41: 703-707. Campbell T.C. and Campbell T.M. (2008) The China Study. Benbella books, Inc. USA Collins F.S. (2009) The Language of Life. Harper, NY, USA. Conrad D.F. et al. (2010) Origins and functional impact of copy number variation in the human genome. Nature 464: 704-712. Dermitzakis E.T. and Clark A.G. (2009) Life after GWA studies.Science 326: 239-240. Gundle K. et al. (2010) ‘To prove this is industry’s best hope’: big tobacco’s support of research on the genetics of nicotine addiction. Addiction 105: 974-983. Eichler, E. et al. (2010) Missing heritability and strategies for finding the underlying causes of complex disease. Nature Genetics 11: 446. Essselstyn C.S., Ellis S.G. Medendorp S.V., Crowe T.D. (1995) A strategy to arrest and reverse coronary artery disease: a 5-year longitudinal study of a single physician’s practice. J. Family Practice 41: 560-568. Fraser G.S. and Shavlik D.J. (2001) Ten Years of Life: Is It a Matter of Choice? Arch. Int. Medicine 161: 1645-52. Hindorff L. et al. (2009) Potential etiologic and functional implications of genome-wide association loci for human diseases and traits. Proc. Natl. Acad. Sci 106: 9362-67. Ioannidis, JPA, Castaldi P. and Evangelou E. (2010) A Compendium of Genome-Wide Associations for Cancer: Critical Synopsis and Reappraisal. J. of the National Cancer Institute 102: 846-858. Lewontin R.C., Rose S. and Kamin L.J. (1984) Not in our genes. Pantheon books New York, USA. Lewontin R.C. (1993) Biology as ideology Penguin books New York, USA. Maller J. et al. (2006) Common variation in three genes, including a non-coding variant in CFH, strongly influences risk of age-related macular degeneration. Nature Genet. 38: 1055-59. Manolio T. et al. (2009) Finding the missing heritability of complex diseases. Nature 461: 747-753. Miki Y. et al. (1994) A strong candidate for the breast and ovarian cancer susceptibility gene BRCA1. Science: 266 66-71. Mozaffarian D., Kamineni A., Carnethon M., Djoussé L., Mukamal K.J., and Siscovick D. (2009) Lifestyle Risk Factors and New-Onset Diabetes Mellitus in Older Adults. Arch. Int. Med. 169: 798–807. Morgan I. (2003) The biological basis of myopic refractive error. Clinical and Experimental Optometry 86: 276-288. Petronis A. (2010) Epigenetics as a unifying principle in the aetiology of complex traits and diseases. Nature 465: 721-727. Quinn G.E. et al. (1999) Myopia and ambient lighting at night. Nature 399: 113-14. Reiman E. et al. (1996) Preclinical Evidence of Alzheimer’s Disease in Persons Homozygous for the 4 Allele for Apolipoprotein E. New England Journal of Medicine 334: 752-758. Rose K.A. Morgan I.G. Smith W. Mitchell P. (2002) High heritability of myopia does not preclude rapid changes in prevalence. Clinical and Experimental Opthalmology 30: 168-172. Tokuhata G.K. and Lilienfeld A.M. (1963) Familial aggregation of lung cancer in humans. J. of the National Cancer Institute 30: 289-312. Visscher P.M. et al. (2008) Heritability in the genomics era-concepts and misconceptions. Nature Rev. Genetics 9:255-266. The Wellcome Trust Consortium (2010) Genome-wide association study of CNV in 16,000 cases of eight common diseases and 3,000 shared controls. Nature 464: 713-720. Wright S. (1920) The Relative Importance of Heredity and Environment in Determining the Piebald Pattern of Guinea Pigs. Proc. Natl. Acad. Sci 6:320-332. Ziman J. (2000) Real Science: What It Is and What It Means. Cambridge University Press. UK
Do a search for the term “carcinogenesis” (literally “cancer creation”) on the product insert for any vaccine. You’ll see the following common phrase regarding most vaccines, “This vaccine has not been evaluated for itscarcinogenicor mutagenic potentials or impairment of fertility.”
Huh? The inserts say that vaccines are not tested to see if they are carcinogenic. Seriously?
How can we inject these dangerous, toxic concoctions into our infants and children without knowing whether or not they cause cancer?
So, since specific vaccines are not tested to see if they cause cancer, let’s take a look at some common vaccine ingredients and look at the medical literature relating to these specific ingredients.Following are a few cancer-causing ingredients, neurotoxins, and immune-destroying ingredients that are in most vaccines:
Formaldehyde (embalming fluid) – This is classified as a known carcinogen. Both the EPA and the International Agency for Research on Cancer admit this. So, why do vaccines still contain formaldehyde? Formaldehyde has been linked to several types of cancers, including Leukemia. It has also been used in the Anthrax vaccine, DT, DTaP, Hib, HPV, Hep A, Hep B, Influenza, Meningitis, Polio, and more.
Aluminum – Aluminum salts are in childhood vaccines such as the DTaP vaccine, the pneumococcal conjugate vaccine, hepatitis B vaccine, and more. Aluminum hydroxide (just one of the aluminum salts that is used) has been linked to autism, Alzheimer’s disease, and yes, cancer. Aluminum causes cancer by displacing iron from its protective proteins, raising the level of free iron in the body, and triggering intense inflammation, free radical generation, and lipid peroxidation.
Mercury (from Thimerosal) – This preservative is in almost every vaccine and has been linked to autism, mental retardation, and yes, cancer. The search for the association between mercury and cancer reveals 643 scientific papers. Wow!
The Physicians’ Desk Reference (PDR) states that in addition to mercury, aluminum, and formaldehyde, vaccines may also contain SV40, bovine serums, latex, neomycin and other known carcinogens and allergens. Many of these toxins are injected into the bloodstream of babies as young as one day old. In fact, no long term testing for efficacy or safety for any vaccination has ever been done. Essentially, Big Pharma is using our babies as their own guinea pigs.
What an arch-criminal cabal, truly criminal beyond comprehension, it is: this vaccine-injection, terminally corrupt medical profession. Those responsible for this are rather than responsible for prevention of disease are, in fact, disease generators. The vast number of diseases induced through vaccination is virtually beyond count. No doubt, modern vaccines, mere inoculations of animal-based matter, along with a litany of noxious chemicals, are among the most potent carcinogens known. Moreover, no one can deny that they are carcinogens. The fact is the proof is absolute.
All modern vaccines have carcinogenic actions. In fact, these inoculations are largely responsible for the cancer epidemic. It’s big business. Unfathomable amounts of money are made by the pharmaceutical houses and their compliant physician population as a result of this disease, the incidence of which would be greatly reduced in the event that vaccination was curtailed.
A century after they were discovered killing bacteria in the feces of World War I soldiers, the viruses known as bacteriophages, or simply phages, are drawing new attention for the role they might play within the human body. Phages have been found most everywhere, from oceans to soils. Now, a study suggests that people absorb up to 30 billion phages every day through their intestines.
Though where the viruses end up is unclear, those data and other recent studies have scientists wondering whether a sea of phages within the body—a “phageome”—might influence our physiology, perhaps by regulating our immune systems. “Basic biology teaching says that phages don’t interact with eukaryotic cells,” says phage researcher Jeremy Barr of Monash University in Melbourne, Australia, who led the study published this week in mBio. He’s now convinced “that’s complete BS.”
For decades, most medical research on phages focused on turning these bacterial parasites into antibiotics. There have been some compelling success stories, but phage therapy has struggled to become a dependable treatment.
Yet Barr’s earlier research showed that phages might naturally help protect us from pathogens. Studying animals ranging from corals to humans, he found that phages are more than four times as abundant in mucus layers, like the ones that protect our gums and gut, as they are in the adjacent environment. The protein shell of a phage, it turned out, can bind mucins, large secreted molecules that together with water make up mucus.
This works out well for both phages and mucusmaking animals. Sticking to mucus enables the phages to encounter more of their bacterial prey. And as a result, Barr showed in a series of in vitro studies, the viruses protect the underlying cells from potential bacteria pathogens, providing an additional layer of immunity.
Now, he has found evidence that these viruses can make their way from the gut’s mucus into the body. In a lab dish, his team showed that human epithelial cells such as those that line our guts, lungs, and the capillaries surrounding the brain take up phages and transport them across their interior. The transport mechanism remains unknown, but the researchers spotted the viruses enclosed in vesicles within the cells.
What’s more, the cells consistently took up phages on the side that in the body faces outward, for example toward the gut lumen, and released them on the opposite, inward-facing side. From the rate at which the epithelial cells took up phages in the lab, the researchers estimated that a person might absorb up to 30 billion in a day.
The new study “nicely” shows how phages might get into the body, says molecular biologist Krystyna Dąbrowska of the Polish Academy of Sciences’s Institute of Immunology and Experimental Therapy in Wrocław. But she cautions that a lab dish is different from the gut of a living animal, and some of the cells used for Barr’s assays are cancer cells, which might have different rates of phage uptake compared with normal cells.
If phages do get into our tissues, what—if anything—do they do there? Only a few studies address the issue. In 2004, researchers led by Dąbrowska reported that a specific type of phage can bind the membrane of cancer cells, reducing tumor growth and spread in mice. A few years later, Dąbrowska's graduate adviser, phage expert Andrzej Gorski, showed that phages can affect the mouse immune system when injected, ramping down T-cell proliferation and antibody production. In mice, they can even prevent the immune system from attacking transplanted tissues.
Barr thinks that in humans, a steady influx of the viruses creates an “intrabody phageome,” which may modulate immune responses. Work published this year by a team of Belgium-based researchers may back up this idea: When white blood cells taken from healthy people were exposed to five different phage species, the cells produced mainly immune molecules known to reduce flulike symptoms and inflammation. And in another hint of an immune link, a group led by immunologist Herbert Virgin at Washington University School of Medicine in St. Louis in Missouri found that people with two autoimmune conditions, type 1 diabetes and inflammatory bowel disease, have altered gut phageomes.
Virgin cautions that his findings are only associations. But Barr goes on to speculate that the phageome might also alert the immune system to the presence of potential pathogens. A bacterial infection would bring a wave of new phages into the body—the parasites of the invading bacteria—which might somehow touch off an inflammatory response that could target the bacteria.
Once scientists understand the role of the human phageome, they could start thinking about using phages to manipulate the bacterial communities within our body and maybe even control our own cells, Barr says. But he, too, is cautious, noting that “phage biology is an inch wide and a mile deep.” Given our current ignorance of phageeukaryote interactions, says Barr, medical uses “are probably decades away.”