点击显示 收起
1 From the Department of Health and Exercise Science, Colorado State University, Fort Collins (LC); the Departments of Radiology and Anthropology, Emory University, Atlanta (SBE); the Department of Medicine and UCSF/Moffitt General Clinical Research Center, University of California, San Francisco (AS); the Department of Food Science, RMIT University, Melbourne, Australia (NM); the Department of Medicine, Lund University, Sweden (SL); the Department of Food Science, Lipid Chemistry and Molecular Biology Laboratory, Purdue University, West Lafayette, IN (BAW); the Mid America Heart Institute, Cardiovascular Consultants, Kansas City, MO (JHO); and the Human Nutrition Unit, Department of Biochemistry, University of Sydney, Australia (JB-M)
2 Address reprint requests to L Cordain, Department of Health and Exercise Science, Colorado State University, Fort Collins, CO 80523. E-mail: cordain{at}cahs.colostate.edu
ABSTRACT
There is growing awareness that the profound changes in the environment (eg, in diet and other lifestyle conditions) that began with the introduction of agriculture and animal husbandry 10000 y ago occurred too recently on an evolutionary time scale for the human genome to adjust. In conjunction with this discordance between our ancient, genetically determined biology and the nutritional, cultural, and activity patterns of contemporary Western populations, many of the so-called diseases of civilization have emerged. In particular, food staples and food-processing procedures introduced during the Neolithic and Industrial Periods have fundamentally altered 7 crucial nutritional characteristics of ancestral hominin diets: 1) glycemic load, 2) fatty acid composition, 3) macronutrient composition, 4) micronutrient density, 5) acid-base balance, 6) sodium-potassium ratio, and 7) fiber content. The evolutionary collision of our ancient genome with the nutritional qualities of recently introduced foods may underlie many of the chronic diseases of Western civilization.
Key Words: Westernized diets chronic disease processed foods genetic discordance hunter-gatherers human evolution
EVOLUTIONARY DISCORDANCE
Evolution acting through natural selection represents an ongoing interaction between a species genome and its environment over the course of multiple generations. Genetic traits may be positively or negatively selected relative to their concordance or discordance with environmental selective pressures (1). When the environment remains relatively constant, stabilizing selection tends to maintain genetic traits that represent the optimal average for a population (2). When environmental conditions permanently change, evolutionary discordance arises between a species genome and its environment, and stabilizing selection is replaced by directional selection, moving the average population genome to a new set point (1, 2). Initially, when permanent environmental changes occur in a population, individuals bearing the previous average status quo genome experience evolutionary discordance (2, 3). In the affected genotype, this evolutionary discordance manifests itself phenotypically as disease, increased morbidity and mortality, and reduced reproductive success (1-3).
Similar to all species, contemporary humans are genetically adapted to the environment of their ancestorsthat is, to the environment that their ancestors survived in and that consequently conditioned their genetic makeup (1-3). There is growing awareness that the profound environmental changes (eg, in diet and other lifestyle conditions) that began with the introduction of agriculture and animal husbandry 10000 y ago occurred too recently on an evolutionary time scale for the human genome to adapt (2-5). In conjunction with this discordance between our ancient, genetically determined biology and the nutritional, cultural, and activity patterns in contemporary Western populations, many of the so-called diseases of civilization have emerged (2-12).
CHRONIC DISEASE INCIDENCE
In the United States, chronic illnesses and health problems either wholly or partially attributable to diet represent by far the most serious threat to public health. Sixty-five percent of adults aged 20 y in the United States are either overweight or obese (13), and the estimated number of deaths ascribable to obesity is 280184 per year (14). More than 64 million Americans have one or more types of cardiovascular disease (CVD), which represents the leading cause of mortality (38.5% of all deaths) in the United States (15). Fifty million Americans are hypertensive; 11 million have type 2 diabetes, and 37 million adults maintain high-risk total cholesterol concentrations (>240 mg/dL) (15). In postmenopausal women aged 50 y, 7.2% have osteoporosis and 39.6% have osteopenia (16). Osteoporotic hip fractures are associated with a 20% excess mortality in the year after fracture (17). Cancer is the second leading cause of death (25% of all deaths) in the United States, and an estimated one-third of all cancer deaths are due to nutritional factors, including obesity (18).
HOMININ DIETARY CHARACTERISTICS
In the 57 million-year period since the evolutionary emergence of hominins (bipedal primates within the taxonomic tribe hominini; note that the newer term hominin supplants the previous term, hominid) 20 species may have existed (Figure 1) (19). Similar to historically studied hunter-gatherers (20, 21), there would have been no single universal diet consumed by all extinct hominin species. Rather, diets would have varied by geographic locale, climate, and specific ecologic niche. However, there are universal characteristics of preagricultural hominin diets that are useful in understanding how the current Western diet may predispose modern populations to chronic disease. Increasingly, clinical trials and interventions that use dietary treatments with nutritional characteristics similar to those found in preindustrial and preagricultural diets have confirmed the beneficial health consequences predicted by the template of evolutionary discordance theory.
FIGURE 1.. The hominin fossil record. Species are indicated with the dates of the earliest and latest fossil record. Adapted from Wood (19).
NUTRITIONAL CHARACTERISTICS OF PRE- AND POSTAGRICULTURAL DIETS
Before the development of agriculture and animal husbandry hominin dietary choices would have been necessarily limited to minimally processed, wild plant and animal foods. With the initial domestication of plants and animals, the original nutrient characteristics of these formerly wild foods changed, subtly at first but more rapidly with advancing technology after the Industrial Revolution. Furthermore, with the advent of agriculture, novel foods were introduced as staples for which the hominin genome had little evolutionary experience. More importantly, food-processing procedures were developed, particularly following the Industrial Revolution, which allowed for quantitative and qualitative food and nutrient combinations that had not previously been encountered over the course of hominin evolution.
In contrasting pre- and postagricultural diets, it is important to consider not only the nutrient qualities and types of foods that likely would have been consumed by preagricultural hominins but to also recognize the types of foods and their nutrient qualities that could not have been regularly consumed before the development of agriculture, industrialization, and advanced technology. Food types that would have generally been unavailable to preagricultural hominins are listed in Table 1 (22-24). Although dairy products, cereals, refined sugars, refined vegetable oils, and alcohol make up 72.1% of the total daily energy consumed by all people in the United States, these types of foods would have contributed little or none of the energy in the typical preagricultural hominin diet (20). Additionally, mixtures of foods listed in Table 1 make up the ubiquitous processed foods (eg, cookies, cake, bakery foods, breakfast cereals, bagels, rolls, muffins, crackers, chips, snack foods, pizza, soft drinks, candy, ice cream, condiments, and salad dressings) that dominate the typical US diet.
View this table:
TABLE 1. Food and food types found in Western diets generally unavailable to preagricultural hominins1
Dairy foods
Hominins, like all mammals, would have consumed the milk of their own species during the suckling period. However, after weaning, the consumption of milk and milk products of other mammals would have been nearly impossible before the domestication of livestock because of the inherent difficulties in capturing and milking wild mammals. Although sheep were domesticated by 11000 before present (BP) (25) and goats and cows by 10000 BP (26, 27), early direct chemical evidence for dairying dates to 6100 to 5500 BP from residues of dairy fats found on pottery in Britain (28). Taken together, these data indicate that dairy foods, on an evolutionary time scale (Figure 1), are relative newcomers to the hominin diet.
Cereals
Because wild cereal grains are usually small, difficult to harvest, and minimally digestible without processing (grinding) and cooking, the appearance of stone processing tools in the fossil record represents a reliable indication of when and where cultures systematically began to include cereal grains in their diet (7). Ground stone mortars, bowls, and cup holes first appeared in the Upper Paleolithic (from 40000 y ago to 12000 y ago) (29), whereas the regular exploitation of cereal grains by any worldwide hunter-gatherer group arose with the emergence of the Natufian culture in the Levant 13000 BP (30). Domestication of emmer and einkorn wheat by the descendants of the Natufians heralded the beginnings of early agriculture and occurred by 1011000 BP from strains of wild wheat localized to southeastern Turkey (31). During the ensuing Holocene (10000 y ago until the present), cereal grains were rarely consumed as year round staples by most worldwide hunter-gatherers (32, 33), except by certain groups living in arid and marginal environments (32, 34). Hence, as was the case with dairy foods, before the Epi-Paleolithic (1000011000 y ago) and Neolithic (10000 to 5500 y ago) periods, there was little or no previous evolutionary experience for cereal grain consumption throughout hominin evolution.
In Table 1, it is shown that 85.3% of the cereals consumed in the current US diet are highly processed refined grains. Preceding the Industrial Revolution, all cereals were ground with the use of stone milling tools, and unless the flour was sieved, it contained the entire contents of the cereal grain, including the germ, bran, and endosperm (35). With the invention of mechanized steel roller mills and automated sifting devices in the latter part of the 19th century (35), the nutritional characteristics of milled grain changed significantly because the germ and bran were removed in the milling process, leaving flour comprised mainly of endosperm of uniformly small particulate size (35, 36). Accordingly, the widespread consumption of highly refined grain flours of uniformly small particulate size represents a recent secular phenomenon dating to the past 150200 y (35).
Refined sugars
The per capita consumption of all refined sugars in the United States in 2000 was 69.1 kg, whereas in 1970 it was 55.5 kg (24). This secular trend for increased sugar consumption in the United States in the past 30 y reflects a much larger worldwide trend that has occurred in Western nations since the beginning of the Industrial Revolution some 200 y ago (37). The per capita refined sucrose consumption in England steadily rose from 6.8 kg in 1815 to 54.5 kg in 1970 (38), as shown in Figure 2. Similar trends in refined sucrose consumption have been reported during the Industrial Era for the Netherlands, Sweden, Norway, Denmark, and the United States (39).
FIGURE 2.. Per capita consumption of sucrose in England from 1815 to 1970. WWI, World War I; WWII, World War II. Adapted from Cleave (38).
The first evidence of crystalline sucrose production appears about 500 BC in northern India (37). Before this time, honey would have represented one of the few concentrated sugars to which hominins would have had access. Although honey likely was a favored food by all hominin species, seasonal availability would have restricted regular access. Studies of contemporary hunter-gatherers show that gathered honey represented a relatively minor dietary component over the course of a year, despite high intakes in some groups during short periods of availability. In the Anbarra Aborigines of northern Australia, average honey consumption over four 1-mo periods, chosen to be representative of the various seasons, was 2 kg per person per year (40). In the Ache Indians of Paraguay, honey represented 3.0% of the average total daily energy intake over 1580 consumer days (41). Consequently, current population-wide intakes of refined sugars in Westernized societies represent quantities with no precedent during hominin evolution.
In the past 30 y, qualitative features of refined sugar consumption have changed concurrently with the quantitative changes. With the advent of chromatographic fructose enrichment technology in the late 1970s, it became economically feasible to manufacture high-fructose corn syrup (HFCS) in mass quantity (42). The rapid and striking increase in HFCS use that has occurred in the US food supply since its introduction in the 1970s is indicated in Figure 3. HFCS is available in 2 main forms, HFCS 42 and HFCS 55, both of which are liquid mixtures of fructose and glucose (42% fructose and 53% glucose and 55% fructose and 42% glucose, respectively) (42). Increases in HFCS occurred simultaneously, whereas sucrose consumption declined (Figure 3). On digestion, sucrose is hydrolyzed in the gut into its 2 equal molecular moieties of glucose and fructose. Consequently, the total per capita fructose consumption (fructose from HFCS and fructose from the digestion of sucrose) increased from 23.1 kg in 1970 to 28.9 kg in 2000. As was the case with sucrose, current Western dietary intakes of fructose could not have occurred on a population-wide basis before industrialization and the introduction of the food-processing industry.
FIGURE 3.. Per capita consumption of refined sugars in the United States from 1970 to 2000. Adapted from the US Department of Agriculture (24).
Refined vegetable oils
In the United States, during the 90-y period from 1909 to 1999, a striking increase in the use of vegetable oils occurred (Figure 4). Specifically, per capita consumption of salad and cooking oils increased 130%, shortening consumption increased 136%, and margarine consumption increased 410% (22). These trends occurred elsewhere in the world and were made possible by the industrialization and mechanization of the oil-seed industry (43). To produce vegetable oils from oil-bearing seeds, 3 procedures can be used: 1) rendering and pressing, 2) expeller pressing, and 3) solvent extraction (43). Oils made from walnuts, almonds, olives, sesame seeds, and flax seeds likely were first produced via the rendering and pressing process between 5000 and 6000 y ago. However, except for olive oil, most early use of oils seems to have been for nonfood purposes such as illumination, lubrication, and medicine (43).
FIGURE 4.. Per capita consumption of vegetable oils in the United States from 19091919 to 19901999. Adapted from Gerrior and Bente (22).
The industrial advent of mechanically driven steel expellers and hexane extraction processes allowed for greater world-wide vegetable oil productivity, whereas new purification procedures permitted exploitation of nontraditionally consumed oils, such as cottonseed (43). New manufacturing procedures allowed vegetable oils to take on atypical structural characteristics. Margarine and shortening are produced by solidifying or partially solidifying vegetable oils via hydrogenation, a process first developed in 1897 (44). The hydrogenation process produces novel trans fatty acid isomers (trans elaidic acid in particular) that rarely, if ever, are found in conventional human foodstuffs (44). Consequently, the large-scale addition of refined vegetable oils to the worlds food supply after the Industrial Revolution significantly altered both quantitative and qualitative aspects of fat intake.
Alcohol
In contrast with dairy products, cereal grains, refined sugars, and oils, alcohol consumption in the typical US diet represents a relatively minor contribution (1.4%) to the total energy consumed. The earliest evidence for wine drinking from domesticated vines comes from a pottery jar dated 74007100 y BP from the Zagros Mountains in northern Iran (45), whereas the earliest archaeologic indication of the brewing of beer and beer consumption dates to the late fourth millennium BC from the Godin site in southern Kurdistan in Iran (46). The incorporation of distilled alcoholic beverages into the human diet came much later. During the period from 800 to 1300 AD, various populations in Europe, the Near East, and China learned to distill alcoholic beverages (47).
The fermentation process that produces wine takes place naturally and, without doubt, must have occurred countless times before humans learned to control the process. As grapes reach their peak of ripeness in the fall, they may swell in size and burst, thereby allowing the sugars in the juice to be exposed to yeasts growing on the skins and to produce carbon dioxide and ethanol (48). Because of seasonal fluctuations in fruit availability and the limited liquid storage capacity of hunter-gatherers, it is likely that fermented fruit drinks, such as wine, would have made an insignificant or nonexistent contribution to total energy in hominin diets before the Neolithic (49).
Salt
The total quantity of salt included in the typical US diet amounts to 9.6 g/d (Table 1). About 75% of the daily salt intake in Western populations is derived from salt added to processed foods by manufacturers; 15% comes from discretionary sources (ie, cooking and table salt use), and the remainder (10%) occurs naturally in basic foodstuffs (50). Hence, 90% of the salt in the typical US diet comes from manufactured salt that is added to the food supply.
The systematic mining, manufacture, and transportation of salt have their origin in the Neolithic Period. The earliest salt use is argued to have taken place on Lake Yuncheng in the Northern Province of Shanxi, China, by 6000 BC (51). In Europe the earliest evidence of salt exploitation comes from salt mines at Cardona, Spain, dating to 62005600 BP (52). It is likely that Paleolithic (the old stone age which began 2.6 million years ago and ended 1000012000 y ago) or Holocene (10000 y ago to the present) hunter-gatherers living in coastal areas may have dipped food in seawater or used dried seawater salt in a manner similar to nearly all Polynesian societies at the time of European contact (53). However, the inland living Maori of New Zealand lost the salt habit (53), and the most recently studied inland hunter-gatherers add no or little salt to their food on a daily basis (54). Furthermore, there is no evidence that Paleolithic people undertook salt extraction or took interest in inland salt deposits (55). Collectively, this evidence suggests that the high salt consumption (10 g/d) in Western societies has minimal or no evolutionary precedent in hominin species before the Neolithic period.
Fatty domestic meats
Before the Neolithic period, all animal foods consumed by hominins were derived from wild animals. The absolute quantity of fat in wild mammals is dependent on the species body mass larger mammals generally maintain greater body fat percentages by weight than do smaller animals (21, 56). Additionally, body fat percentages in wild mammals typically vary by age and sex and also seasonally in a cyclic waxing and waning manner with changing availability of food sources and the photoperiod (Figure 5) (57, 58). Hence, maximal or peak body fat percentages in wild mammals are maintained only for a few months during the course of a year, even for mammals residing at tropical and southern latitudes (59). In mammals, storage of excess food energy as fat occurs primarily as triacylglycerols in subcutaneous and abdominal fat depots. The dominant (>50% fat energy) fatty acids in the fat storage depots (adipocytes) of wild mammals are saturated fatty acids (SFAs), whereas the dominant fatty acids in muscle and all other organ tissues are polyunsaturated fatty acids (PUFAs) and monounsaturated fatty acids (MUFAs) (11). Because subcutaneous and abdominal body fat stores are depleted during most of the year in wild animals, PUFAs and MUFAs ordinarily constitute most of the total carcass fat (11). MUFAs and PUFAs are the dominant fats in the edible carcass of caribou for all 12 mo of the year, as illustrated in Figure 6 (11, 60-65). Because of the seasonal cyclic depletion of SFAs and enrichment of PUFAs and MUFAs, a year-round dietary intake of high amounts of SFAs would have not been possible for preagricultural hominins preying on wild mammals. Even with selective butchering by hominins, in which much of the lean muscle meat is discarded, MUFAs and PUFAs constitute the greatest percentage (>50% of energy as fat) of edible fatty acids in the carcass of wild mammals throughout most of the year (Figure 6).
FIGURE 5.. Seasonal fluctuations in percentage body fat in caribou. Adapted from Spiess (57).
FIGURE 6.. Seasonal variation in mean percentage body fat for mature male, immature male, and mature female caribou (57). Total body fat and total body protein, as a percentage of energy, were calculated from the respective mean values by weight by using the cubic regression equations developed by Cordain et al (20). The edible carcass mass was calculated by subtracting the mass of the bones (minus marrow), hide, hooves, antlers, blood, urine, and gastrointestinal contents from the total live weight. The mass of the edible organs and tissues were calculated from the allometric relation between body mass and organ and tissue mass (60-63). Edible carcass fatty acid composition was calculated by multiplying tissue and organ mass by fatty acid composition (% mass) in these tissues from values for caribou or similar ruminant species (11, 64, 65).
Beginning with the advent of animal husbandry, it became feasible to prevent or attenuate the seasonal decline in body fat (and hence in SFAs) by provisioning domesticated animals with stored plant foods. Furthermore, it became possible to consistently slaughter the animal at peak body fat percentage. Neolithic advances in food-processing procedures allowed for the storage of concentrated sources of animal SFAs (cheese, butter, tallow, and salted fatty meats) for later consumption throughout the year.
Technologic developments of the early and mid 19th centurysuch as the steam engine, mechanical reaper, and railroadsallowed for increased grain harvests and efficient transport of both grain and cattle, which in turn spawned the practice of feeding grain (corn primarily) to cattle sequestered in feedlots (66). In the United States before 1850, virtually all cattle were free range or pasture fed and were typically slaughtered at 45 y of age (66). By about 1885, the science of rapidly fattening cattle in feedlots had advanced to the point that it was possible to produce a 545-kg steer ready for slaughter in 24 mo and that exhibited "marbled meat" (66). Wild animals and free-range or pasture-fed cattle rarely display this trait (11). Marbled meat results from excessive triacylglycerol accumulation in muscle interfascicular adipocytes. Such meat has a greatly increased SFA content, a lower proportion of n3 fatty acids, and more n6 fatty acids (11, 65).
Modern feedlot operations involving as many as 100000 cattle emerged in the 1950s and have developed to the point that a characteristically obese (30% body fat) (67) 545-kg pound steer can be brought to slaughter in 14 mo (68). Although 99% of all the beef consumed in the United States is now produced from grain-fed, feedlot cattle (69), virtually no beef was produced in this manner as recently as 200 y ago (66). Accordingly, cattle meat (muscle tissue) with a high absolute SFA content, low n3 fatty acid content, and high n6 fatty acid content represents a recent component of human diets (11).
HEALTH RAMIFICATIONS OF FOODS IN THE NEOLITHIC AND INDUSTRIAL ERAS
The novel foods (dairy products, cereals, refined cereals, refined sugars, refined vegetable oils, fatty meats, salt, and combinations of these foods) introduced as staples during the Neolithic and Industrial Eras fundamentally altered several key nutritional characteristics of ancestral hominin diets and ultimately had far-reaching effects on health and well-being. As these foods gradually displaced the minimally processed wild plant and animal foods in hunter-gatherer diets, they adversely affected the following dietary indicators 1) glycemic load, 2), fatty acid composition, 3) macronutrient composition, 4) micronutrient density, 5) acid-base balance, 6) sodium-potassium ratio, and 7) fiber content.
Glycemic load
The glycemic index, originally developed in 1981, is a relative comparison of the blood glucose raising potential of various foods or combination of foods based on equal amounts of carbohydrate in the food (70). In 1997, the concept of glycemic load (glycemic index x the carbohydrate content per serving size) was introduced to assess blood glucose raising potential of a food based on both the quality and quantity of dietary carbohydrate (71). Table 2 shows that refined grain and sugar products nearly always maintain much higher glycemic loads than unprocessed fruits and vegetables. Unrefined wild plant foods like those available to contemporary hunter-gatherers typically exhibit low glycemic indices (73).
View this table:
TABLE 2. Glycemic indexes and glycemic loads of various food groups1
Acute elevations in blood glucose concentrations, along with increases in hormones secreted from the gut, stimulate pancreatic insulin secretion causing an acute rise in blood insulin concentrations. Consumption of mixed meals containing protein and fat combined with carbohydrate may lower the total glycemic and insulinemic response of the carbohydrate food alone (74). Nevertheless, it is established that repeated consumption of high glycemic index, mixed meals results in higher mean 24 h blood glucose and insulin concentrations when compared with low glycemic index, mixed meals of identical caloric content (75, 76).
Within the past 20 y, substantial evidence has accumulated showing that long term consumption of high glycemic load carbohydrates can adversely affect metabolism and health (71, 77, 78). Specifically, chronic hyperglycemia and hyperinsulinemia induced by high glycemic load carbohydrates may elicit a number of hormonal and physiologic changes that promote insulin resistance (71, 77, 78). Chronic hyperinsulinemia represents the primary metabolic defect in the metabolic syndrome (79). Diseases of insulin resistance are frequently referred to as "diseases of civilization" (5, 78, 79) and include: obesity, coronary heart disease (CHD), type 2 diabetes, hypertension, and dyslipidemia [elevated serum triacylglycerols, small-dense, LDL cholesterol and reduced HDL cholesterol]. It is likely that the metabolic syndrome may extend to other chronic illnesses and conditions that are widely prevalent in Western societies, including: myopia (80), acne (81), gout (79), polycystic ovary syndrome, epithelial cell cancers (breast, colon, and prostate), male vertex balding, skin tags and acanthosis nigricans (78). Diseases of insulin resistance are rare or absent in hunter-gatherer and other less westernized societies living and eating in their traditional manner (5, 21, 82, 83).
In addition to high-glycemic-load carbohydrates, other elements of Neolithic and Industrial Era foods may contribute to the insulin resistance underlying metabolic syndrome diseases. Milk, yogurt, and ice cream, despite having relatively low glycemic loads (Table 2), are highly insulinotropic, with insulin indexes comparable with white bread (84). Fructose maintains a low glycemic index of 23 and a low glycemic load, but paradoxically it is routinely used to induce insulin resistance in laboratory rodents at high (3565% of energy) dietary concentrations (85, 86). Diets containing lower concentrations (20% of energy) of fructose worsened insulin sensitivity in hyperinsulinemic men (87); more recently it was shown that fructose infusions in healthy men and women induce insulin resistance (88). Dietary fructose may contribute to insulin resistance via its unique ability among all sugars to cause a shift in balance from oxidation to esterification of serum nonesterified free fatty acids (89, 90).
In the typical US diet, sugars with a high glycemic load (HFCS 42, HFCS 55, sucrose, glucose, honey, and syrups) now supply 18.6% of total energy, whereas refined cereal grains with a high glycemic load supplies 20.4% of energy (Table 1). Hence, 39% of the total energy in the typical US diet is supplied by foods that may promote the 4 proximate causes of insulin resistance: chronic and substantial elevations in plasma glucose (91, 92), insulin (93, 94), VLDL (95), and free fatty acid (96) concentrations. Although sugars and grains with a high glycemic load now represent a dominant element of the modern urban diet, these foods were rarely or never consumed by average citizens as recently as 200 y ago.
Fatty acid composition
Chemically, fats are defined as acylglycerolscompounds in which a fatty acid molecule (acyl group) is linked to a glycerol molecule by an ester bond. Almost all dietary and storage fats are triacylglycerols, compounds in which 3 fatty acid molecules are bound to a single glycerol molecule. Fatty acids fall into 1 of 3 major categories: 1) SFAs, 2) MUFAs, and 3) PUFAs. Additionally, essential PUFAs occur in 2 biologically important families, the n6 PUFAs and the n3 PUFAs. Substantial evidence now indicates that to prevent the risk of chronic disease, the absolute amount of dietary fat is less important than is the type of fat (97). Beneficial health-promoting fats are MUFAs and some PUFAs, whereas most SFAs and trans fatty acids are detrimental when consumed in excessive quantities (97). Furthermore, the balance of dietary n6 and n3 PUFAs is integral in preventing the risk of chronic disease and promoting health (97-99).
The Western diet frequently contains excessive saturated and trans fatty acids and has too little n3 PUFAs than n6 PUFAs (97-99). High dietary intakes of SFAs and trans fatty acids increase the risk of CVD by elevating blood concentrations of total and LDL cholesterol (97, 100-102). n3 PUFAs may reduce the risk of CVD via many mechanisms, including reductions in ventricular arrhythmias, blood clotting, serum triacylglycerol concentrations, growth of atherosclerotic plaques, and blood pressure (98). A 20% reduction in overall mortality and a 45% reduction in sudden death after 3.5 y were reported in subjects with preexisting CVD when given 850 mg n3 fatty acids, either with or without vitamin E (103). Higher dietary intakes of n3 fatty acids are also therapeutic in preventing or ameliorating many inflammatory and autoimmune diseases (99). Low- (22% energy) and high- (39% energy) fat diets that had identical ratios of PUFAs to SFAs, n6 PUFAs to n3 PUFAs, and MUFAs to total fat produced no significant differences in total or LDL cholesterol after a 50-d trial (104). These data support the notion that fat quality is more important than fat quantity in regard to CVD risk.
Although much of the early work on the link between diet and CVD focused primarily on dietary fats and their effect on total and LDL-cholesterol concentrations, there are many other dietary elements that can operate synergistically to promote atherosclerosis. As was previously mentioned, carbohydrates with a high glycemic load encourage a proatherogenic blood profile by elevating triacylglycerols and small-dense LDLs, while reducing HDL cholesterol. Atherosclerosis is not just a "plumbing" problem involving excessive LDL cholesterol in the blood from excessive dietary SFAs, but also from chronic inflammation, which is essential in the formation of atherosclerotic plaques (105). A recent study suggested that the blood concentration of the inflammatory marker C-reactive protein (CRP) is a stronger predictor of CVD than is LDL cholesterol (106). High-glycemic-load diets are associated with increased concentrations of CRP (107), as are low dietary intakes of n3 PUFAs (108), and diets that encourage weight loss reduce CRP (109) concentrations. These studies indicate how multiple interrelated qualities of Western diets and recently introduced Neolithic and Industrial Era foods may drive a variety of mechanisms that promote the development of chronic diseases.
The 6 major sources of SFAs in the United States diet are fatty meats, baked goods, cheese, milk, margarine, and butter (110). Five of these 6 foods would not have been components of hominin diets before the advent of animal husbandry or the Industrial Revolution. Because of the inherently lean nature of wild animal tissues throughout most of the year (Figure 5) and the dominance of MUFAs and PUFAs, high dietary levels of SFAs on a year-round basis (Figure 6) could not have exerted adverse selective pressure on the hominin genome before the development of agriculture.
The advent of the oil-seed processing industry at the beginning of the 20th century significantly raised the total intake of vegetable fat (Figure 4), which directly increased the dietary level of n6 PUFAs at the expense of a lowered level of n3 PUFAs because of the inherently higher concentrations of n6 PUFAs and lower concentrations of n3 PUFAs in most vegetable oils (111). The trend toward a higher ratio of n6 to n3 PUFAs was exacerbated as meat from grain fed cattle and livestock became the norm in the US diet over the past 100 y (11, 66). In the current US diet, the ratio of n6 to n3 PUFAs has risen to 10:1 (112), whereas the ratio in hunter-gatherer diets predominant in wild animal foods (20, 21) has been estimated to be between 2:1 and 3:1 (11, 111).
The invention of the hydrogenation process in 1897 (44) allowed vegetable oils to become solidified and marketed as shortening or margarine and as foods containing hydrogenated vegetable oils. The hydrogenation process introduced a novel trans fatty acid (trans elaidic acid) into the human diet, which elevates blood cholesterol concentrations and leads to an increased risk of CVD (113). trans Fatty acids in the US diet are now estimated to constitute 7.4% of the total fatty acid intake (114).
Macronutrient composition
In the present US diet, the percentage of total food energy derived from the 3 major macronutrients is as follows (23): carbohydrate (51.8%), fat (32.8%), and protein (15.4%). Current advice for reducing the risk of cardiovascular disease and other chronic diseases is to limit fat intake to 30% of total energy, to maintain protein at 15% of total energy, and to increase complex carbohydrates to 5560% of total energy (115, 116). Both the current US macronutrient intakes and suggested healthful levels differ considerably from average levels obtained from ethnographic (20) and quantitative (21) studies of hunter gatherers in which dietary protein is characteristically elevated (1935% of energy) at the expense of carbohydrate (2240% of energy) (20, 21). Although the macronutrient compositions of hominin diets during the Paleolithic period cannot be directly determined, recent isotopic data from Neanderthal (117) and Upper Paleolithic European (118) skeletons support the notion that protein consumption may have been substantially higher than current values.
An increasing body of evidence indicates that high-protein diets may improve blood lipid profiles (119-123) and thereby lessen the risk of CVD. Wolfe and Giovannetti (121) have shown that the isocaloric substitution of protein (23% of energy) for carbohydrate in moderately hypercholesterolemic subjects resulted in significant decreases in total, LDL, and VLDL cholesterol and triacylglycerols and an increase in HDL cholesterol. Similar beneficial blood lipid changes have been observed in type 2 diabetic patients in conjunction with improvements in glucose and insulin metabolism (119, 120). Furthermore, high-protein diets have been shown to improve metabolic control in patients with type 2 diabetes (119, 120, 124). In obese women, hypocaloric, high-protein diets improved insulin sensitivity and prevented muscle loss, whereas hypocaloric, high-carbohydrate diets worsened insulin sensitivity and caused reductions in fat-free mass (125).
Epidemiologic evidence supports the clinical data, which shows a cardiovascular protective effect of dietary protein. Protein intake has been shown to be inversely related to CVD in a cohort of 80082 women (126). Dietary protein is also inversely related to blood homocysteine concentration (127), an independent risk factor for CVD. Meat-eating populations have been shown to maintain lower plasma homocysteine concentrations than nonmeat eaters (128, 129). In numerous population studies, summarized by Obarzanek et al (130), higher blood pressure has been associated with lower intakes of protein. A 4-wk dietary intervention of hypertensive subjects showed that a high-protein diet (25% energy) was effective in significantly lowering blood pressure (131). Furthermore, many population studies have established that stroke mortality is inversely related to protein intake (132, 133).
Because protein has >3 times the thermic effect of either fat or carbohydrate (134) and because it has a greater satiety value than do fat or carbohydrate (134, 135), increased dietary protein may represent an effective weight-loss strategy for the overweight or obese. Recent clinical trials have shown that calorie-restricted, high-protein diets are more effective than are calorie-restricted, high-carbohydrate diets in promoting (136-138) and maintaining (139) weight loss in overweight subjects while producing less hunger and more satisfaction (140).
Micronutrient density
Refined sugars are essentially devoid of any vitamin or mineral (64). Accordingly, the consumption of refined sugar or foods containing refined sugar reduces the total vitamin and mineral (micronutrient) density of the diet by displacing more nutrient-dense foods. A similar situation exists for refined vegetable oils, except that they contain 2 fat-soluble vitamins (vitamin E and vitamin K) (64). Because vegetable oils and refined sugars contribute 36.2% of the energy in a typical US diet (Table 1), the widespread consumption of these substancesor foods made with themhas considerable potential to influence the risk of vitamin and mineral deficiencies.
The vitamins and minerals most frequently lacking in the US diet are listed in Table 3. At least half the US population fails to meet the recommended dietary allowance (RDA) for vitamin B-6, vitamin A, magnesium, calcium, and zinc, and 33% of the population does not meet the RDA for folate. Adequate dietary intake of both folate and vitamin B-6 prevents the accumulation of homocysteine in the bloodstream. Elevated blood concentrations of homocysteine represent an independent risk factor for the development of CVD, stroke, and deep vein thrombosis (141, 142).
View this table:
TABLE 3. Percentages of all individuals aged 2 y not meeting 100% of the 1989 US recommended dietary allowances1
The nutrient density in various food groups for the 13 vitamins and minerals most frequently lacking in the US diet are contrasted in Table 4 (64, 143, 144). Because whole grains and milk maintain the next to the lowest nutrient density rankings, displacement of fruit, vegetables, lean meats, and seafood by these 2 staple food groups lowers the overall micronutrient density in the diet. Wild plant foods known to be consumed by hunter-gatherers generally maintain higher micronutrient concentrations than do their domesticated counterparts (4, 145), as does the muscle meat of wild animals (64). Consequently, the Neolithic introduction of dairy foods and cereal grains as staples would have caused the average micronutrient content of the diet to decline. This situation worsened as cereal milling techniques developed in the Industrial era allowed for the production of bread flour devoid of the more nutrient-dense bran and germ (35). The displacement of more nutrient-dense foods (eg, fruit, vegetables, lean meats, and seafood) by less-dense foods (refined sugars, grains, vegetable oils, and dairy products) and the subsequent decline in dietary vitamin and mineral density has far reaching health implicationsconsequences that not only promote the development of vitamin- deficiency diseases but also numerous infectious and chronic diseases (7).
View this table:
TABLE 4. Mean nutrient density of various foods groups (418-kJ samples)1
Acid-base balance
After digestion, absorption, and metabolism, nearly all foods release either acid or bicarbonate (base) into the systemic circulation (146, 147). As shown in Table 5, fish, meat, poultry, eggs, shellfish, cheese, milk, and cereal grains are net acid producing, whereas fresh fruit, vegetables, tubers, roots, and nuts are net base producing. Legumes yield near-zero mean acid values, which reflects an overlapping distribution from slightly net acid producing to slightly net base producing. Not shown in Table 5 are energy-dense, nutrient-poor foods such as separated fats and refined sugars that contribute neither to the acid nor the base load. Additionally, salt is net acid producing because of the chloride ion (146).
View this table:
TABLE 5. Potential net acid (or base) loads of 17 food groups1
The typical Western diet yields a net acid load estimated to be 50 mEq/d (148). As a result, healthy adults consuming the standard US diet sustain a chronic, low-grade pathogenic metabolic acidosis that worsens with age as kidney function declines (146, 149). Virtually all preagricultural diets were net base yielding because of the absence of cereals and energy-dense, nutrient-poor foodsfoods that were introduced during the Neolithic and Industrial Eras and that displaced base-yielding fruit and vegetables (147). Consequently, a net base-producing diet was the norm throughout most of hominin evolution (147). The known health benefits of a net base-yielding diet include preventing and treating osteoporosis (150, 151), age-related muscle wasting (152), calcium kidney stones (153, 154), hypertension (155, 156), and exercise-induced asthma (157) and slow the progression of age- and disease-related chronic renal insufficiency (158).
Sodium-potassium ratio
The average sodium content (3271 mg/d) of the typical US diet is substantially higher than its potassium content (2620 mg/d) (23). Three dietary factors are primarily responsible for the dietary ratio of sodium to potassium, which is >1.0. First, 90% of the sodium in Western diets comes from manufactured salt (sodium chloride); hence, the sodium content of naturally occurring foods in the average US diet (330 mg) is quite low. Second, vegetable oils and refined sugars, which are essentially devoid of potassium, constitute 36% of the total food energy. The inclusion of these 2 foods into the diet displaces other foods with higher potassium concentrations and thereby reduces the total dietary potassium content. Third, the displacement of vegetables and fruit by whole grains and milk products may further reduce the potassium intake because potassium concentrations in vegetables are 4 and 12 times those in milk and whole grains, respectively, whereas in fruit the potassium concentration is 2 and 5 times that in milk and whole grains (64). Taken together, the addition of manufactured salt to the food supply and the displacement of traditional potassium-rich foods by foods introduced during the Neolithic and Industrial periods caused a 400% decline in the potassium intake while simultaneously initiating a 400% increase in sodium ingestion (4, 12, 159).
The inversion of potassium and sodium concentrations in hominin diets had no evolutionary precedent and now plays an integral role in eliciting and contributing to numerous diseases of civilization. Diets low in potassium and high in sodium may partially or directly underlie or exacerbate a variety of maladies and chronic illnesses, including hypertension, stroke, kidney stones, osteoporosis, gastrointestinal tract cancers, asthma, exercise-induced asthma, insomnia, air sickness, high-altitude sickness, and Menieres Syndrome (ear ringing) (160-170).
Fiber content
The fiber content (15.1 g/d) (23) of the typical US diet is considerably lower than recommended values (2530 g) (116). Refined sugars, vegetable oils, dairy products, and alcohol are devoid of fiber and constitute an average of 48.2% of the energy in the typical US diet (Table 1). Furthermore, fiber-depleted, refined grains represent 85% of the grains consumed in the United States (Table 1), and because refined grains contain 400% less fiber than do whole grains (by energy), they further dilute the total dietary fiber intake. Fresh fruit typically contains twice the amount of fiber in whole grains, and nonstarchy vegetables contain almost 8 times the amount of fiber in whole grains on an energy basis (64). Fruit and vegetables known to be consumed by hunter-gatherers also maintain considerably more fiber than do their domestic counterparts (145). Contemporary diets devoid of cereal grains, dairy products, refined oils and sugars, and processed foods have been shown to contain significantly more fiber (42.5 g/d) than either current or recommended values (159).
Once again, the displacement of fiber-rich plant foods by novel dietary staples, introduced during the Neolithic and Industrial periods, was instrumental in changing the diets that our species had traditionally consumeda diet that would have almost always been high in fiber. Soluble fibers (those found primarily in fruit and vegetables) modestly reduce total and LDL-cholesterol concentrations beyond those achieved by a diet low in saturated fat and fiber, by slowing gastric emptying, may reduce the appetite and help to control caloric intake (171). Diets low in dietary fiber may underlie or exacerbate constipation, appendicitis, hemorrhoids, deep vein thrombosis, varicose veins, diverticulitis, hiatal hernia, and gastroesophageal reflux (172).
SUMMARY
In the United States and most Western countries, diet-related chronic diseases represent the single largest cause of morbidity and mortality. These diseases are epidemic in contemporary Westernized populations and typically afflict 5065% of the adult population, yet they are rare or nonexistent in hunter-gatherers and other less Westernized people. Although both scientists and lay people alike may frequently identify a single dietary element as the cause of chronic disease (eg, saturated fat causes heart disease and salt causes high blood pressure), evidence gleaned over the past 3 decades now indicates that virtually all so-called diseases of civilization have multifactorial dietary elements that underlie their etiology, along with other environmental agents and genetic susceptibility. Coronary heart disease, for instance, does not arise simply from excessive saturated fat in the diet but rather from a complex interaction of multiple nutritional factors directly linked to the excessive consumption of novel Neolithic and Industrial era foods (dairy products, cereals, refined cereals, refined sugars, refined vegetable oils, fatty meats, salt, and combinations of these foods). These foods, in turn, adversely influence proximate nutritional factors, which universally underlie or exacerbate virtually all chronic diseases of civilization: 1) glycemic load, 2) fatty acid composition, 3) macronutrient composition, 4) micronutrient density, 5) acid-base balance, 6) sodium-potassium ratio, and 7) fiber content. However, the ultimate factor underlying diseases of civilization is the collision of our ancient genome with the new conditions of life in affluent nations, including the nutritional qualities of recently introduced foods.
ACKNOWLEDGMENTS
LC, SBE, and SL conceived the article and wrote much of the evolutionary, historical, and background perspective. AS edited the health ramification section on acid-base balance and on the sodium-potassium ratio. BAW and NM edited the health ramification section on fatty acid composition. JHO edited and reviewed all sections concerning the health ramifications of CVD and the metabolic syndrome. JBM reviewed the section of the article on glycemic index and fiber. None of the authors had a financial interest or professional or personal affiliation that compromised the scientific integrity of this work.
REFERENCES