Full
Full
A COMPREHENSIVE
FOUNDATION
Heriberto Cabezas
Georgia College and State University
Heriberto Cabezas
Sustainability - A Comprehensive
FoundGeorgia College and State University
ation
This text is disseminated via the Open Education Resource (OER) LibreTexts Project ([Link] and like the hundreds
of other texts available within this powerful platform, it is freely available for reading, printing and "consuming." Most, but not all,
pages in the library have licenses that may allow individuals to make changes, save, and print this book. Carefully
consult the applicable license(s) before pursuing such effects.
Instructors can adopt existing LibreTexts texts or Remix them to quickly build course-specific resources to meet the needs of their
students. Unlike traditional textbooks, LibreTexts’ web based origins allow powerful integration of advanced features and new
technologies to support learning.
The LibreTexts mission is to unite students, faculty and scholars in a cooperative effort to develop an easy-to-use online platform
for the construction, customization, and dissemination of OER content to reduce the burdens of unreasonable textbook costs to our
students and society. The LibreTexts project is a multi-institutional collaborative venture to develop the next generation of open-
access texts to improve postsecondary education at all levels of higher learning by developing an Open Access Resource
environment. The project currently consists of 14 independently operating and interconnected libraries that are constantly being
optimized by students, faculty, and outside experts to supplant conventional paper-based books. These free textbook alternatives are
organized within a central environment that is both vertically (from advance to basic level) and horizontally (across different fields)
integrated.
The LibreTexts libraries are Powered by NICE CXOne and are supported by the Department of Education Open Textbook Pilot
Project, the UC Davis Office of the Provost, the UC Davis Library, the California State University Affordable Learning Solutions
Program, and Merlot. This material is based upon work supported by the National Science Foundation under Grant No. 1246120,
1525057, and 1413739.
Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not
necessarily reflect the views of the National Science Foundation nor the US Department of Education.
Have questions or comments? For information about adoptions or adaptions contact info@[Link]. More information on our
activities can be found via Facebook ([Link] Twitter ([Link] or our blog
([Link]
This text was compiled on 03/12/2025
TABLE OF CONTENTS
Foreword
Licensing
Preface
4: Biosphere
4.1: Prelude to the Biosphere
4.2: Biogeochemical Cycles and the Flow of Energy in the Earth System
4.3: Biodiversity, Species Loss, and Ecosystem Function
4.4: Soil and Sustainability
1 [Link]
6.4: Environmental Valuation
6.5: Evaluating Projects and Policies
6.6: Solutions- Property Rights, Regulations, and Incentive Policies
Index
Glossary
2 [Link]
Detailed Licensing
3 [Link]
Foreword
Foreword
Sustainability is derived from two Latin words: sus which means up and tenere which means to hold. In its modern form it is a
concept born out of the desire of humanity to continue to exist on planet Earth for a very long time, perhaps the indefinite future.
Sustainability is, hence, essentially and almost literally about holding up human existence. Possibly, the most succinct articulation
of the issue can be found in the Report of the World Commission on Environment and Development. The report entitled “Our
1
Common Future ” primarily addressed the closely related issue of Sustainable Development. The report, commonly know as the
Brundtland Report after the Commission Chair Gro Harlem Brundtland, stated that “Humanity has the ability to make development
sustainable to ensure that it meets the needs of the present without compromising the ability of future generations to meet their own
needs.” Following the concept of Sustainable Development, the commission went on to add ” Yet in the end, sustainable
development is not a fixed state of harmony, but rather a process of change in which the exploitation of resources, the direction of
investments, the orientation of technological development, and institutional change are made consistent with future as well as
present needs. We do not pretend that the process is easy or straightforward. Painful choices have to be made. Thus, in the final
analysis, sustainable development must rest on political will.” Sustainability and the closely related concept of Sustainable
Development are, therefore, very human constructs whose objective is to insure the very survival of humanity in a reasonably
civilized mode of existence. Here, however, I will focus primarily on Sustainability.
The seriousness of the issue of Sustainability has become increasingly important and obvious over the last fifty years driven by an
increasing human population with increasing per capita resource consumption on a planet which is after all finite. Note that the
2
World population increased from approximately 2.5 billion in 1950 to about 7.0 billion in 2012. Furthermore, total World
3
consumption expenditures rose from about 171 Billion in 1960 to approximately 44,000 billions in 2010 expressed in 2012 U.S.
dollars. This is not to say that consumption is necessarily bad, but rather that there are so many people consuming so many
resources that both the World environment and human consumption will have to be managed with far more care and delicacy than
has been necessary in all of the historical past.
A text such as the one being presented here is of paramount importance because it will help to educate the next generation of
students on the very important subject of sustainability. Now sustainability is not exactly a discipline such as, for example, physics.
Rather it is truly a metadiscipline drawing on nearly all of existing human knowledge in approximately equal parts and with more
or less equal importance. This is not to say that different disciplines have not in the past drawn ideas from each other, creating
hybrid disciplines such as, for instance, biophysics - a fusion of physics and biology. Rather, in Sustainability the range of ideas and
issues reach from the depth of biological sciences to the physical sciences and to the social sciences, including politics.
Additionally, the relative importance of each of these aspects seems to be about the same. The reasons for this inherent, perhaps
unprecedented complexity, is that sustainability is about sustaining human existence which requires many things to be sustained
including functioning economic, social, and political systems along with a supportive physical and biological environment and
more.
Hence, the effort to produce a text covering the breadth of sustainability must by necessity come from a comprehensive group of
specialists as is the case here. This allows each field of study to bring its own unique perspective and shed its own light on a very
complex and important subject which could otherwise be intractable. The authors very interestingly point out in the preface that the
text does not necessarily present a self-consistent set of ideas. Rather, a degree of diversity is accepted within the overall rubric of
Sustainability and Science itself. This may be unusual for an academic text, but it is necessary here. The reason is that
environmental problems of our time are both time-sensitive and evolving, and a complete understanding does not exist and may
never exist. But the issues still have to be addressed in good faith, in a timely manner, with the best science on hand. With the
reader’s indulgence, I would like to draw an analogy to a physician who has the responsibility of healing or attempting to heal
patients using the best available medical science in a timely manner, knowing that a complete understanding of medical science
does not exist and, in fact, may never exist.
It is my sincerest hope this work shared freely and widely will be an educational milestone as humanity struggles to understand and
solve the enormous environmental challenges of our time. Further, the text “Sustainability: A comprehensive Foundation,” helps to
provide the intellectual foundation that will allow students to become the engines that move and maintain society on the path of
Sustainability and Sustainable Development through the difficult process of change alluded to in “Our Common Future.”
1 [Link]
Licensing
A detailed breakdown of this resource's licensing can be found in Back Matter/Detailed Licensing.
1 [Link]
Preface
This text is designed to introduce the reader to the essential concepts of sustainability. This subject is of vital importance – seeking
as it does to uncover the principles of the long-term welfare of all the peoples of the planet – but is only peripherally served by
existing college textbooks.
The content is intended to be useful for both a broad-based introductory class on sustainability and as a useful supplement to
specialist courses which wish to review the sustainability dimensions of their areas of study. By covering a wide range of topics
with a uniformity of style, and by including glossaries, review questions, case studies, and links to further resources, the text has
sufficient range to perform as the core resource for a semester course. Students who cover the material in the book will be
conversant in the language and concepts of sustainability, and will be equipped for further study in sustainable planning, policy,
economics, climate, ecology, infrastructure, and more.
Furthermore, the modular design allows individual chapters and sections to be easily appropriated – without the purchase of a
whole new text. This allows educators to easily bring sustainability concepts, references, and case studies into their area of study.
This appropriation works particularly well as the text is free – downloadable to anyone who wishes to use it. Furthermore, readers
are encouraged to work with the text. Provided there is attribution to the source, users can adapt, add to, revise and republish the
text to meet their own needs.
Because sustainability is a cross-disciplinary field of study, producing this text has required the bringing together over twenty
experts from a variety of fields. This enables us to cover all of the foundational components of sustainability: understanding our
motivations requires the humanities, measuring the challenges of sustainability requires knowledge of the sciences (both natural
and social), and building solutions requires technical insight into systems (such as provided by engineering, planning, and
management).
Readers accustomed to textbooks that present material in a unitary voice might be surprised to find in this one statements that do
not always agree. Here, for example, cautious claims about climate change stand beside sweeping pronouncements predicting
future social upheaval engendered by a warming world. And a chapter that includes market-based solutions to environmental
problems coexists with others that call for increased government control. Such diversity of thought characterizes many of the fields
of inquiry represented in the book; by including it, we invite users to engage in the sort of critical thinking a serious study of
sustainability requires.
It is our sincerest hope that this work is shared freely and widely, as we all struggle to understand and solve the enormous
environmental challenges of our time.
1 [Link]
CHAPTER OVERVIEW
This page titled 1: Introduction to Sustainability Humanity and the Environment is shared under a CC BY license and was authored, remixed,
and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
1
1.1: What is Sustainability?
In 1983 the United Nations General Assembly passed resolution 38/161 entitled “Process of Preparation of the Environmental Perspective to the Year 2000
and Beyond,” establishing a special commission whose charge was:
a. To propose long-term environmental strategies for achieving sustainable development to the year 2000 and beyond;
b. To recommend ways in which concern for the environment may be translated into greater co-operation among developing countries and between countries
at different stages of economic and social development and lead to the achievement of common and mutually supportive objectives which take account of
the interrelationships between people, resources, environment and development;
c. To consider ways and means by which the international community can deal more effectively with environmental concerns, in the light of the other
recommendations in its report;
d. To help to define shared perceptions of long-term environmental issues and of the appropriate efforts needed to deal successfully with the problems of
protecting and enhancing the environment, a long-term agenda for action during the coming decades, and inpirational goals for the world community,
taking into account the relevant resolutions of the session of a special character of the Governing Council in 1982.
The commission later adopted the formal name “World Commission on Environment and Development” (WCED) but became widely known by the name of
its chair Gro Harlem Brundtland, a medical doctor and public health advocate who had served as Norway’s Minister for Environmental Affairs and
subsequently held the post of Prime Minister during three periods. The commission had twenty-one members drawn from across the globe, half representing
developing nations. In addition to its fact-finding activities on the state of the global environment, the commission held fifteen meetings in various cities
around the world seeking firsthand experiences on the how humans interact with the environment. The Brundtland Commission issued its final report “Our
Common Future” in 1987.
Although the Brundtland Report did not technically invent the term “sustainability,” it was the first credible and widely-disseminated study that probed its
meaning in the context of the global impacts of humans on the environment. Its main and often quoted definition refers to sustainable development as “…
development that meets the needs of the present without compromising the ability of future generations to meet their own needs.” The report uses the terms
“sustainable development,” “sustainable,” and “sustainability” interchangeably, emphasizing the connections among social equity, economic productivity, and
environmental quality. The pathways for integration of these may differ nation by nation; still these pathways must share certain common traits: “the essential
needs of the world's poor, to which overriding priority should be given, and the idea of limitations imposed by the state of technology and social organization
on the environment's ability to meet present and future needs.”
Thus there are three dimensions that sustainability seeks to integrate: economic, environmental, and social (including sociopolitical). Economic interests
define the framework for making decisions, the flow of financial capital, and the facilitation of commerce, including the knowledge, skills, competencies and
other attributes embodied in individuals that are relevant to economic activity. Environmental aspects recognize the diversity and interdependence within
living systems, the goods and services produced by the world’s ecosystems, and the impacts of human wastes. Socio-political refers to interactions between
institutions/firms and people, functions expressive of human values, aspirations and well-being, ethical issues, and decision-making that depends upon
collective action. The report sees these three elements as part of a highly integrated and cohesively interacting, if perhaps poorly understood, system.
The Brundtland Report makes it clear that while sustainable development is enabled by technological advances and economic viability, it is first and foremost
a social construct that seeks to improve the quality of life for the world’s peoples: physically, through the equitable supply of human and ecological goods and
services; aspirationally, through making available the widespread means for advancement through access to education, systems of justice, and healthcare; and
strategically, through safeguarding the interests of generations to come. In this sense sustainability sits among a series of human social movements that have
occurred throughout history: human rights, racial equality, gender equity, labor relations, and conservation, to name a few.
1.1.1 [Link]
Figure 1.1.1 Overlapping Themes of the Sustainability Paradigm A depiction of the sustainability paradigm in terms of its three main components,
showing various intersections among them. Source: International Union for the Conservation of Nature
The intersection of social and economic elements can form the basis of social equity. In the sense of enlightened management, "viability" is formed through
consideration of economic and environmental interests. Between environment and social elements lies “bearability,” the recognition that the functioning of
societies is dependent on environmental resources and services. At the intersection of all three of these lies sustainability.
The US Environmental Protection Agency (US EPA) takes the extra step of drawing a distinction between sustainability and sustainable development, the
former encompassing ideas, aspirations and values that inspire public and private organizations to become better stewards of the environment and that
promote positive economic growth and social objectives, the latter implying that environmental protection does not preclude economic development and that
economic development must be ecologically viable now and in the long run.
Chapter four presents information on how the three components that comprise sustainability have influenced the evolution of environmental public policy.
Chapter 12 explores in greater detail the ethical basis for sustainability and its cultural and historical significance.
Glossary
sustainable development
Development that meets the needs of the present without compromising the ability of future generations to meet their own needs.
This page titled 1.1: What is Sustainability? is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto Cabezas (GALILEO Open Learning
Materials) .
1.1.2 [Link]
1.2: The IPAT Equation
As attractive as the concept of sustainability may be as a means of framing our thoughts and goals, its definition is rather broad and
difficult to work with when confronted with choices among specific courses of action. Chapter 11 is devoted to various ways of
measuring progress toward achieving sustainable goals, but here we introduce one general way to begin to apply sustainability
concepts: the IPAT equation.
As is the case for any equation, IPAT expresses a balance among interacting factors. It can be stated as
I = P ×A×T
where I represents the impacts of a given course of action on the environment, P is the relevant human population for the problem
at hand, A is the level of consumption per person, and T is impact per unit of consumption. Impact per unit of consumption is a
general term for technology, interpreted in its broadest sense as any human-created invention, system, or organization that serves to
either worsen or uncouple consumption from impact. The equation is not meant to be mathematically rigorous; rather it provides a
way of organizing information for a “first-order” analysis.
Suppose we wish to project future needs for maintaining global environmental quality at present day levels for the mid-twenty-first
century. For this we need to have some projection of human population (P) and an idea of rates of growth in consumption (A).
graph showing world population growth
Figure 1.2.1 World Population Growth. source: U.S. Census Bureau, International Data Base, December 2010 Update
Figure \(\PageIndex{1]\) suggests that global population in 2050 will grow from the current 6.8 billion to about 9.2 billion, an
increase of 35%. Global GDP (Gross Domestic Product, one measure of consumption) varies from year to year but, using Figure
1.2.2 as a guide, an annual growth rate of about 3.5% seems historically accurate (growth at 3.5%, when compounded for forty
years, means that the global economy will be four times as large at mid-century as today).
Figure 1.2.2 Worldwide growth of Gross Domestic Product Source: CIA World Factbook, Graph from IndexMundi
Thus if we wish to maintain environmental impacts (I) at their current levels (i.e. I2050 = I2010), then
P2010 × A2010 × T2010 = P2050 × A2050 × T2050
or
T2050 P2010 A2010 1 1 1
( ) =( )×( ) =( )×( ) =( )
P2050 P2050 A2050 1.35 4 5.4
This means that just to maintain current environmental quality in the face of growing population and levels of affluence, our
technological decoupling will need to reduce impacts by about a factor of five. So, for instance, many recently adopted “climate
action plans” for local regions and municipalities, such as the Chicago Climate Action Plan, typically call for a reduction in
greenhouse gas emissions (admittedly just one impact measure) of eighty percent by mid-century. The means to achieve such
reductions, or even whether or not they are necessary, are matters of intense debate; where one group sees expensive remedies with
little demonstrable return, another sees opportunities for investment in new technologies, businesses, and employment sectors, with
collateral improvements in global and national well-being.
1.2.1 [Link]
Glossary
Sustainable Development
Development that meets the needs of the present without compromising the ability of future generations to meet their own needs.
This page titled 1.2: The IPAT Equation is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto Cabezas
(GALILEO Open Learning Materials) .
1.2.2 [Link]
1.3: Human Consumption Patterns and the “Rebound” Effect
In 1865 William Jevons (1835-1882), a British economist, wrote a book entitled “The Coal Question,” in which he presented data
on the depletion of coal reserves yet, seemingly paradoxically, an increase in the consumption of coal in England throughout most
of the 19th century. He theorized that significant improvements in the efficiency of the steam engine had increased the utility of
energy from coal and, in effect, lowered the price of energy, thereby increasing consumption. This is known as the Jevons Paradox,
the principle that as technological progress increases the efficiency of resource utilization, consumption of that resource will
increase. Increased consumption that negates part of the efficiency gains is referred to as “rebound,” while overconsumption is
called “backfire.” Such a counter-intuitive theory has not been met with universal acceptance, even among economists (see, for
example, “The Efficiency Dilemma”). Many environmentalists, who see improvements in efficiency as a cornerstone of
sustainability, openly question the validity of this theory. After all, is it sensible to suggest that we not improve technological
efficiency?
Whether or not the paradox is correct, the fact that it has been postulated gives us pause to examine in somewhat greater depth
consumption patterns of society. If we let Q be the quantity of goods and services delivered (within a given time period) to people,
and R be the quantity of resources consumed in order to deliver those goods and services, then the IPAT equation can be rewritten
in a slightly different way as seen in 1.3.1.
GDP Q R I
I = P ×[ ] ×[ ] ×[ ] ×[ ] (1.3.1)
P GDP Q R
where [
R
Q
] represents the “resource intensity,” and [
I
R
] is the impact created per unit of resources consumed. Rearranging this
version of the equation gives 1.3.2.
R
R = Q ×[ ] (1.3.2)
Q
which says simply that resources consumed are equal to the quantity of goods and services delivered times the resource intensity.
Q
The inverse of resource intensity is called the resource use efficiency, also known as “resource productivity” or “eco-efficiency,”
R
an approach that seeks to minimize environmental impacts by maximizing material and energy efficiencies of production. Thus we
can say:
1
R = Q ×[ ]
Eco − ef f iciency
that is, resources consumed are equal to goods and services delivered divided by eco-efficiency. Whether or not gains in eco-
efficiency yield genuine savings in resources and lower environmental impacts depends on how much, over time, society consumes
ΔQ
of a given product or service (i.e. the relative efficiency gain, must outpace the quantity of goods and services delivered
Δe
e
. In Q
ΔQ
the terms of Jevons paradox, if Q
≥ Δe
e
then the system is experiencing “backfire.”
Part of the problem in analyzing data pertaining to whether or not such “overconsumption” is happening depends on the specific
good or service in question, the degree to which the data truly represent that good or service, and the level of detail that the data
measure. Table 1.3.1 summarizes some recent findings from the literature on the comparative efficiency and consumption for
several activities over extended periods of observation. Taken collectively these activities capture several basic enabling aspects of
modern society: major materials, transportation, energy generation, and food production. In all cases the data show that over the
ΔQ
long term, consumption outpaces gains in efficiency by wide margins, (i.e., ≥ ). It should also be noted that in all cases, the
Q
Δe
increases in consumption are significantly greater than increases in population. The data of Table 1.3.1 do not verify Jevons
paradox; we would need to know something about the prices of these goods and services over time, and examine the degree to
which substitution might have occurred (for instance aluminum for iron, air travel for automobile travel). To see if such large
increases in consumption have translated into comparable decreases in environmental quality, or declines in social equity, other
information must be examined. Despite this, the information presented does show a series of patterns that broadly reflect human
consumption of goods and services that we consider essential for modern living and for which efficiency gains have not kept pace;
in a world of finite resources such consumption patterns cannot continue indefinitely.
Table 1.3.1 Historical Efficiency and Consumption Trends in the United States. source: Dahmus and Gutowski 2011
1.3.1 [Link]
Avg. Annual Efficiency Avg. Annual Increase in Consumption/Efficienc
Activity Time Period
Improvement (%) Ratio Consumption (%) y
Our consumption of goods and services creates a viable economy, and also reflects our social needs. For example, most of us
consider it a social good that we can travel large distances rather quickly, safely, and more or less whenever we feel the need.
Similarly, we realize social value in having aluminum (lightweight, strong, and ductile) available, in spite of its energy costs,
because it makes so many conveniences, from air travel to beverage cans, possible. This is at the center of the sustainability
paradigm: human behavior is a social and ethical phenomenon, not a technological one. Whether or not we must “overconsume” to
realize social benefits is at the core of sustainable solutions to problems.
Resources
For more information about eco-efficiency, see the World Business Council for Sustainable Development report titled "Eco-
Efficiency: Creating more value with less impact"
References
Dahmus, J. B., and T. G. Gutowski (2011) “Can Efficiency Improvements Reduce Resource Consumption? A Historical Analysis
of Ten Activities” Journal of Industrial Ecology (accepted for publication).
Glossary
eco-efficiency
An approach that seeks to minimize environmental impacts by maximizing material and energy efficiencies of production.
Jevons paradox
The principle that as technological progress increases the efficiency of resource utilization, consumption of that resource will
increase.
overconsumption
A long-term result in which the increase in consumption is greater than the efficiency improvement
This page titled 1.3: Human Consumption Patterns and the “Rebound” Effect is shared under a CC BY license and was authored, remixed, and/or
curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
1.3.2 [Link]
1.4: Challenges for Sustainability
The concept of sustainability has engendered broad support from almost all quarters. In a relatively succinct way it expresses the
basis upon which human existence and the quality of human life depend: responsible behavior directed toward the wise and
efficient use of natural and human resources. Such a broad concept invites a complex set of meanings that can be used to support
divergent courses of action. Even within the Brundtland Report a dichotomy exists: alarm over environmental degradation that
typically results from economic growth, yet seeing economic growth as the main pathway for alleviating wealth disparities.
The three main elements of the sustainability paradigm are usually thought of as equally important, and within which tradeoffs are
possible as courses of action are charted. For example, in some instances it may be deemed necessary to degrade a particular
ecosystem in order to facilitate commerce, or food production, or housing. In reality, however, the extent to which tradeoffs can be
made before irreversible damage results is not always known, and in any case there are definite limits on how much substitution
among the three elements is wise (to date, humans have treated economic development as the dominant one of the three). This has
led to the notion of strong sustainability, where tradeoffs among natural, human, and social capital are not allowed or are very
restricted, and weak sustainability, where tradeoffs are unrestricted or have few limits. Whether or not one follows the strong or
weak form of sustainability, it is important to understand that while economic and social systems are human creations, the
environment is not. Rather, a functioning environment underpins both society and the economy.
This inevitably leads to the problem of metrics: what should be measured and how should the values obtained be interpreted, in
light of the broad goals of the sustainability paradigm? A future chapter addresses this in detail, but presented here is a brief
summary of the findings of the Millennium Ecosystem Assessment (MEA), a project undertaken by over a thousand internationally
recognized experts, from 2001-2005, who assessed the state of the world’s major ecosystems and the consequences for humans as a
result of human-induced changes. In its simplest form, a system is a collection of parts that function together. The MEA presents
findings as assessments of ecosystems and ecosystem services: provisioning services such as food and water; regulating services
such as flood control, drought, and disease; supporting services such as soil formation and nutrient cycling; and cultural services
such as recreational, spiritual, religious and other nonmaterial benefits. MEA presents three overarching conclusions:
“Approximately 60% (15 out of 24) of the ecosystem services examined are being degraded or used unsustainably, including fresh
water, capture fisheries, air and water purification, and the regulation of regional and local climate, natural hazards, and pests. The
full costs of the loss and degradation of these ecosystem services are difficult to measure, but the available evidence demonstrates
that they are substantial and growing. Many ecosystem services have been degraded as a consequence of actions taken to increase
the supply of other services, such as food. These trade-offs often shift the costs of degradation from one group of people to another
or defer costs to future generations.” “There is established but incomplete evidence that changes being made are increasing the
likelihood of nonlinear changes in ecosystems (including accelerating, abrupt, and potentially irreversible changes) that have
important consequences for human well-being. Examples of such changes include disease emergence, abrupt alterations in water
quality, the creation of “dead zones” in coastal waters, the collapse of fisheries, and shifts in regional climate.” “The harmful
effects of the degradation of ecosystem services are being borne disproportionately by the poor, are contributing to growing
inequities and disparities across groups of people, and are sometimes the principal factor causing poverty and social conflict. This
is not to say that ecosystem changes such as increased food production have not also helped to lift many people out of poverty or
hunger, but these changes have harmed other individuals and communities, and their plight has been largely overlooked. In all
regions, and particularly in sub-Saharan Africa, the condition and management of ecosystem services is a dominant factor
influencing prospects for reducing poverty.”
Organizations such as the World Commission on Environment and Development, the Millennium Ecosystem Assessment, and
several others including the Intergovernmental Panel on Climate Change, the Organization for Economic Cooperation and
Development, and the National Academy Report to Congress have all issued reports on various aspects of the state of society and
the environment. The members of these groups are among the best experts available to assess the complex problems facing human
society in the 21st century, and all have reached a similar conclusion: absent the enactment of new policies and practices that
confront the global issues of economic disparities, environmental degradation, and social inequality, the future needs of humanity
and the attainment of our aspirations and goals are not assured.
Review Questions
1. What are the essential aspects of “sustainability” as defined in the Brundtland Report?
2. Define “strong” and “weak” sustainability and give examples of each.
3. State, in your own words, the meaning of the “IPAT” equation?
1.4.1 [Link]
4. What is the “rebound” effect and how is it related to human patterns of consumption?
Glossary
ecosystems
Dynamic systems of human, plant, animal, and microorganism communities and the nonliving environment that interact as a
functional unit
ecosystem services
The benefits humans receive from ecosystems
strong sustainability
All forms of capital must be maintained intact independent of one another. The implicit assumption is that different forms of
capital are mainly complementary; that is, all forms are generally necessary for any form to be of value. Produced capital used
in harvesting and processing timber, for example, is of no value in the absence of stocks of timber to harvest. Only by
maintaining both natural and produced capital stocks intact can non-declining income be assured.
weak sustainability
All forms of capital are more or less substitutes for one another; no regard has to be given to the composition of the stock of
capital. Weak sustainability allows for the depletion or degradation of natural resources, so long as such depletion is offset by
increases in the stocks of other forms of capital (for example, by investing royalties from depleting mineral reserves in
factories).
References
Dahmus, J. B., and T. G. Gutowski (2011) “Can Efficiency Improvements Reduce Resource Consumption? A Historical Analysis
of Ten Activities” Journal of Industrial Ecology (accepted for publication).
This page titled 1.4: Challenges for Sustainability is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto
Cabezas (GALILEO Open Learning Materials) .
1.4.2 [Link]
CHAPTER OVERVIEW
This page titled 2: The Evolution of Environmental Policy in the United States is shared under a CC BY license and was authored, remixed,
and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
1
2.1: The Evolution of Environmental Policy in the United States - Chapter
Introduction
Introduction
It is not uncommon to think of the sustainability paradigm as being a recent interpretation of environmental policy, one that was
given credence by the United Nations report "Our Common Future" (the Brundtland Report) when it was first presented in 1987.
Certainly the period during the final decade of the twentieth century was witness to significant growth in our understanding of the
complexity and global reach of many environmental problems and issues, and as discussed in Chapter 2, the Brundtland report
gave a clear voice to these concerns through its analysis of human dependency and quality of life on ecological systems, social
networks, and economic viability—systems that are closely intertwined and that require more integrated approaches to solving the
many problems that confront humanity at this time. It is also true that it was among the first widely disseminated writings to define
and use the modern meaning of the term "sustainable" through the often-quoted concept of "sustainable development." However, it
would be a mistake to conclude that sustainability as a mental construct and policy framework for envisioning the relationship of
humans and nature came into being suddenly and at a single moment in time. Most environmental historians who have studied U.S.
policy have discerned at least three distinct periods during which new concepts and ideas, scientific understandings, technological
advances, political institutions, and laws and regulations came or were brought into being in order to understand and manage
human impacts on the environment. These were (1) the American conservation movement, (2) the rise of environmental risk
management as a basis for policy, and (3) the integration of social and economic factors to create what we now refer to as the
sustainability paradigm. In this chapter we will explore the roots of modern sustainability (Module 4.2), see how our thinking about
the environment has shifted (Module 4.3), and examine the ways that our environmental public policies have changed through time
(Module 4.4). Along the way it is important to understand that this has been an evolutionary process and that these environmental
"eras," while reflecting the norms, attitudes, and needs of the day, are still very much embodied within the modern concept of
sustainability.
This page titled 2.1: The Evolution of Environmental Policy in the United States - Chapter Introduction is shared under a CC BY license and was
authored, remixed, and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
2.1.1 [Link]
2.2: The American Conservation Movement
Introduction
To most early colonists who immigrated to North America, for whom the concept of “wastage” had no specific meaning, the
continent was a land of unimaginably vast resources in which little effort was made to treat, minimize, or otherwise manage. This is
not surprising, when one stand of trees was consumed for housing or fuel, another was nearby; when one field was eroded to the
point of limited fertility, expansion further inland was relatively simple; when rivers became silted so that fisheries were impaired,
one moved further upstream; and when confronted with endless herds of wild animals, it was inconceivable that one might over-
consume to the point of extinction. European-settled America was a largely agrarian society and, apart from the need to keep
spaces productive and clear of debris, there was little incentive to spend time and energy managing discharges to the “commons”
(see Module 8.2). These attitudes persisted well into the 19th century and aspects of them are still active in the present day. While
such practices could hardly be said to constitute an “environmental policy,” they did serve the purpose of constellating a number of
groups into rethinking the way we went about managing various aspects of our lives, in particular our relationship to the land and
the resources it contained or provided. As early as the mid-18th century, Jared Eliot (1685-1763) of Connecticut, a minister, doctor,
and farmer, wrote a series of treatises on the need for better farming methods. He summarized:
“When our fore-Fathers settled here, they entered a Land which probably never had been Ploughed since the Creation, the Land
being new they depended upon the natural Fertility of the Ground, which served their purpose very well, and when they had worn
out one piece they cleared another, without any concern to amend their Land…(Carman, Tugwell, & True, 1934, p. 29).”
Although Eliot avidly instructed his fellow farmers on better methods of “field husbandry,” there is little evidence that his writings
had a lasting effect (he is most known for advances in the design of the “drill plough,” an early planter that produced even rows of
crops, increasing yields).
By 1850, the population of the United States was approaching 25 million and increasing at the rate of three to four percent per year
(for comparison the population of England was about 26 million, of France 36 million, and Germany about 40 million). Although
the westward migration across North America was well underway, most people still lived within a relatively narrow strip of land
along the east coast. By modern measures the United States was not densely populated, and yet the perception of the country as
“big” and on the international stage was in contrast to the mentality just a few decades before of a new world that had broken with
the old, one of endless open spaces and inexhaustible resources. The country was also becoming more urbanized (about 15 percent
of the population lived in cities, three times the proportion of just fifty years before), and increasingly literate.
Thus by the mid-19th century the American public was prepared to listen to the messages of various groups who had become
concerned about the impacts of growth on society. Three groups in particular, of considerably different sympathies and character,
came to have profound influences on the way we thought of ourselves in relation to the environment, on our land use policies, and
on providing environmental goods and services to the growing population: the “resource efficiency” group, the transcendentalist
movement, and organized industrial interests.
Resource Efficiency
As typified by the concerns of Jared Eliot nearly a century before, there were always some who were alarmed at widespread
agricultural practices that were wasteful, inefficient and, using the modern terminology, unsustainable. By the early 1800s the
cumulative impacts of soil erosion and infertility, decreasing crop yields, and natural barriers to expansion such as terrain and poor
transportation to markets led to an organized effort to understand the causes of these problems, invent and experiment with new,
more soil-conserving and less wasteful practices, communicate what was being learned to the public, and begin to build
government institutions to promote better stewardship of the land and its resources. Although initial conservation concerns were
associated with farming, the same approach soon found its way into the management of forests and timbering, wastes from mining
and smelting, and by the end of the century the control of human disease outbreaks (most commonly associated with cholera and
typhoid) and the impact of chemical exposure on workers. There were many individuals who contributed to understanding the
scientific underpinnings of the environment and educating practitioners: Eugene Hilgard (agricultural science), John Wesley Powell
(water rights), George Perkins Marsh (ecological science), Franklin Hough and Gifford Pinchot (sustainable forestry), J. Sterling
Morton (forestry and environmental education; co-founder of Arbor Day), Frederick Law Olmsted (landscape architecture), and
Alice Hamilton (industrial hygiene), to name a few. These resource conservationists were instrumental in applying scientific
methods to solving the problems of the day, problems that were rooted in our behavior toward the environment, and that had
serious consequences for the well-being of people. It was as a result of these efforts that the basis for the fields of environmental
2.2.1 [Link]
science and engineering, agronomy and agricultural engineering, and public health was established. Over time these fields have
grown in depth and breadth, and have led to the establishment of new areas of inquiry.
Just as importantly, several federal institutions were created to oversee the implementation of reforms and manage the
government’s large land holdings. Legislation forming the Departments of the Interior (1849), and Agriculture (1862), the U.S.
Forest Service (1881), the Geological Survey (1879), and the National Park Service (1916) were all enacted during this period. It
was also the time when several major conservation societies, still active today, came into being: the Audubon Society (1886), the
Sierra Club (1892), and the National Wildlife Federation (1935). Arbor Day was first celebrated in 1872, and Bird Day in 1894.
2.2.2 [Link]
Figure 2.2.1 The painting titled "The Kindred", dated 1849, depicts the artist, Thomas Cole, and poet, William Cullen Bryant.
Source: Asher Brown Durand via Wikimedia Commons
Yet, it is difficult to fully appreciate Emerson’s vision of humans and nature through language alone. As might be expected, the
counter-reaction to the state of society and its attitudes toward the environment found expression in other media as well, in
particular the rise of a cadre of American landscape artists. The camera had not yet been perfected, and of course there was no
electronic media to compete for people’s attention, thus artists’ renditions of various scenes, especially landscapes, were quite
popular. Figure 2.2.2, a rendering by A.B. Durand (1796-1886) of an artist and a poet out for a hike amid a lush forest scene
captures much of the essence of transcendental thought, which had strongly influenced Durand’s style. The offset of the human
subjects, to left-of-center, is purposeful: the main subject is nature, with humans merely a component. This theme carried through
many of the landscapes of the period, and helped to define what became known, among others, as the “Hudson River School,”
whose artists depicted nature as an otherwise inexpressible manifestation of God. This is further expressed in the painting, In the
Heart of the Andes, by Frederic Church (Figure 2.2.2). Here, the seemingly sole theme is the landscape itself, but closer inspection
(see detail in red square) reveals a small party of people, perhaps engaged in worship, again offset and virtually invisible amid the
majesty of the mountains.
Figure 2.2.2 In the Heart of the Andes, The painting, dated 1859, depicts a majestic landscape and closer inspection reveals a small
party of people near the bottom left. Source: Frederic Edwin Church via Wikimedia Commons.
Other notable contributors to the transcendental movement were Henry David Thoreau (1817-1862), abolitionist and author of
Walden and Civil Disobedience, Margaret Fuller (1810-1850), who edited the transcendental journal “The Dial” and wrote Woman
in the Nineteenth Century, widely considered the first American feminist work, and Walt Whitman (1819-1892) whose volume of
poetry Leaves of Grass celebrates both the human form and the human mind as worthy of praise.
It is important to recognize that the transcendental redefinition of our social contract with the environment was holistic. Within it
can be found not only a new appreciation of nature, but also the liberation of the human mind from convention and formalism,
attacks on slavery, the need for racial equality, concern for universal suffrage and women’s rights, and gender equity. In many ways
it was a repositioning of the ideals of the enlightenment that had figured so prominently in the founding documents of the republic.
These social concerns are represented today within the sustainability paradigm in the form of such issues as environmental justice,
consumer behavior, and labor relations.
2.2.3 [Link]
Transcendentalism as a formal movement diminished during the latter half of the 19th century, but it had a far-reaching influence on
the way society perceived itself relative to the environment. Perhaps no one is more responsible for translating its aspirations into
environmental public policy than John Muir (1838-1914), a Scottish-born immigrant who was heavily influenced by Emerson’s
writings (it is said that the young Muir carried with him a copy of Nature from Scotland). The two first met in 1871 during a
camping trip to the Sierra Mountains of California. Upon learning of Emerson’s planned departure, Muir wrote to him on May 8,
1871 hoping to convince him to stay longer, “I invite you join me in a months worship with Nature in the high temples of the great
Sierra Crown beyond our holy Yosemite. It will cost you nothing save the time & very little of that for you will be mostly in
Eternity” (Chou, 2003).
Muir was a naturalist, author, organizer (founder of the Sierra Club), and as it turns out a remarkably effective political activist and
lobbyist. His association with Theodore Roosevelt (1858-1919, 26th president of the United States), began with a 1903 campaign
visit by Roosevelt to California, where he specifically sought out Muir, whose reputation was by then well known, as a guide to the
Yosemite area (see Figure 2.2.3).
Figure 2.2.3 "Roosevelt and Muir". Theodore Roosevelt and John Muir at Yosemite National Park in 1903.
It was one of Muir’s special talents that he could bridge across their rather different views on the environment (he a strict
preservationist, Roosevelt a practical outdoorsman). By all accounts they had frank but cordial exchanges; for example, upon
viewing the giant Sequoias, Muir remarked to Roosevelt, “God has cared for these trees…but he cannot save them from fools –
only Uncle Sam can do that.” Roosevelt was so taken with his companion that he insisted they avoid political crowds and camp
together overnight in the mountains.
The subsequent legacy of the Roosevelt administration in the name of conservation, even by today’s standards, was significant.
Known as the “conservation president,” Roosevelt was responsible for 225 million acres of land added to the U.S. Forest Service,
and the creation of 50 wildlife refuges and 150 national forests representing, in total, 11 percent of the total land area of the 48
contiguous states.
2.2.4 [Link]
materials into mostly waste (defined here as a quantity of material that no one values, as opposed to salable products) are simply
inefficient and reduce profitability.
Figure 2.2.4 "The Oregon Trail" the painting, dated 1869, depicts the westward migration of settlers via wagon trains, on
horseback, and by foot. Source: Albert Bierstadt via Wikimedia Commons.
As noted above, industrial activities during this time were responsible for significant environmental degradation. Policy reformers
of the day, such as Carl Schurz (as secretary of the Interior) turned their attention in particular to land reforms, which impacted the
expansion of railroads, and forest preservation. And yet, industry played an unquestionable role as enablers of societal shifts
occurring in America by making goods and services available, increasing the wealth of the emerging middle class, and in particular
providing relatively rapid access to previously inaccessible locations – in many cases the same locations that preservationists were
trying to set aside. Reading, hearing stories about, and looking at pictures of landscapes of remote beauty and open spaces was
alluring and stirred the imagination, but being able to actually visit these places firsthand was an educational experience that had
transformative powers. Alfred Bierstadt’s The Oregon Trail (Figure 2.2.4), painted in 1868, depicts the westward migration of
settlers via wagon trains, on horseback, and simply walking – a journey, not without peril, that took about six months. The next
year saw the completion of the transcontinental railroad, and within a few years it became possible to complete the same journey in
as little as six days in comparative comfort and safety.
The movement to designate certain areas as national parks is an illustrative example of the role of industry in promoting land
conservation, thereby setting in motion subsequent large conservation set-asides that reached their zenith during the Roosevelt
administration. It began, in 1864, with the efforts of several California citizens to have the U.S. Congress accept most of Yosemite,
which had been under the “protection” of the State of California as a national preserve. The petition cited its value “for public use,
resort, and recreation,” reasoning that already reflected the combined interests of the resource efficiency group, preservationists,
and business opportunists. Frederick Law Olmsted (1822-1903), the landscape architect most well known for the design of New
York’s Central Park, and an ardent believer in the ability of open spaces to improve human productivity, oversaw the initial efforts
to manage the Yosemite area. Although the effort was infused with renewed vigor after John Muir’s arrival in the late 1860s, it
wasn’t until 1906 that the park was officially designated.
In the meantime, similar interests had grown to name Yellowstone as a national park, with the same basic justification as for
Yosemite. Since there were no states as yet formed in the region the pathway was more straightforward, and was made
considerably easier by the lack of interest by timber and mining companies to exploit (the area was thought to have limited
resource value), and the railroads who, seeing potential for significant passenger traffic, lobbied on its behalf. Thus the first
national park was officially designated in 1872, only three years after the completion of the transcontinental railroad. Indeed, in
relatively rapid succession the Union Pacific Railroad got behind the Yosemite efforts, and the Northern Pacific Railroad lobbied
heavily for the creation of parks at Mount Rainier (1899) and Glacier (1910). By 1916, when the National Park Service was
formed, sixteen national parks had been created. States too began to see value in creating and, to a degree, preserving open spaces,
as evidenced by New York’s Adirondack Park (1894), still the largest single section of land in the forty-eight contiguous states
dedicated to be “forever wild.”
2.2.5 [Link]
Results of the American Conservation Movement
With the advent of the First World War, and subsequent political, social, and economic unrest that lasted for another thirty years,
actions motivated by the conservation movement declined. The coalition between the resource efficiency group and those wishing
to preserve nature, always uncomfortable, was further eroded when it became clear that the main reason Congress was “setting
aside” various areas was mainly to better manage commercial exploitation. And yet, the period from 1850 to 1920 left a remarkable
legacy of environmental reform, and laid the foundation for future advances in environmental policy. In summary, the conservation
movement accomplished the following:
Redefined the social contract between humans and the environment, establishing a legacy of conservation as part of the
American character, and a national model for the preservation of natural beauty.
Invented the concept of national parks and forests, wildlife refuges, and other sites for commercial and recreational uses by
society.
Developed the first scientific understanding of how the environment functioned, integrating the scientific approach to resource
management into government policy.
Pioneered technological practices to improve resource management.
Established the major federal institutions with responsibility for land and resource conservation.
Communicated the impact of pollution on human health and welfare.
Through publications and travel, exposed many to the beauty of the natural environment and the consequences of human
activities.
Finally, although sustainability as a way of envisioning ourselves in relation to the environment was still many years away,
already its three principal elements, imperfectly integrated at the time, are seen clearly to be at work.
References
Carman, H.J., Tugwell, R.G., & True, R.H. (Eds.). (1934). Essays upon field husbandry in New England, and other papers, 1748-
1762, by Jared Eliot. New York: Columbia University Press.
Chou, P.Y. (Ed.). (2003). Emerson & John Muir. WisdomPortal. Retrieved December 11, 2011 from
[Link]
This page titled 2.2: The American Conservation Movement is shared under a CC BY license and was authored, remixed, and/or curated by
Heriberto Cabezas (GALILEO Open Learning Materials) .
2.2.6 [Link]
2.3: Environmental Risk Management
General Definitions
For most people, the concept of risk is intuitive and, often, experiential; for instance most people are aware of the considerably
greater likelihood of suffering an injury in an automobile accident (116/100 million vehicle miles) versus suffering an injury in a
commercial airplane accident (0.304/100 million airplane miles). Environmental risk can be defined as the chance of harmful
effects to human health or to ecological systems resulting from exposure to any physical, chemical, or biological entity in the
environment that can induce an adverse response (see Module 9.5 for more detail on the science of risk assessment). Environmental
risk assessment is a quantitative way of arriving at a statistical probability of an adverse action occurring. It has four main steps:
1. Identification of the nature and end point of the risk (e.g. death or disability from hazardous chemicals, loss of ecological
diversity from habitat encroachment, impairment of ecosystem services, etc.)
2. Development of quantitative methods of analysis (perturbation-effect, dose-response)
3. Determination of the extent of exposure (i.e. fate, transport, and transformation of contaminants to an exposed population), and
4. Calculation of the risk, usually expressed as a statistical likelihood.
Risk management is distinct from risk assessment, and involves the integration of risk assessment with other considerations, such
as economic, social, or legal concerns, to reach decisions regarding the need for and practicability of implementing various risk
reduction activities. Finally, risk communication consists of the formal and informal processes of communication among various
parties who are potentially at risk from or are otherwise interested in the threatening agent/action. It matters a great deal how a
given risk is communicated and perceived: do we have a measure of control, or are we subject to powerful unengaged or arbitrary
forces?
2.3.1 [Link]
Figure 2.3.1 Texas Dust Storm. Photograph shows a dust storm approaching Stratford, TX in 1935. Source: NOAA via Wikimedia
Commons
Figure 2.3.2 Clear Cutting, Lousiana 1930. Typical cut-over longleaf pine area, on Kisatchie National Forest. Areas of this type
were the first to be planted on this forest. Circa 1930s. Source: Wait, J.M. for U.S. Forest Service. U.S. Forest Service photo
courtesy of the Forest History Society, Durham, N.C.
In the aftermath of the World War II, economic and industrial activity in the United States accelerated, and a consumer-starved
populace sought and demanded large quantities of diverse goods and services. Major industrial sectors, primary metals, automotive,
chemical, timber, and energy expanded considerably; however there were still few laws or regulations on waste management, and
the ones that could and often were invoked (e.g. the Rivers and Harbors Act of 1899) were devised in earlier times for problems of
a different nature. The Module 9.2 provides a more detailed accounting of the current framework for managing waste. Here we
recount the circumstances that eventually resulted in the promulgation of environmental risk as a basis for public policy, with
subsequent passage of major environmental legislation.
Figure 2.3.3 Zinc Smelter. Photograph shows a local smelter in a small valley town in Pennsylvania with, essentially, uncontrolled
emissions. Source: The Wire Mill, Donora, PA, taken by Bruce Dresbach in 1910. Retrieved from the Library of Congress
If there were any doubts among American society that the capacity of the natural environment to absorb human-caused
contamination with acceptably low risk was indeed infinite, these were dispelled by a series of well-publicized incidents that
occurred during the period 1948-1978. Figure 2.3.3 shows a local smelter in a small valley town in Pennsylvania with, essentially,
uncontrolled emissions. During periods of atmospheric stability (an inversion), contaminants became trapped, accumulated, and
caused respiratory distress so extraordinary that fifty deaths were recorded. Figure 2.3.4 illustrates the dramatically poor air quality,
in the form of reduced visibility, during this episode. Such incidents were not uncommon, nor were they limited to small American
towns. A well-documented similar episode occurred in London, England in 1952 with at least 4000 deaths, and 100,000 illnesses
resulting.
2.3.2 [Link]
Figure 2.3.4 Noon in Donora. Photograph, dated October 29, 1948, illustrates the extremely poor air quality in the Pennsylvania
town at the time. Source: NOAA
The generally poor state of air quality in the United States was initially tolerated as a necessary condition of an industrialized
society. Although the risks of occupational exposure to chemicals was becoming more well known, the science of risk assessment
as applied to the natural environment was in its infancy, and the notion that a polluted environment could actually cause harm was
slow to be recognized, and even if true it was not clear what might be done about it. Nevertheless, people in the most contaminated
areas could sense the effects of poor air quality: increased incidence of respiratory disease, watery eyes, odors, inability to enjoy
being outside for more than a few minutes, and diminished visibility.
Figure 2.3.5 Cuyahoga River Fire, 1969. Photograph illustrates a 1969 fire on the Cuyahoga River, one of many fires during the
time period. Source: NOAA.
2.3.3 [Link]
Figure 2.3.6 Love Canal. The Love Canal region of Niagara Falls, NY, 1978 showing the local grade school and neighboring
houses. Source: New York State Department of Health (1981, April). Love Canal: A special report to the Governor and Legislature,
p. 5.
2.3.4 [Link]
Every four to six years the U.S. EPA releases its Report on the Environment, a collection of data and analysis of trends on
environmental quality. It is quite comprehensive; reporting on an array of measures that chart progress, or lack thereof, on human
impacts on the environment and, in turn, the effects of our actions on human health. It is difficult to summarize all the information
available in a concise way, however most measures of human exposure to toxic chemicals, dating in many cases back to the late
1980s, show clear downward trends, in some cases dramatically so (for example DDT in human tissues, lead in blood serum,
exposure to hazardous wastes from improper disposal, exposure to toxic compounds emitted to the air). In addition, many of other
indicators of environmental quality such as visibility, drinking water quality, and the biodiversity of streams, show improvement.
These are success stories of the risk management approach to environmental quality. On the other hand, other measures, such as
hypoxia in coastal waters, quantities of hazardous wastes generated, and greenhouse gases released are either not improving or are
getting worse.
References
National Environmental Policy Act of 1970, 42 U.S.C., 4321, et seq. (1970). [Link]/compliance/basics/[Link]
This page titled 2.3: Environmental Risk Management is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto
Cabezas (GALILEO Open Learning Materials) .
2.3.5 [Link]
2.4: Sustainability and Public Policy
Complex Environmental Problems
NEPA, both in tone and purpose, was in sharp contrast to the many environmental laws that followed in the 1970s and 1980s that
defined increasingly proscriptive methods for controlling risks from chemical exposure (this is sometimes termed the "command-
and-control" approach to environmental management). In many ways these laws and regulations are ill-suited to the types of
environmental problems that have emerged in the past twenty years. Whereas the focus of our environmental policy has been on
mitigating risk from local problems that are chemical – and media – (land, water, or air) specific, the need has arisen to address
problems that are far more complex, multi-media, and are of large geographic, sometimes global, extent.
An early example of this type of shift in the complexity of environmental problems is illustrated by the phenomenon of acidic
rainfall, a regional problem that occurs in many areas across the globe. Although the chemical cause of acid rain is acidic gases
(such as sulfur dioxide and nitrogen oxides) released into the atmosphere from combustion processes (such as coal burning), the
problem was made considerably worse because of the approach to problem solving typical of the day for episodes such as the
Donora disaster (see Figures 3.3.3 and 3.3.4 in the last section).
Figure shows the distribution in rainfall pH in the United States for the year 1996. Source: National Atmospheric Deposition
Program/National Trends Network via National Park Service.
In order to prevent the local accumulation of contaminants, emission stacks were made much taller, effectively relying on the
diluting power of the atmosphere to disperse offending pollutants. The result was a significant increase in the acidity of rainfall
downwind of major sources, with associated impacts on aquatic and forest resources. Figure 2.4.1 shows this pattern for the eastern
United States in 1996. A more comprehensive solution to this problem (short of replacing coal as a fuel source), has involved
integrated activity on many fronts: science to understand the impacts of acid rain, technology to control the release of acidic gases,
politics in the form of amendments to the Clean Air Act, social equity that defined the role of regional responsibilities in the face of
such large geographic disparities, and economics to understand the total costs of acid rain and design markets to spread the costs of
control. Although acidic rainfall is still an issue of concern, its impacts have been mitigated to a significant degree.
2.4.1 [Link]
Global climate change, and its resultant impacts (increases in temperature and storm and flooding frequency, ocean
acidification, displacement of human populations, loss of biodiversity, sea-level rise), caused by the human-induced emission of
greenhouse gases.
Problems such as these, which require highly integrated solutions that include input from many disciplines and stakeholders, have
been termed "wicked" (Batie, 2008; Kreuter, DeRosa, Howze, & Baldwin, 2004). Wicked problems have certain key
characteristics:
There is not universal agreement on what the problem is – different stakeholders define it differently.
There is no defined end solution, the end will be assessed as "better" or "worse."
The problem may change over time.
There is no clear stopping rule – stakeholders, political forces and resource availability will make that determination on the
basis of "judgments."
The problem is associated with high uncertainty of both components and outcomes.
Values and societal goals are not necessarily shared by those defining the problem or those attempting to make the problem
better.
Wicked problems are not confined to environmental issues, for example the same characteristics arise for problems such as food
safety, health care disparities, and terrorism, but in the context of environmental policy they create the need to reassess policy
approaches and goals, laws and regulations, as well as methods and models for integrated research.
Table 2.4.1 summarizes the major attributes of U.S. environmental policy as it has evolved over the past two centuries. To most
observers it would seem to be true that advances in public policy, in any realm, are driven by problems, real and perceived, that
require systemic solutions. Environmental policy is no exception. Early conservationists were alarmed at the inefficiencies of
human resource management and the encroachment of humans on unspoiled lands. During the 20th century many groups: scientists,
economists, politicians, and ordinary citizens, became alarmed and fearful of the consequences of toxic pollutant loads to the
environment that included localized effects on human health and well-being. And now, as we proceed into the 21st century, an array
of complex problems that have the potential to alter substantially the structure and well-being of large segments of human societies,
calls for a renewal and reassessment of our approach to environmental policy. This has, thus far, proven to be a difficult transition.
Many of these complex problems have multiple causes and impacts, affect some groups of people more than others, are
economically demanding, and are often not as visibly apparent to casual observers as previous impacts, nor are the benefits
perceived to be commensurate with costs. Devising a regulatory strategy for such problems requires an adaptive and flexible
approach that current laws do not foster.
Table 2.4.1 Table summarizes the major attributes of U.S. environmental policy as it has evolved over the past two centuries. Source: T. Theis
adapted from Fiksel, Graedel, Hecht, Rejeski, Saylor, Senge, et al. (2009).
1850-1920 1960-1990 1990-present
2.4.2 [Link]
References
Batie, S. S. (2008, December). Wicked problems and applied economics. American Journal of Agricultural Economics, 90, 1176-
1191 doi: 10.1111/j.1467-8276.2008.01202.x
Fiksel, J., Graedel, T., Hecht, A. D., Rejeski, D., Saylor, G. S., Senge, P. M., Swackhamer, D. L., & Theis, T. L. (2009). EPA at 40:
Bringing environmental protection into the 21st century. Environmental Science and Technology, 43, 8716-8720. doi:
10.1021/es901653f
Kreuter, M. W., DeRosa, C., Howze, E. H., & Baldwin, G. T. (2004, August). Understanding wicked problems: A key to advancing
environmental health promotion. Health, Education and Behavior, 31, 441-54. doi: 10.1177/1090198104265597
This page titled 2.4: Sustainability and Public Policy is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto
Cabezas (GALILEO Open Learning Materials) .
2.4.3 [Link]
2.5: Public Health and Sustainability
Introduction
“Much discussion about sustainability treats the economy, livelihoods, environmental conditions, our cities and infrastructure, and
social relations as if they were ends in themselves; as if they are the reason we seek sustainability. Yet their prime value is as the
foundations upon which our longer-term health and survival depend.” (McMichael, 2006)
Ecological sustainability is more than just continuing the resource flows of the natural world to sustain the economic machine,
while maintaining diversity of species and ecosystems. It is also about sustaining the vast support systems for health and life which
could be considered the real bottom line of sustainability. Before examining the public health effects of non-sustainable
development, we should define public health.
The website for UIC’s School of Public Health says “we are passionate about improving the health and well-being of the people
of Chicago, the state of Illinois, the nation and the world.”
The Illinois Department of Public Health is responsible for protecting the state's 12.4 million residents, as well as countless
visitors, through the prevention and control of disease and injury.”
The New Zealand Ministry of Health defines it as “the science and art of promoting health, preventing disease and prolonging
life through organized efforts of society.”
The National Resources Defense Council an NGO devoted to environmental action, states that public health is “the health or
physical well-being of a whole community.”
Obesity
If our communities are not walkable or bikeable, we need to drive to schools, shops, parks, entertainment, play dates, etc. Thus we
become more sedentary. A sedentary lifestyle increases the risk of overall mortality (2 to 3-fold), cardiovascular disease (3 to 5-
fold), and some types of cancer, including colon and breast cancer. The effect of low physical fitness is comparable to that of
hypertension, high cholesterol, diabetes, and even smoking (Wei et al., 1999; Blair et al., 1996).
Economic Segregation
Walkable and safe communities provide sidewalks, bike paths, proximity, and connections to community services such as grocery
stores, schools, health care, parks, and entertainment. Community design that creates a segregated housing environment with only
expensive housing and no affordable housing segregates people by socio-economic level (i.e. poor from non-poor) and this
generally leads to segregation by race. Lack of physical activity will occur in neighborhoods with no good green and safe
recreational sites. If we have poor public transit systems partly due to lack of density (only more expensive, low-density housing)
and our love of the automobile, then we have increased emissions that contribute to global warming.
2.5.1 [Link]
Figure 2.5.1 Minutes Americans Walk per day Source: National Household Travel Survey, 2001, USDOT
Water Quality
Increasing numbers of roads and parking lots are needed to support an automobile transportation system, which lead to increased
non-point source water pollution and contamination of water supplies (road runoff of oil/gas, metals, nutrients, organic waste, to
name a few) with possible impacts on human health. Increased erosion and stream siltation causes environmental damage and may
affect water treatment plants and thus affect water quality.
Social Capital
On the social sustainability side, we can look at social capital otherwise defined as the “connectedness” of a group built through
behaviors such as social networking and civic engagement, along with attitudes such as trust and reciprocity. Greater social capital
has been associated with healthier behaviors, better self-rated health, and less negative results such as heart disease. However,
social capital has been diminishing over time. Proposed causes include long commute times, observed in sprawling metropolitan
areas. Past research suggests that long commute times are associated with less civic participation; Robert Putnam suggests that
every ten additional minutes of commuting predicts a 10% decline in social capital (Besser, Marcus, & Frumkin, 2008). Urban
Sprawl is considered the reason for most long commutes.
As of 2011, according to an article in the Chicago Tribune, Chicago commuting times are some of the worst – with Chicagoans
spending 70 hours per year more on the road than they would if there was no congestion – up from 18 hours in 1982. They have an
average commute time of 34 minutes each way. These drivers also use 52 more gallons per year per commuter, increasing their
costs and pollution.
Residents of sprawling counties were likely to walk less during leisure time, weigh more, and have greater prevalence of
hypertension than residents of compact counties (Ewing, Schmid, Killingsworth, Zlot, & Raudenbush, 2003).
While more compact development is found to have a negative impact on weight, we also find that individuals with low BMI are
more likely to select locations with dense development. This suggests that efforts to curb sprawl, and thereby make communities
more exercise-friendly, may simply attract those individuals who are predisposed to physical activity (Plantinga & Bernell, 2007).
2.5.2 [Link]
Figure 2.5.2 How Climate Change Affects Population. Diagram summarizing the main pathways by which climate change affects
population health. Source:Created by Cindy Klein-Banai, based on McMichael et al., 2006
Measurement of health effects from climate change can only be very approximate. One major study, by the World Health
Organization (WHO), was a quantitative assessment of some of the possible health impacts that looked at the effects of the climate
changes since the mid-1970s and determined that this may have resulted in over 150,000 deaths in 2000. The study concluded that
the effects will probably grow in the future (World Health Organization, 2009).
Extreme Weather
Climate change can influence heat-related morbidity and mortality, generally a result of the difference between temperature
extremes and mean climate in a given area. Higher temperatures in the summer increase mortality. Studies on the effects of heat
waves in Europe indicate that half of the excess heat during the European heat wave of 2003 was due to global warming and, by
inference, about half of the excess deaths during that heat wave could be attributed to human-generated greenhouse gas emissions
(see Haines, Kovats, Campbell-Lendrum, & Corvalan, 2006; Hellmann, Lesht, & Nadelhoffer, 2007; McMichael, 2006). Urban
centers are more susceptible due to the urban heat island effect that produces higher temperatures in urban areas as compared to the
near-by suburbs and rural areas. Lack of vegetation or evaporation, and large areas of pavement, in cities result in an “Urban Heat
Island,” where urban areas are warmer than the neighboring suburban and rural areas (See Figure 2.5.3). Adaptation can help
reduce mortality through greater prevention awareness and by providing more air-conditioning and cooling centers.
Figure 2.5.3 Sketch of an Ubran Heat-Island Profile Source: Heat Island Group.
The reduction of extreme cold due to global warming, could reduce the number of deaths due to low temperatures. Unlike for heat,
those deaths are usually not directly related to the cold temperature itself but rather to influenza. Also, deaths related to cold spells
would increase to a lesser extent by (1.6%), while heat waves increase them by 5.7%.
Since volatile organic compounds are precursors of ozone, and VOC emissions increase with temperature, this could lead to an
increase in ozone concentrations. For fifteen cities in the eastern United States, the average number of days exceeding the health-
based eight-hour ozone standard is projected to increase by 60 percent (from twelve to almost twenty days each summer) by the
2050s because of warmer temperatures (Lashof, & Patz, 2004). Pollen levels may increase with increased CO2 levels since that
promotes growth and reproduction in plants. This will increase the incidence of allergic reactions. Similarly, poison ivy will grow
more and be more toxic.
Infectious diseases are influenced by climate as pathogen survival rates are strongly affected by temperature change. Diseases
carried by birds, animals, and insects (vector-born) – such as malaria, dengue fever, and dengue hemorrhagic fever – may be
influenced by temperature as mosquitoes are sensitive to climate conditions such as temperature humidity, solar radiation, and
rainfall. For example, there has been a strengthening of the relationship between the El Nino global weather cycle and cholera
2.5.3 [Link]
outbreaks in Bangladesh. Increases in malaria in the highlands of eastern Africa may be associated with local warming trends.
Temperature also affects the rate of food-born infectious disease. In general, however, it is hard to isolate the effects of climate
change that affect the transmission rate and geographic boundaries of infectious disease from other social, economic, behavioral,
and environmental factors (see McMichael et al., 2006). Increased precipitation from extreme rainfall events can cause flooding
which, especially in cities with combined sewer and stormwater systems can be contaminated by sewage lines. This can happen
when the deep tunnels that carry stormwater in Chicago reach capacity and untreated sewage then must be released into Lake
Michigan. E. Coli levels in the lake then increase, forcing beaches to close to prevent the spread of infection.
Diseases are re-emerging and emerging infectious due to intensified food production in “factory” farms. Examples include mad
cow disease (1980s in Britain); the encroachment on rain forest by pig farmers exposed pigs and farmers to the “Nipah” virus
carried by rainforest bats that were seeking food from orchards around the pig farms – driven by deforestation and the drought of El
Nino. This caused infection of pigs which lead to human illness and more than one hundred deaths. Poultry farming (avian
influenza viruses) - crowded ‘factory farming’ may increase the likelihood of viral virulence when there is no selective advantage
in keeping the host bird alive. Other food related issues are discussed in the next section.
Food Production
Climate change can influence regional famines because droughts and other extreme climate conditions have a direct influence on
food crops and also by changing the ecology of plant pathogens (Patz et al., 2005).
There are likely to be major effects of climate change on agricultural production and fisheries. This can be both positive and
negative depending on the direct effects of temperature, precipitation, CO2, extreme climate variations, and sea-level rise. Indirect
effects would have to do with changes in soil quality, incidence of plant diseases and weed and insect populations. Food spoilage
will increase with more heat and humidity. Persistent drought has already reduced food production in Africa. There could be
reduction in nutritional quality due to a reduction in the amount of nitrogen crops incorporate when CO2 levels increase.
Malnutrition will be increased due to drought, particularly poorer countries. Increasing fuel costs also increase the cost of food, as
we are already seeing in 2011. Again, this incremental cost rise affects those who already spend a large portion of their income on
food and can contribute to malnutrition. About one-third, or 1.7 billion, of all people live in water-stressed countries and this is
anticipated to increase to five billion by 2025. Frequency of diarrhea and other diseases like conjunctivitis that are associated with
poor hygiene and a breakdown in sanitation may increase.
Figure 2.5.4 Projection for Future EHW-like Summers in Chicago. The average number of summers per decade with mortality
rates projected to equal those of the Chicago analog to the European Heat Wave of 2003. Values shown are the average of three
climate models for higher (orange) and lower (yellow) emission scenarios for each decade from 1980 to 2090 Source: Hellmann et
al., 2007.
Various studies suggest that increases in population at risk from malnutrition will increase from 40-300 million people over the
current 640 million by 2060 (Rosenzweig, Parry, Fischer & Frohberg, 1993). A more recent study said that today 34% of the
population is at risk and by 2050 this value would grow to 64-72%. Climate change is associated with decreased pH (acidification)
2.5.4 [Link]
of oceans due to higher CO2 levels. Over the past 200 years ocean pH has been reduced by 0.1 units and the IPCC predicts a drop
of 0.14 to 0.35 units by 2100. This may affect shell-forming organisms and the species that depend on them. There could be a
reduction in plankton due to the North Atlantic Gulf Stream (Pauly & Alder, 2005). With already overexploited fish populations, it
will be harder for them to recover.
Natural disasters like floods, droughts, wildfires, tsunamis, and extreme storms have resulted in millions of deaths over the past 25
years and negatively affected the lives of many more. Survivors may experience increased rates of mental health disorders such as
post-traumatic stress disorder. Wildfires reduce air quality, increasing particulate matter that provokes cardiac and respiratory
problems. Sea level rise will increase flooding and coastal erosion. Indirect effects of rising sea levels include the infiltration of salt
water and could interfere with stormwater drainage and sewage disposal. This could force coastal communities to migrate and
create refugees with health burdens such as overcrowding, homelessness, and competition for resources. Air pollution is likely to
be worse with climate change. It can also lead to mobilization of dangerous chemicals from storage or remobilize chemicals that
are already in the environment.
Specific regional effects have may be more severe. Vulnerable regions include temperate zones predicted to experience
disproportionate warming, areas around the Pacific and Indian Oceans that are currently subject to variability in rainfall, and large
cities where they experience the urban heat island effect (Patz et al., 2005). The Chicago area is one urban area where analysis has
been performed to determine the specific health effects that are projected due to climate change (see Figure 2.5.4). Those effects
are similar to the ones described above.
An evaluation of the reductions in adverse health effects that could be achieved by 2020 in four major cities with a total population
of 45 million found that GHG mitigation would “reduce particulate matter and ozone ambient concentrations by about 10% and
avoid some 64,000 premature deaths, 65,000 person-chronic bronchitis case, and 37 million days of restricted activities (Cifuentes,
Borja-Aburto, Gouveia, Thurston & Davis, 2001). The cities’ ozone levels are estimated to increase under predicted future climatic
conditions, and this effect will be more extreme in cities that already suffer from high pollution. The estimates of elevated ozone
levels could mean a 0.11% to 0.27% increase in daily total mortality (Bell et al., 2007). Therefore, reduction of GHG emissions,
along with actions to mitigate the effects of climate change are likely to reduce the public health outcomes associated with climate
change.
Conclusions
The implications of climate change on public health are broad and vast. The interconnectedness of all of earth’s systems and human
health is an area that is a challenge to study; the climate change scenarios are variable. Public health is directly tied to the human
ecosystem that we create through our unsustainable activities. The deterioration of public health on this planet is perhaps the most
important consequence of our own unsustainable choices. Without good public health outcomes, human life on this planet is
threatened and ultimately our actions could cause significant changes in human health, well-being and longevity. It is not the earth
that is at stake - it is humanity.
Review Questions
1. Think about the major sources of energy: coal, nuclear and petroleum. Name some health effects that are associated with each,
as portrayed in recent world events. Find one popular and one scientific source to support this.
2. Describe three health impacts of climate change.
3. Modern farming practices are meant to increase productivity and feed the world solving the problems of malnutrition and
starvation. How would you argue for or against this?
4. What are some outcomes that could be measured to determine if a community is healthy?
Resources
Health Impacts of Climate Change – Society of Occupational and Environmental Health [Link]
v=aLfhwaS677c
References
Bell, M. L., Goldberg, R., Hogrefe, C., Kinney, P. L., Knowlton, K., Lynn, B., . . . Patz, J. A. (2007). Climate change, ambient
ozone, and health in 50 US cities. Climatic Change, 82, 61-76.
Besser L. M., & Dannenberg A. L. (2005, November). Walking to public transit steps to help meet physical activity
recommendations. American Journal of Preventive Medicine, 29(4), 273-280.
2.5.5 [Link]
Besser, L. M., Marcus, M., & Frumkin, H. (2008, March). Commute time and social capital in the U.S. American Journal of
Preventive Medicine, 34(3), 207-211.
Blair S. N., Kampert, J. B., Kohl III, H. W., Barlow, C. E., Macera, C. A., Paffenbarger, Jr, R. S., & Gibbons, L. W. (1996).
Influences of cardiorespiratory fitness and other precursors on cardiovascular disease and all-cause mortality in men and women.
Journal of American Medical Association, 276(3), 205-210.
Cifuentes, L., Borja-Aburto, V. H., Gouveia, N., Thurston, G., & Davis, D. L. (2001). Hidden health benefits of greenhouse gas
mitigation. Science, 293(5533), 1257-1259.
Ewing, R., Schmid, T., Killingsworth, R., Zlot, A., & Raudenbush, S. (2003, September/October). Relationship between urban
sprawl and physical activity, obesity, and morbidity. American Journal of Health Promotion, 18(1), 49-57.
Friedman, M. S., Powell, K. E., Hutwagner, L., Graham, L. M., & Teague, W. G. (2001). Impact of changes in transportation and
commuting behaviors during the 1996 Summer Olympic Games in Atlanta on air quality and childhood asthma. JAMA: The
Journal of the American Medical Association, 285(7), 897–905.
Haines, A., Kovats, R. S., Campbell-Lendrum, D., & Corvalan, C. (2006). Climate change and human health: Impacts,
vulnerability and public health. Journal of the Royal Institute of Public Health. 120, 585-596.
Hellmann, J., Lesht, B., & Nadelhoffer, K. (2007). Chapter Four – Health. In Climate Change and Chicago: Projections and
Potential Impacts. Retrieved from
[Link]/filebin/pdf/report/Chicago_climate_impacts_report_Chapter_Four_Health.pdf
Jentes, E. S., Davis, X. M., MacDonald, S., Snyman, P. J., Nelson, H., Quarry, D., . . . & Marano, N. (2010). Health risks and travel
preparation among foreign visitors and expatriates during the 2008 Beijing Olympic and Paralympic Games. American Journal of
Tropical Medical Hygene, 82, 466–472.
Lashof, D. A., & Patz, J. (2004). Heat advisory: How global warming causes more bad air days. Retrieved from
[Link]
McMichael, A. J. (2006) Population health as the ‘bottom-line’ of sustainability: A contemporary challenge for public health
researchers. European Journal of Public Health, 16(6), 579–582.
McMichael, A. J., Woodruff, R. E., & Hales, S. (2006). Climate change and human health: Present and future risks. Lancet, 367,
859-869.
Patz, J. A., Campbell-Lendrum, D., Holloway, T., & Foley, J. A. (2005). Impact of regional climate change on human health.
Nature, 438, 310-317.
Pauly, D., & Alder, J. (2005). Marine Fisheries Systems. In R. Hassan, R. Scholes, & N. Ash (eds.), Ecosystems and Human Well -
being: Current State and Trends . (Vol. 1). Washington, D.C., Island Press.
Plantinga, A. J., & Bernell, S. (2007). The association between urban sprawl and obesity: Is it a two-way street?, Journal of
Regional Science, 47(5), 857-879.
Rosenzweig, C., Parry, M. L., Fischer, G., & Frohberg, K. (1993). Climate change and world food supply. Research Report No. 3.
Oxford, U.K. : Oxford University, Environmental Change Unit.
Wang, X., Westerdahl, D., Chen, L., Wu, Y., Hao, J., Pan, X., Guo, X., & Zhang, K. M. (2009). Evaluating the air quality impacts
of the 2008 Beijing Olympic Games: On-road emission factors and black carbon profiles. Atmospheric Environment, 43, 4535–
4543.
Wei, M., Kampert, J. B. , Barlow, C. E. , Nichaman, M. Z. , Gibbons, L. W., Paffenbarger, Jr., R. S., & Blair, S. N. (1999).
Relationship between low cardiorespiratory fitness and mortality in normal-weight, overweight, and obese men. Journal of the
American Medical Association, 282(16), 1547-1553.
World Health Organization. (2009). Climate change and human health. Fact sheet, July 2005. Retrieved from
[Link]
Glossary
morbidity
2.5.6 [Link]
The relative frequency of occurrence of a disease.
mortality
The number of deaths that occur at a specific time, in a specific group, or from a specific cause.
urban sprawl
Any environment characterized by (1) a population widely dispersed in low density residential development; (2) rigid
separation of homes, shops, and workplaces; (3) a lack of distinct, thriving activity centers, such as strong downtowns or
suburban town centers; and (4) a network of roads marked by large block size and poor access from one place to another) has
been found to correlate with increased body mass index.
This page titled 2.5: Public Health and Sustainability is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto
Cabezas (GALILEO Open Learning Materials) .
2.5.7 [Link]
CHAPTER OVERVIEW
This page titled 3: Climate and Global Change is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto Cabezas
(GALILEO Open Learning Materials) .
1
3.1: Climate and Global Change – Chapter Introduction
The Earth’s climate is changing. The scientific consensus is that by altering the composition of the atmosphere humans are
increasing the average temperature of the Earth’s surface. This process has already begun – the planet is measurably warmer than it
was at the start of the last century – but scientists predict the change that will occur over the 21st century will be even greater. This
increase will have unpredictable impacts on weather patterns around the globe. We are all experiencing climate change. Our
descendants will likely experience far more.
This chapter focuses on the science of climate change. We recognize that climate change can be a controversial subject, and that
prescriptions for solutions quickly take on a political character, which can raise suspicions of bias. Some argue that the climate is
too complicated to predict, and others suggest that natural variations can explain the observed changes in the climate.
These objections have some merit. It should be no surprise that the Earth’s climate is a complicated subject. First, the atmosphere is
vast: it extends over 600 km (370 miles) above the ground, and it weighs over five quadrillion tons (that’s a five followed by 15
zeros). Second, the atmosphere is a dynamic system, creating blizzards, hurricanes, thunderstorms, and all the other weather we
experience. And it is true that this dynamic system is largely controlled by natural processes – the Earth’s climate has been
changing continually since the atmosphere was produced.
And yet scientists can still confidently say that humans are responsible for the current warming. We can do this because this
complicated system obeys underlying principles. In the modules 5.2 and 5.3, we will describe how these principles work, and how
they have been observed by scientists. We can then use these principles to understand how, by producing greenhouse gases, humans
are altering the physical properties of the atmosphere in such a way as to increase its ability to retain heat.
In the module 5.4 we show how this theoretical prediction of a warming world is borne out by ever stronger evidence.
Temperatures have been measured, and are shown to be increasing. These increases in temperatures are significant and have
observable effects on the world: glaciers are shrinking, sea ice is retreating, sea levels are rising – even cherry blossoms are
blooming earlier in the year.
In the module 5.5, we describe how we can attempt to predict the climate of the future. This is a doubly difficult problem, as it
involves not only physics, but, harder yet, people. What will today’s societies do with the foreknowledge of the consequences of
our actions? The climate has become yet another natural system whose immediate fate is connected to our own. The reader may
find it either reassuring or frightening that we hold the climate’s future in our hands.
This page titled 3.1: Climate and Global Change – Chapter Introduction is shared under a CC BY license and was authored, remixed, and/or
curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
3.1.1 [Link]
3.2: Climate Processes; External and Internal Controls
Learning Objectives
After reading this module, students should be able to
define both "climate" and "weather" and explain how the two are related
use the Celsius temperature scale to describe climate and weather
discuss the role and mechanisms of the major controls on Earth's climate using the concepts of insolation, albedo and
greenhouse gases
identify and describe the mechanisms by which major external and internal changes to the climate (including solar output
variation, volcanoes, biological processes, changes in glacial coverage, and meteorite impacts) operate
know that the Earth's climate has changed greatly over its history as a result of changes in insolation, albedo, and
atmospheric composition
describe the processes that can lead to a "Snowball Earth" using the "positive feedback" concept, and be able to contrast the
climate factors that influenced this period of Earth's history with others, including the dominant factors that operated during
the Cretaceous
state the major ways in which carbon dioxide is both added to and removed from the atmosphere, and be able to describe
why levels of carbon dioxide and other greenhouse gases can be kept in balance
Introduction
The Earth's climate is continually changing. If we are to understand the current climate and predict the climate of the future, we
need to be able to account for the processes that control the climate. One hundred million years ago, much of North America was
arid and hot, with giant sand dunes common across the continent's interior. Six hundred and fifty million years ago it appears that
the same land mass—along with the rest of the globe—was covered in a layer of snow and ice. What drives these enormous
changes through Earth's history? If we understand these fundamental processes we can explain why the climate of today may also
change.
In discussing climate in this chapter, we will be using degrees Celsius (oC) as the unit of temperature measurement.
Figure 3.2.1 This thermometer shows how the two scales compare for typical atmospheric temperatures. A change of one degree
Celsius (1 oC) is equivalent to a change of one and four fifths degrees Fahrenheit (1.8 oF). Source: Michiel1972 at [Link].
The Celsius scale is the standard international unit for temperature that scientists use when discussing the climate. In the Celsius
scale, water freezes at 0 oC and boils at 100 oC. A comfortable room might be heated to 20 oC (which is equivalent to 68 oF).
Temperatures can be converted from the Celsius scale to the Fahrenheit scale with the following equation:
9
°F = °C + 32 (3.2.1)
5
Weather describes the short term state of the atmosphere. This includes such conditions as wind, air pressure, precipitation,
humidity and temperature. Climate describes the typical, or average, atmospheric conditions. Weather and climate are different as
3.2.1 [Link]
the short term state is always changing but the long-term average is not. On The 1st of January, 2011, Chicago recorded a high
temperature of 6 oC; this is a measure of the weather. Measurements of climate include the averages of the daily, monthly, and
yearly weather patterns, the seasons, and even a description of how often extraordinary events, such as hurricanes, occur. So if we
consider the average Chicago high temperature for the 1st of January (a colder 0.5 oC) or the average high temperature for the entire
year (a warmer 14.5 oC) we are comparing the city's weather with its climate. The climate is the average of the weather.
Figure 3.2.3 Insolation effect. A cartoon of how latitude is important in determining the amount of insolation. The same amount of
sunlight (yellow bars) is spread out over twice the planet's surface area when the rays strike the Earth at an angle (compare the
length of the dark lines at the equator and at the poles). Source: Jonathan H. Tomkin.
Figure 3.2.4 shows both the equatorial and seasonal impacts of insolation. High levels of insolation are shown in warm colors (red
and pink) and low levels of insolation are shown in cold colors (blue). Notice that in January (top map) the maximum levels of
insolation are in the Southern Hemisphere, as this is when the Southern Hemisphere is most directly facing the sun. The Arctic
receives very little insolation at this time of year, as it experiences its long polar night. The reverse is true in April (bottom map).
3.2.2 [Link]
Figure 3.2.4 Insolation. Average insolation over ten years for the months of January (top) and April (bottom). Source: Roberta
DiPasquale, Surface Meteorology and Solar Energy Project, NASA Langley Research Center, and the ISCCP Project. Courtesy of
NASA's Earth Observatory.
The equator always receives plenty of sunlight, however, and has a much higher average temperature as a consequence; compare
the average temperature of the equator with that of the poles in Figure 3.2.5.
Figure 3.2.5 The Earth's average annual temperature. Source: Robert A. Rohde for Global Warming Art.
The level of insolation affecting Earth depends on the amount of light (or solar radiation) emitted by the Sun. Over the current
geologic period, this is very slowly changing—solar radiation is increasing at a rate of around 10% every billion years. This change
is much too slow to be noticeable to humans. The sun also goes through an 11-year solar cycle, in which the amount of solar
radiation increases and decreases. At the solar cycle peak, the total solar radiation is about 0.1% higher than it is at the trough.
The Earth's orbit is not perfectly circular, so sometimes the Earth is closer to or further from the Sun than it is on average. This also
changes the amount of insolation, as the closer the Earth is to the Sun the more concentrated the solar radiation. As we shall see in
the next section, these orbital variations have made a big difference in conditions on the Earth during the period in which humans
have inhabited it.
In addition to considering how much energy enters the Earth system via insolation, we also need to consider how much energy
leaves. The climate of the Earth is controlled by the Earth's energy balance, which is the movement of energy into and out of the
Earth system. Energy flows into the Earth from the Sun and flows out when it is radiated into space. The Earth's energy balance is
determined by the amount of sunlight that shines on the Earth (the insolation) and the characteristics of the Earth's surface and
atmosphere that act to reflect, circulate and re-radiate this energy. The more energy in the system the higher the temperature, so
either increasing the amount of energy arriving or decreasing the rate at which it leaves would make the climate hotter.
One way to change how quickly energy exits the Earth system is to change the reflectivity of the surface. Compare the difference in
dark surface of tilled soil (Figure 3.2.6a) with the blinding brightness of snow-covered ice (Figure 3.2.6b).
3.2.3 [Link]
Figure 3.2.6 The reflectivity of the surface of the earth.
The dark soil is absorbing the sun's rays and in so doing is heating the Earth surface, while the brilliant snow is reflecting the
sunlight back into space. Albedo is a measure of how reflective a surface is. The higher the albedo the more reflective the material:
a perfectly black surface has zero albedo, while a perfectly white surface has an albedo of 1 - it reflects 100% of the incident light.
If a planet has a high albedo, much of the radiation from the Sun is reflected back into space, lowering the average temperature.
Today, Earth has an average albedo of just over 30%, but this value depends on how much cloud cover there is and what covers the
surface. Covering soil with grass increases the amount of light reflected from 17% to 25%, while adding a layer of fresh snow can
increase the amount reflected to over 80%. Figure 3.2.7 is a composite photograph of the Earth with the cloud cover removed. As
you can see, forests and oceans are dark (low albedo) while snow and deserts are bright (high albedo).
Figure 3.2.7 The surface of the Earth with cloud cover removed. The poles and deserts are much brighter than the oceans and
forests. Source: NASA Goddard Space Flight Center Image by Reto Stöckli. Courtesy of NASA's Earth Observatory.
Changes in albedo can create a positive feedback that reinforces a change in the climate. A positive feedback is a process which
amplifies the effect of an initial change. If the climate cools, (the initial change), snow covers more of the surface of the land, and
sea-ice covers more of the oceans. Because snow has a higher albedo than bare ground, and ice has a higher albedo than water, this
initial cooling increases the amount of sunlight that is reflected back into space, cooling the Earth further (the amplification, or
positive feedback). Compare the brightness of Figure 3.2.7 with a similar photo montage from February (Figure 3.2.7): the extra
snow has increased the Earth's albedo. Imagine what would happen if the Earth produced even more snow and ice as a result of this
further cooling. The Earth would then reflect more sunlight into space, cooling the planet further and producing yet more snow. If
such a loop continued for long enough, this process could result in the entire Earth being covered in ice! Such a feedback loop is
3.2.4 [Link]
known as the Snowball Earth hypothesis, and scientists have found much supporting geological evidence. The most recent period
in Earth's history when this could have occurred was around 650 Million years ago. Positive feedbacks are often described as
"runaway" processes; once they are begun they continue without stopping.
Figure 3.2.8 This image shows the surface of the Earth in February (the Northern Hemisphere winter) with cloud cover removed.
The seasonal snow cover is brighter (and so has a higher albedo) than the land surface it covers. Source: NASA Goddard Space
Flight Center Image by Reto Stöckli. Courtesy of NASA's Earth Observatory
Albedo does not explain everything, however. The Earth and the Moon both receive the same amount of insolation. Although the
Moon is only slightly more reflective than the Earth, it is much colder. The average temperature on Earth is 15 oC, while the
Moon's average temperature is -23 oC. Why the difference? A planet's energy balance is also regulated by its atmosphere. A thick
atmosphere can act to trap the energy from sunlight, preventing it from escaping directly into space. Earth has an atmosphere while
the Moon does not. If the Earth did not have an atmosphere, it would have an average temperature of -18 oC; slightly warmer than
the Moon since it has a lower albedo.
How does the atmosphere trap the energy from the Sun? Shouldn't the Earth's atmosphere reflect as much incoming radiation as it
traps? It is true the atmosphere reflects incoming solar radiation—in fact, only around half the insolation that strikes the top of the
atmosphere reaches the Earth's surface. The reason an atmosphere generally acts to warm a planet is that the nature of light
radiation changes as it reaches the planet's surface. Atmospheres trap more light than they reflect.
Humans see the Earth's atmosphere as largely transparent; that is, we can see a long way in air. This is because we see light in the
visible spectrum, which is the light radiation in the range of wavelengths the human eye is able to perceive, and visible light is able
to travel a long way through the Earth's atmosphere before it is absorbed. Light is also transmitted in wavelengths we can't see,
such as in the infrared spectrum, which is sometimes referred to as infrared light, heat, or thermal radiation. Compared to visible
light, infrared light cannot travel very far in the Earth's atmosphere before it is absorbed. Solar radiation striking the Earth is largely
in the visible part of the spectrum. The surface of the Earth absorbs this energy and re-radiates it largely in the infrared part of the
spectrum. This means that solar radiation enters the Earth in the form of visible light, unhindered, but tries to leave in the form of
infrared light, which is trapped. Thicker atmospheres keep this infrared radiation trapped for longer, and so warm the Earth—just
like an extra blanket makes you warmer in bed.
This effect is shown in Figure 3.2.9. The visible light radiation enters the atmosphere, and quickly exits as infrared radiation if
there is no atmosphere (top Earth). With our atmosphere (the middle Earth), visible light enters unhindered but the infrared light is
partially reflected back to the surface, increasing the amount of energy and thus the temperature at the Earth's surface. If the
atmosphere is made thicker (bottom Earth) the infrared radiation is trapped for longer, further warming the planet's surface.
3.2.5 [Link]
Figure 3.2.9 A cartoon of the greenhouse effect. (Top) Visible light radiation emitted by the sun (yellow arrows) strikes the Earth
and reflects as infrared radiation (orange arrow); (middle) an atmosphere reflects some of the infrared radiation back toward the
planet; (bottom) a thickened atmosphere reflects greater amounts of infrared radiation. Source: Jonathan H. Tomkin.
Figure 3.2.10 The Mt. Bromo volcano in Indonesia emitting gas into the atmosphere. Source: Jan-Pieter Nap, taken on July 11,
2004.
While volcanoes can warm the Earth by adding carbon dioxide to the atmosphere, which produces a greenhouse effect, they can
also cool the Earth by injecting ash and sulfur into the atmosphere. These additions raise the albedo of the atmosphere, allowing
3.2.6 [Link]
less sunlight to reach the surface of the Earth. The effect lasts until the particles settle out of the atmosphere, typically within a few
years. Volcanic eruptions have impacted human societies throughout history; the Mt. Tambora eruption in 1815 cooled the Earth so
much that snow fell during June in New England, and the more recent Mt. Pinatubo eruption in 1991 (see Figure 3.2.11) ejected so
much sulfuric acid into the atmosphere that global temperatures were lowered by about 0.5 oC in the following year.
Figure 3.2.11 The 1991 eruption of Mt. Pinatubo. Source: U.S. Geological Survey photograph, by Richard P. Hoblitt.
Evidence from the geologic past indicates that similar events have caused mass extinctions wherein a significant fraction of all
species on Earth were wiped out in a relatively short amount of time. Sustained outgassing from continuous volcanic eruptions is
thought to have produced so much ash and aerosols that light sufficient to support photosynthesis in plants was unable to penetrate
the atmosphere, causing the food chain to collapse. The ash particles produced by extended eruptions would also have increased the
Earth's albedo, making conditions inhospitably cool for plants and animals adapted to a warmer environment.
Asteroid impacts can also cause the climate to suddenly cool. When large asteroids strike the Earth, ash is ejected into the
atmosphere, which increases albedo in the same way as volcanic eruptions. Everyday clouds (made up of water droplets) both cool
and warm the Earth. They can cool the Earth by increasing the planet's albedo, reflecting sunlight into space before it reaches the
surface. Clouds can also warm the Earth, by reflecting infrared radiation emitted by the surface back towards the planet. Different
types of clouds, and different conditions, determine which effect predominates. On a hot summer's day, for example, clouds cool us
by shielding us from the sun's rays, but on a winter's night a layer of cloud can act as a warming blanket.
The composition of the Earth's atmosphere is not fixed; greenhouse gases can be added to and removed from the atmosphere over
time. For example, carbon dioxide is added by volcanoes and the decay or burning of organic matter. It is removed by
photosynthesis in plants, when it is dissolved in the oceans and when carbonate sediments (a type of rock) are produced. Over
geologic time, these processes have significantly reduced the proportion of carbon dioxide in the atmosphere. Atmospheric carbon
dioxide levels just prior to the industrial revolution are thought to have been only one twentieth of those of 500 million years ago.
Natural processes also remove carbon dioxide added by human activity, but only very slowly. It is estimated that it would take the
Earth around a thousand years to naturally remove most of the carbon dioxide released by the industrial consumption of fossil fuels
up to the present.
Greenhouse gases other than carbon dioxide are shorter-lived: methane is removed from the atmosphere in around a decade, and
chlorofluorocarbons break down within a century. Individual water molecules spend only a few days at a time in the atmosphere,
but unlike the other greenhouse gases, the total amount of water vapor in the atmosphere remains constant. Water evaporated from
the oceans replaces water lost by condensation and precipitation.
Changing the composition of the Earth's atmosphere also changes the climate. Do you remember the snowball earth? — how
increasing ice cover also increased the Earth's albedo, eventually covering the entire planet in ice and snow? Today's climate is
temperate—so we must have escaped this frozen trap. But how? The leading hypothesis is that the composition of the Earth's
atmosphere changed, with volcanoes slowly adding more and more carbon dioxide to it. Without access to the oceans, plants, or
surface rocks, this carbon dioxide was not removed from the atmosphere and so continued to build up over millions of years.
Eventually, the additional warming caused by the increase in greenhouse gases overcame the cooling caused by the snow's high
albedo, and temperatures rose enough to melt the ice, freeing the Earth.
3.2.7 [Link]
For most of Earth's history, carbon dioxide concentrations have been higher than they are today. As a consequence, past climates
have often been very warm. During the late stage of the dinosaur era (the cretaceous era, a period that lasted between 65 and 145
million years ago), carbon dioxide levels were about 5 times higher than they are today, and the average global temperatures were
more than 10 oC higher than today's. There were no large ice sheets, and dinosaur fossils from this period have been found as far
north as Alaska. These animals would not survive the cold conditions found in the arctic today. Further south, fossil crocodiles
from 60 million years ago have been found in North Dakota. The modern average winter temperature in North Dakota is around
-10 oC –but being cold-blooded, crocodiles are most at home when the air temperature is around 30 oC! The climate was warmer in
the past when the amount of carbon dioxide was higher.
Review Questions
1. The text describes how the high albedo of snow acts as a positive feedback—if the Earth is made cooler, the highly reflective
snow can act to further cool the Earth. Today, part of the Earth is covered with snow and ice. Can you describe a mechanism by
which warmer temperatures would also produce a positive feedback—this time heating the Earth further—through a similar
albedo mechanism?
2. Mars is colder than the Earth. Venus, on the other hand, is much hotter, with average surface temperatures of around 450 oC.
Venus is closer to the Sun than the Earth is, and so receives about twice as much solar radiation. Venus's atmosphere is also
different than Earth's, as it is much thicker and mainly consists of carbon dioxide. Using the terms insolation and greenhouse
gasses, can you suggest reasons why Venus is so hot?
3. Oxygen makes up over 20% of Earth's atmosphere, while carbon dioxide makes up less than 0.04%. Oxygen is largely
transparent to both visible and infrared light. Explain why carbon dioxide is a more important greenhouse gas in the Earth's
atmosphere than oxygen, even though there is much more oxygen than carbon dioxide.
4. Figure 3.2.4 shows the insolation at the surface of the Earth. The Earth is spherical, so we would expect the values to be the
same for places of the same latitude. But notice that this is not true – compare, for example, central Africa with the Atlantic
Ocean at the same latitude. What feature of the atmosphere might explain this variation, and why?
Resources
The National Aeronautical and Space Administration (NASA) Earth Observatory website has an array of climate resources. For a
more in-depth discussion of Earth's energy budget, go to [Link]
Are you interested in finding more about the controversial Snowball Earth hypothesis? The National Science foundation and
Harvard University have set up a website with more about the hypothesis and the evidence. Go to [Link]
Glossary
albedo
A measure of how reflective a surface is. A perfectly black surface has an albedo of 0, while a perfectly white surface has an
albedo of 1.
climate
The average of the weather.
cretaceous period
The period between 65 and 145 million years ago, which was the final period of Earth's history that included dinosaurs.
greenhouse effect
The process by which the atmosphere acts to trap heat, warming the climate.
greenhouse gases
Those gases in the atmosphere that warm the climate, most importantly, water vapor, carbon dioxide, methane, and ozone.
infrared spectrum
The light radiation just below the range of wavelengths visible to the human eye. Also referred to as thermal radiation.
insolation
3.2.8 [Link]
The measure of the amount of solar radiation falling on a surface.
positive feedback
A runaway process which amplifies the effect of an initial change.
snowball earth
A condition in which the entire planet is covered in ice, last thought to have happened 650 million years ago.
solar radiation
The energy emitted by the sun in the form of light.
visible spectrum
The light radiation that is in the range of wavelengths that is visible to the human eye.
weather
A description of the short term state of the atmosphere.
This page titled 3.2: Climate Processes; External and Internal Controls is shared under a CC BY license and was authored, remixed, and/or
curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
3.2.9 [Link]
3.3: Milankovitch Cycles and the Climate of the Quaternary
Learning Objectives
After reading this module, students should be able to
describe the changing climate of the Quaternary
explain why Milankovitch cycles explain the variations of climate over the Quaternary, in terms of the similar periods of
orbital variations and glacial cycles
explain how the glacier/climate system is linked via albedo feedbacks
describe how sediment and ice cores provide information about past climates
use the mechanisms that cause stable isotope fractionation to predict the impact of changing climate on stable isotope
records
Introduction
In Module 5.2 we saw the major drivers of the climate—the energy that comes from the Sun (insolation) and the properties of the
planet that determine how long that energy stays in the Earth system (albedo, greenhouse gases). In this section, we will look at the
recent natural changes in Earth's climate, and we will use these drivers to understand why the climate has changed.
The most recent period of Earth's geologic history—spanning the last 2.6 million years—is known as the quaternary period. This is
an important period for us because it encompasses the entire period over which humans have existed—our species evolved about
200,000 years ago. We will examine how the climate has changed over this period in detail. By understanding recent natural
processes of climate change, we will be able to better understand why scientists attribute the currently observed changes in global
climate as being the result of human activity.
3.3.1 [Link]
however; some are "light" and some are "heavy". These different types of oxygen are called isotopes, which are atoms that have
same number of protons but different numbers of neutrons. The heavy isotope of oxygen (oxygen-18, or 18O) is more than 10%
heavier than the light isotope (oxygen-16 or 16O). This means that some water molecules weigh more than others. This is important
because lighter water molecules are more easily evaporated from the ocean, and once in the atmosphere, heavier water molecules
are more likely to condense and fall as precipitation. As we can see from Figure 3.3.2, the water in the ice sheets is lighter (has a
higher proportion of 16O relative to 18O) than the water in the oceans.
The process of differentiation between heavy and light water molecules is temperature dependent. If the atmosphere is warm, there
is more energy available to evaporate and hold the heavier 18O water in the atmosphere, so the snow that falls on the polar ice
sheets is relatively higher in 18O. When the atmosphere is cold, the amount of energy is less, and so less 18O makes it to the poles
to be turned into glacial ice. We can compare the amount of 18O in different parts of the ice core to see how the atmosphere's
temperature—the climate—has changed.
Figure 3.3.2 Water becomes lighter as it travels toward the poles. The heavy (18O) water drops out of the atmosphere (as rain or
snow) before reaching the ice sheet. This means that the snow that forms the glacial ice is lighter than the ocean water (has more
16O than 18O, compared to ocean water). Source: Robert Simmon, NASA GSFC, NASA Earth Observatory
Figure 3.3.3 shows what this record looks like over the last 400,000 years. The blue and green lines depict two different Antarctic
ice cores (taken from ice about 350 miles apart) and the variations in oxygen isotopes are converted into temperature changes. The
y-axis shows temperature change; today's climate is at zero—the dashed line. Notice that the Earth's climate has not been stable!
Sometimes the temperature is higher than it is today—the blue and green lines are higher than the dashed about 120,000 years ago,
for example. Most of the time the climate is much colder than today's, however: the most common value is around -6 oC (-13 oF).
On average, the earth's temperature between 25,000 and 100,000 years ago was about 6 oC lower than it is today. These changes
can be double-checked by measuring the temperature of the ice in the cores directly. Ice that is 30,000 years old is indeed colder
than the ice made today, just as the isotope data predicts.
Figure 3.3.3 Ice age temperature. The blue and green lines depict two different Antarctic ice cores (taken from ice about 350 miles
apart) and the variations in oxygen isotopes are converted into temperature changes. The red line depicts global ice volume. The y-
axis shows temperature change; today's climate is at zero – the dashed line. Source: Robert A. Rohde
3.3.2 [Link]
Figure 3.3.4 A comparison of the age of sediment (x-axis) and the change in temperature over time (left y-axis) as derived from
oxygen isotope ratios (right y-axis). The dashed line shows today's climate. Note that the climate is cooling over the last few
million years, but it is highly variable. In the last one million years the climate alternates between warm and cool conditions on a
100,000-year time scale ("100 kyr cycle"), before this it alternated on a 41,000 year cycle. Both these period lengths are the same
as Milankovitch cycles. These cores suggest that today's temperature is higher than almost all of that of the Quaternary (the last 2.6
Million years). Source: Jo Weber
The changes in climate recorded in ice sheets are thought to be worldwide. The same climate changes observed in Antarctica are
also found in cores taken from Greenland, which is on the other side of the Earth. Isotope data can also be taken from sediment
cored from the ocean floor—all over the planet—and these cores also show the same changes in climate, alternating between cold
and warm. Because ocean sediment is deposited over millions of years, the sediment can give an indication of the climate across
the whole of the Quaternary and beyond. Figure 3.3.4 shows how temperature has changed over time (blue line), compared with
today (dashed line). The temperature has, on average, gotten colder over the Quaternary, but it also appears to oscillate between
warm and cold periods. We'll investigate these periodic changes in the next section of this chapter.
As falling snow accumulates on the ground, tiny bubbles of air become trapped in it. These bubbles are retained as the snow
transforms to ice, and constitute tiny samples of the ancient atmosphere that can be analyzed to find out if the changes in
temperature (as recorded in the oxygen isotopes) are related to changes in the atmosphere. The temperature recorded by the
isotopes in the ice is directly related to the amount of carbon dioxide in the trapped air (Figure 3.3.5): the times with higher carbon
dioxide are also times of high temperature.
Falling snow also captures and entombs atmospheric dust, which is topsoil born aloft by the wind, and which is especially
prevalent during droughts. The fact that more dust occurs in the ice accumulated during cold periods suggests that the glacial
climate was dry, as well as cold.
Figure 3.3.5 Vostok Petit Data. These graphs depict how changes in temperature—inferred from changes in isotope ratios (blue
line)--correspond to changes in atmospheric carbon dioxide (green line) and dust (red line) over the last 400,000 years as recorded
in an ice core extracted from Antarctica. Carbon dioxide varies directly with temperature – the warmer the climate the higher the
carbon dioxide level. Atmospheric dust is highest during the coolest periods (such as 25,000 and 150,000 years ago). Source:
William M. Connolley produced figure using data from the National Oceanic and Atmospheric Administration, U.S. Department of
Commerce, Paleoclimatology branch, Vostok Ice Core Data.
3.3.3 [Link]
Figure 3.3.6 Ice Age Earth- an artist's impression of the Earth during an ice age. Note that the Northern parts of North America and
Europe (including Canada and Scandinavia) are entirely covered by ice-sheets. Source: Itti
During the Quaternary, the Earth has cycled between glacial periods (sometimes referred to as "ice ages") and interglacial periods.
The ice was at its most recent extreme around 20,000 years ago in a period known as the last glacial maximum, or LGM. As we can
see from the ice core record, the Quaternary climate is usually cold (see Figure 3.3.3), with long periods of cold punctuated with
shorter (10,000 year long, or so) periods of warmer conditions, like those we experience today. In many ways, our current climate
is exceptional—for most of human existence, the Earth has been a much colder place.
What was the Earth like during these glacial periods? Almost all the world was cold; average temperatures were around 6 oC (-13
o
F) colder than today. Such conditions allow ice sheets to grow—much of North America, Asia and Europe were covered under
mile-thick ice (see Figure 3.3.3). Because this ice was made of water that was once in the oceans, sea levels were much lower. At
the LGM, sea level was about 120 meters (or about 400 feet) lower than it is today. As the seas retreated, the continents grew
larger, creating land bridges that joined Asia with North America, Britain with Europe, and Australia with Papua New Guinea.
During glacial periods the climate was also much drier, as evidenced by the increase in atmospheric dust (Figure 3.3.5). The lands
at and near the poles were covered with ice, and dry grasslands occupied areas where temperate forests occur today. Deserts were
much larger than they are now, and tropical rainforests, having less water and less warmth, were small. The animals and plants of
glacial periods were different in their distribution than they are today, as they were adapted to these different conditions. Fossils of
Mastodons (Figure 3.3.7) have been found from all across what is now the United States, including from Florida, which currently
enjoys a subtropical climate.
Figure 3.3.7 An artist's impression of a Knight Mastodon, an elephant-like mammal with a thick wooly coat. Mastodon fossils
dating from past glacial periods have been found across North America—from Florida to Alaska. Source: Charles R. Knight
During glacial periods humans would have been unable to occupy the globe as they do today because all landmasses experienced
different climactic conditions. Some countries of the present could not exist, as they would be almost completely covered by ice.
As examples, look for Canada, Iceland and The United Kingdom in Figure 3.3.9.
Milankovitch Cycles
Why has the Earth cycled through hot and cold climates throughout the Quaternary? As we learned in the previous module, the
Earth's climate is controlled by several different factors—insolation, greenhouse gases, and albedo are all important. Scientists
believe that changes in insolation are responsible for these climate swings, and the insolation varies as a result of wobbles in the
Earth's orbit.
3.3.4 [Link]
The Earth's orbit is not fixed – it changes regularly over time. These periodic changes in Earth's orbit named are referred to as
Figure 3.3.8, and are illustrated in Figure 3.3.8. Changes in the Earth's orbit alter the pattern of insolation that the Earth receives.
There are three principle ways in which the Earth's orbit varies:
1. Eccentricity (or Orbital shape): The Earth's orbit is not perfectly circular, but instead follows an ellipse. This means that the
Earth is, through the course of the year, sometimes closer and sometimes further away from the Sun. Currently, the Earth is
closest to the Sun in early January, and furthest from the Sun in Early July. This changes the amount of insolation by a few
percent, so Northern Hemisphere seasons are slightly milder than they would be if the orbital was circular (cooler summers and
warmer winters). The orbital shape changes over time: the Earth moves between being nearly circular and being mildly
elliptical. There are two main periods over which this change occurs, one takes around 100,000 years (this is the time over
which the orbit goes from being circular, to elliptic, and back to circular), another takes around 400,000 years.
2. Axial Tilt (or Obliquity): The Earth axis spins at an angle to its orbit around the Sun – currently this angle is 23.5 degrees (this
angle is known as the axial tilt). This difference in orbit creates the seasons (as each hemisphere takes turns being tilted towards
and away from the Sun over the course of the year). If the axis of spin lined up with the direction of the Earth's orbit (so that the
tilt angle was zero) there would be no seasons! This axial tilt also changes over time, varying between 22.1 and 24.5 degrees.
The larger the angle, the larger the temperature difference between summer and winter. It takes about 41,000 year for the axial
tilt to change from one extreme to the other, and back again. Currently, the axial tilt is midway between the two extremes and is
decreasing—which will make the seasons weaker (cooler summers and warmer winters) over the next 20,000 years.
3. Axial Precession: The direction of Earth's axis of rotation also changes over time relative to the stars. Currently, the North Pole
points towards the star Polaris, but the axis of rotation cycles between pointing to that star and the star Vega. This impacts the
Earth's climate as it determines when the seasons occur in Earth's orbit. When the axis is pointing at Vega, the Northern
Hemisphere's peak summer is in January, not July. If this were true today, it would mean that the Northern Hemisphere would
experience more extreme seasons, because January is when the Earth is closest to the Sun (as discussed above in eccentricity).
This cycle takes around 20,000 years to complete.
Figure 8. Milankovitch Cycles Illustration of the three variables in Earth's orbit, with periods of variation marked. Source:
COMET® at the University Corporation for Atmospheric Research (UCAR) pursuant to a Cooperative Agreements with the
National Oceanic and Atmospheric Administration, U.S. Department of Commerce. ©1997-2009 University Corporation for
Atmospheric Research. All Rights Reserved.
The three cycles described above have different periods, all of which are long by human standards: 20,000, 40,000, 100,000 and
400,000 years. If we look at the temperature data from ice and sediment cores, we see that these periods are reflected in Earth's
climate. In the last million or so years, the 100,000-year eccentricity in the orbit has determined the timing of glaciations, and
before that the 40,000-year axial tilt was dominant (Figure 3.3.8). These cycles have been important for a long time; geologists
have even found evidence of these periods in rocks that are hundreds of millions of years old.
But how do the Milankovitch Cycles change our climate? These orbital cycles do not have much impact on the total insolation the
Earth receives: they change only the timing of that insolation. Since the total insolation does not change, these orbital variations
have the power to make the Earth's seasons stronger or weaker, but the average annual temperature should stay the same. The best
explanation for long term changes in average annual temperature is that the Milankovitch cycles initiate a positive feedback that
amplifies the small change in insolation.
3.3.5 [Link]
Insolation and the Albedo Feedback
Today, the Earth's orbit is not very eccentric (it is almost circular), but at the beginning of each of the recent ice age periods, the
orbit was much more elliptical. This meant that the Earth was further away from the sun during the northern hemisphere summers,
reducing the total insolation. Lower insolation meant that the summer months were milder than they would otherwise be, with
cooler temperatures. Summer temperatures were also lower when the Earth's axial tilt was smaller, so the two different orbital
parameters could reinforce one another's effects, in this case producing especially mild summers.
It is thought that these mild northern summers produced an albedo feedback that made the whole planet slip into an ice age. The
northern hemisphere has continents near the poles—Europe, Asia, and North America. Today, these continents have largely
temperate climates. During the winter, snow falls across much of the land (see Figure 3.3.9 in the previous module) only to melt
during the summer months. If the summers are not hot enough to melt all the snow and ice, glaciers can advance, covering more of
the land. Because ice has a high albedo, more sunlight is reflected than before, and the Earth is made cooler. This creates a positive
feedback, as the cooler conditions allow the ice to advance further—which, in turn, increases the albedo and cools the Earth!
Eventually, a large proportion of the northern continents became covered in ice (Figure 3.3.9).
Figure 3.3.9 Glacial coverage (light blue) of the northern hemisphere during the ice ages. Source: Hannes Grobe
This positive feedback process works in the other direction, as well. The interglacial periods are ushered in when the orbital
parameters create summers that are unusually warm, which melts some of the ice. When the ice sheets shrink, the Earth's albedo
decreases, which further warms the system. The giant northern ice sheets shriveled up in a few thousand years as warm summers
and decreasing albedo worked together.
These cycles of alternating cooling and warming are also related to changes in the amount of greenhouse gases in the atmosphere.
As we observed in Figure 3.3.5, the climate contains higher levels of carbon dioxide during interglacial periods. Although this
appears to make sense—carbon dioxide is a greenhouse gas, and so should produce warmer climates—it is also a puzzle, because it
is not clear how changes in Milankovitch cycles lead to higher levels of carbon dioxide in the atmosphere. It is clear that these
changes in carbon dioxide are important in making the change in temperature between interglacial and glacial periods so extreme.
Several different hypotheses have been proposed to explain why glacial periods produce lower levels of carbon dioxide (it may be
related to how the physical changes influence the Earth's ecosystems ability to absorb carbon dioxide: perhaps lower sea levels
increase the nutrient supply in the ocean, or the drop in sea level destroys coral reefs, or iron-rich dust from new deserts fertilizes
the oceans) but further work on this question remains to be done.
It is a concern for all of us that there are gaps in our understanding of how the feedbacks between insolation, albedo and
greenhouse gases operate, as it makes it hard to predict what the consequences of any changes in the climate system might lead to.
The current level of atmospheric carbon dioxide is unprecedented in human experience; it is at the highest level ever recorded in
the Quaternary. Will the current increase in greenhouse gases lead to a positive feedback, warming the Earth even more?
Review Questions
1. In the text, we discuss how polar ice has a smaller 18O to 16O ratio (that is, it has proportionally less heavy isotope water) than
ocean water does. Hydrogen also has isotopes, the two most common being hydrogen-1 (1H) and hydrogen-2 (2H, also known
as deuterium). Water is made up of both hydrogen and oxygen, and scientists analyze both elements when examining ice cores.
3.3.6 [Link]
Do you predict that polar ice sheets would have a higher ratio or a lower ratio of 1H to 2H than ocean water? Will colder global
temperatures increase or decrease the amount of 2H in polar ice?
2. In the text, we discuss how polar ice has a smaller 18O to 16O ratio (that is, it has proportionally less heavy-isotope water) when
the climate is cooler. We also discuss how changes in the ratio of 18O to 16O ratio in sediment cores can also be used to
determine the climate's average temperature. In ocean sediments, the ratio of 18O to 16O increases when the climate is cooler
(that is, it has proportionally more heavy isotope water). Explain why isotope ratios in ocean sediment have the opposite
reaction to those in polar ice.
3. There are three different ways in which the Earth's orbit changes through time. What combination of orbital parameters would
be most likely to start an ice age? (Hint: Ice ages require cool northern summers.)
Resources
Do you want to know more about how ice cores are extracted and analyzed? NASA's Earth Observatory has details about the
practical issues of drilling ice cores (deep ice needs to "relax" for as long as a year at the surface before being cut open – or it can
shatter!) and how chemical data is interpreted. Go to [Link] for an
in-depth article with great links.
Glossary
axial precession
The movement in the axis of rotation, which change in the direction of Earth's axis of rotation relative to the stars.
axial tilt
The angle between a planet's axis of rotation and the line perpendicular to the plane in which it orbits. The Earth's current axial
tilt is 23.5 degrees.
eccentricity
A measure of how much an ellipse departs from circularity.
glacial period
A long period of time in which ice -sheets and glaciers are advanced in their extent.
ice sheets
Glaciers big enough to cover a continent. Currently, ice sheets are found in Antarctica and Greenland, but during glacial
periods, ice sheets have covered other land masses, including North America.
interglacial period
The warm periods of the Quaternary in which glaciers and ice-sheets retreat. These occur between the longer glacial periods.
isotopes
Atoms that have same number of protons but different numbers of neutrons. This means that they are the same element (e.g.
oxygen), have the same chemical properties, but different masses.
milankovitch cycles
Periodic variations in the Earth's orbit that influence its climate. These cycles are named after Milutin Milankovitch, a
mathematician who quantified the theory.
quaternary period
The most recent geological period, spanning the time from 2.6 million years ago to today.
obliquity
See Axial Tilt.
3.3.7 [Link]
proxy data
Information about the climate that accumulates through natural phenomena.
This page titled 3.3: Milankovitch Cycles and the Climate of the Quaternary is shared under a CC BY license and was authored, remixed, and/or
curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
3.3.8 [Link]
3.4: Modern Climate Change
Learning Objectives
After reading this module, students should be able to
assess long-term global temperature records and place recent climate change into the context of historical temperature
observations
explain how changes in the Sun's energy output have impacted the last 1300 years of global temperature records
analyze the human impact on the planetary albedo and relate these changes to recent climate change
predict the response of the global average temperature when large volcanic eruptions occur
explain the enhanced greenhouse effect
discuss how recent observations of change measured within regional ecosystems are related to global climate change
Introduction
In previous modules, an examination of the geologic record of the earth’s climate in the Quaternary Period revealed the primary
drivers of climate change. The most important conclusions to be drawn from the Modules 5.2 and 5.3 are the following:
1. In the past, Earth has been significantly warmer (and mostly ice free) and significantly colder (especially during the so-called
“Snowball Earth” eras) than it is today.
2. Climate change occurs when there are changes in insolation, albedo, and composition of the atmosphere.
3. Climate is the average of weather, and changes to the earth’s climate occur on long time scales.
Recent climate change, which has occurred during the modern instrument era, is the focus of this module. It is through the lens of
long-term climate change (occurring on thousands to millions of years) that we will view earth’s current climate and recent climate
change. The goal is to investigate how the principles listed above are shaping current climate events.
Mechanisms
Temperature Records
Figure 3.4.1 clearly shows that the current global average temperature reflects an interglacial warm period. If we focus in on the
end of this record we can observe some of the fine scale changes in the global temperature records. Figure 3.4.1 combines proxy
data (i.e., information from ice cores and tree rings) with the modern instrument record to create a graph showing the last 1300
years of Northern Hemisphere (hereafter, NH) temperatures. Each line on the top two panels represents a different temperature data
set collected in the NH and the bottom panel color codes the percentage of overlap among these data sets.
3.4.1 [Link]
Figure 3.4.1 Northern Hemisphere Surface Area. Panel (a) – Northern Hemisphere surface air temperature data from the modern
instrument era from various sources. Panel (b) – Northern Hemisphere surface air temperature reconstruction dating back 1300
years from various sources. Panel (c) - Percent of overlap between the various sources of Panel (b). Source: Climate Change 2007:
The Physical Science Basis: Contribution of Working Group I to the Fourth Assessment Report of the Intergovernmental Panel on
Climate Change, Cambridge University Press
Major features in these data include the medieval warm period approximately 1,000 years ago and the little ice age approximately
400 years ago. Even with these events, the bottom panel shows that most of the variability in the NH temperature fits within a
0.5°C temperature range. Rarely has the temperature exceeded the 1961-1990 average, which is the dividing line on this graph. The
only major fluctuation outside of this range is during the modern instrument era of the last 300 years, where confidence between
the data sets is high. Beginning in the 1800s, the solid black line in each panel traces out approximately a 1°C increase in global
temperatures. It is this increase that is the central focus in recent climate change science. Remember from the previous chapter that
a 1°C change in the earth’s temperature is a large change; reduce the global average by 4°C to 6°C and much of the NH will be
covered with ice as it was 20,000 years ago.
There has been much debate over recent climate change, especially in the news media and among political parties around the
world. This debate is centered on the cause of the recent 1°C increase–is it a part of the natural variability in the climate system or
have anthropogenic, which simply means human caused, influences played a major role? In a recent survey given to more than
3,000 college students at the University of Illinois at Urbana-Champaign, it was found the approximately two thirds of those
surveyed agreed that recent climate change was due to reasons beyond natural variability in the climate system. (see Figure 3.4.2)
Approximately 20% reported that the climate change is due to natural changes and the remainder was undecided. Let’s investigate
both sides of this argument!
3.4.2 [Link]
Figure 3.4.2 Recent climate change student survey results from 3,000+ college students at the University of Illinois at Urbana-
Champaign when asked if climate was changing beyond natural variability. Source: Snodgrass, E.
Recall from the Module 5.3 that global climate will change as a response to changes in insolation, albedo and the composition of
the atmosphere. It was shown that the amount of energy entering the earth-atmosphere system from the sun varies less than 0.1%
during the 11-year solar cycle in sunspot activity. Outside of this cycle, the amount of energy from the sun has increased 0.12 watts
per square meter (W/m2) since 1750. Is this enough excess energy to produce the 1°C increase in global temperatures that has been
observed since the 1800s? As it turns out, the climate system needs nearly 8 times that amount of energy to warm by 1°C. This
essentially eliminates fluctuations in solar output as the culprit for recent climate change.
Has the earth’s albedo changed since the 1800s? As we know from the Module 5.2, increases in the Earth’s albedo lead to global
cooling and decreases lead to warming. The net effect of human existence on Earth is to brighten the surface and increase the
global albedo. This change is primarily accomplished through intensive agriculture where forest, marshland, and open prairie are
cut down and crops like soybeans, corn, wheat, cotton, and rice are grown in their place. Add this to the current high rates of
deforestation in South America and Africa and the evidence is clear that mankind has increased the Earth’s albedo, which should
have led to global cooling. (see Figure 3.4.3)
Figure 3.4.3 Deforestation in the Amazon (2010) Satellite image shows the extent of deforestation in the Amazon as of 2010.
Source: NASA Earth Observatory
Outside of human influence, planetary albedo can also be changed by major volcanic eruptions. When volcanoes erupt, they spew
enormous amounts of soot, ash, dust, sulfur, and other aerosols into the atmosphere. During major eruptions, like that of Mt.
Pinatubo in 1991, some particles of this debris find their way into the stratosphere, where they reside for a few years. (see Figure
3.4.4) The presence of these particles high in the earth’s atmosphere acts like a shield that prevents sunlight from penetrating
through the lower atmosphere to warm the earth’s surface. Instead, the energy is either absorbed by the particles or reflected and
scattered away. The net effect is that large volcanic eruptions can cool the planet for a few years by changing the earth’s albedo.
3.4.3 [Link]
Figure 3.4.4 Mt. Pinatubo Erupting in 1991. Photograph of Mt. Pinatubo erupting in the Philippines in 1991. Source:
USGS/Cascades Volcano Observatory
3.4.4 [Link]
Source: Climate Change 2007: The Physical Science Basis: Contribution of Working Group I to the Fourth Assessment Report of
the Intergovernmental Panel on Climate Change, Cambridge University Press
Figure 3.4.6 Annual Global Temperature Anomalies. Global average surface temperature from 1880 to 2007. Source: National
Climate Data Center
Long before the term “global warming” became a common household phrase, nineteenth-century Irish physicist John Tyndall said,
“Remove for a single summer-night the aqueous vapor from the air which overspreads this country, and you would assuredly
destroy every plant capable of being destroyed by a freezing temperature.” This now famous quote reveals the importance of
greenhouse gases, like water vapor, in maintaining a balance between the incident solar radiation and the emitted terrestrial
radiation. Tyndall understood that without greenhouse gases, water vapor being the most abundant, the earth’s temperature would
be markedly cooler. The global average surface temperature is approximately 15°C (59°F) but if the greenhouse gases were
removed, the average global temperature would plummet to -18°C (0°F). Remember that these gases make up a small fraction of
the composition of the atmosphere! Therefore, adjustments to their concentration will produce dramatic effects.
To understand why these gases are so efficient at keeping the planet warm, let’s examine Figure 3.4.7. The top panel of this figure
shows the normalized intensity of the radiation emitted by both the sun and earth as a function of wavelength. The middle panel
shows the total atmospheric absorption spectrum and the bottom panel shows the individual gas absorption spectrum (excluding
Nitrogen and Argon). Notice from the top panel that the sun’s peak energy emission falls within the visible portion of the spectrum
and suffers very little atmospheric absorption (middle panel). The peak emission wavelength for the earth is in the thermal infrared
(IR), and it is effectively absorbed by water vapor (H20), carbon dioxide (CO2), methane (CH4) and nitrous oxide (N02). The
primary purpose of this figure is to show that the gases in the earth’s atmosphere are transparent to the sun’s peak energy emission
(visible light) but not the earth’s peak emission (thermal IR). It is through the absorption of the earth’s outgoing thermal infrared
radiation that the global average temperature warms approximately 60°F over what it would be without greenhouse gases.
3.4.5 [Link]
Figure 3.4.7 Atmospheric Transition. top graph – normalized spectral intensity (radiant energy) emitted by the earth and sun as a
function of wavelength. Middle graph – total atmospheric absorption as a function of wavelength. Bottom graph – individual gas
absorption as a function of wavelength. Source: R.A. Rohde for Global Warming Art Project
Are humans altering the natural greenhouse effect? Based upon our assessment so far, this is the final mechanism by which the
global climate can be changed. Let’s look into the alteration of the chemistry and composition of the earth’s atmosphere. First are
humans increasing the amount of water vapor, the most abundant but also weakest greenhouse gas in the atmosphere? As the air
temperature increases, the amount of water vapor the atmosphere can hold also increases. However, a closer investigation of the
water cycle is needed to understand what will happen to this increase in water vapor. In this cycle, the amount of evaporation must
equal the amount of condensation and thus precipitation on a global scale. This equilibrium must be achieved or else water would
end up entirely in its liquid form or in its vapor form. Also due to the speed at which the hydrological cycle operates, a large
increase in water vapor would be quickly precipitated out of the atmosphere.
Other greenhouse gases progress through their respective cycles much more slowly than water. There are vast amounts of carbon
and carbon dioxide in the earth-atmosphere system. Most carbon is locked up in rocks, where it may remain for millions of years.
The carbon dioxide that is mobile, however, is mostly found in other places: the ocean, soils, vegetation, fossil fuels like coal, oil,
and natural gas, and also in small concentrations in the atmosphere. These reservoirs of CO2 can exchange mass like oceans and
clouds do in the water cycle, but with one extremely important difference–the exchange rate is much slower. That means the system
can get out of balance and remain out of balance for a long time, hundreds or thousands of years. There are two primary
mechanisms for sequestering carbon dioxide that is released into the atmosphere: it can be captured by the respiration of plants, or
dissolved in the ocean.
However, the rate at which plants and oceans can take CO2 out of the atmosphere is fixed. Therefore, if a surplus of CO2 is added
to the atmosphere, it will stay there for a long time. This has major implications, given the fact that CO2 is a powerful greenhouse
gas. The question then to ask becomes, “is this exchange rate out of balance?”
The current average concentration of CO2 in the atmosphere is about 390 parts per million (PPM), which means there are 390 parts
of CO2 per million parts of air. That does not seem like very much, but if that small amount of carbon dioxide were removed from
the air, the global average temperature would plummet. Has this concentration been changing? To answer the question, we will turn
to the findings of Richard Keeling, whose life’s work was the observation of CO2 concentrations at the Mauna Loa Observatory in
Hawaii. Beginning in the early 1950s, observations of CO2, a well mixed gas in our atmosphere, have shown a remarkable climb in
concentration. (see Figure 3.4.8) The “Keeling Curve,” as it is sometimes called, clearly shows that since the 1950s CO2
concentrations have increased steadily from 315 ppm to 390 ppm. The zigzag nature of this graph is due to life cycle of plants in
the NH. The NH has much more land area that the SH, so when spring and summer arrive in the NH, the abundance of new plant
life reduces the CO2 concentrations in the atmosphere. When the plants die or become dormant in the fall and winter, CO2
concentrations spike again.
3.4.6 [Link]
Figure 3.4.8 CO2 Concentrations at the Mauna Loa Observatory. The “Keeling Curve” of CO2 concentrations measured in Mauna
Loa, Hawaii, since the 1950s. Source: NASA Earth Observatory
What is troublesome about this figure is that the carbon cycle is out of its normal rhythm and a surplus of CO2, a known
greenhouse gas, is building in the earth’s atmosphere. Where is this surplus coming from? To answer this question, let’s look at two
historical records of CO2 concentrations taken from ice core deposits. The top panel in Figure 3.4.9 shows the past 10,000 years of
atmospheric CO2 concentrations. Before 1750, the amount of CO2 in the atmosphere was relatively steady at 280 ppm. Since 1750
there has been a dramatic increase in CO2concentrations.
Figure 3.4.9 Changes in Greenhouse Gases from Ice Core and Modern Data. Top panel shoes CO2concentrations (ppm) over the
last 10,000 years. Source: Climate Change 2007: The Physical Science Basis: Contribution of Working Group I to the Fourth
Assessment Report of the Intergovernmental Panel on Climate Change
If we look even further back in time, over the last half million years, we see a similar story. (see Figure 3.4.10) The current
concentration of CO2 in the earth’s atmosphere is higher than at any time in the past half million years. Where is this abundance of
3.4.7 [Link]
CO2 coming from? Which reservoirs are being depleted of their CO2 while the atmosphere takes on more? The answer lies in the
burning of fossil fuels and in the deforestation of significant chunks of the earth’s forest biomes. Notice the spike in CO2
concentrations beginning around 1750. This time period marks the beginning of the industrial revolution, when fossil fuels
overtook wood as the primary energy source on our planet. Over the subsequent two and a half centuries, oil, coal, and natural gas
have been extracted from their underground reservoirs and burned to generate electricity and power modern forms of
transportation. The exhaust from this process is currently adding 30 billions of tons, or gigatons (Gt), of carbon dioxide to the
atmosphere each year. Combine this addition of CO2, a known greenhouse gas, to the subtraction of one of the sinks of CO2
through deforestation and the imbalance grows even further.
Figure 3.4.10 Evidence of climate change. CO2 concentrations over the last 400,000+ years. Source: NASA/NOAA
What is the end result? By examining the earth’s climate, both current and past and by investigating the three ways in which
climate can change, we have arrived at the conclusion that the current warming is being caused by an imbalance in the carbon cycle
that has been induced by human activity, namely the burning of fossil fuels. The record warmth over the last 1,300 years is very
likely to have been caused by human decisions that have lead to a change in the chemistry of the atmosphere, and which has altered
the natural climate variability toward warmer global temperatures. We are essentially changing the climate faster and in a different
direction than natural processes have intended.
3.4.8 [Link]
Figure 3.4.11 Cherry blossoms.
Figure 3.4.12 Fire Ants. Photograph of fire ants on a piece of wood. Source: Scott Bauer of the Agricultural Research Service, U.S.
Department of Agriculture via Wikimedia Commons
3.4.9 [Link]
Figure 3.4.13 Mosquitoes. Photograph of a mosquito on skin. Source: Centers for Disease Control and Prevention
3.4.10 [Link]
Figure 3.4.15 2007 Sea Oce Extent in the Artic Source: NASA Goddard Space Flight Center
Figure 3.4.16 Coral Bleaching. A part of coral that has experienced coral bleaching. Source: NOAA
Finally, as the planet has adjusted to warmer temperatures the proliferation of drought conditions in some regions has dramatically
affected human populations. The Sahel, for example, is a border region between the Sahara Desert in the north of Africa and the
tropical rainforests that occupy the central part of the continent. (see Figure 3.4.17) This region is experiencing desertification as
the Sahara steadily expands southward. Since the 1970s, the amount of precipitation in this region has been steadily below normal.
The combination of over irrigation and recent climate change has made the region uninhabitable and forced millions to relocate.
3.4.11 [Link]
Figure 3.4.17 The Sahel in Africa Source: NASA Earth Observatory
Review Questions
1. In Figure 3.4.1 the dividing line on the graph is the 1961-1990 average temperature. Explain the relevance of this line to the
data presented in this figure.
2. Explain how deforestation can lead to both a warming effect and cooling effect for global temperatures.
3. In Figure 3.4.7, which gas is contributing the most to the absorption of ultra-violet light? If this gas were removed from the
atmosphere, how might global temperatures respond?
4. If the surface of the Greenland Ice Sheet continues to melt, how will this impact the albedo of this region and what impact will
this have on the air temperature there?
5. When sea ice melts, what happens to global sea level?
References
Walther, G. R., Post, E., Convey, P., Menzel, A., Parmesan, C., Beebee, T. J .C., et al. (2002, March 28). Ecological responses to
recent climate change. Nature, 416, 389-395. doi: 10.1038/416389a
Glossary
anthropogenic
Caused or produced by humans.
hydrological cycle
The continuous movement of water on, above and below the surface of the earth. This cycle is dominated by the global
equilibrium in evaporation and condensation.
northwest passage
A sea route for commerce through the Arctic Ocean north of Canada.
permafrost
Soil that has a temperature that has remained below freezing (0°C or 32°F) for at least two years.
3.4.12 [Link]
radiative forcing
Change{a in net irradiance (an energy flux) measured at some boundary. For this text the boundary is typically at the surface of
the earth or the top of atmosphere. A positive change indicates warming and a negative change indicates cooling.
well-mixed gas
A gas that can be found at the same concentration throughout the lower atmosphere regardless of location.
This page titled 3.4: Modern Climate Change is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto Cabezas
(GALILEO Open Learning Materials) .
3.4.13 [Link]
3.5: Climate Projections
Learning Objectives
After reading this module, students should be able to
assess global CO2 emissions and determine which countries and regions are responsible for the greatest emissions currently
and historically
explain the relationship between fossil fuel usage and CO2 emissions
link variables such as wealth, population, fuel imports, and deforestation to CO2 emissions
use IPCC future climate projections to assess future global temperature scenarios
distinguish between weather events and climate change, and discuss the differences between weather forecasting and
climate projections
analyze the anthropogenic impact on climate by examining climate change without people
assess the regional and global impacts of climate change on air temperature and precipitation
Introduction
In the Module 5.4 we discovered that the global warming of approximately 1°C over the past 200 years was human induced
through an enhancement of the natural greenhouse effect. We learned that the burning of fossil fuels has upset the natural carbon
cycle, which has steadily increased the amount of carbon dioxide (CO2) in the atmosphere since the 1750s. Finally we looked at
ancillary evidence of this warming to see the immediate impact of these changes. In this module we will investigate the findings of
the Intergovernmental Panel on Climate Change (IPCC) and look at future climate projections. We will inspect these findings and
analyze their impacts on a global scale.
3.5.1 [Link]
Figure 3.5.1 CO2 emissions for the United States and China. CO2 emission in millions of metric tons graphed against time for the
United States and China. Source: Snodgradd, E. created graphs using data from the US Energy Information Association.
Rather than point the finger at individual countries, let’s examine the bigger problem. The maps in Figure 3.5.2 distort the size of
each country based on a certain variable, like CO2 emissions, with respect to the rest of the world. In the upper left panel, the map
is based on population, which is why China and India appear so large. The upper right map distorts the size of the country based
upon fuel imports. Notice that the United States, much of Europe, and Japan are expanded the most, while Africa, the Middle East,
and much of South America are barely visible. Compare these two maps with absolute wealth and carbon emissions and the story is
quite clear. The industrialized and wealthy nations are responsible for the largest quantities of carbon emissions and fuel imports.
These societies are built on the foundation of energy production through the consumption of fossil fuels.
The bottom two panels tell another aspect of this story. Focus first on the graph in the lower right, which shows forest loss by
country. The world’s forest biomes are a large part of the CO2 cycle and with deforestation, a large sink for atmospheric CO2 is
taken away. Notice that deforestation is most prevalent in Africa, South America, and Indonesia while the United States is barely
visible on this map. In the United States, reforestation is practiced, but in the rainforests of the world, which are those areas in
South America, Africa, and Indonesia that are ballooned on this map, deforestation is commonplace.
Figure 3.5.2 Global Influence Maps. The variable labeled on each map are used to distort the size of each country to show their
global influence. Source: Worldmapper. Copyright SASI Group (University of Sheffield) and Mark Newman (University of
Michigan)
The last graph in Figure 3.5.2 distorts each country’s size according to poverty. Much of Asia and Africa are distorted the most,
and it is in these regions that we need to pay close attention over the upcoming years. Many of the nations found within these
countries are what economists and politicians call “emerging economies.” Although much of the current abundance of CO2 in the
atmosphere is from developed countries such as the United States, CO2 emissions from these countries are not increasing with time
according to a 2008 report from the Energy Information Administration. In Figure 3.5.3, the world’s CO2 emissions from coal
combustion in billions of metric tons are plotted against time. Notice that countries of the Organization for Economic Co-operation
and Development (OECD), which comprises a large group of developed and industrialized nations, have not increased their CO2
3.5.2 [Link]
emissions from coal combustion since 1990, and future projections also reveal a flat line in CO2 emissions. Compare this to the
non-OECD countries, many of which are emerging economies like China and India, and you see that CO2 emissions are set to triple
in the next 25 years. There is much debate over information like this now that recent climate change has been linked so closely to
anthropogenic emission of CO2. This debate revolves around the fact that developed nations used coal, oil, and natural gas during a
time when the impacts of CO2 and climate change were not well researched. This meant that during the time these countries,
including the United States, industrialized there were no regulations on the emissions of CO2. Now that CO2 emissions have been
shown to cause global warming, pressure is being applied to these emerging economies to regulate and control their CO2emissions.
This is subject of much of the debate at the international climate summits at Kyoto, Bali, and Copenhagen. What is important to
remember when discussing developed countries vs. emerging economies is that the per capita emissions of CO2 in emerging
economies are approximately one third of those for developed countries.
Figure 3.5.3 Global CO2 Emissions from Coal Combustion The world's CO2 emissions from coal combustion in billions of metric
tons are plotted against time for OECD countries and non-OECD countries. Source: US Energy Information Administration (Oct
2008)
Climate Projections
One of the greatest obstacles climate scientists face in educating the public on issues of climate change is time. Most people take
weather and climate to be one branch of scientific study. In reality, this could not be further from the truth. Weather and climate are
two separate fields of study that are joined only by one definition—climate is the average of weather. It is important to understand
this statement because people—news reporters, broadcast meteorologists, politicians, and even scientists—often make the mistake
of attributing weather events, such as Hurricane Katrina (2005), to global climate change. Katrina was a weather event and, as
such, it cannot be said to have been caused by global climate change. Are the ingredients for stronger hurricanes present in greater
frequency now than they have been in the past? That’s the type of question climate scientists seek to answer, and they do so by
analyzing decades worth of data. Thirty years is the lowest value used in the denominator of climate calculations. In other words,
30 years is the shortest time period over which weather can be averaged to extract climate information. Therefore, it is impossible
to blame a single weather event on climate change—this branch of science does not work that way.
To better understand the differences between weather and climate, take a look at Figure 3.5.4 which shows in red the actual high
temperatures for each day in 2005 in Champaign, Illinois, compared to the average high temperature in black over a period
beginning in 1899 and ending in 2009. It is completely normal for the temperature to vary ±20°F around this average. In 2005 there
were only a handful of days where the actual high was the same as the average high. This graph shows the highly variable and
chaotic behavior of weather. But, when data from a long span of time is averaged, the climatological mean emerges.
3.5.3 [Link]
Figure 3.5.4 High Temperature vs Low Temperature, Champaign, IL. The average high temperature for Champaign- Urbana
Illinois in black (1899-2009). The 2005 actual high temperature are graphed in red. Source: E. Snodgass using data from the
National Climate Data Center.
To think of it another way, imagine you are in a large lecture hall with over 300 college students. If the professor were to call out
one student and try to predict the course of her life over the next 70 years it would be nearly impossible! It would even be difficult
to predict when that person would eat her next meal. However, the professor could project with great confidence that on average,
most of the people in the room will eat dinner at 6:30PM on a given night. Beyond this meal, most of them will graduate from
college by the time they are 22 years old. Many will be married by 27 years old and have their first children at 30. Most will have a
job by the time they are 24 and most will have a job they like by 34. Most will have a total of 2.1 children by the time they are 36,
and by the time they are 50 most will have gone to the doctor to have their first routine procedure. Most will retire at 67, and since
they are college grads in the United States, there is a safe bet that they will retire with over a million dollars in assets. On average,
the men in the room will die at 85 years old and most of the women will die before their ninetieth birthday. Now, if the professor
were to single out one individual, the chances that her life would follow this path exactly are small, but when an entire class of 300
is averaged, this is the result. Weather is like the individual life. Climate is like the average of 300 lives. Weather and Climate are
two separate branches of study in atmospheric science.
In addition to keeping in mind the difference between weather and climate, remember that the focus of this chapter is global
climate change. It is tempting to forget the global nature of this problem because it is happening very slowly on a large scale and it
is averaged over long time periods. Recall the differences between weather and climate and remember that in conjunction with
global warming there can still be weather events that take temperatures far below normal. The temptation during these events is to
discount the science behind global climate change. For example, during the winter of 2009-2010, the weather patterns were such
that the east coast of the United States experienced repeated record-setting snowstorms and cold air outbreaks. Many television
news reports, weather broadcasts, and newspaper headlines scoffed at the idea of global warming during this winter and proclaimed
that it was not happening or that it was a hoax. The shortsightedness of such responses is evidenced by the fact that globally, 2009
and 2010 were among the warmest years during the instrument record: 2009 ranked seventh, and 2010 tied for first. These were
likely two of the warmest years of the last 1,300.
3.5.4 [Link]
Figure 3.5.5 Petascale Computing Facility. The petascale computing facility "Blue Waters" located at the University of Illinois at
Urbana-Champaign. Source: HorsePunchKid via Wikimedia Commons.
What are these climate models predicting will happen by the year 2100? First, we will look at the global average surface
temperature projections. Figure 3.5.6 plots global surface warming against time with the present day in the middle of this chart.
Recall that over the last 200 years, there has been a 1°C increase in global temperatures, and that the rate of change has been
extremely fast compared to natural changes in the earth’s climate. The graphs in Figure 3.5.6 show the range of model projections
from different climate simulation scenarios based upon various greenhouse gas emission scenarios (left graph). Focus on the top
and bottom curves in the right panel, which show the most dramatic warming and the most conservative warming. The worst-case
scenario, found in the top line, shows the “business as usual” projections. If nothing is done to mitigate the emission of greenhouse
gases into the atmosphere, these climate models are predicting a 4°C to 6°C increase in global average temperature by 2100. The
best-case scenario, from a climate change perspective, would be for a cessation of CO2 emissions or for the current emission rates
to not increase. In this case, there would still be a warming of 0.5° to 2°C by 2100 as indicated by the bottom curves.
Figure 3.5.6 Climate Simulation Scenarios. Left- multiple climate model projections (or scenarios) of greenhouse gas emissions
(including CO2, CH4, and N2O) emissions in Gt-CO2 equivalent through 2100. Right- Multiple climate model projections of
globally averaged surface air temperature through 2100. Source: Climate Change 2007: Synthesis Report, Contribution of Working
Groups I, II, and III to the Fourth Assessment Report of the Intergovernmental Panel on Climate Change,
In addition to predicting warming of the atmosphere, climate models also suggest that sea level will continue to rise. Since 1880,
sea level has risen 20 cm (approximately 8 inches) as seen in Figure 3.5.7. This rise has been primarily the result of the water
thermally expanding as it warms along with the atmosphere. Polar ice cap melt from land-based ice sheets and glaciers has also
added to increase in sea level. The current projection is that sea level will rise at the rate of at least 2 mm per year over the next
century, with an overall increase ranging from 15 to 60 inches cm.
3.5.5 [Link]
Figure 5.13, pg 410.
How much confidence can we place in predictions about temperature and sea level by climate scientists? Let’s take a little detour
before we address this important question directly. Imagine you are contemplating signing up with a psychic so you can better plan
for the future—why save for retirement if money is tight and you’re not sure how long you’ll live? But you are uncertain about
whether she can really see what lies ahead. You could pay her $20 a week for her predictions, and discover over time whether they
come true or not. The trouble is, during this trial period you wouldn’t know whether to spend your money as fast as you make it or
put some aside. But you come up with a better plan. You’ll pay the psychic $20 one time, but instead of asking her to predict your
future, you’ll ask her to tell what has happened to you in the past week. If she gets that right, she gets your business.
Along similar lines, climate scientists assess the trustworthiness of their models by checking how well they “predict” the past. In
Figure 3.5.8, 58 different climate model simulations were tasked with predicting the past climate from 1900 to 2005. By comparing
the model simulations to the observed temperature record the scientists with the IPCC tested the accuracy of their models. In Figure
3.5.8, the yellow lines in the top panel trace out the individual model simulations, the red line shows the model ensemble mean,
and the black line represents the actual observed mean temperature. The models performed exceedingly well, as evidenced by the
very small variability around the observed temperature. The success of this test demonstrates the high-quality construction of these
models and shows they are capable of accurately projecting the earth’s future climate.
Figure 3.5.8 Model Simulations. Top panel - Climate model simulations of the global mean surface temperature compared to the
observed global mean surface temperature in black. Each yellow line is one of 58 climate model simulations of which the red line
is the ensemble mean. Bottom panel - 19 climate model simulations in blue with the ensemble mean in dark blue. These
simulations were run without anthropogenic influences. The thick black line is the observed global mean surface temperature.
Source: Climate Change 2007: The Physical Science Basis: Contribution of Working Group I to the Fourth Assessment Report of
the Intergovernmental Panel on Climate Change, Cambridge University Press
The bottom of Figure 3.5.8 and Figure 3.5.9 presents the most compelling argument that current climate change is caused in large
part by humans. The bottom panel of Figure 3.5.8 shows 19 climate model simulations between 1900 and 2000 with human
influences left out of the simulations. The thick black line represents the observed global mean surface temperature over this time.
Compare this figure with that of Figure 3.5.9, which depicts a series of graphs that plot temperature anomalies against time from
3.5.6 [Link]
the early 1900s to 2000. The blue color shading on these graphs shows the computer model projections without anthropogenic
effects, while the pink shading includes them. The black line represents the actual measured air temperatures in each of the
locations over which the inlaid graphs are positioned. Notice that without humans the blue shading stays level or decreases with
time. Compared with the pink shading and the black line, which both increase with time, and we find that these climate simulations
cannot accurately represent the past climate without anthropogenic effects. Simply put, these models are unable to represent our
current climate without greenhouse contributions from humans. Rigorous testing like this proves these models are robust and well-
designed to simulate future climate conditions.
Figure 3.5.10 Observed vs Projected Temperatures. Observed global average surface temperatures (black line) overlaid on the
temperature projections of the first IPCC report (FAR), second report (SAR) and third report (TAR).
3.5.7 [Link]
occur at high northerly latitudes. If these projections hold true, ice and snow cover will continue to retreat and enhance the ice-
albedo effect discussed in Module 5.2. Since the 1980s, NH snow-covered area has shrunk by 3 million square kilometers, and
many northerly lakes are spending less time each year covered in ice.
Figure 3.5.11 Projected Temperature Increases. IPCC projected temperature increases for the years 2020-2029.
Aside from air temperature, global precipitation patterns and amounts are expected to change. As the atmosphere warms, its ability
to hold water vapor increases, which leads to more evaporation from water on the earth’s surface. As this water condenses in the
earth’s atmosphere to form clouds and precipitation, the distribution of the precipitation will vary greatly. Current projections
forecast an increase in precipitation in the tropics and polar latitudes, with drier conditions over the mid-latitudes. Even though
there will be more water vapor in the atmosphere, the distribution of precipitation may be such that large regions formerly unused
to drought may be subjected to prolonged dry periods. Focus on the middle panels of Figure 3.5.12, which shows the winter (top)
and summer (bottom) precipitation anomalies. Notice that the tropics and polar regions are expected to have above normal
precipitation, while the mid-latitudes have below normal precipitation. Although more areas are expected to experience prolonged
drought, these projections suggest that when it does rain, rainfall will arrive in much greater amounts over shorter time periods.
This will lead to increased flash flooding, the deadliest weather phenomenon in the United States.
Figure 3.5.12 Winter and Summer Precipitation Anomalies. Global temperature and precipitation projections for 2080-2099 using
the A1B scenario. Top panels are temperature (left), precipitation (middle) and sea level pressure (right) for December-January-
February. Bottom panels shows the same variables for June-July-August.
The goal of climate science is not to craft public policy on global warming. It is to provide the public and policymakers alike with
reasonable projections about future climate conditions. This information should be used to show the potential impacts of our
presence on the climate system so as to form the best possible mitigation plans. Current projections show that if we are able to slow
greenhouse gas emissions, the climate system will respond with the least amount of warming. They also suggest that if we continue
with "business as usual" the change in the global climate will be great in magnitude and occur very quickly—both beyond past
"natural" change.
Review Questions
1. How much CO2 does the average world citizen release each year into the atmosphere? Assume a population of 7 billion people.
Compare this number to the United States, China and Qatar.
2. Explain why the April 2011 tornado outbreak, which set the record for the most tornadoes in a singe 24-hour period, cannot be
blamed on climate change.
3.5.8 [Link]
3. In Illinois, during the summer of 2009 only two days topped 90°F. In total it was the seventh coolest summer on record. Does
this disprove climate change? In what context should we view this cold summer in Illinois?
4. Why will there still be global warming if there is a complete cessation of CO2 emissions?
5. Carbon cap and trade is one of many solutions proposed to reduce CO2 emissions. Make a list of pros and cons to a federally
mandated cap and trade system. Be sure to consider what will happen to consumers, businesses and the federal government.
Resources
For further reading on global climate change, read A Rough Guide to Climate Change: The Symptoms, The Science, The Solutions,
by Robert Henson (Penguin, 2011, ISBN-13: 978-1843537113)
For more information about the:
U.S. Global Change Research Program, visit [Link]
Global temperatures in the year 2010, visit [Link]
year/
This page titled 3.5: Climate Projections is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto Cabezas
(GALILEO Open Learning Materials) .
3.5.9 [Link]
CHAPTER OVERVIEW
4: Biosphere
4.1: Prelude to the Biosphere
4.2: Biogeochemical Cycles and the Flow of Energy in the Earth System
4.3: Biodiversity, Species Loss, and Ecosystem Function
4.4: Soil and Sustainability
This page titled 4: Biosphere is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto Cabezas (GALILEO Open
Learning Materials) .
1
4.1: Prelude to the Biosphere
Introduction
Humanity and the natural world are inextricably linked. A growing appreciation for the importance of this fact led to the formation
and publication of the Millennium Ecosystem Assessment by the United Nations in 2005. It defines key concepts necessary for
understanding how sustainable development can be achieved. In the terms of the Assessment, an ecosystem is a dynamic complex
of plant, animal, and microorganism communities and the nonliving environment interacting as a functional unit, while ecosystem
services are “the benefits people obtain from ecosystems.” Ecosystem services are critical to human well-being and sufficiently
diverse and numerous to justify classification into four major categories (Figure 4.1.1). Provisioning ecosystem services are
actively harvested by us from the natural world to meet our resource needs, e.g. food, water, timber, and fiber. Regulating
ecosystem processes are processes in the Earth system that control key physical and biological elements of our environment, e.g.
climate regulation, flood regulation, disease regulation, water purification. Cultural ecosystem services reflect the aesthetic and
spiritual values we place on nature, as well as the educational and recreational activities dependent on ecosystems. Finally,
supporting ecosystem services are the biogeochemical cycles, as well as biological and physical processes that drive ecosystem
function, e.g. soil formation, nutrient cycling, and photosynthesis.
Figure shows the linkages between ecosystem services and human well-being. Source: Millennium Ecosystem Assessment, 2005.
Ecosystems and Human Well-being: Synthesis. Island Press, Washington, DC.
We benefit from the services associated with both pristine, natural ecosystems, such as tropical rain forests or arctic tundra, and
highly managed ecosystems, such as crop fields or urban landscapes. In all cases, ecosystems contribute to human well-being by
influencing the attainability of basic material needs (e.g. food and shelter), health (e.g. clean air and water), good social relations
and security (i.e. sufficient resources to avoid conflict, tolerate natural and man-made disasters, provide for children, and maintain
social cohesion), as well as freedom of choice and action (an inherent component of the other elements of well-being is the right to
live as one chooses). Linkages between some ecosystem services and human well-being vary in strength depending on socio-
economic status (see Figure 4.1.1). For example, many people in developed countries can always afford to buy imported food
without dependence on the yields of locally grown crops, thereby avoiding shortages when yields are low because of bad weather.
However, in other cases our ability to control the impact of losing an ecosystem service on human well-being is limited. For
example, despite major engineering efforts flooding still causes considerable human and economic damage in developed countries.
The challenge of sustainable development stems from the need to benefit from and manage ecosystem services without causing
damage to the ecosystems and Earth system that will reduce their value in the longer term. People have long recognized that some
ways of using natural resources are unsustainable, especially where ecosystems are rapidly exploited to the maximum extent
possible and further access to the ecosystem services can be achieved only by moving on to previously unexploited areas, as in the
case of slash and burn agriculture. Only more recently have we come to appreciate that human activity is altering global-scale
phenomena, such as climate regulation, and this understanding raises a host of difficult questions. That is because the benefit of an
ecosystem service may be realized by people in one locale, while the costs (in the form of negative environmental consequences)
are imposed on people who live elsewhere, and who may be less equipped to withstand them.
4.1.1 [Link]
The following sections discuss: (1) the natural biogeochemical cycling of carbon, water and nitrogen, the ecosystem services we
derive from these biogeochemical cycles and human activities that are disturbing them; (2) species extinctions and ecosystem
changes being caused by human activity; and (3) soil, how it is formed, its value to society, and practices that diminish or degrade
it.
Glossary
Cultural Ecosystem Services
The aesthetic and spiritual values we place on nature as well as the educational and recreational activities dependent on
ecosystems.
Ecosystem
A dynamic complex of plant, animal, and microorganism communities and the nonliving environment interacting as a
functional unit.
Ecosystem Services
The benefits people obtain from ecosystems.
This page titled 4.1: Prelude to the Biosphere is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto Cabezas
(GALILEO Open Learning Materials) .
4.1.2 [Link]
4.2: Biogeochemical Cycles and the Flow of Energy in the Earth System
Learning Objectives
After reading this module, students should be able to
explain the concept of a biogeochemical cycle, incorporating the terms "pool" and "flux"
describe the natural cycles of carbon, water, and nitrogen
name some of the important ways human activity disrupts those cycles
Introduction
If people are to live sustainably, they will need to understand the processes that control the availability and stability of the
ecosystem services on which their well-being depends. Chief among these processes are the biogeochemical cycles that
describehow chemical elements (e.g. nitrogen, carbon) or molecules (e.g. water) are transformed and stored by both physical and
biological components of the Earth system. Storage occurs in pools, which are amounts of material that share some common
characteristic and are relatively uniform in nature, e.g. the pool of carbon found as carbon dioxide (CO2) in the atmosphere.
Transformations or flows of materials from one pool to another in the cycle are described as fluxes, for example, the movement of
water from the soil to the atmosphere resulting from evaporation is a flux. Physical components of the earth system are nonliving
factors such as rocks, minerals, water, climate, air, and energy. Biological components of the earth system include all living
organisms, e.g. plants, animals and microbes. Both the physical and biological components of the earth system have varied over
geological time. Some landmark changes include the colonization of the land by plants (~400 million years ago), the evolution of
mammals (~200 million years ago), the evolution of modern humans (~200 thousand years ago) and the end of the last ice age (~10
thousand years ago). The earth system and its biogeochemical cycles were relatively stable from the end of the last ice age until the
Industrial Revolution of the eighteenth and nineteenth centuries initiated a significant and ongoing rise in human population and
activity. Today, anthropogenic (human) activities are altering all major ecosystems and the biogeochemical cycles they drive. Many
chemical elements and molecules are critical to life on earth, but the biogeochemical cycling of carbon, water, and nitrogen are
most critical to human well-being and the natural world.
4.2.1 [Link]
Figure 4.2.1 The carbon cycle. This illustrates the carbon cycle on, above, and below the Earth's surface. Source: U.S. Department
of Energy Genomic Science Program.
4.2.2 [Link]
Precipitation is the conversion of atmospheric water from vapor into liquid (rain) or solid forms (snow, hail) that then fall to Earth's
surface. Some water from precipitation moves over the land surface by surface runoff and streamflow, while other water from
precipitation infiltrates the soil and moves below the surface as groundwater discharge. Water vapor in the atmosphere is
commonly moved away from the source of evaporation by wind and the movement of air masses. Consequently, most water falling
as precipitation comes from a source of evaporation that is located upwind. Nonetheless, local sources of evaporation can
contribute as much as 25-33% of water in precipitation.
Figure 4.2.2 The Water Cycle. This figure illustrates the water cycle on, above, and below the Earth's surface. Source: U.S.
Department of the Interior and U.S. Geological Survey, The Water Cycle.
4.2.3 [Link]
is significant because nitrogen is an essential component of all cells—for instance, in protein, RNA, and DNA—and nitrogen
availability frequently limits the productivity of crops and natural vegetation. Atmospheric nitrogen is made available to plants in
two ways. Certain microbes are capable of biological nitration fixation, whereby N2 is converted into ammonium, a form of
nitrogen that plants can access. Many of these microbes have formed symbiotic relationships with plants—they live within the
plant tissue and use carbon supplied by the plant as an energy source, and in return they share ammonia produced by nitrogen
fixation. Well-known examples of plants that do this are peas and beans. Some microbes that live in the soil are also capable of
nitrogen fixation, but many are found in a zone very close to roots, where significant carbon sources are released from the plant.
Together these biological nitrogen fixing processes on land, coupled with others that take place at sea, generate an annual flux out
of the atmosphere of approximately 200 MtN (megatonnnes of nitrogen or 200,000,000 tonnes of nitrogen). Lightning causes
nitrogen and oxygen in the atmosphere to react and produce nitrous oxides that fall or are washed out of the atmosphere by rain and
into the soil, but the is flux is much smaller (30 MtN per year at most) than biological nitrogen fixation.
While the inputs of nitrogen from the atmosphere to the biosphere are important, the majority (90%) of nitrogen used by plants for
growth each year comes from ammonification of organic material. Organic material is matter that comes from once-living
organisms. Ammonification (or mineralization) is the release of ammonia by decomposers (bacteria and fungi) when they break
down the complex nitrogen compounds in organic material. Plants are able to absorb (assimilate) this ammonia, as well as nitrates,
which are made available by bacterial nitrification. The cycle of nitrogen incorporation in growing plant tissues and nitrogen
release by bacteria from decomposing plant tissues is the dominant feature of the nitrogen cycle and occurs very efficiently.
Nitrogen can be lost from the system in three main ways. First, denitrifying bacteria convert nitrates to nitrous oxide or N2 gases
that are released back to the atmosphere. Denitrification occurs when the bacteria grow under oxygen-depleted conditions, and is
therefore favored by wet and waterlogged soils. Denitrification rates almost match biological nitrogen fixation rates, with wetlands
making the greatest contribution. Second, nitrates are washed out of soil in drainage water (leaching) and into rivers and the ocean.
Third, nitrogen is also cycled back into the atmosphere when organic material burns.
Figure illustrates the nitrogen cycle on, above, and below the Earth's surface. Source: Physical Geography Fundamentals eBook.
4.2.4 [Link]
an unusually high supply of nitrogen, and it has knock-on effects, including the following: certain plant species out-competing
other species, leading to biodiversity loss and altered ecosystem function; algal blooms that block light and therefore kill aquatic
plants in rivers, lakes, and seas; exhaustion of oxygen supplies in water caused by rapid microbial decomposition at the end of algal
blooms, which kills many aquatic organisms. Excess nitrates in water supplies have also been linked to human health problems.
Efforts to reduce nitrogen pollution focus on increasing the efficiency of synthetic fertilizer use, altering feeding of animals to
reduce nitrogen content in their excreta, and better processing of livestock waste and sewage sludge to reduce ammonia release. At
the same time, increasing demand for food production from a growing global population with a greater appetite for meat is driving
greater total fertilizer use, so there is no guarantee that better practices will lead to a reduction in the overall amount of nitrogen
pollution.
Review Questions
1. There is approximately 2,000 cubic kilometers of water stored in rivers around the world. Using the terms water cycle, flux and
pool, describe under what conditions removing 1000 cubic kilometers per year from rivers for human use could be sustainable.
2. Each year, around a quarter of the carbon dioxide found in the atmosphere is turned into plant matter via photosynthesis. Does
this mean that, in the absence of human activity, all carbon dioxide would be removed from the atmosphere in around four
years? Explain your answer.
3. The water, carbon, and nitrogen cycles are all influenced by human activity. Can you describe a human activity that impacts all
three cycles? In your example, which of the cycles is most significantly altered?
References
Le Quere, C., Raupach, M. R., Canadell, J. G., Marland, G., Bopp, L., Ciais, P., et al. (2009, December). Trends in the sources and
sinks of carbon dioxide. Nature Geoscience, 2, 831-836. doi: 10.1038/ngeo689
Millennium Ecosystem Assessment (2005). Ecosystems and Human Well-Being: Synthesis. Washington DC. Retrieved from
[Link]/en/[Link]
Glossary
ammonification
The release of ammonia by decomposerswhen they break down the complex nitrogen compounds in organic material
assimilation
Acquisition and incorporation of nutrients or resources by plants e.g. nitrogen or carbon.
biogeochemical cycles
A concept describing how chemical elements (e.g., nitrogen, carbon) or molecules (e.g. water) are transformed and stored by
both physical and biological components of the Earth system.
decomposers
Bacteria and fungi that break down rotting organic material, releasing component elements in the process.
denitrifying bacteria
Microbes that convert nitrates to nitrous oxide or N2 gases that are released back to the atmosphere.
eutrophication
Accelerated plant growth and decay caused by nitrogen pollution.
4.2.5 [Link]
evaporation
The process whereby water is converted from a liquid into a vapor, as a result of absorbing energy (usually from solar
radiation).
evapotranspiration
Evaporation from vegetated land that includes water transpired by plants as well as evaporation from open water and soils.
fluxes
Transformations or flow of materials from one pool to another in a biogeochemical cycle.
greenhouse gases
Gases in Earth's atmosphere that absorb long-wave radiation and retain heat.
groundwater discharge
Flow of water from below-ground into rivers, lakes, or the ocean.
infiltration
Flow of water from the land surface into soils and rocks.
leaching
Loss of nitrates from soil in drainage water
nitrification
Conversion of ammonia into nitrates by microbes.
photosynthesis
The process in which plants use energy from sunlight to combine CO2 from the atmosphere with water to make sugars, and in
turn build biomass.
pools
Amounts of material in biogeochemical cycles that share some common characteristic and are relatively uniform in nature.
precipitation
The conversion of atmospheric water from vapor into liquid (rain) or solid forms (snow, hail) that then fall to Earth's surface.
primary producers
The primary entry point of carbon into the biosphere—in nearly all land and aquatic ecosystems plants perform this role by
virtue of photosynthesis.
respiration
Metabolic process in all organisms that generates energy and synthesizes biomass while releasing CO2 as a by-product.
streamflow
Flow of water in streams.
surface runoff
Flow of water over the land surface.
4.2.6 [Link]
This page titled 4.2: Biogeochemical Cycles and the Flow of Energy in the Earth System is shared under a CC BY license and was authored,
remixed, and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
4.2.7 [Link]
4.3: Biodiversity, Species Loss, and Ecosystem Function
Learning Objectives
After reading this module, students should be able to
define biodiversity
articulate current trends in biodiversity loss with reference to species and ecosystems
explain some of the ways human activity affects biodiversity
explain how biodiversity loss concerns people
What is Biodiversity?
You're probably familiar with the word, biodiversity, whether or not you can give an exact definition of it. It's common on the signs
at zoos, parks, and nature centers, and it's often used without explanation or definition. Most people understand biodiversity in
general terms as the number and mix of plant and animal species that occurs in a given place. Scientists are more precise and
include more in their definition. The International Union for the Conservation of Nature, which coordinates efforts to catalogue and
preserve biodiversity worldwide, defines biodiversity as "the variability among living organisms from all sources including
terrestrial, marine and other aquatic ecosystems, and the ecological complexes of which they are part; this includes diversity within
species, between species, and of ecosystems." Rather than just species, biodiversity therefore includes variation from the level of
genes and genomes to that of ecosystems to biomes.
Even within a single ecosystem, the numbers of species can be impressive. For example, there is a large region of dry forest and
savanna in Brazil known as the Cerrado (see Figure 4.3.1). This ecosystem alone hosts over 10,000 species of plants, almost 200
species of mammals, over 600 species of birds, and about 800 species of fish.
Figure 4.3.1 Cerrado Forest Photograph of the Cerrado Forest. Source: C2rik via Wikimedia Commons.
Generally, biodiversity is greatest in tropical areas–especially "rainforests"—but there are terrestrial biodiversity "hotspots" on all
the major continents. (View an interactive map of hotspots.)
4.3.1 [Link]
Figure 4.3.2 Summary of Threatened Species. Table lists the numbers and proportions of species assessed as threatened on the
2008 IUCN Red List by major taxonomic group. Source: IUCN Red List, Wildlife in a Changing World 2008, p. 17. Please see
IUCN Terms of Use for copyright restrictions.
Vertebrates
Scientists know much more about the state of vertebrates—especially mammals, birds, and amphibians—than they do about other
forms of animal life. Every one of the 5,488 species of mammals that have been described, for example, has been evaluated for
purposes of the Red List. Of them, 76 species have become extinct since 1500, and two, Pere David's deer, which is native to
China, and the scimitar oryx from Africa survive only in managed facilities. Another 29 of the mammal species listed as critically
endangered are also tagged as "possibly extinct;" they are very likely gone, but the sort of exhaustive surveys required to confirm
that fact have not been conducted. Overall, approximately 22% of mammal species worldwide are known to be threatened or
extinct. (In the terms of the Red List, the broad designation "threatened" includes three levels of risk for extinction in the wild:
Vulnerable [high], Endangered [higher], and Critically Endangered [highest].)
The Red List categorizes a smaller proportion of the world's 9,990 described bird species—14%—as threatened or extinct. But the
raw number of species lost since 1500 is at least 134, and four more species persist only in zoos. Another 15 species of birds are
considered possibly extinct. The fact that 86% of bird species are categorized as "not threatened" is good news in the context of the
Red List.
4.3.2 [Link]
Figure 4.3.3 Passenger Pigeons North American passenger pigeons lived in enormous flocks and were once the most numerous
birds on earth. Market hunting on a massive scale and habitat destruction combined to extinguish them as a species in the early
twentieth century. Source: Ltshears via Wikimedia Commons
Among the well-studied vertebrates, amphibians are faring especially poorly. Of the more than 6,000 known species of amphibians,
38 have become extinct worldwide since 1500, and another one, the Wyoming toad, survives only in captivity. Another 120 species
are considered possibly extinct. Overall, 2,030, or one-third of the world's amphibian species are known to be threatened or extinct.
More troubling still, many amphibian species—42.5%—are reported to be declining, and that number is probably low, since trend
information is unavailable for 30.4% of species.
Figure 4.3.4 Monteverde Golden Toad. The golden toad of Monteverde, Costa Rica, is one of 11 species of amphibians to become
extinct since 1980. Habitat loss and chytrid fungus. Source: U.S. Fish and Wildlife Service via Wikimedia Commons.
Only small proportions of the world's species of reptiles and fish have been evaluated for purposes of the Red List. Among those,
the numbers of species that fall into the threatened category are very high: 1,275 of the 3,481 evaluated species, or 37%, for fish;
and 423 of 1,385 evaluated species, or 31%, for reptiles. It should be noted, however, that these percentages are likely
overestimates, since species of concern are more likely to be selected for evaluation than others.
Invertebrates
The category "invertebrates" lumps together the vast majority of multi-cellular animals, an estimated 97% of all species. It includes
everything from insects and arachnids, to mollusks, crustaceans, corals, and more. Few of these groups have been assessed in a
comprehensive way, and so as with fish and reptiles, the Red List percentages of threatened species are skewed high. But
assessments within some groups call attention to disturbing, large-scale trends. For example, 27% of the world's reef-building
corals are already considered threatened, and many more of them are experiencing rates of decline that move them toward
threatened status. The demise of reef-building corals has magnified ecological impacts, since so much other marine life depends on
them.
4.3.3 [Link]
Figure 4.3.5 Pink and Soft Coral with Reef Fish. Photograph shows some pink, soft coral with reef fish nearby. Source: Linda
Wade via National Oceanic & Atmospheric Administration (NOAA).
It should be understood that information about familiar creatures such as amphibians, mammals, and birds is just a beginning, and
that even with the inclusion of some invertebrates the Red List does not provide a comprehensive picture of life on Earth. Scientists
have described fewer than 2 million of the 8-9 million species of organisms thought to exist, most of which are insects. And of
those 2 million, the status of only 44,838 has been assessed by IUCN.
In addition, it should be understood that among the species that have been assessed so far, there is a strong bias toward terrestrial
vertebrates and plants, especially the ones that occur where biologists have visited frequently. Red List assessments also tend to
focus on species that are likely to be threatened, since the effort also has the aim of enabling people to conserve species.
Whereas extinction is the global loss of a species, the elimination of species at a local level–known as extirpation– also poses
threats to the integrity and sustainability of ecosystems. Widespread extirpation obviously leads to threatened or endangered status,
but absence of species, even at a local scale, can affect ecosystem function. For example, by the mid-1920s wolves had been
extirpated from Yellowstone National Park, although they continued to thrive elsewhere. When wolves were reintroduced to the
park in the mid-1990s, numbers of elk (a main prey item) decreased significantly. This, in turn, reduced browsing pressure and had
a significant effect on the vegetation and plant communities. What mattered for ecosystem function in Yellowstone was whether
wolves were present there, not just whether the species survived somewhere.
The human activities that account for extinction and extirpation vary considerably from one species to another, but they fall into a
few broad categories: habitat destruction and fragmentation; intentional and unintentional movement of species that become
invasive (including disease-causing organisms); over-exploitation (unsustainable hunting, logging, etc.); habitat/ecosystem
degradation (e.g. pollution, climate change).
4.3.4 [Link]
Species and Ecosystem Loss in Perspective
To understand why biologists talk about ongoing losses of species and ecosystems as the "biodiversity crisis," it is useful to put
current and projected rates of species loss into historical perspective. Over the history of life on Earth—a span of 3.5 billion years
—nearly all species that existed eventually became extinct. This, of course, is coupled with the processes of speciation and
biological diversification. Rates of extinction and diversification have fluctuated significantly over geologic time. For extinction,
paleontologists have detected five episodes of mass extinction over the last 540 million years. These periods contrast with the
relatively constant "background rate" of extinction observed over the geologic record, and include the relatively well-known event
65 million years ago when most of the extant dinosaurs went extinct. By definition, these episodes are characterized by the
comparatively rapid loss of at least three-fourths of the species thought to exist at the onset of the event.
Recently, the question has been posed whether present-day rates of species loss constitute a sixth episode of mass extinction
(Barnosky, et al., 2011). Even with caveats about uncertainty in how many species there are today (only a fraction of the estimated
total have been described, especially for plants, invertebrates, and microbes) and about comparisons of the fossil record with
modern data, it appears that estimated rates of loss in the near future could rival those of past mass extinctions. Some estimates
indicate that we will see a 30% loss of species within decades. Put another way, forecasted rates of species loss could be as much as
1000 to 10,000 times higher than background rates.
Review Questions
1. What is the difference between extinction and extirpation?
2. What are some human activities that impact species diversity and ecosystem function?
3. Does the loss of one species lead to loss of ecosystem function? Why or why not?
4. How does biodiversity promote sustainability?
References
Barnosky, A.D., Matzke, N., Tomiya, S., Wogan, G.O.U., Swartz, B., Quental, T.B., et al. (2011, March). Has the Earth's sixth mass
extinction already arrived? Nature, 471, 51-57. doi:10.1038/nature09678
Vie, J-C, Hilton-Taylor, C. & Stuart S.N. (Eds.). (2009). Wildlife in a Changing World: An Analysis of the 2008 IUCN Red List of
Threatened Species™. Gland, Switzerland: IUCN. Retrieved from [Link]/dbtw-wpd/edocs/[Link].
4.3.5 [Link]
Glossary
biodiversity
The number of different species within an ecosystem (or globally). Biodiversity is also considered a metric of ecosystem health.
ecosystem function
Processes such as decomposition, production, nutrient cycling, and fluxes of nutrients and energy that allow an ecosystem to
maintain its integrity as a habitat.
extinction
The death of all individuals within a species. A species may be functionally extinct when a low number of surviving individuals
are unable to reproduce.
extirpation
Local extinction of a species; elimination or removal of a species from the area of observation.
ecological services
Ecosystem functions that are essential to sustaining human health and well-being. Examples include provisioning services such
as food, fiber and water; regulating services such as climate, flood, and disease control; cultural services such as spiritual and
recreational benefits, and supporting services such as nutrient cycling. Also called ecosystem services
This page titled 4.3: Biodiversity, Species Loss, and Ecosystem Function is shared under a CC BY license and was authored, remixed, and/or
curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
4.3.6 [Link]
4.4: Soil and Sustainability
Learning Objectives
After reading this module, students should be able to
define soil and comment on its importance to society
describe how soil profiles form
explain the importance of soil constituents for plant growth and nutrient uptake
understand the importance of soil to agricultural sustainability and ecological processes
Figure 4.4.1 Soil Profile. Photograph shows a soil profile from South Dakota with A, E, and Bt horizons. The yellow arrows
symbolize translocation of fine clays to the Bt horizon. The scale is in feet. Source: University of Idaho and modified by D.
Grimley.
We will use this definition in this chapter. In common usage, the term soil is sometimes restricted to only the dark topsoil in which
we plant our seeds or vegetables. In a more broad definition, civil engineers use the term soil for any unconsolidated (soft when
wet) material that is not considered bedrock. Under this definition, soil can be as much as several hundred feet thick! Ancient soils,
sometimes buried and preserved in the subsurface, are referred to as paleosols (see Figure 4.4.2) and reflect past climatic and
environmental conditions.
4.4.1 [Link]
Figure 4.4.2 Modern versus Buried Soil Profiles. A buried soil profile, or paleosol (above geologist 's head), represents soil
development during the last interglacial period. A modern soil profile (Alfisol) occurs near the land surface. Source: D. Grimley.
From a somewhat philosophical standpoint, soil can be viewed as the interface between the atmosphere and the earth's crust, and is
sometimes referred to as the skin of the earth. Soil also incorporates aspects of the biosphere and the hydrosphere. From a physical
standpoint, soil contains solid, liquid, and gaseous phases. The solid portion of the soil consists predominantly of mineral matter,
but also contains organic matter (humus) and living organisms. The pore spaces between mineral grains are filled with varying
proportions of water and air.
Importance of Soil
Soil is important to our society as it provides the foundation for most of the critical aspects of civilization. Our building structures
and homes, food, agricultural products, and wood products all rely on soil. Forests, prairies, and wetlands all have a dependence on
soil. Of course, soil is also a critical component for terrestrial life on earth, including most animals, plants, and many
microorganisms.
Soil plays a role in nearly all natural cycles on the earth's surface. Global cycling of key nutrients, such as Carbon (C), Nitrogen
(N), Sulfur (S), and Phosphorous (P), all pass through soil. In the hydrologic cycle, soil helps to mediate the flow of precipitation
from the land surface into the groundwater or can control stormwater runoff into lakes, streams, bays, and oceans. Soil
microorganisms or microflora can help to modify or destroy environmental pollutants.
Climate
The role of climate in soil development includes aspects of temperature and [Link] in very cold areas with permafrost
conditions (Gelisols) tend to be shallow and weakly developed due to the short growing season. Organic rich surface horizons are
common in low-lying areas due to limited chemical decomposition. In warm, tropical soils (Ultisols, Oxisols), other factors being
equal, soils tend to be thicker, with extensive leaching and mineral alteration. In such climates, organic matter decomposition and
chemical weathering occur at an accelerated rate.
Organisms
4.4.2 [Link]
Animals, plants, and microorganisms all have important roles in soil development processes, in providing a supply of organic
matter, and/or in nutrient cycling. Worms, nematodes, termites, ants, gophers, moles, crayfish, etc. all cause considerable mixing of
soil and help to blend soil, aerate and lighten the soil by creating porosity, and create characteristic natural soil structure over time.
Animal life, such as insects and mammals, can cause irregularities in the soil horizons.
Plant life provides much organic matter to soil and helps to recycle nutrients with uptake by roots in the subsurface. The type of
plant life that occurs in a given area, such as types of trees or grasses, depends on the climate, along with parent material and soil
type. So there are clearly feedbacks among the soil forming factors. With the annual dropping of leaves and needles, trees tend to
add organic matter to soil surfaces, helping to create a thin, organic-rich A or O horizon over time. Grasses, on the other hand, have
a considerable root mass, in addition to surficial organic material, that is released into the soil each fall for annuals and short-lived
perennials. For this reason, grassland soils (Mollisols) have much thicker A horizons with higher organic matter contents, and are
more agriculturally productive than forest soils. Grasses release organic matter to soils that is more rich in base cations, whereas
leaf and needle litter result in release of acids into the soil.
Microorganisms aid in the oxidation of organic residues and in production of humus material. They also play a role in iron
oxidation-reduction cycles, fine-grained mineral dissolution (providing nutrients to soil solutions), and mineral neoformation. New
research is continually expanding our knowledge of the role of microorganisms in plant growth, nutrient cycling, and mineral
transformations.
Parent Material
The parent material of a soil is the material from which the soil has developed, whether it be river sands, lake clays, windblown
loess, shoreline deposits, glacial deposits, or various types of bedrock. In youthful soils, the parent material has a clear connection
to the soil type and has significant influence. Over time, as weathering processes deepen, mix, and alter the soil, the parent material
becomes less recognizable as chemical, physical, and biological processes take their effect. The type of parent material may also
affect the rapidity of soil development. Parent materials that are highly weatherable (such as volcanic ash) will transform more
quickly into highly developed soils, whereas parent materials that are quartz-rich, for example, will take longer to develop. Parent
materials also provide nutrients to plants and can affect soil internal drainage (e.g. clay is more impermeable than sand and impedes
drainage).
Time
In general, soil profiles tend to become thicker (deeper), more developed, and more altered over time. However, the rate of change
is greater for soils in youthful stages of development. The degree of soil alteration and deepening slows with time and at some
point, after tens or hundreds of thousands of years, may approach an equilibrium condition where erosion and deepening (removals
and additions) become balanced. Young soils (< 10,000 years old) are strongly influenced by parent material and typically develop
horizons and character rapidly. Moderate age soils (roughly 10,000 to 500,000 years old) are slowing in profile development and
deepening, and may begin to approach equilibrium conditions. Old soils (>500,000 years old) have generally reached their limit as
far as soil horizonation and physical structure, but may continue to alter chemically or mineralogically.
4.4.3 [Link]
To be sure, soil development is not always continual. Geologic events can rapidly bury soils (landslides, glacier advance, lake
transgression), can cause removal or truncation of soils (rivers, shorelines) or can cause soil renewal with additions of slowly
deposited sediment that add to the soil (wind or floodplain deposits). Biological mixing can sometimes cause soil regression, a
reversal or bump in the road for the normal path of increasing development over time.
Figure illustrates the uptake of nutrients by plants in the forest" soil ecosystem. Source: U.S. Geological Survey.
Plants mainly obtain nutrients from dissolved soil solutions. Though many aspects of soil are beneficial to plants, excessively high
levels of trace metals (either naturally occurring or anthropogenically added) or applied herbicides can be toxic to some plants.
The ratio of solids/water/air in soil is also critically important to plants for proper oxygenation levels and water availability. Too
much porosity with air space, such as in sandy or gravelly soils, can lead to less available water to plants, especially during dry
seasons when the water table is low. Too much water, in poorly drained regions, can lead to anoxic conditions in the soil, which
may be toxic to some plants. Hydrophytic vegetation can handle anoxic conditions and is thus suitable to poorly drained soils in
wetland areas.
4.4.4 [Link]
they are in an extractable form in soil solutions, such as an exchangeable cation, rather than in a solid mineral grain. As nutrients
are used up in the microenvironment surrounding a plant's roots, the replenishment of nutrients in soil solution is dependent on
three aspects: (a) the rate of dissolution/alteration of soil minerals into elemental constituents, (b) the release rate of organically
bound nutrients, and (c) the rate of diffusion of nutrients through the soil solution to the area of root uptake.
Many nutrients move through the soil and into the root system as a result of concentration gradients, moving by diffusion from high
to low concentrations. However, some nutrients are selectively absorbed by the root membranes, such that elemental concentrations
of solutions within plants may differ from that in soil solutions. Most nutrients exist as exchangeable cations that are acquired by
roots from the soil solution—rather than from mineral or particle surfaces. Inorganic chemical processes and organic processes,
such as the action of soil microorganisms, can help to release elemental nutrients from mineral grains into the soil environment.
4.4.5 [Link]
compaction issues can occur with excessive passage of equipment during times when the soil has a high water content. The
problem with soil compaction is that increased soil density limits root penetration depth and may inhibit proper plant growth.
Current practices generally encourage minimal tillage or no tillage in order to reduce the number of trips across the field. With
proper planning, this can simultaneously limit compaction, protect soil biota, reduce costs (if performed correctly), promote water
infiltration, and help to prevent topsoil erosion (see below). Tillage of fields does help to break up clods that were previously
compacted, so best practices may vary at sites with different soil textures and composition. Crop rotation can also help to reduce
bulk density with planting of crops with different root depth penetration. Another aspect of soil tillage is that it may lead to more
rapid decomposition of organic matter due to greater soil aeration. Over large areas of farmland, this has the unintended
consequence of releasing more carbon and nitrous oxides (greenhouse gases) into the atmosphere, thereby contributing to global
warming effects. In no-till farming, carbon can actually become sequestered into the soil. Thus, no-till farming may be
advantageous to sustainability issues on the local scale and the global scale.
Soil Erosion
Accelerated erosion of topsoil due to human activities and poor agricultural land management is a potentially serious issue. The
areas most vulnerable to soil erosion include locations with thin organic (A and O) horizons and hilly terrains (see Figure 4.4.4).
Figure shows a global map of soil erosion vulnerability and includes a photograph of water and wind erosion. Source: U.S.
Department of Agriculture, National Resource Conservation Service, Rodney Burton via Wikimedia Commons, and Jim Bain via
Wikimedia Commons.
Some amount of soil erosion is a natural process along sloping areas and/or in areas with soft or noncohesive materials susceptible
to movement by water, wind, or gravity. For instance, soil material can be mobilized in strong windstorms, along the banks of
rivers, in landslides, or by wave action along coastlines. Yet most topsoil erosion results from water influenced processes such as in
rivers, creeks, ravines, small gullies, and overland flow or sheetwash from stormwater runoff. Although some soil erosion is
natural, anthropogenic (human-induced) processes have greatly accelerated the erosion rate in many areas. Construction and
agriculture are two of the more significant activities in our modern society that have increased erosion rates. In both cases, the
erosion of topsoil can be significant if poor land management practices are used or if the area is geologically sensitive. For
instance, in the 1930's, drought conditions and poor land management methods (lack of cover crops and rotation) combined to
result in severe wind erosion and dust storms in the Great Plains of the United States, which came to be known as the Dust Bowl.
Deep plowing of soil and displacement of the original prairie grasses (that once held the soil together) also contributed to the crisis.
Once the natural topsoil is eroded by wind or water, it is only slowly renewable to its former pre-eroded condition. It may take
anywhere from several decades to hundreds of years to millennia, under replanted native vegetation, to restore the soil to a
relatively natural (pre-disturbed) state with its original physical, chemical, and biological characteristics. Furthermore, when soil is
eroded, the particles become sedimented downstream in streams, rivers, lakes, and reservoirs. If rapid, this sedimentation can
deteriorate the water quality with sediment and agricultural chemicals. Better land management practices, such as more limited
tillage or no-till practices, can help to greatly limit soil erosion to a rate that is sustainable over the long term. Practices today are
somewhat improved overall, but more improvement in agricultural practices are needed over large areas of farmland in the United
States and other countries to bring us on a path to long-term sustainability of agricultural lands.
Deforestation due to logging, construction, or increased fire occurrences can also cause significant increases in soil erosion in many
areas globally and may be a particular problem in developing countries. Removal of the natural cover of vegetation enhances
erosion since plant foliage tends to buffer the intensity of rainfall and roots hold soil together and prevent breakup and erosion.
4.4.6 [Link]
Furthermore, decomposing plant material provides a protective cover of organic material on the soil surface. Watersheds with large
areas of construction or deforestation can experience several times the natural erosion rate. In such watersheds, streams can become
clogged with unwanted sediment that disturbs the natural ecosystem and infills valuable wetland areas, in addition to the problem
of valuable topsoil loss from upland areas.
Figure 4.4.5 Aquatic Dead Zones. The red circles show the size of many of our planet's dead (hypoxia) zones, whereas the plain
black dots are dead zones of unknown size. Darker blue colors show high concentrations of particulate organic matter, an indication
of overly fertile waters (high in N and P). Most dead zones occur in downriver of agricultural areas (with overused fertilizer) or
areas of high population density with poorly treated wastewater. Source: NASA Earth Observatory via Wikimedia Commons and
Lamiot via Wikimedia Commons.
One of the most notorious dead zones (second to the Baltic Sea) is an 8,500 square mile region in the Gulf of Mexico (see Figure
4.4.5). The Mississippi River dumps high-nutrient runoff from its drainage basin that includes vast agricultural lands in the
American Midwest. Increased algal growth produced by these nutrients has affected important shrimp fishing grounds in the Gulf.
The primary source of the nutrients is the heavily tile-drained areas of farmland in the Midwest corn and soybean belt (SW
Minnesota, N Iowa, NE Illinois, N Indiana and NW Ohio). Improved soil drainage systems over the past century or more have
allowed for effective transport of nitrate compounds as stormwater runoff into drainage basins (Ohio River, Wabash River, Illinois
River, Missouri River, etc.) that feed into the Mississippi River. In other words, the same drainage tiles that allow for the
agricultural benefit of having rich bottomland/wetland soils in production, have the disadvantage of increased and more rapid
movements of nitrate solutes to the Gulf of Mexico. Such large-scale problems, across state governmental boundaries, can only be
fully addressed in the future with a national system of incentives, regulations, or laws.
4.4.7 [Link]
In addition to fertilizers, Nitrogen inputs to watersheds can also include atmospheric deposition, livestock waste, and sewage, but
nitrogen fertilizers comprise a significant majority of the input to monitored streams, particularly in springtime when much
fertilizer is applied. Possible solutions to this problem include encouraging farmers to apply a more limited quantity of fertilizer in
the spring (only as much as necessary), rather than in the fall, to allow for considerably less time for stormwater or meltwater
runoff. Other solutions include maintaining cover crops, or restoring wetlands in key locations to contain nitrate losses. An overall
strategy that limits the excess capacity of nutrients can simultaneously benefit farmers (by limiting cost), the ecology of stream
watersheds and coastal ecosystems (also locally stressed by oil spills and other pollution). Over the long term, more efforts will
need to be made in the Mississippi River Basin, and globally in similarly stressed agricultural or urban watersheds (see Figure
4.4.5), to improve the health and sustainability of our soil, land, and aquatic ecosystems.
Review Questions
1. What is the importance of soil to our society today?
2. How has human activity changed the physical, chemical, or biological character of native soil?
3. What practices can be used to improve the long-term sustainability of soil health?
This page titled 4.4: Soil and Sustainability is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto Cabezas
(GALILEO Open Learning Materials) .
4.4.8 [Link]
CHAPTER OVERVIEW
This page titled 5: Physical Resources- Water, Pollution, and Minerals is shared under a CC BY license and was authored, remixed, and/or curated
by Heriberto Cabezas (GALILEO Open Learning Materials) .
1
5.1: Physical Resources- Water, Pollution, and Minerals - Chapter Introduction
Introduction
Water, air, and food are the most important natural resources to people. Humans can live only a few minutes without oxygen, about
a week without water, and about a month without food. Water also is essential for our oxygen and food supply. Plants, which
require water to survive, provide oxygen through photosynthesis and form the base of our food supply. Plants grow in soil, which
forms by weathering reactions between water and rock.
Water is the most essential compound for Earth’s life in general. Human babies are approximately 75% water and adults are 50–
60% water. Our brain is about 85% water, blood and kidneys are 83% water, muscles are 76% water, and even bones are 22%
water. We constantly lose water by perspiration; in temperate climates we should drink about 2 quarts of water per day and people
in hot desert climates should drink up to 10 quarts of water per day. Loss of 15% of body-water usually causes death. Earth is truly
the “Water Planet” (see Figure 5.1.1). The abundance of water on Earth distinguishes us from other bodies in the solar system.
About 70% of Earth's surface is covered by oceans and approximately half of Earth's surface is obscured by clouds at any time.
There is a very large volume of water on our planet, about 1.4 billion cubic kilometers (km3) (330 million cubic miles) or about 53
billion gallons per person on Earth. All of Earth’s water could cover the United States to a depth of 145 km (90 mi). From a human
perspective, the problem is that over 97% of it is seawater, which is too salty to drink or use for irrigation. The most commonly
used water sources are rivers and lakes, which contain less than 0.01% of the world’s water!
graphic of spinning earth
Figure 5.1.1 Planet Earth from Space Source: Created by Marvel, based on a Nasa image via Wikimedia Commons
One of our most important environmental goals is to provide a clean, sufficient, and sustainable water supply for the world.
Fortunately, water is a renewable resource, and it is difficult to destroy. Evaporation and precipitation combine to replenish our
fresh water supply constantly and quickly; however, water availability is complicated by its uneven distribution over the Earth.
Arid climate and densely populated areas have combined in many parts of the world to create water shortages, which are projected
to worsen significantly in the coming years. Human activities such as water overuse and water pollution have compounded the
water crisis that exists today. Hundreds of millions of people lack access to safe drinking water, and billions of people lack access
to improved sanitation as simple as a pit latrine. As a result, nearly two million people die every year from diarrheal diseases and
90% of those deaths occur among children under the age of 5. Most of these are easily prevented deaths.
Although few minerals are absolutely essential for human life, the things that define modern society require a wide range of them:
iron ore for steel, phosphate minerals for fertilizer, limestone rock for concrete, rare earth elements for night-vision goggles and
phosphors in computer monitors, and lithium minerals for batteries in our laptops, cell phones, and electric cars. As global
population grows and emerging large economies expand, we will face a crisis in the supply of many important minerals because
they are nonrenewable, which is to say we consume them far more quickly than nature creates them. As we consume minerals from
larger and lower grade mineral deposits there will be greater environmental impacts from mineral mining and processing. The
impending mineral crisis may be more challenging to address than the water crisis.
This chapter introduces basic principles in water supply, water pollution, and mineral resources. The emphasis, however, is on
environmental issues and sustainable solutions for each problem.
This page titled 5.1: Physical Resources- Water, Pollution, and Minerals - Chapter Introduction is shared under a CC BY license and was
authored, remixed, and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
5.1.1 [Link]
5.2: Water Cycle and Fresh Water Supply
Learning Objectives
After reading this module, students should be able to
understand how the water cycle operates
understand the principles controlling groundwater resources and how they also can affect surface water resources
know the causes and effects of depletion in different water reservoirs
understand how we can work toward solving the water supply crisis
Figure 5.2.1 Earth's Water Resevoir. Bar chart Distribution of Earth’s water including total global water, fresh water, and surface
water and other fresh water and Pie chart Water usable by humans and sources of usable water. Source: United States Geographical
Survey Igor Skiklomanov's chapter "World fresh water resources" in Peter H. Gleick (editor), 1993, Water in Crisis: A Guide to the
World's Fresh Water Resources
Figure 5.2.2 The Water Cycle. Arrows depict movement of water to different reservoirs located above, at, and below Earth’s
surface. Source: United States Geological Survey
5.2.1 [Link]
An important part of the water cycle is how water varies in salinity, which is the abundance of dissolved ions in water. Ocean water
is called salt water because it is highly saline, with about 35,000 mg of dissolved ions per liter of seawater. Evaporation (where
water changes from liquid to gas at ambient temperatures) is a distillation process that produces nearly pure water with almost no
dissolved ions. As water vaporizes, it leaves the dissolved ions in the original liquid phase. Eventually, condensation (where water
changes from gas to liquid) forms clouds and sometimes precipitation (rain and snow). After rainwater falls onto land, it dissolves
minerals, which increases its salinity. Most lakes, rivers, and near-surface groundwater have a relatively low salinity and are called
fresh water. The next several sections discuss important parts of the water cycle relative to fresh water resources.
Figure 5.2.3 World Rainfall Map. The false-color map above shows the amount of rain that falls around the world. Areas of high
rainfall include Central and South America, western Africa, and Southeast Asia. Since these areas receive so much rainfall, they are
where most of the world's rainforests grow. Areas with very little rainfall usually turn into deserts. The desert areas include North
Africa, the Middle East, western North America, and Central Asia. Source: United States Geological Survey Earth Forum, Houston
Museum Natural Science
vegetation, topography, land use, and soil characteristics. Soon after a heavy rainstorm, river discharge increases due to surface
runoff. The steady normal flow of river water is mainly from groundwater that discharges into the river. Gravity pulls river water
downhill toward the ocean. Along the way the moving water of a river can erode soil particles and dissolve minerals, creating the
river’s load of moving sediment grains and dissolved ions. Groundwater also contributes a large amount of the dissolved ions in
river water. The geographic area drained by a river and its tributaries is called a drainage basin. The Mississippi River drainage
basin includes approximately 40% of the U.S., a measure that includes the smaller drainage basins (also called watersheds), such as
the Ohio River and Missouri River that help to comprise it. Rivers are an important water resource for irrigation and many cities
around the world. Some of the world’s rivers that have had international disputes over water supply include the Colorado (Mexico,
southwest U.S.), Nile (Egypt, Ethiopia, Sudan), Euphrates (Iraq, Syria, Turkey), Ganges (Bangladesh, India), and Jordan (Israel,
Jordan, Syria).
5.2.2 [Link]
Figure 5.2.4 Surface runoff, part of overland flow in the water cycle Source: James M. Pease at Wikimedia Commons
Figure 5.2.5 Groundwater seepage can be seen in Box Canyon in Idaho, where approximately 10 cubic meters per second of
seepage emanates from its vertical headwall. Source: NASA
Figure 5.2.6 Colorado River, U.S.. Rivers are part of overland flow in the water cycle and an important surface water resource.
Source: Gonzo fan2007 at Wikimedia Commons
Lakes can also be an excellent source of fresh water for human use. They usually receive water from surface runoff and
groundwater discharge. They tend to be short-lived on a geological time-scale because they are constantly filling in with sediment
supplied by rivers. Lakes form in a variety of ways including glaciation (Great Lakes, North America, See Figure 5.2.7), recent
tectonic uplift (Lake Tanganyika, Africa), and volcanic eruptions (Crater Lake, Oregon). People also create artificial lakes
(resevoirs) by damming rivers. Large changes in climate can result in major changes in a lake’s size. As Earth was coming out of
the last Ice Age about fifteen thousand years ago, the climate in the western U.S. changed from cool and moist to warm and arid,
which caused more than100 large lakes to disappear. The Great Salt Lake in Utah is a remnant of a much larger lake called Lake
Bonneville.
5.2.3 [Link]
Figure 5.2.7 Great Lakes from Space. The Great Lakes hold 21% of the world's surface fresh water. Lakes are an important surface
water resource. Source: SeaWiFS Project, NASA/Goddard Space Flight Center, and ORBIMAGE
Although glaciers represent the largest reservoir of fresh water, they generally are not used as a water source because they are
located too far from most people (see Figure 5.2.8). Melting glaciers do provide a natural source of river water and groundwater.
During the last Ice Age there was as much as 50% more water in glaciers than there is today, which caused sea level to be about
100 m lower. Over the past century, sea level has been rising in part due to melting glaciers. If Earth’s climate continues to warm,
the melting glaciers will cause an additional rise in sea level.
Figure 5.2.8 Mountain Glacier in Argentina. Glaciers are the largest reservoir of fresh water but they are not used much as a water
resource directly by society because of their distance from most people. Source: Luca Galuzzi - [Link]
Groundwater Resources
Although most people in the U.S. and the world use surface water, groundwater is a much larger reservoir of usable fresh water,
containing more than 30 times more water than rivers and lakes combined. Groundwater is a particularly important resource in arid
climates, where surface water may be scarce. In addition, groundwater is the primary water source for rural homeowners, providing
98% of that water demand in the U.S.. Groundwater is water located in small spaces, called pore space, between mineral grains and
fractures in subsurface earth materials (rock or sediment, i.e., loose grains). Groundwater is not located in underground rivers or
lakes except where there are caves, which are relatively rare. Between the land surface and the depth where there is groundwater is
the unsaturated zone, where pore spaces contain only air and water films on mineral grains (see Figure 5.2.9)1. Below the
unsaturated zone is the saturated zone, where groundwater completely fills pore spaces in earth materials. The interface between
the unsaturated zone and saturated zone is the water table. Most groundwater originates from rain or snowmelt, which infiltrates the
ground and moves downward until it reaches the saturated zone. Other sources of groundwater include seepage from surface water
(lakes, rivers, reservoirs, and swamps), surface water deliberately pumped into the ground, irrigation, and underground wastewater
5.2.4 [Link]
treatment systems, i.e., septic tanks. Recharge areas are locations where surface water infiltrates the ground rather than running off
into rivers or evaporating. Wetlands and flat vegetated areas in general are excellent recharge areas.
Figure 5.2.9 Subsurface Water Terminology. Groundwater in pore spaces and fractures of earth materials, saturated zone,
unsaturated zone, and water table, which follows land surface but in a more subdued way. Source: United States Geological Survey
Groundwater is in constant motion due to interconnection between pore spaces. Porosity is the percentage of pore space in an earth
material and it gives a measure of how much groundwater an earth material can hold. Permeability is a measure of the speed that
groundwater can flow through an earth material, and it depends on the size and degree of interconnection among pores. An earth
material that is capable of supplying groundwater from a well at a useful rate—i.e., it has relatively high permeability and medium
to high porosity—is called an aquifier. Examples of aquifers are earth materials with abundant, large, well-connected pore spaces
such as sand, gravel, uncemented sandstone, and any highly fractured rock. An earth material with low hydraulic conductivity is an
aquitard. Examples of aquitards include clay, shale (sedimentary rock with abundant clay), and igneous and metamorphic rock, if
they contain few fractures.
As discussed above, groundwater flows because most earth materials near the surface have finite (nonzero) porosity and
permeability values. Another reason for groundwater movement is that the surface of the water table commonly is not completely
flat but mimics the topography of the land surface, especially in humid climates. There is "topography" to the water table because
groundwater moves slowly through rock and soil, so it builds up in higher elevation areas. In fact, when groundwater flows slowly
through aquitards and deep underground, it can take many thousands of years to move relatively short distances. An unconfined
aquifier has no aquitard above it and, therefore, it is exposed to the atmosphere and surface waters through interconnected pores
(See Figure 5.2.10). In an unconfined aquifer, groundwater flows because of gravity to lower water table levels, where it eventually
may discharge or leave the groundwater flow system. Discharge areas include rivers, lakes, swamps, reservoirs, water wells, and
springs (see Figure 5.2.11). Springs are rivers that emerge from underground due to an abrupt intersection of the land surface and
the water table caused by joints, caves, or faults that bring permeable earth materials to the surface. A confined aquifier is bounded
by aquitards below and above, which prevents recharge from the surface immediately above. Instead, the major recharge occurs
where the confined aquifer intercepts the land surface, which may be a long distance from water wells and discharge areas (see
Figure 5.2.12). Confined aquifers are commonly inclined away from recharge areas, so groundwater in a confined aquifer is under
greater-than-atmospheric pressure due to the weight of water in the upslope direction. Similar to river discharge, groundwater
discharge describes the volume of water moving through an aquifer over time. Total groundwater discharge depends on the
permeability of the earth material, the pressure that drives groundwater flow, and the size of the aquifer. It is important to determine
groundwater discharge to evaluate whether an aquifer can meet the water needs of an area.
5.2.5 [Link]
Figure 5.2.10 Flowing Groundwater. Blue lines show the direction of groundwater in unconfined aquifers, confined aquifers, and
confining beds. Deep groundwater moves very slowly especially through low permeability [Link]: United States Geological
Survey
Figure 5.2.11 Fatzael Springs in Jordan Valley. A spring is a river that emerges from underground due to an abrupt intersection of
the water table with the land surface such as alongside a hill. Source: Hanay at Mediawiki Commons
5.2.6 [Link]
Figure 5.2.12 Schematic Cross Section of Aquifer Types. This figure shows different types of aquifers and water wells, including
unconfined aquifer, confined aquifer, water table well, artesian well, and flowing artesian well. Point of triangle is water level in
each well and water table in other parts of figure. Water level in artesian well is at potentiometric surface and above local water
table (dashed blue line) due to extra pressure on groundwater in confined aquifer. Water in flowing artesian well moves above land
surface. Source: Colorado Geological Survey
Most shallow water wells are drilled into unconfined aquifers. These are called water table wells because the water level in the well
coincides with the water table (See Figure 5.2.13). 90% of all aquifers for water supply are unconfined aquifers composed of sand
or gravel. To produce water from a well, you simply need to drill a hole that reaches the saturated zone and then pump water to the
surface. Attempting to pump water from the unsaturated zone is like drinking root beer with a straw immersed only in the foam at
the top.
To find a large aquifer for a city, hydrogeologists (geologists who specialize in groundwater) use a variety of information including
knowledge of earth materials at the surface and sub-surface as well as test wells. Some people search for water by dowsing, where
someone holds a forked stick or wire (called a divining rod) while walking over an area. The stick supposedly rotates or deflects
downward when the dowser passes over water. Controlled tests show that a dowser's success is equal to or less than random
chance. Nevertheless, in many areas water wells are still drilled on dowser’s advice sometimes for considerable money. There is no
scientific basis to dowsing.
Wells into confined aquifers typically are deeper than those into unconfined aquifers because they must penetrate a confining layer.
The water level in a well drilled into a confined aquifer, which is an artesian well, (see Figure 5.2.12), moves above the local water
table to a level called the potentiometric surface because of the greater pressure on the groundwater. Water in a flowing well (see
Figure 5.2.14) moves all of the way to the land surface without pumping.
5.2.7 [Link]
Figure 5.2.13 A Flowing Well. Flowing artesian well where water moves above the land surface due to extra pressure on the
groundwater in a confined aquifer. Source: Environment Canada
A confined aquifer tends to be depleted from groundwater pumping more quickly than an unconfined aquifer, assuming similar
aquifer properties and precipitation levels. This is because confined aquifers have smaller recharge areas, which may be far from
the pumping well. Conversely, an unconfined aquifer tends to be more susceptible to pollution because it is hydrologically
connected to the surface, which is the source of most pollution.
Groundwater and surface water (rivers, lakes, swamps, and reservoirs) are strongly interrelated because both are part of the same
overall resource. Major groundwater removal (from pumping or drought) can lower the levels of surface water and vice versa. We
can define two types of streams: gaining (effluent) streams and losing (influent) streams (see Figure 5.2.14). Gaining streams tend
to be perennial (flow year round), are characteristic of humid climates, have the water table sloping towards the river, and therefore
gain water from groundwater discharge. Losing streams tend to be ephemeral (flow only after significant rain), are characteristic of
arid climates, are located above the water table (which slopes away from the river), and therefore lose water to groundwater
recharge. Pollution that is dumped into a losing stream will tend to move into the ground and could also contaminate local
groundwater.
5.2.8 [Link]
Figure 5.2.14 Interaction of Streams and Ground Water. A) Gaining stream where water table slopes toward river and groundwater
discharges into river, B) Losing stream where water table slopes away from river and river water discharges into groundwater, C)
Losing stream where water table is separated from and below river. Source: United States Geological Survey
5.2.9 [Link]
Figure 5.2.15 Trends in Total Water Withdrawals by Water-use Category, 1950-2005. Trends in total water withdrawals in the U.S.
from 1950 to 2005 by water use category, including bars for thermoelectric power, irrigation, public water supply, and rural
domestic and livestock. Thin blue line represents total water withdrawals using vertical scale on right. Source: United States
Geological Survey
Figure 5.2.16 Trends in Source of Fresh Water Withdrawals in the U.S. from 1950 to 2005. Trends in source of fresh water
withdrawals in the U.S. from 1950 to 2005, including bars for surface water, groundwater, and total water. Red line gives U.S.
population using vertical scale on right. Source: United States Geological Survey
In contrast to trends in the U.S., global total water use is steadily increasing at a rate greater than world population growth (see
Figure 5.2.17). During the twentieth century global population tripled and water demand grew by a factor of six. The increase in
global water demand beyond the rate of population growth is due to improved standard of living without an offset by water
conservation. Increased production of goods and energy entails a large increase in water demand. The major global offstream water
uses are irrigation (68%), public supply (21%), and industry (11%).
5.2.10 [Link]
Figure 5.2.17 Trends in World Water Use from 1900 to 2000 and Projected to 2025. For each water major use category, including
trends for agriculture, domestic use, and industry. Darker colored bar represents total water extracted for that use category and
lighter colored bar represents water consumed (i.e., water that is not quickly returned to surface water or groundwater system) for
that use category. Source: Igor A. Shiklomanow, State Hydrological Institute (SHI, St. Petersburg) and United Nations Educational,
Scientific and Cultural Organisation (UNESCO, Paris), 1999
5.2.11 [Link]
Figure 5.2.18 Formation of a Cone of Depression around a Pumping Water Well. Source: Fayette County Groundwater
Conservation District, TX
Figure 5.2.19 Drop in Water Table in a Confined Aquifer in the Area of Chicago, Illinois and Milwaukee, Wisconsin, U.S. from
1864 - 1980. Source: United States Geological Survey
Another water resource problem associated with groundwater mining is saltwater intrusion, where overpumping of fresh water
aquifers near ocean coastlines causes saltwater to enter fresh water zones. Saltwater intrusion is a significant problem in many
coastal areas of the U.S. including Long Island, New York; Cape Cod, Massachusetts; and southeastern and Gulf Coastal states.
The drop of the water table around a cone of depression in an unconfined aquifer can change the regional groundwater flow
direction, which could send nearby pollution toward the pumping well instead of away from it. Finally, problems of subsidence
(gradual sinking of the land surface over a large area) and sinkholes (rapid sinking of the land surface over a small area) can
develop due to a drop in the water table.
5.2.12 [Link]
significantly worse than water stress levels in 1995. In general, water stress is greatest in areas with very low precipitation (major
deserts) or large population density (e.g., India) or both. Future global warming could worsen the water crisis by shifting
precipitation patterns away from humid areas and by melting mountain glaciers that recharge rivers downstream. Melting glaciers
will also contribute to rising sea level, which will worsen saltwater intrusion in aquifers near ocean coastlines. Compounding the
water crisis is the issue of social injustice; poor people generally get less access to clean water and commonly pay more for water
than wealthy people.
Figure 5.2.20 Countries Facing Water Stress in 1995 and Projected in 2025. Water stress is defined as having a high percentage of
water withdrawal compared to total available water in the area. Source: Philippe Rekacewicz (Le Monde diplomatique), February
2006
According to a 2006 report by the United Nations Development Programme, in 2005, 700 million people (11% of the world’s
population) lived under water stress with a per capita water supply below 1,700 m3/year2 (Watkins, 2006). Most of them live in the
Middle East and North Africa. By 2025, the report projects that more than 3 billion people (about 40% of the world’s population)
will live in water-stressed areas with the large increase coming mainly from China and India. The water crisis will also impact food
production and our ability to feed the ever-growing population. We can expect future global tension and even conflict associated
with water shortages and pollution. Historic and future areas of water conflict include the Middle East (Euphrates and Tigris River
conflict among Turkey, Syria, and Iraq; Jordan River conflict among Israel, Lebanon, Jordan, and the Palestinian territories), Africa
(Nile River conflict among Egypt, Ethiopia, and Sudan), Central Asia (Aral Sea conflict among Kazakhstan, Uzbekistan,
Turkmenistan, Tajikistan, and Kyrgyzstan), and south Asia (Ganges River conflict between India and Pakistan).
5.2.13 [Link]
Figure 5.2.21 Hoover Dam, Nevada, U.S.- Hoover Dam, Nevada, U.S.. Behind the dam is Lake Mead, the largest reservoir in U.S..
White band reflects the lowered water levels in the reservoir due to drought conditions from 2000 - 2010. Source: Cygnusloop99 at
Wikimedia Commons
Figure 5.2.22 The California Aqueduct. California Aqueduct in southern California, U.S. Source: David Jordan at [Link]
5.2.14 [Link]
Figure 5.2.23 Map of California Aqueducts. Map of California aqueducts that bring water to southern California from central and
northern California and from the Colorado River to the east. Source: Central Basin Municipal Water District
The Colorado River, probably the most exploited river in the U.S., has many dams, some huge reservoirs, and several large
aqueducts so that it can provide large amounts of fresh water to 7 states in the arid southwestern U.S. and Mexico. The primary use
for the water is for a few large cities (Las Vegas, Phoenix, and Tuscon) and irrigation. Allocation of Colorado River water is strictly
regulated. Fortunately, not all states use all of their water allocation because the total amount of allocated water is more than the
typical Colorado River discharge. Colorado River water gets so saline due to evaporation along its course that the U.S. was forced
to build a desalination plant near the border with Mexico so that it could be used for drinking and irrigation. The wetlands of the
Colorado River delta and its associated ecosystem have been sadly degraded by the water overuse; some years, no river flow even
reaches the ocean.
One method that actually can increase the amount of fresh water on Earth is desalination, which involves removing dissolved salt
from seawater or saline groundwater. There are several ways to desalinate seawater including boiling, filtration, electrodialysis, and
freezing. All of these procedures are moderately to very expensive and require considerable energy input, making the produced
water much more expensive than fresh water from conventional sources. In addition, the processes create highly saline wastewater,
which must be disposed of. Desalination is most common in the Middle East, where energy from oil is abundant but water is
scarce.
Conservation means using less water and using it more efficiently. Around the home, conservation can involve both engineered
features, such as high-efficiency clothes washers and low-flow showers and toilets, as well as behavioral decisions, such as
growing native vegetation that require little irrigation in desert climates, turning off the water while you brush your teeth, and
fixing leaky faucets. Rainwater harvesting involves catching and storing rainwater for reuse before it reaches the ground. Efficient
irrigation is extremely important because irrigation accounts for a much larger water demand than public water supply. Water
conservation strategies in agriculture include growing crops in areas where the natural rainfall can support them, more efficient
irrigation systems such as drip systems that minimize losses due to evaporation, no-till farming that reduces evaporative losses by
covering the soil, and reusing treated wastewater from sewage treatment plants. Recycled wastewater has also been used to
recharge aquifers. There are a great many other specific water conservation strategies. Sustainable solutions to the water crisis must
use a variety of approaches but they should have water conservation as a high priority.
Review Questions
1. What is the water cycle and why is it important to fresh water resources?
2. What are the relative merits of using surface water vs. groundwater as a water resource?
3. What should society learn from the case history of the Aral Sea?
4. Why is society facing a crisis involving water supply and how can we solve it?
5.2.15 [Link]
References
Watkins, K. (2006). Beyond scarcity: Power, poverty and the global water crisis. Human Development Report 2006, United Nations
Development Programme. Retrieved from [Link]
Footnotes
1. Groundwater is the name for water in the saturated zone and soil moisture describes water in the unsaturated zone. Therefore,
groundwater is the underground water resource used by society but soil moisture is the principal water supply for most plants
and is an important factor in agricultural productivity.
2. Although 1,700 m3/year sounds like a lot of water for every person, it is the minimum amount that hydrologists consider is
needed to grow food, support industry, and maintain the environment in general.
Glossary
aqueduct
An aqueduct is a water supply or navigable channel constructed to convey water. In modern engineering, the term is used for
any system of pipes, ditches, canals, tunnels, and other structures used for this purpose.
aquifer
Rock or sediment that is capable of supplying groundwater from a well at a useful rate.
aquitard
Earth material with low hydraulic conductivity.
artesian well
Water well drilled into a confined aquifer where the water level in the well moves above the local water table.
condensation
Change in the physical state of water where it goes from gas to liquid.
cone of depression
A localized drop in the water table around a pumping well.
confined aquifer
An aquifer that is bounded by aquitards below and above.
dam
A barrier built across a river to obstruct the flow of water.
desalination
Removing dissolved salt from seawater or saline groundwater.
discharge area
Location on Earth where groundwater leaves the groundwater flow system.
drainage basin
Geographic area drained by a river and its tributaries.
evaporation
Where water changes from liquid to gas at ambient temperatures.
groundwater
5.2.16 [Link]
Water located in small spaces between mineral grains and fractures in subsurface rock or sediment.
groundwater mining
A depletion in groundwater resources caused by a large number of water wells that pumped water for a long time.
permeability
Measure of the speed that groundwater can flow through rock or sediment.
pore space
Small spaces between mineral grains in subsurface rock or sediment.
porosity
Percentage of pore space in rock or sediment.
rainwater harvesting
Catching and storing rainwater for reuse before it reaches the ground.
recharge area
Location on Earth where surface water infiltrates into the ground rather than runs off into rivers or evaporates.
reservoir
Large artificial lake used as a source of water.
river discharge
Volume of water moving through a river channel over time.
saltwater intrusion
Saltwater that enters an aquifer due to overpumping of freshwater aquifers near ocean coastlines.
saturated zone
Subsurface area where groundwater completely fills pore spaces in rock or sediment.
soil moisture
Water in the unsaturated zone.
spring
River that emerges from underground due to an abrupt intersection of the water table with the land surface.
surface runoff
Unchannelized overland flow of water.
transpiration
Loss of water by plants to the atmosphere.
unconfined aquifer
Aquifer with no aquitard above it.
unsaturated zone
Subsurface area where pore spaces contain only air and water films on mineral grains.
5.2.17 [Link]
water conservation
Using less water and using it more efficiently
water crisis
A global situation where people in many areas lack access to sufficient water or clean water or both.
water cycle
The continuous movement of water through water reservoirs located on, above, and below Earth’s surface.
water table
Interface between the unsaturated zone and saturated zone.
This page titled 5.2: Water Cycle and Fresh Water Supply is shared under a CC BY license and was authored, remixed, and/or curated by
Heriberto Cabezas (GALILEO Open Learning Materials) .
5.2.18 [Link]
5.3: Case Study- The Aral Sea - Going, Going, Gone
The Aral Sea is a lake located east of the Caspian Sea between Uzbekistan and Kazakhstan in central Asia (see Figure 5.3.1). This
area is part of the Turkestan desert, which is the fourth largest desert in the world; it is produced from a rain shadow effect by
Afghanistan's high mountains to the south. Due to the arid and seasonally hot climate there is extensive evaporation and limited
surface waters in general. Summer temperatures can reach 60° C (140° F)! The water supply to the Aral Sea is mainly from two
rivers, the Amu Darya and Syr Darya, which carry snowmelt from mountainous areas. In the early 1960s the then-Soviet Union
diverted the Amu Darya and Syr Darya Rivers for irrigation of one of the driest parts of Asia to produce rice, melons, cereals, and
especially cotton. The Soviets wanted cotton or “white gold” to become a major export. They were successful and today
Uzbekistan is one of the world's largest exporters of cotton. Unfortunately this action essentially eliminated any river inflow to the
Aral Sea and caused it to disappear almost completely.
Figure 5.3.1 Map of Aral Sea Area. Map shows lake size in 1960 and political boundaries of 2011. Countries in yellow are at least
partially in Aral Sea drainage basin. Source: Wikimedia Commons
In 1960 Aral Sea was the fourth largest inland water body; only the Caspian Sea, Lake Superior, and Lake Victoria were larger.
Since then, it has progressively shrunk due to evaporation and lack of recharge by rivers (see Figure 5.3.2). Before 1965 the Aral
Sea received 20–60 km3 of fresh water per year from rivers and by the early 1980s it received none. By 2007 the Aral Sea shrank to
about 10% of its original size and its salinity increased from about 1% dissolved salt to about 10% dissolved salt, which is 3 times
more saline than seawater. These changes caused an enormous environmental impact. A once thriving fishing industry is dead as
are the 24 species of fish that used to live there; the fish could not adapt to the more saline waters. The current shoreline is tens of
kilometers from former fishing towns and commercial ports. Large fishing boats lie in the dried up lakebed of dust and salt (see
Figure 5.3.2). A frustrating part of the river diversion project is that many of the irrigation canals were poorly built, allowing
abundant water to leak or evaporate. An increasing number of dust storms blow salt, pesticides, and herbicides into nearby towns
causing a variety of respiratory illnesses including tuberculosis.
5.3.1 [Link]
Figure 5.3.2 Shrinking Aral Sea. Blue area gives size of Aral Sea in 1960, 1970, 1980, 1990, 2000, 2004, 2008, and 2009 Source:
NordNordWest at Wikimedia Commons
Figure 5.3.3 An Abandoned Ship. This abandoned ship lies in a dried up lake bed that was the Aral Sea near Aral, Kazakhstan
Source: Staecker at Wikimedia Commons
The wetlands of the two river deltas and their associated ecosystems have disappeared. The regional climate is drier and has greater
temperature extremes due to the absence of moisture and moderating influence from the lake. In 2003 some lake restoration work
began on the northern part of the Aral Sea and it provided some relief by raising water levels and reducing salinity somewhat. The
southern part of the Aral Sea has seen no relief and remains nearly completely dry. The destruction of the Aral Sea is one of the
planet’s biggest environmental disasters and it is caused entirely by humans. Lake Chad in Africa is another example of a massive
lake that has nearly disappeared for the same reasons as the Aral Sea. Aral Sea and Lake Chad are the most extreme examples of
5.3.2 [Link]
large lakes destroyed by unsustainable diversions of river water. Other lakes that have shrunk significantly due to human diversions
of water include the Dead Sea in the Middle East, Lake Manchar in Pakistan, and Owens Lake and Mono Lake, both in California.
This page titled 5.3: Case Study- The Aral Sea - Going, Going, Gone is shared under a CC BY license and was authored, remixed, and/or curated
by Heriberto Cabezas (GALILEO Open Learning Materials) .
5.3.3 [Link]
5.4: Water Pollution
Learning Objectives
After reading this module, students should be able to
understand the major kinds of water pollutants and how they degrade water quality
understand how and why the lack of safe drinking water in some parts of the world is a major problem
know what sewage treatment does and why it is important
know why it is more difficult to remediate groundwater pollution than surface water pollution
understand how we can work toward solving the crisis involving water pollution
Figure 5.4.1 Proportion of Population by Country Using Improved Drinking Water Sources in 2008. Improved drinking water
sources, e.g., household connections, public standpipes, boreholes, protected dug wells and springs, and rainwater collections, are
defined as those more likely to provide safe water than unimproved water sources, e.g., unprotected wells and springs, vendor-
provided water, bottled water (unless water for other uses is available from an improved source), and tanker truck-provided water.
Source: World Health Organization
5.4.1 [Link]
Figure 5.4.2 Proportion of Population by Country Using Improved Sanitation Facilities in 2008. Improved sanitation facilities, e.g.,
connection to public sewers or septic systems, pour-flush latrines, pit latrines, and ventilated improved pit latrines, are defined as
those more likely to be sanitary than unimproved facilities, e.g., bucket latrines, public latrines, and open pit latrines. Source: World
Health Organization
Figure 5.4.3 Deaths by Country from Diarrhea Caused by Unsafe Water, Unimproved Sanitation, and Poor Hygiene in Children
Less than 5 Years Old, 2004. Source: World Health Organization
Figure 5.4.4 Percentage of Impaired Water Bodies in a Watershed by State in USA Based on US EPA Data in 2000. Map of
watersheds containing impaired water bodies from the U.S. Environmental Protection Agency's 1998 list of impaired waters
Source: U.S. Geological Survey
5.4.2 [Link]
it has a net positive charge on the side with the two hydrogen atoms and a net negative charge on the oxygen side (see Figure
5.4.5). This separation of the electrical charge within a water molecule results in hydrogen bondswith other water molecules,
mineral surfaces (hydrogen bonding produces the water films on minerals in the unsaturated zone of the subsurface), and dissolved
ions (atoms with a negative or positive charge). Many minerals and pollutants dissolve readily in water because water forms
hydration shells (spheres of loosely coordinated, oriented water molecules) around ions.
Figure 5.4.5 Structure of Water, Polar Charge of Water, and Hydrogen Bonds between Water Molecules. Source: Michal Maňas at
Wikimedia Commons
Any natural water contains dissolved chemicals; some of these are important human nutrients, while others can be harmful to
human health. The abundance of a water pollutant is commonly given in very small concentration units such as parts per million
(ppm) or even parts per billion (ppb). An arsenic concentration of 1 ppm means 1 part of arsenic per million parts of water. This is
equivalent to one drop of arsenic in 50 liters of water. To give you a different perspective on appreciating small concentration units,
converting 1 ppm to length units is 1 cm (0.4 in) in 10 km (6 miles) and converting 1 ppm to time units is 30 seconds in a year.
Total dissolved solids (TDS) represent the total amount of dissolved material in water. Average TDS (salinity) values for rainwater,
river water, and seawater are about 4 ppm, 120 ppm, and 35,000 ppm. As discussed in Module 5.2, the most important processes
that affect the salinity of natural waters are evaporation, which distills nearly pure water and leaves the dissolved ions in the
original water, and chemical weathering, which involves mineral dissolution that adds dissolved ions to water. Fresh water is
commonly defined as containing less than either 1,000 or 500 ppm TDS, but the US Environmental Protection Agency (EPA)
recommends that drinking water not exceed 500 ppm TDS or else it will have an unpleasant salty taste.
5.4.3 [Link]
Figure 5.4.6 Water Pollution. Obvious water pollution in the form of floating debris; invisible water pollutants sometimes can be
much more harmful than visible ones. Source: Stephen Codrington at Wikimedia Commons
Figure 5.4.7 Sources of Water Contamination. Sources of some water pollutants and movement of pollutants into different water
reservoirs of the water cycle. Source: U.S. Geological Survey
Pollutants enter water supplies from point sources, which are readily identifiable and relatively small locations, or nonpoint
sources, which are large and more diffuse areas. Point sources of pollution include animal “factory” farms that raise a large number
and high density of livestock such as cows, pigs, and chickens (see Figure 5.4.8) and discharge pipes from a factories or sewage
treatment plants. Combined sewer systems that have a single set of underground pipes to collect both sewage and storm water
runoff from streets for wastewater treatment can be major point sources of pollutants. During heavy rain, storm water runoff may
exceed sewer capacity, causing it to back up and spilling untreated sewage into surface waters (see Figure 5.4.9). Nonpoint sources
of pollution include agricultural fields, cities, and abandoned mines. Rainfall runs over the land and through the ground, picking up
pollutants such as herbicides, pesticides, and fertilizer from agricultural fields and lawns; oil, antifreeze, car detergent, animal
waste, and road salt from urban areas; and acid and toxic elements from abandoned mines. Then, this pollution is carried into
surface water bodies and groundwater. Nonpoint source pollution, which is the leading cause of water pollution in the U.S., is
5.4.4 [Link]
usually much more difficult and expensive to control than point source pollution because of its low concentration, multiple sources,
and much greater volume of water.
Figure 5.4.8 A Commercial Meat Chicken Production House. This chicken factory farm is a possible major point source of water
pollution. Source: Larry Rana at Wikimedia Commons
Figure 5.4.9 Combined Sewer System. A combined sewer system is a possible major point source of water pollution during heavy
rain due to overflow of untreated sewage. During dry weather (and small storms), all flows are handled by the publicly owned
treatment works (POTW). During large storms, the relief structure allows some of the combined stormwater and sewage to be
discharged untreated to an adjacent water body. Source: U.S. Environmental Protection Agency at Wikimedia Commons
C H 2O + O2 → C O2 + H 2O
Too much decaying organic matter in water is a pollutant because it removes oxygen from water, which can kill fish, shellfish, and
aquatic insects. The amount of oxygen used by aerobic (in the presence of oxygen) bacterial decomposition of organic matter is
called biochemical oxygen demand (BOD). The major source of dead organic matter in most natural waters is sewage; grass and
leaves are smaller sources. An unpolluted water body with respect to oxygen is a turbulent river that flows through a natural forest.
Turbulence continually brings water in contact with the atmosphere where the O2 content is restored. The dissolved oxygen content
in such a river ranges from 10 to 14 ppm O2, BOD is low, and clean-water fish, e.g., bass, trout, and perch dominate. A polluted
water body with respect to oxygen is a stagnant deep lake in an urban setting with a combined sewer system. This system favors a
high input of dead organic carbon from sewage overflows and limited chance for water circulation and contact with the
5.4.5 [Link]
atmosphere. In such a lake, the dissolved O2 content is ≤5 ppm O2, BOD is high, and low O2-tolerant fish, e.g., carp and catfish
dominate.
Excessive plant nutrients, particularly nitrogen (N) and phosphorous (P), are pollutants closely related to oxygen-demanding waste.
Aquatic plants require about 15 nutrients for growth, most of which are plentiful in water. N and P are called limiting nutrients,
because they usually are present in water at low concentrations and therefore restrict the total amount of plant growth. This
explains why N and P are major ingredients in most fertilizer. High concentrations of N and P from human sources (mostly
agricultural and urban runoff including fertilizer, sewage, and P-based detergent) can cause cultural eutrophication, which involves
the rapid growth of aquatic plants, particularly algae, called an algal bloom. Thick mats of floating and rooted green or sometimes
red algae (see Figure 5.4.10) create water pollution, damage the ecosystem by clogging fish gills and blocking sunlight, and
damage lake aesthetics by making recreation difficult and creating an eyesore. A small percentage of algal species produce toxins
that can kill fish, mammals, and birds, and may cause human illness; explosive growths of these algae are called harmful algal
blooms (see Figure 5.4.11). When the prolific algal layer dies, it becomes oxygen-demanding waste, which can create very low O2
water (<~2 ppm O2), called hypoxia or dead zone because it causes death to organisms that are unable to leave that environment.
An estimated 50% of lakes in North America, Europe, and Asia are negatively impacted by cultural eutrophication. In addition, the
size and number of marine hypoxic zones have grown dramatically over the past 50 years (see Figure 5.4.12), including a very
large dead zone located offshore Louisiana in the Gulf of Mexico. Cultural eutrophication and hypoxia are difficult to combat,
because they are caused primarily by nonpoint source pollution, which is difficult to regulate, and N and P, which are difficult to
remove from wastewater.
Figure 5.4.10 Algal Bloom in River in Sichuan, China. Algal blooms can present problems for ecosystems and human society.
Source: Felix Andrews via Wikimedia Commons
5.4.6 [Link]
Figure 5.4.11 Harmful Algal Bloom. Harmful algal bloom with deep red color. Source: Kai Schumann via National Oceanic and
Atmospheric Administration
Figure 5.4.12 Aquatic Dead Zones. Zones of hypoxia shown as red circles. Black dots show hypoxia zones of unknown size,
brown shading shows population density, and blue shading shows density of particulate organic carbon, an indicator of organic
productivity. Source: Robert Simmon & Jesse Allen at NASA Earth Observatory via Wikimedia Commons
Pathogens are disease-causing microorganisms, e.g., viruses, bacteria, parasitic worms, and protozoa, which cause a variety of
intestinal diseases such as dysentery, typhoid fever, hepatitis, and cholera. Pathogens are the major cause of the water pollution
crisis discussed at the beginning of this section. Unfortunately nearly a billion people around the world are exposed to waterborne
pathogen pollution daily and around 1.5 million children mainly in underdeveloped countries die every year of waterborne diseases
from pathogens (see Figure 5.4.3). Pathogens enter water primarily from human and animal fecal waste due to inadequate sewage
treatment. In many underdeveloped countries, sewage is discharged into local waters either untreated or after only rudimentary
treatment. In developed countries untreated sewage discharge can occur from overflows of combined sewer systems, poorly
managed livestock factory farms, and leaky or broken sewage collection systems (see Figure 5.4.14). Water with pathogens can be
remediated by adding chlorine or ozone, by boiling, or by treating the sewage in the first place.
5.4.7 [Link]
Figure 5.4.13 Overflowing Sanitary Sewer. A manhole cover blown off by a June 2006 sanitary sewer overflow in Rhode Island.
Source: U.S. Environmental Protection Agency via Wikimedia Commons
Oil spills are another kind of organic pollution. Oil spills can result from supertanker accidents such as the Exxon Valdez in 1989,
which spilled 10 million gallons of oil into the rich ecosystem of offshore south Alaska and killed massive numbers of animals. The
largest marine oil spill was the Deepwater Horizon disaster, which began with a natural gas explosion (see Figure 5.4.14 ) at an oil
well 65 km offshore of Louisiana and flowed for 3 months in 2010, releasing an estimated 200 million gallons of oil. The worst oil
spill ever occurred during the Persian Gulf war of 1991, when Iraq deliberately dumped approximately 200 million gallons of oil in
offshore Kuwait and set more than 700 oil well fires that released enormous clouds of smoke and acid rain for over nine months.
During an oil spill on water, oil floats to the surface because it is less dense than water, and the lightest hydrocarbons evaporate,
decreasing the size of the spill but polluting the air. Then, bacteria begin to decompose the remaining oil, in a process that can take
many years. After several months only about 15% of the original volume may remain, but it is in thick asphalt lumps, a form that is
particularly harmful to birds, fish, and shellfish. Cleanup operations can include skimmer ships that vacuum oil from the water
surface (effective only for small spills), controlled burning (works only in early stages before the light, ignitable part evaporates but
also pollutes the air), dispersants (detergents that break up oil to accelerate its decomposition, but some dispersants may be toxic to
the ecosystem), and bioremediation (adding microorganisms that specialize in quickly decomposing oil, but this can disrupt the
natural ecosystem).
Figure 5.4.14 Deepwater Horizon Explosion. Boats fighting the fire from an explosion at the Deepwater Horizon drilling rig in
Gulf of Mexico offshore Louisiana on April 20, 2010. Source: United States Coast Guard via Wikimedia Commons
Toxic chemicals involve many different kinds and sources, primarily from industry and mining. General kinds of toxic chemicals
include hazardous chemicals, which are a wide variety of synthetic organic and inorganic chemicals such as acids, bases, cyanide,
and a class of compounds called persistent organic pollutants that includes DDT (pesticide), dioxin (herbicide by-product), and
5.4.8 [Link]
PCBs (polychlorinated biphenyls, which were used as a liquid insulator in electric transformers). Persistent organic pollutants are
long-lived in the environment, accumulate through the food chain (bioaccumulation), and can be toxic. Another category of toxic
chemicals includes radioactive materials such as cesium, iodine, uranium, and radon gas, which can result in long-term exposure to
radioactivity if it gets into the body. A final group of toxic chemicals is heavy metals such as lead, mercury, arsenic, cadmium, and
chromium, which can accumulate through the food chain. Heavy metals are commonly produced by industry and at metallic ore
mines. Arsenic and mercury are discussed in more detail below. The US EPA regulates 83 contaminants in drinking water to ensure
a safe public water supply. Similarly, at the international level the World Health Organization has drinking water standards for a
variety of contaminants.
Arsenic (As) has been famous as an agent of death for many centuries. In large doses arsenic causes cancer and can be fatal. Only
recently have scientists recognized that health problems can be caused by drinking small arsenic concentrations in water over a
long time. It attacks the central nervous system and can damage the respiratory system, bladder, lungs, liver, and kidneys. It enters
the water supply naturally from weathering of As-rich minerals and from human activities such as coal burning and smelting of
metallic ores. The worst case of arsenic poisoning occurred in the densely populated impoverished country of Bangladesh, which
had experienced 100,000s of deaths from diarrhea and cholera each year from drinking surface water contaminated with pathogens
due to improper sewage treatment. In the 1970s the United Nations provided aid for millions of shallow water wells, which resulted
in a dramatic drop in pathogenic diseases. Unfortunately, many of the wells produced water naturally rich in arsenic. Tragically,
there are an estimated 77 million people (about half of the population) who inadvertently may have been exposed to toxic levels of
arsenic in Bangladesh as a result. The World Health Organization has called it the largest mass poisoning of a population in history.
Mercury (Hg) is used in a variety of electrical products, such as dry cell batteries, fluorescent light bulbs, and switches, as well as
in the manufacture of paint, paper, vinyl chloride, and fungicides. In the methylmercury form (CH3Hg+) it is highly toxic; ≥ 1 ppb
of methylmercury represents water contaminated with mercury. Mercury concentrates in the food chain, especially in fish, in a
process caused biomagnification (see Sidebar Biomagnification). It acts on the central nervous system and can cause loss of sight,
feeling, and hearing as well as nervousness, shakiness, and death. Like arsenic, mercury enters the water supply naturally from
weathering of Hg-rich minerals and from human activities such as coal burning and metal processing. A famous mercury poisoning
case in Minamata, Japan involved methylmercury-rich industrial discharge that caused high Hg levels in fish. People in the local
fishing villages ate fish up to three times per day for over 30 years, which resulted in over 2,000 deaths. During that time the
responsible company and national government did little to mitigate, help alleviate, or even acknowledge the problem.
Biomagnification
Biomagnification represents the processes in an ecosystem that cause greater concentrations of a chemical, such as methylmercury,
in organisms higher up the food chain. Mercury and methylmercury are present in only very small concentrations in seawater;
however, at the base of the food chain algae absorb methylmercury. Then, small fish eat the algae, large fish and other organisms
higher in the food chain eat the small fish, and so on. Fish and other aquatic organisms absorb methylmercury rapidly but eliminate
it slowly from the body. Therefore, each step up the food chain increases the concentration from the step below (see Figure 5.4.15).
Largemouth bass can concentrate methylmercury up to 10 million times over the water concentration and fish-eating birds can
concentrate it even higher. Other chemicals that exhibit biomagnification are DDT, PCBs, and arsenic.
illustration of
Biomagnification
Figure 5.4.15 An illustrative example of biomagnification of mercury from water through the food chain and into a bird's egg.
Source: U.S. Geological Survey
Other water pollutants include sediment and heat. Muddy water is bad for drinking but even worse for underwater plants that need
sunlight for photosynthesis. Much of the sediment in water bodies is derived from the erosion of soil, so it also represents a loss of
agricultural productivity. Thermal pollution involves the release of heated waters from power plants and industry to surface water,
causing a drop in the dissolved O2 content, which can stress fish.
Hard water contains abundant calcium and magnesium, which reduces its ability to develop soapsuds and enhances scale (calcium
and magnesium carbonate minerals) formation on hot water equipment. Water softeners remove calcium and magnesium, which
allows the water to lather easily and resist scale formation. Hard water develops naturally from the dissolution of calcium and
magnesium carbonate minerals in soil; it does not have negative health effects in people.
Groundwater pollution can occur from underground sources and all of the pollution sources that contaminate surface waters.
Common sources of groundwater pollution are leaking underground storage tanks for fuel, septic tanks, agricultural activity, and
landfills. Common groundwater pollutants include nitrate, pesticides, volatile organic compounds, and petroleum products.
5.4.9 [Link]
Polluted groundwater can be a more serious problem than polluted surface water because the pollution in groundwater may go
undetected for a long time because usually it moves very slowly. As a result, the pollution in groundwater may create a
contaminant plume, a large body of flowing polluted groundwater (see Figure 5.4.15), making cleanup very costly. By the time
groundwater contamination is detected, the entity responsible for the pollution may be bankrupt or nonexistent. Another
troublesome feature of groundwater pollution is that small amounts of certain pollutants, e.g., petroleum products and organic
solvents, can contaminate large areas. In Denver, Colorado 80 liters of several organic solvents contaminated 4.5 trillion liters of
groundwater and produced a 5 km long contaminant plume. Most groundwater contamination occurs in shallow, unconfined
aquifers located near the contamination source. Confined aquifers are less susceptible to pollution from the surface because of
protection by the confining layer. A major threat to groundwater quality is from underground fuel storage tanks. Fuel tanks
commonly are stored underground at gas stations to reduce explosion hazards. Before 1988 in the U.S. these storage tanks could be
made of metal, which can corrode, leak, and quickly contaminate local groundwater. Now, leak detectors are required and the metal
storage tanks are supposed to be protected from corrosion or replaced with fiberglass tanks. Currently there are around 600,000
underground fuel storage tanks in the U.S. and over 30% still do not comply with EPA regulations regarding either release
prevention or leak detection.
Figure 5.4.16 Contaminant Plume in Groundwater. Mapping how a contaminant plume will migrate once it reaches groundwater
requires understanding of the pollutant's chemical properties, local soil characteristics, and how permeable the aquifer is. Source:
United States Geological Survey
5.4.10 [Link]
underground caverns and also use abandoned rock quarries to hold storm sewer overflow. After the rain stops, the stored water goes
to the sewage treatment plant for processing.
Figure 5.4.17 Steps at a Sewage Treatment Plant. The numerous processing steps at a conventional sewage treatment plant include
pretreatment (screening and removal of sand and gravel), primary treatment (settling or floatation to remove organic solids, fat, and
grease), secondary treatment (aerobic bacterial decomposition of organic solids), tertiary treatment (bacterial decomposition of
nutrients and filtration), disinfection (treatment with chlorine, ozone, ultraviolet light, or bleach), and either discharge to surface
waters (usually a local river) or reuse for some other purpose, such as irrigation, habitat preservation, and artificial groundwater
recharge. Source: Leonard [Link] Wikipedia
A septic tank system is an individual sewage treatment system for homes in rural and even some urban settings. The basic
components of a septic tank system (see Figure 5.4.18) include a sewer line from the house, a septic tank (a large container where
sludge settles to the bottom and microorganisms decompose the organic solids anaerobically), and the drain field (network of
perforated pipes where the clarified water seeps into the soil and is further purified by bacteria). Water pollution problems occur if
the septic tank malfunctions, which usually occurs when a system is established in the wrong type of soil or maintained poorly.
Figure 5.4.18 Septic System. Septic tank system for sewage treatment. Source: United States Geological Survey
For many developing countries, financial aid is necessary to build adequate sewage treatment facilities; however, the World Health
Organization estimates an estimated cost savings of between $3 and $34 for every $1 invested in clean water delivery and
sanitation (Water for Life, 2005). The cost savings are from health care savings, gains in work and school productivity, and deaths
prevented. Simple and inexpensive techniques for treating water at home include chlorination, filters, and solar disinfection.
5.4.11 [Link]
Another alternative is to use constructed wetlands technology (marshes built to treat contaminated water), which is simpler and
cheaper than a conventional sewage treatment plant.
Bottled water is not a sustainable solution to the water crisis, despite exponential growth in popularity in the U.S. and the world.
Bottled water is not necessarily any safer than the U.S. public water supply, it costs on average about 700 times more than U.S. tap
water, and every year it uses approximately 200 billion plastic and glass bottles that have a relatively low rate of recycling.
Compared to tap water, it uses much more energy, mainly in bottle manufacturing and long-distance transportation. If you don’t
like the taste of your tap water, then please use a water filter instead of bottled water!
Figure 5.4.19 Storm Drain. Curbside storm drain receiving urban runoff. Source: By Robert Lawton via Wikimedia Commons
Additional sustainable solutions to the water pollution crisis include legislation to eliminate or greatly reduce point sources of water
pollution. In the U.S., the Clean Water Act of 1972 and later amendments led to major improvements in water quality (see Sidebar
Clean Water Act). Nonpoint sources of water pollution, e.g., agricultural runoff and urban runoff (see Figure 5.4.19), are much
harder to regulate because of their widespread, diffuse nature. There are many construction and agricultural practices that reduce
polluted runoff including no-till farming and sediment traps. Artificial aeration or mechanical mixing can remediate lakes with
oxygen depletion. Specific things that we can do to reduce urban runoff include the following: keep soil, leaves, and grass clippings
off driveways, sidewalks, and streets; don't pour used motor oil, antifreeze, paints, pesticides, or any household hazardous chemical
down the storm sewer or drain; recycle used motor oil; use hazardous waste disposal programs offered by the community; compost
your organic waste; don't use fertilizers and herbicides on your lawn; and flush pet waste down the toilet.
5.4.12 [Link]
Figure 5.4.20 Cuyahoga River on Fire. Source: National Oceanic and Atmospheric
Sometimes slow flow through a soil can naturally purify groundwater because some pollutants, such as P, pesticides, and heavy
metals, chemically bind with surfaces of soil clays and iron oxides. Other pollutants are not retained by soil particles: These include
N, road salt, gasoline fuel, the herbicide atrazine, tetrachloroethylene (a carcinogenic cleaning solvent used in dry cleaning), and
vinyl chloride. In other cases, slow groundwater flow can allow bacteria to decompose dead organic matter and certain pesticides.
There are many other ways to remediate polluted groundwater. Sometimes the best solution is to stop the pollution source and
allow natural cleanup. Specific treatment methods depend on the geology, hydrology, and pollutant because some light
contaminants flow on top of groundwater, others dissolve and flow with groundwater, and dense contaminants can sink below
groundwater. A common cleanup method called pump and treat involves pumping out the contaminated groundwater and treating it
by oxidation, filtration, or biological methods. Sometimes soil must be excavated and sent to a landfill. In-situ treatment methods
include adding chemicals to immobilize heavy metals, creating a permeable reaction zone with metallic iron that can destroy
organic solvents, or using bioremediation by adding oxygen or nutrients to stimulate growth of microorganisms.
Review Questions
1. What are the major kinds of water pollutants and how do they degrade water quality?
2. How would you rank the water pollution problems described in this chapter? Why?
3. Why is untreated sewage such an important water pollutant to remediate?
4. What should society learn from the case history of Love Canal?
5. Why are people facing a crisis involving water pollution and how can we solve it?
References
Water for Life: Making it Happen (2005) World Health Organization and UNICEF. Retrieved from
[Link]
World Health Statistics (2010) World Health Organization. Retrieved from
[Link]
Glossary
arsenic
A type of water pollutant that can be fatal in large doses and can cause health problems in small doses over a long time.
bioremediation
Method of groundwater remediation involving the addition oxygen or nutrients. to stimulate growth of microorganisms, which
decompose an organic pollutant.
bottled water
5.4.13 [Link]
Drinking water packaged in plastic bottles or glass bottles, bottled water is not a sustainable solution to the water crisis because
of the nonrenewable energy and material resources involved in manufacturing and transporting it.
constructed wetland
Marsh built to treat contaminated water.
contaminant plume
A large body of flowing polluted groundwater.
cultural eutrophication
Rapid aquatic plant growth, particularly algae, in a surface water body.
hard water
Water with abundant calcium and magnesium, which reduces its ability to develop soapsuds and enhances scale; hard water
does not have negative health effects in people.
heat
A type of water pollutant that causes a drop in the dissolved oxygen content, which can stress fish.
heavy metal
A type of water pollutant involving elements such as lead, mercury, arsenic, cadmium, and chromium, which can accumulate
through the food chain.
hypoxia
Very low oxygen water due to prolific growth of algae, algal death, and then decomposition, also called dead zone.
mercury
A type of water pollutant that acts on the central nervous system and can cause loss of sight, feeling, and hearing as well as
nervousness, shakiness, and death.
oil spill
A type of organic water pollutant involving the release of liquid petroleum into the environment due to human activity.
oxygen-demanding waste
A type of water pollutant involving abundant dead organic matter.
pathogens
Disease-causing microorganisms, e.g., viruses, bacteria, parasitic worms, and protozoa, which cause a variety of intestinal
diseases such as dysentery, typhoid fever, hepatitis, and cholera.
5.4.14 [Link]
sediment
A type of water pollutant that degrades drinking water and can kill underwater plants that need sunlight for photosynthesis.
sludge
Concentrated organic solid produced during primary and secondarytreatment of sewage treatment.
solvent
Capacity of a liquid such as water to dissolve soluble minerals.
toxic chemical
A type of organic water pollutant involving chemicals with a severe human health risk.
water pollution
Contamination of water by an excess amount of a substance that can cause harm to human beings and the ecosystem.
This page titled 5.4: Water Pollution is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto Cabezas
(GALILEO Open Learning Materials) .
5.4.15 [Link]
5.5: Case Study- The Love Canal Disaster
One of the most famous and important examples of groundwater pollution in the U.S. is the Love Canal tragedy in Niagara Falls,
New York. It is important because the pollution disaster at Love Canal, along with similar pollution calamities at that time (Times
Beach, Missouri and Valley of Drums, Kentucky), helped to create Superfund, a federal program instituted in 1980 and designed to
identify and clean up the worst of the hazardous chemical waste sites in the U.S.
Love Canal is a neighborhood in Niagara Falls named after a large ditch (approximately 15 m wide, 3–12 m deep, and 1600 m
long) that was dug in the 1890s for hydroelectric power. The ditch was abandoned before it actually generated any power and went
mostly unused for decades, except for swimming by local residents. In the 1920s Niagara Falls began dumping urban waste into
Love Canal, and in the 1940s the U.S. Army dumped waste from World War II there, including waste from the frantic effort to
build a nuclear bomb. Hooker Chemical purchased the land in 1942 and lined it with clay. Then, the company put into Love Canal
an estimated 21,000 tons of hazardous chemical waste, including the carcinogens benzene, dioxin, and PCBs in large metal barrels
and covered them with more clay. In 1953, Hooker sold the land to the Niagara Falls school board for $1, and included a clause in
the sales contract that both described the land use (filled with chemical waste) and absolved them from any future damage claims
from the buried waste. The school board promptly built a public school on the site and sold the surrounding land for a housing
project that built 200 or so homes along the canal banks and another 1,000 in the neighborhood (see Figure 5.5.1). During
construction, the canal’s clay cap and walls were breached, damaging some of the metal barrels.
5.5.1 [Link]
government paid $8 million for pollution by the Army. The total clean up cost was estimated to be $275 million. The only good
thing about the Love Canal tragedy is that it helped to create Superfund, which has analyzed tens of thousands of hazardous waste
sites in the U.S. and cleaned up hundreds of the worst ones. Nevertheless, over 1,000 major hazardous waste sites with a significant
risk to human health or the environment are still in the process of being cleaned.
Glossary
superfund
A federal program created in 1980 and designed to identify and clean up the worst of the hazardous chemical waste sites in the
U.S.
This page titled 5.5: Case Study- The Love Canal Disaster is shared under a CC BY license and was authored, remixed, and/or curated by
Heriberto Cabezas (GALILEO Open Learning Materials) .
5.5.2 [Link]
5.6: Mineral Resources- Formation, Mining, Environmental Impact
Learning Objectives
After reading this module, students should be able to
know the importance of minerals to society
know factors that control availability of mineral resources
know why future world mineral supply and demand is an important issue
understand the environmental impact of mining and processing of minerals
understand how we can work toward solving the crisis involving mineral supply
Importance of Minerals
Mineral resources are essential to our modern industrial society and they are used everywhere. For example, at breakfast you drink
some juice in a glass (made from melted quartz sand), eat from a ceramic plate (created from clay minerals heated at high
temperatures), sprinkle salt (halite) on your eggs, use steel utensils (from iron ore and other minerals), read a magazine (coated
with up to 50% kaolinite clay to give the glossy look), and answer your cellphone (containing over 40 different minerals including
copper, silver, gold, and platinum). We need minerals to make cars, computers, appliances, concrete roads, houses, tractors,
fertilizer, electrical transmission lines, and jewelry. Without mineral resources, industry would collapse and living standards would
plummet. In 2010, the average person in the U.S. consumed more than 16,000 pounds of mineral resources1 (see Table 5.6.1). With
an average life expectancy of 78 years, that translates to about1.3 million pounds of mineral resources over such a person’s
lifetime. Here are a few statistics that help to explain these large values of mineral use: an average American house contains about
250,000 pounds of minerals (see Figure 5.6.1 for examples of mineral use in the kitchen), one mile of Interstate highway uses 170
million pounds of earth materials, and the U.S. has nearly 4 million miles of roads. All of these mineral resources are
nonrenewable, because nature usually takes hundreds of thousands to millions of years to produce mineral deposits. Early hominids
used rocks as simple tools as early as 2.6 million years ago. At least 500,000 years ago prehistoric people used flint (fine-grained
quartz) for knives and arrowheads. Other important early uses of minerals include mineral pigments such as manganese oxides and
iron oxides for art, salt for food preservation, stone for pyramids, and metals such as bronze (typically tin and copper), which is
stronger than pure copper and iron for steel, which is stronger than bronze.
Figure 5.6.1 Mineral Use in the Kitchen. Source: U.S. Geological Survey
Table 5.6.1 Per capita consumption of nonenergy related minerals and metals in the U.S. for 2010 and for a lifetime of 78.3 years assuming 2010
mineral consumption rates Sources: US Geological Survey, National Mining Association, and U.S. Census Bureau
Per Capita Consumption of Minerals – Per Capita Consumption of Minerals -
Mineral
2010 (Pounds per Person) Lifetime (Pounds Per Person)
5.6.1 [Link]
Copper 12 939.6
Lead 11 861
Manganese 5 392
Potash 37 2,897
Sulfur 86 6,734
Zinc 6 470
Figure 5.6.2 Giant Gypsum Crystals. Giant gypsum crystals in the Cave of Crystals in Naica, Mexico. There are crystals up to 11
m long in this cave, which is located about 1 km underground. Source: National Geographic via Wikipedia
Table 5.6.2 Enrichment Factor. Approximate enrichment factors of selected metals needed before profitable mining is possible. Source: US
Geological Survey Professional Paper 820, 1973
Average Concentration in Concentration Needed for Approximate Enrichment
Metal
Crust (%) Economic Mine (%) Factor
Aluminum 8 35 4
Iron 5
Copper 0.005
5.6.2 [Link]
Lead 0.0015 4 2,500
Figure \(\PageIndex{3}\) A Diamond Mine Udachnaya Pipe, an open-pit diamond mine in Russia, is more than 600 meters (1,970
ft) deep, making it the third deepest open-pit mine in the world. Source: Stapanov Alexander via Wikimedia Commons
5.6.3 [Link]
Figure 5.6.3 Black Smoker. A billowing discharge of superheated mineral-rich water at an oceanic ridge, in the Atlantic Ocean.
Black “smoke” is actually from metallic sulfide minerals that form modern ore deposits. Source: P. Rona of U.S. National Oceanic
and Atmospheric Administration via Wikimedia Commons
5.6.4 [Link]
Figure 5.6.4 Open pit mines. Bingham Canyon copper mine in Utah, USA. At 4 km wide and 1.2 km deep, it is the world’s deepest
open-pit mine. It began operations in 1906. Source: Tim Jarrett via Wikimedia Commons
Figure 5.6.5 Demand for Nonfuel Minerals Materials. US mineral consumption from 1900 - 2006, excluding energy-related
minerals Source: U.S. Geological Survey
Table 5.6.3 Strategic MineralsUses, world production in 2010, and estimated projected lifetime of reserves (ore that is profitable to mine under
current conditions) for selected minerals Source: US Geological Survey Mineral Commodity Summaries, 2011
5.6.5 [Link]
2010 2010
Estimated Life of
Mineral Uses Production(thousands Reserves(thousands of
Reserves(years)
of metric tons) metric tons)
catalysts, alloys,
Rare earths electronics, phosphors, 130 110,000 846
magnets
catalysts, electronics,
Platinum Group 0.4 66 178
glass, jewelry
Al cans, airplanes,
Aluminum ore 211,000 28,000,000 133
building, electrical
electrical wire,
Copper electronics, pipes, 16,200 630,000 39
ingredient in brass
electrical, cans,
Tin 261 5,200 20
construction,
A more complex analysis of future depletions of our mineral supplies predicts that 20 out of 23 minerals studied will likely
experience a permanent shortfall in global supply by 2030 where global production is less than global demand (Clugston, 2010).
Specifically this study concludes the following: for cadmium, gold, mercury, tellurium, and tungsten—they have already passed
their global production peak, their future production only will decline, and it is nearly certain that there will be a permanent global
supply shortfall by 2030; for cobalt, lead, molybdenum, platinum group metals, phosphate rock, silver, titanium, and zinc—they are
likely at or near their global production peak and there is a very high probability that there will be a permanent global supply
5.6.6 [Link]
shortfall by 2030; for chromium, copper, indium, iron ore, lithium, magnesium compounds, nickel, and phosphate rock—they are
expected to reach their global production peak between 2010 and 2030 and there is a high probability that there will be a permanent
global supply shortfall by 2030; and for bauxite, rare earth minerals, and tin—they are not expected to reach their global production
peak before 2030 and there is a low probability that there will be a permanent global supply shortfall by 2030. It is important to
note that these kinds of predictions of future mineral shortages are difficult and controversial. Other scientists disagree with
Clugston’s predictions of mineral shortages in the near future. Predictions similar to Clugston were made in the 1970s and they
were wrong. It is difficult to know exactly the future demand for minerals and the size of future mineral reserves. The remaining
life for specific minerals will decrease if future demand increases. On the other hand, mineral reserves can increase if new mineral
deposits are found (increasing the known amount of ore) or if currently unprofitable mineral deposits become profitable ones due to
either a mineral price increase or technological improvements that make mining or processing cheaper. Mineral resources, a much
larger category than mineral reserves, are the total amount of a mineral that is not necessarily profitable to mine today but that has
some sort of economic potential.
Mining and processing ore can have considerable impact on the environment. Surface mines can create enormous pits (see Figure
Open pit mine) in the ground as well as large piles of overburden and tailings that need to be reclaimed, i.e., restored to a useful
landscape. Since 1977 surface mines in U.S. are required to be reclaimed, and commonly reclamation is relatively well done in this
country. Unfortunately, surface mine reclamation is not done everywhere, especially in underdeveloped countries, due to lack of
regulations or lax enforcement of regulations. Unreclaimed surface mines and active surface mines can be major sources of water
and sediment pollution. Metallic ore minerals (e.g., copper, lead, zinc, mercury, and silver) commonly include abundant sulfide,
and many metallic ore deposits contain abundant pyrite (iron sulfide). The sulfide in these minerals oxidizes quickly when exposed
to air at the surface producing sulfuric acid, called acid mine drainage. As a result streams, ponds, and soil water contaminated with
this drainage can be highly acidic, reaching pH values of zero or less (see Figure Acid Mine Drainage)! The acidic water can leach
heavy metals such as nickel, copper, lead, arsenic, aluminum, and manganese from mine tailings and slag. The acidic contaminated
water can be highly toxic to the ecosystem. Plants usually will not regrow in such acidic soil water, and therefore soil erosion rates
skyrocket due to the persistence of bare, unvegetated surfaces. With a smaller amount of tailings and no overburden, underground
mines usually are much easier to reclaim, and they produce much less acid mine drainage. The major environmental problem with
underground mining is the hazardous working environment for miners primarily caused by cave-ins and lung disease due to
prolonged inhalation of dust particles. Underground cave-ins also can damage the surface from subsidence. Smelting can be a
major source of air pollution, especially SO2 gas. The case history below examines the environmental impact of mining and
processing gold ore.
Figure 5.6.6 Acid Mine Drainage. The water in Rio Tinto River, Spain is highly acidic (pH = ~2) and the orange color is from iron
in the water. A location along this river has been mined beginning some 5,000 years ago primarily for copper and more recently for
silver and gold. Source: Sean Mack of NASA via Wikimedia Commons
5.6.7 [Link]
Sustainable Solutions to the Mineral Crisis?
Providing sustainable solutions to the problem of a dwindling supply of a nonrenewable resource such as minerals seems
contradictory. Nevertheless, it is extremely important to consider strategies that move towards sustainability even if true
sustainability is not possible for most minerals. The general approach towards mineral sustainability should include mineral
conservation at the top of the list. We also need to maximize exploration for new mineral resources while at the same time we
minimize the environmental impact of mineral mining and processing.
Conservation of mineral resources includes improved efficiency, substitution, and the 3 Rs of sustainability, reduce, reuse, and
recycle. Improved efficiency applies to all features of mineral use including mining, processing, and creation of mineral products.
Substituting a rare nonrenewable resource with either a more abundant nonrenewable resource or a renewable resource can help.
Examples include substituting glass fiber optic cables for copper in telephone wires and wood for aluminum in construction.
Reducing global demand for mineral resources will be a challenge, considering projections of continuing population growth and the
rapid economic growth of very large countries such as China, India, and Brazil. Historically economic growth is intimately tied to
increased mineral consumption, and therefore it will be difficult for those rapidly developing countries to decrease their future
demand for minerals. In theory, it should be easier for countries with a high mineral consumption rate such as the U.S. to reduce
their demand for minerals but it will take a significant change in mindset to accomplish that. Technology can help some with some
avenues to reducing mineral consumption. For example, digital cameras have virtually eliminated the photographic demand for
silver, which is used for film development. Using stronger and more durable alloys of steel can translate to fewer construction
materials needed. Examples of natural resource reuse include everything at an antique store and yard sale. Recycling can extend the
lifetime of mineral reserves, especially metals. Recycling is easiest for pure metals such as copper pipes and aluminum cans, but
much harder for alloys (mixtures of metals) and complex manufactured goods, such as computers. Many nonmetals cannot be
recycled; examples include road salt and fertilizer. Recycling is easier for a wealthy country because there are more financial
resources to use for recycling and more goods to recycle. Additional significant benefits of mineral resource conservation are less
pollution and environmental degradation from new mineral mining and processing as well as reductions in energy use and waste
production.
Because demand for new minerals will likely increase in the future, we must continue to search for new minerals, even though we
probably have already found many of the “easy” targets, i.e., high-grade ore deposits close to the surface and in convenient
locations. To find more difficult ore targets, we will need to apply many technologies including geophysical methods (seismic,
gravity, magnetic, and electrical measurements, as well as remote sensing, which uses satellite-based measurements of
electromagnetic radiation from Earth’s surface), geochemical methods (looking for chemical enrichments in soil, water, air, and
plants), and geological information including knowledge of plate tectonics theory. We also may need to consider exploring and
mining unconventional areas such as continental margins (submerged edges of continents), the ocean floor (where there are large
deposits of manganese ore and other metals in rocks called manganese nodules), and oceanic ridges (undersea mountains that have
copper, zinc, and lead ore bodies).
Finally, we need to explore for, mine, and process new minerals while minimizing pollution and other environmental impacts.
Regulations and good engineering practices are necessary to ensure adequate mine reclamation and pollution reduction, including
acid mine drainage. The emerging field of biotechnology may provide some sustainable solutions to metal extraction. Specific
methods include biooxidation (microbial enrichment of metals in a solid phase), bioleaching (microbial dissolution of metals),
biosorption (attachment of metals to cells), and genetic engineering of microbes (creating microorganisms specialized in extracting
metal from ore).
Review Questions
1. Name some important ways mineral resources are used. Why are they important to society?
2. What are the major environmental issues associated with mineral resources?
3. What should society learn from the case history of gold?
4. Why is society facing a crisis involving mineral supply and how might we work to solve it?
References
Clugston, C. (2010) Increasing Global Nonrenewable Natural Resource Scarcity - An Analysis, The Oil Drum. Retrieved from
[Link]
Craig J, Vaughan D, and Skinner B (2011) Earth Resources and the Environment (4th ed.). Pearson Prentice Hall, p. 92
5.6.8 [Link]
Footnotes
Americans also consumed more than 21,000 pounds of energy resources from the Earth including coal, oil, natural gas, and
uranium.
Economic concentration value for gold comes from Craig, Vaughan, Skinner (2011).
Glossary
acid mine drainage
Surface water or groundwater that is highly acidic due to oxidation of sulfide minerals at a mineral mine.
bioleaching of minerals
Microbial dissolution of metals.
biological processes
Processes of ore formation that involve the action of living organisms. Examples include the formation of pearls in oysters, as
well as phosphorous ore in the feces of birds and the bones and teeth of fish.
biooxidation of minerals
Microbial enrichment of metals in a solid phase.
biosorption of minerals
Attachment of metals to cells.
black smoker
Discharge of mineral-rich waters up to 350°C from cracks in oceanic crust; these waters precipitate a variety of metallic sulfide
ore minerals that make the water appear black.
enrichment factor
Ratio of the metal concentration needed for an economic ore deposit over the average abundance of that metal in Earth’s crust.
hydrothermal
Ore forming process involving hot salty water that dissolves metallic elements from a large area and then precipitates ore
minerals in a smaller area, commonly along rock fractures and faults.
igneous crystallization
Ore forming process where molten rock cools to form igneous rock.
igneous rock
Forms by cooling and solidification of hot molten rock.
leaching
Using chemicals to dissolve metal from a large volume of crushed rock.
metamorphic rock
Forms when a preexisting rock changes the shape or type of minerals due to intense heat and pressure deep within the Earth.
metamorphism
Process of ore formation that occurs deep in the earth under very high temperature and pressure and produces several building
stones, including marble and slate, as well as some nonmetallic ore, including asbestos, talc, and graphite.
mineral
Naturally occurring inorganic solid with a defined chemical composition and crystal structure.
5.6.9 [Link]
mineral conservation
Method of extending the mineral supply that includes improved efficiency, substitution, reduce, reuse, and recycle.
mineral recycling
Method of extending the mineral supply that involves processing used Minerals into new products to prevent waste of
potentially useful materials.
mineral reserves
The known amount of ore in the world.
mineral resources
Total amount of a mineral used by society that is not necessarily profitable to Mine today but has some sort of economic
potential.
mineral reuse
Method of extending the mineral supply that involves using a mineral multiple times.
mineral substitution
Method of extending the mineral supply; involves substituting a rare nonrenewable resource with either a more abundant
nonrenewable resource or a renewable resource.
open-pit mine
Type of surface mineral mine which commonly involve large holes that extract relatively low-grade metallic ore.
ore
Rock with an enrichment of minerals that can be mined for profit.
ore deposit
Location with abundant ore.
placer deposit
Ore forming process where dense gold particles and diamonds are concentrated by flowing water in rivers and at beaches.
placer mine
Type of surface mineral mine which extracts gold or diamonds from river and beach. sediment by scooping up the sediment and
then separating the ore by density.
reclaimed mine
Mineral mine restored to a useful landscape.
rock
A solid coherent piece of planet Earth.
sedimentary processes
Processes of ore formation that occur in rivers and concentrate sand and gravel (used in construction), as well as dense gold
particles and diamonds that weathered away from bedrock.
sedimentary rock
Forms by hardening of layers of sediment (loose grains such as sand or mud) deposited at Earth's surface or by mineral
precipitation, i.e., formation of minerals in water from dissolved mineral matter.
slag
Glassy unwanted by-product of smelting ore.
smelting
5.6.10 [Link]
Heating ore minerals with different chemicals to extract the metal.
strategic mineral
Mineral considered essential to a country for some military, industrial, or commercial purpose but the country must import the
mineral to meet its needs.
strip mine
Type of surface mineral mine which extracts horizontal layers of ore or rock.
surface mine
Mineral mine that occurs at Earth’s surface.
tailings
Fine-grained waste produced from processing ore.
underground mine
Mineral mine that involves a network of tunnels to access and extract the ore.
weathering
Ore forming process where soil water in a tropical rain forest environment concentrates insoluble elements such as aluminum
(bauxite) by dissolving away the soluble elements.
This page titled 5.6: Mineral Resources- Formation, Mining, Environmental Impact is shared under a CC BY license and was authored, remixed,
and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
5.6.11 [Link]
5.7: Case Study- Gold- Worth its Weight?
Gold is a symbol of wealth, prestige, and royalty that has attracted and fascinated people for many thousands of years (see Figure
5.7.1). Gold is considered by many to be the most desirable precious metal because it has been sought after for coins, jewelry, and
other arts since long before the beginning of recorded history. Historically its value was used as a currency standard (the gold
standard) although not anymore. Gold is very dense but also very malleable; a gram of gold can be hammered into a 1 m2 sheet of
gold leaf. Gold is extremely resistant to corrosion and chemical attack, making it almost indestructible. It is also very rare and
costly to produce. Today the primary uses of gold are jewelry and the arts, electronics, and dentistry. The major use in electronics is
gold plating of electrical contacts to provide a corrosion-resistant conductive layer on copper. Most gold is easily recycled except
for gold plating due to combinations with other compounds such as cyanide. About half of the world’s gold ever produced has been
produced since 1965 (see Figure 5.7.2). At the current consumption rate today’s gold reserves are expected to last only 20 more
years.
photograph of native gold
Figure 5.7.1 Native Gold A collage of 2 photos, showing 3 pieces of native gold. The top piece is from the Washington mining
district, California, and the bottom two are from Victoria, Australia. Source: Aram Dulyan via Wikimedia Commons
chart showing world gold production
Figure 5.7.2 World Gold Production World gold production from 1900 to 2009 including annual (blue line) and cumulative data
(gray line) Source: Realterm via Wikimedia Commons
There are two types of gold ore deposits: (1) hydrothermal, where magma-heated groundwater dissolves gold from a large volume
of rock and deposits it in rock fractures and (2) placer, where rivers erode a gold ore deposit of hydrothermal origin and deposit the
heavy gold grains at the bottom of river channels. Although gold’s resistance to chemical attack makes it extremely durable and
reusable, that same property also makes gold difficult to extract from rock. As a result, some gold mining methods can have an
enormous environmental impact. The first discovered gold ore was from placer deposits, which are relatively simple to mine. The
method of extracting gold in a placer deposit involves density settling of gold grains in moving water, similar to how placer
deposits form. Specific variations of placer mining include hushing (developed by the ancient Romans where a torrent of water is
sent through a landscape via an aqueduct), sluice box (where running water passes through a wooden box with riffles on the
bottom), panning (a hand-held conical metal pan where water swirls around) and hydraulic (where high pressure hoses cut into
natural landscapes, see Figure 5.7.3. Figure 5.7.3, developed during the California Gold Rush in the middle 1800s, can destroy
natural settings, accelerate soil erosion, and create sediment-rich rivers that later flood due to sediment infilling the channel. The
largest gold ore body ever discovered is an ancient, lithified (i.e., hardened) placer deposit. Nearly half of the world’s gold ever
mined has come from South Africa’s Witwatersrand deposits, which also have the world’s deepest underground mine at about
4,000 m. To increase the efficiency of gold panning, liquid mercury is added to gold pans because mercury can form an alloy with
gold in a method called mercury amalgamation. The mercury-gold amalgam is then collected and heated to vaporize the mercury
and concentrate the gold. Although mercury amalgamation is no longer used commercially, it is still used by amateur gold panners.
Unfortunately, considerable mercury has been released to the environment with this method, which is problematic because mercury
bioaccumulates and it is easily converted to methylmercury, which is highly toxic.
photograph of gold hydraulic mining
Figure 5.7.3 Gold hydraulic mining in New Zealand, 1880s Source: James Ring via Wikimedia Commons
Today most gold mining is done by a method called heap leaching, where cyanide-rich water percolates through finely ground gold
ore and dissolves the gold over a period of months; eventually the water is collected and treated to remove the gold. This process
revolutionized gold mining because it allowed economic recovery of gold from very low-grade ore (down to 1 ppm) and even from
gold ore tailings that previously were considered waste. On the other hand, heap leaching is controversial because of the toxic
nature of cyanide. The world’s largest cyanide spill to date occurred at Baia Mare in northern Romania (see Figure 5.7.4). In
January 2000 after a period of heavy rain and snowmelt, a dam surrounding a gold tailings pond collapsed and sent into the
drainage basin of the Danube River 100,000 m3 (100 million liters) of water with 500 - 1,000 ppm cyanide, killing more than a
thousand metric tons of fish (see Figure 5.7.4). Considering the large environmental impact of gold mining, this may take some of
the glitter from gold.
map of Baia Mare
Figure 5.7.4 Baia Mare Map of Tisza River drainage basin with pollution hot spots including Baia Mare, Romania, which is the
location of a cyanide spill disaster in 2000 Source: United Nations Environment Program - GRID-Arendal
photograph of dead fish on the shores of Baia Mare
Figure 5.7.5 Baia Mare Cyanide Spill Dead fish from cyanide spill disaster Baia Mare, Romania, the location of a in 2000 Source:
Toxipedia
5.7.1 [Link]
Glossary
Heap leaching
Method of gold mining where cyanide-rich water percolates through finely ground gold ore and dissolves the gold over a period of
months; eventually the water is collected and treated to remove the gold.
Hushing
Method of placer mining developed by the ancient Romans where a torrent of water is sent through a landscape via an aqueduct.
Hydraulic mining
Method of placer mining where high pressure hoses cut into natural landscapes.
Mercury amalgamation
Method of gold panning where liquid mercury is added to gold pans because mercury can form an alloy with gold.
Panning
Method of placer mining where water in a hand-held conical metal pan swirls around.
Sluice box
Method of placer mining where running water passes through a wooden box with riffles on the bottom.
5.7: Case Study- Gold- Worth its Weight? is shared under a not declared license and was authored, remixed, and/or curated by LibreTexts.
5.7.2 [Link]
CHAPTER OVERVIEW
This page titled 6: Environmental and Resource Economics is shared under a CC BY license and was authored, remixed, and/or curated by
Heriberto Cabezas (GALILEO Open Learning Materials) .
1
6.1: Environmental and Resource Economics - Chapter Introduction
Introduction
The field of environmental and natural resource economics sounds to many like an oxymoron. Most people think economists study
money, finance, and business—so what does that have to do with the environment? Economics is really broadly defined as the
study of the allocation of scarce resources. In other words, economics is a social science that helps people understand how to make
hard choices when there are unavoidable tradeoffs. For example, a company can make and sell more cars, which brings in revenue,
but doing so also increases production costs. Or a student can choose to have a part-time job to reduce the size of the loan she needs
to pay for college, but that reduces the time she has for studying and makes it harder for her to get good grades. Some economists
do study business, helping companies and industries design production, marketing, and investment strategies that maximize their
profits. Other economists work to understand and inform the choices individuals make about their investments in education and
how to divide their time between work, leisure, and family in order to make themselves and their families better off. Environmental
and natural resource economists study the tradeoffs associated with one of the most important scarce resources we have—nature.
Economists contribute to the study of environmental problems with two kinds of work. First, they do normative studies of how
people should manage resources and invest in environmental quality to make themselves and/or society as well off as possible.
Second, they do positive analyses of how human agents—individuals, firms, and so forth—actually do behave. Normative studies
give recommendations and guidance for people and policy makers to follow. Positive studies of human behavior help us to
understand what causes environmental problems and which policies are most likely to work well to alleviate them.
This chapter gives an overview of a few of the key ideas that have been developed in this field. First, we will learn the economic
theories that help us understand where environmental problems come from and what makes something a problem that actually
needs to be fixed. This section of the chapter will introduce the concepts of externalities, public goods, and open access resources,
and explain how in situations with those features we often end up with too much pollution and excessive rates of natural resource
exploitation. Second, we will learn the tools economists have developed to quantify the value of environmental amenities. It is very
difficult to identify a monetary value for things like clean air and wildlife, which are not traded in a marketplace, but such value
estimates are often helpful inputs for public discussions about environmental policies and investments. Third, we will discuss a set
of approaches economists use to evaluate environmental policies and projects. We want to design policies that are “good,” but what
exactly does that mean? Finally, we will learn about the different policy tools that can be used to solve problems of excess
environmental degradation and resource exploitation, including a set of incentive policies that were designed by economists to
work with rather than against the way that people really behave, and we will discuss the strengths and weaknesses of those different
tools.
Glossary
normative analysis
A study of how things should be.
positive analysis
A study of how things are.
This page titled 6.1: Environmental and Resource Economics - Chapter Introduction is shared under a CC BY license and was authored, remixed,
and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
6.1.1 [Link]
6.2: Tragedy of the Commons
Learning Objectives
After reading this module, students should be able to
know how economists define environmental outcomes that make society as well off as possible.
understand what externalities are, and how they can lead to outcomes with too much pollution and resource exploitation.
be able to define public goods and common-property resources, and understand how those things are prone to under-
provision and over-exploitation, respectively.
Introduction
To identify and solve environmental problems, we need to understand what situations are actually problems (somehow formally
defined) and what circumstances and behaviors cause them. We might think that it is easy to recognize a problem—pollution is
bad, saving natural resources is good. However, critical thinking often reveals snap judgments to be overly simplistic. Some
examples help to illustrate this point.
Running out! Oil is a depletable resource, and many people worry that rapid extraction and use of oil might cause us to run out.
But would it really be a bad thing to use up all the oil as long as we developed alternative energy technologies to which we
could turn when the oil was gone? Is there any intrinsic value to keeping a stock of oil unused in the ground? Running out of oil
someday may not be a problem. However, subsidies for oil extraction might cause us to run out more quickly than is socially
optimal. Other inefficiencies arise if multiple companies own wells that tap the same pool of oil, and each ends up racing to
extract the oil before the others can take it away—that kind of race can increase total pumping costs and reduce the total amount
of oil that can be gleaned from the pool.
Biological pollution! Horror stories abound in the news about the havoc raised by some nonnative animal and plant species in
the United States. Zebra mussels clog boats and industrial pipes, yellow star thistle is toxic to horses and reduces native
biodiversity in the American West, and the emerald ash borer kills ash trees as it marches across the landscape. From the current
tone of much media and scientific discourse about nonnative species, one could conclude that all nonnative species are
problems. But does that mean we should forbid farmers in the U.S from growing watermelons, which come from Africa? Or
should we ship all the ring-necked pheasants back to Eurasia whence they originally came, and tell North Dakota to choose a
new state bird? The costs and benefits of nonnative species vary greatly – one policy approach is not likely to apply well to
them all.
This section first explains the way economists think about whether an outcome is good. Then it describes some of the features of
natural resources and environmental quality that often trigger problematic human behaviors related to the environment.
6.2.1 [Link]
that last unit, the marginal benefit, which means that the net benefits in the market are maximized. Regular goods are supplied by
industry such that supply is equivalent to the marginal production costs to the firms, and they are demanded by consumers in such a
way that we can read the marginal benefit to consumers off the demand curve; when the market equilibrates at a price that causes
quantity demanded to equal quantity supplied at that price (Qmarket in Figure 6.2.1), it is also true that marginal benefit equals
marginal cost.
Figure 6.2.1 Market Equilibrium. A private market equilibrates at a price such that the quantity supplied equals the quantity
demanded, and thus private marginal cost equals private marginal benefit. Source: Amy Ando
Even depletable resources such as oil would be used efficiently by a well-functioning market. It is socially efficient to use a
depletable resource over time such that the price rises at the same rate as the rate of interest. Increasing scarcity pushes the price up,
which stimulates efforts to use less of the resource and to invest in research to make “backstop” alternatives more cost-effective.
Eventually, the cost of the resource rises to the point where the backstop technology is competitive, and the market switches from
the depletable resource to the backstop. We see this with copper; high prices of depletable copper trigger substitution to other
materials, like fiber optics for telephone cables and plastics for pipes. We would surely see the same thing happen with fossil fuels;
if prices are allowed to rise with scarcity, firms have more incentives to engage in research that lowers the cost of backstop
technologies like solar and wind power, and we will eventually just switch.
Unfortunately, many conditions can lead to market failure such that the market outcome does not maximize social welfare. The
extent to which net benefits fall short of their potential is called deadweight loss. Deadweight loss can exist when not enough of a
good is produced, or too much of a good is produced, or production is not done in the most cost-effective (least expensive) way
possible, where costs include environmental damages. Some types of market failures (and thus deadweight loss) are extremely
common in environmental settings.
Externalities
In a market economy, people and companies make choices to balance the costs and benefits that accrue to them. That behavior can
sometimes yield outcomes that maximize total social welfare even if individual agents are only seeking to maximize their own
personal well-being, because self-interested trades lead the market to settle where aggregate marginal benefits equal aggregate
marginal costs and thus total net benefits are maximized.
However, people and companies do not always bear the full costs and benefits associated with the actions they take. When this is
true economists say there are externalities, and individual actions do not typically yield efficient outcomes.
A negative externality is a cost associated with an action that is not borne by the person who chooses to take that action. For
example, if a student cheats on an exam, that student might get a higher grade. However, if the class is graded on a curve, all the
other students will get lower grades. And if the professor learns that cheating happened, she might take steps to prevent cheating on
the next exam that make the testing environment more unpleasant for all the students (no calculators allowed, no bathroom breaks,
id checks, etc.). Negative externalities are rampant in environmental settings:
6.2.2 [Link]
Companies that spill oil into the ocean do not bear the full costs of the resulting harm to the marine environment, which include
everything from degraded commercial fisheries to reduced endangered sea turtle populations).
Commuters generate emissions of air pollution, which lowers the ambient quality of the air in areas they pass through and
causes health problems for other people.
Developers who build houses in bucolic exurban settings cause habitat fragmentation and biodiversity loss, inflicting a cost on
the public at large.
Figure 6.2.2 Negative Externality: Smog. A NASA photograph of the atmosphere over upstate New York, with Lake Eire (top) and
Lake Ontario (bottom) featured. Both natural, white clouds and man-made smog (grey clouds below) are visible. The smog is an
example of a negative externality, as the cost of the pollution is borne by everyone in the region, not just by the producers. Source:
Image Science and Analysis Laboratory, NASA-Johnson Space Center
In situations where an action or good has a negative externality, the private marginal cost that shapes the behavior of an agent is
lower than the marginal cost to society as a whole, which includes the private marginal cost and the external environmental
marginal cost. The efficient outcome would be where the social marginal cost equals the social marginal benefit (labeled Qefficient in
Figure 6.2.3). Unfortunately, the free-market outcome (labeled Qmarket in Figure 6.2.3) will tend to have more of the good or
activity than is socially optimal because the agents are not paying attention to all the costs. Too much oil will be shipped, and with
insufficient care; people will drive too many miles on their daily commutes; developers will build too many new homes in sensitive
habitats. Thus, there is deadweight loss (the shaded triangle in the figure); the marginal social cost associated with units in excess
of the social optimum is greater than the marginal benefit society gets from those units. Public policy that reduces the amount of
the harmful good or activity could make society as a whole better off.
6.2.3 [Link]
Figure 6.2.3 Inefficiency from Negative Externality. When there is a negative externality, the market equilibrates where the total
social marginal cost exceeds the marginal benefit of the last unit of a good and society is not as well off as it could be if less were
produced. Source: Amy Ando
Conversely, a positive externality is a benefit associated with an action that is not borne by the person who chooses to take that
action. Students who get flu shots in October, for example, gain a private benefit because they are less likely to get the flu during
the winter months. However, their classmates, roommates, and relatives also gain some benefit from that action because inoculated
students are less likely to pass the flu along to them. Positive externalities exist in the world of actions and products that affect the
environment:
A homeowner who installs a rain barrel to collect unchlorinated rainwater for her garden also improves stream habitat in her
watershed by reducing stormwater runoff.
A delivery company that re-optimizes its routing system to cut fuel costs also improves local air quality by cutting its vehicle air
pollution emissions.
A farmer who plants winter cover crops to increase the productivity of his soil will also improve water quality in local streams
by reducing erosion.
In situations where an action or good has a positive externality, the private marginal benefit that shapes the behavior of an agent is
lower than the marginal benefit to society as a whole, which includes the private marginal benefit and the external environmental
marginal benefit. The efficient outcome would be where the social marginal cost equals the social marginal benefit (labeled
Qefficient in Figure 6.2.4). In the presence of a positive externality, the free-market outcome will tend to promote less of the good or
activity than is socially optimal because the agents do not reap all the benefits. Too few rain barrels will be installed; not enough
delivery routes will be re-optimized; too few acres of agricultural fields will have cover crops in the winter months. Again there is
deadweight loss (the shaded triangle in the figure), but this time because the marginal social benefit associated with some of the
units not produced would have been greater than the marginal costs of producing them. Just because an externality is positive rather
than negative doesn’t mean there isn’t a problem; public policy could still make society as a whole better off.
6.2.4 [Link]
Figure 6.2.4 Positive Externality. When there is a a positive externality, the market equilibrates where the total social marginal
benefit exceeds the marginal cost of the last unit of a good and society is not as well off as it could be if more were produced.
Source: Amy Ando
6.2.5 [Link]
Swordfish in the ocean can be caught by anyone with the right boat and gear, and the more fish are caught by one fleet of boats,
the fewer remain for other fishers to catch.
Old growth timber in a developing country can be cut down by many people, and slow regrowth means that the more timber
one person cuts the less there is available for others.
One person’s use of a common-pool resource has negative effects on all the other users. Thus, these resources are prone to
overexploitation. One person in Indonesia might want to try to harvest tropical hardwood timber slowly and sustainably, but the
trees they forebear from cutting today might be cut down by someone else tomorrow. The difficulty of managing common-pool
resources is evident around the world in rapid rates of tropical deforestation, dangerous overharvesting of fisheries (see 8.3), and
battles fought over mighty rivers that have been reduced to dirty trickles.
The tragedy of the commons occurs most often when the value of the resource is great, the number of users is large, and the users
do not have social ties to one another, but common-pool resources are not always abused. Elinor Ostrom’s Nobel prize-winning
body of work, for example, has studied cases of common-pool resources that were not over-exploited because of informal social
institutions.
Review Questions
1. What does it mean for an outcome to be efficient?
2. How do externalities cause market outcomes not to be efficient?
3. How are the free rider problem and the common pool resource problem related to basic problems of externalities?
Glossary
common pool resource
A resource that is open to all users, but which is highly rival in use.
cost-effective
As inexpensive as possible; cost minimizing.
deadweight loss
The extent to which net benefits are lower than they could be.
efficient
Having the feature that net benefits are maximized.
free rider
A person who does not contribute to a public good in hopes that they can benefit from provision by other people.
marginal benefit
The additional benefit of doing one more unit of something.
marginal cost
The additional cost of doing one more unit of something.
market failure
A condition that causes a market not to yield the efficient outcome.
negative externality
A cost that is borne by someone who did not agree to the activity that caused the cost.
net benefits
The difference between total benefits and total costs.
normative analysis
A study of how things should be.
6.2.6 [Link]
positive analysis
A study of how things are.
positive externality
A benefit that accrues to someone who did not agree to the activity that caused the benefit.
public good
A good with two features: (i) it has a benefit that does not diminish with the number of people enjoying it, and (ii) no one can
be excluded from consuming it.
welfare
Broadly defined, welfare is well-being.
This page titled 6.2: Tragedy of the Commons is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto Cabezas
(GALILEO Open Learning Materials) .
6.2.7 [Link]
6.3: Case Study- Marine Fisheries
Fisheries are classic common-pool resources. The details of the legal institutions that govern access to fisheries vary around the
globe. However, the physical nature of marine fisheries makes them prone to overexploitation. Anyone with a boat and some gear
can enter the ocean. One boat’s catch reduces the fish available to all the other boats and reduces the stock available to reproduce
and sustain the stock available in the following year. Economic theory predicts that the market failure associated with open access
to a fishery will yield socially excessive levels of entry into the fishery (too many boats) and annual catch (too many fish caught)
and inefficiently low stocks of fish (Beddington, Agnew, & Clark, 2007).
6.3.1 [Link]
Figure 6.3.2 Marine Fisheries: Fishing Boats. Alaskan waters have been fished by people for thousands of years, but they are under
pressure from modern fishing technologies and large-scale extraction. Source: National Oceanic and Atmospheric Administration
Resource economists developed the idea of a tradable permit scheme to help manage fisheries. Individual tradable quota (ITQ)
schemes are cap-and-trade policies for fish, where total catch is limited but fishers in the fishery are given permits that guarantee
them a right to a share of that catch. Players in the fishery can sell their quota shares to each other (helping the catch to flow
voluntarily to the most efficient boats in the industry) and there is no incentive for captains to buy excessively large boats or fish
too rapidly to beat the other boats to the catch. ITQ policies have rationalized the Alaskan halibut fishery completely: the fish stock
is thriving, overcapitalization is gone, and the fish catch is spread out over time (Levy, 2010). ITQs have also been implemented in
the fisheries of New Zealand, yielding large improvements in the biological status of the stocks (Annala, 1996). There is some
general evidence that ITQ systems have been relatively successful in improving fishery outcomes (Costello, Gaines, & Lynham et
al. 2008), though other research implies that evidence of the superiority of the ITQ approach is more mixed (Beddington 2007)
Scholars and fishery managers continue to work to identify the details of ITQ management that make such systems work most
effectively, and to identify what needs to be done to promote more widespread adoption of good fishery management policy
worldwide.
References
Acheson, J. M. (1988). The Lobster Gangs of Maine. Lebanon, NH: University of New England Press.
Annala, J. H. (1996). New Zealand’s ITQ system: have the first eight years been a success or a failure? Reviews in Fish Biology
and Fisheries. 6(1), 43–62. doi: 10.1007/BF00058519
Beddington, J. R., Agnew, D.J., & Clark, C. W. (2007). Current problems in the management of marine fisheries. Science,
316(5832), 1713-1716. doi:10.1126/science.1137362
Costello, C., Gaines, S. D., & Lynham, J. (2008). Can catch shares prevent fisheries collapse? Science, 321(5896), 1678 – 1681.
doi10.1126/science.1159478
Levy, S. (2010). Catch shares management. BioScience, 60(10), 780–785. doi:10.1525/bio.2010.60.10.3
Montaigne, F. (2007, April). Still waters, the global fish crisis. National Geographic Magazine. Retrieved from
[Link]
Worm, B., Barbier, E. B., Beaumont, N., Duffy, J. E., Folke, C., Halpern, B. S., Hackson, J. B. C., Lotze, H. K., Micheli, F.,
Palumbi, S. R., Sala, E., Selkoe, K. A., Stachowicz, J. J., & Watson, R. (2006). Impacts of biodiversity loss on ocean ecosystem
services. Science, 314(5800), 787 – 790. doi:10.1126/science.1132294
6.3.2 [Link]
This page titled 6.3: Case Study- Marine Fisheries is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto
Cabezas (GALILEO Open Learning Materials) .
6.3.3 [Link]
6.4: Environmental Valuation
Learning Objectives
After reading this module, students should be able to
understand why it might be useful to develop estimates of the values of environmental goods in dollar terms.
know the difference between the two economic measures of value, willingness to pay and willingness to accept.
be familiar with valuation methods in all three parts of the environmental valuation toolkit: direct, revealed preference, and
stated preference methods.
understand the strengths and weaknesses of those valuation methods.
Use Values
Externality, public good, and common-pool resource problems yield suboptimal levels of environmental quality and excessive rates
of resource exploitation. Many factors complicate the process of deciding what to do about these problems. One is that
environmental goods are not traded in any marketplace, and hence analysts struggle to identify quantitative measures of their
values to society.
Environmental valutaion is controversial. Some environmentalists object to efforts to place dollar values on elements of the
environment that might be viewed as priceless. Such values are important, however, for making sure that society does not fail to
take the value of nature into account when making policy and investment choices. All U.S. government regulations, for example,
are subjected to benefit-cost analyses to make sure that government actions don’t inadvertently make society worse off (see Module
8.5). If we do not have dollar values for the environmental benefits of things like clean water and air, then estimates of the benefits
of pollution control will be consistently lower than the true social benefits, and government policy will chronically underinvest in
efforts to control pollution.
Environmental and natural resource economists have worked for decades to develop valuation methods that can be used to generate
reasonable estimates of the dollar values of environmental amenities. Thousands of journal articles have been published in this
effort to refine valuation methodology. In the early years of valuation studies, most of the work was focused on generating
estimates of the social values of water and air quality. Over time, economists broadened their focus to study how to value a broader
range of amenities such as wetland habitat and endangered species.
The United Nations launched an international effort in 2000 called the Millennium Ecosystem Assessment which was to evaluate
the current state of earth’s ecosystems (and the services that flow from nature to humans) and identify strategies for conservation
and sustainable use. Reports from this effort) have helped scientists and policy makers develop a new framework for thinking about
how nature has value to humans by providing a wide range of ecosystem services. Since then, a surge of multidisciplinary research
has emerged to quantify the physical services provided by the environment and estimate the values to humanity of those services.
Economists recognize two broad categories of environmental values: use and non-use. Use values flow from services that affect
people directly, such as food production, flood regulation, recreation opportunities, and potable water provision. Non-use values are
less tangible: the desire for endangered tigers to continue to exist even on the part of people who will never see them in the wild;
concern about bequeathing future generations a planet with healthy fish populations; a sense that people have an ethical
responsibility to be good stewards of the earth. Economic valuation methods exist to capture all of these environmental values.
6.4.1 [Link]
Figure 6.4.1 Use Value: Recreational Angler. A fisherman takes advantage of the use value of the natural environment of Four
Springs Lake, Tasmania. Source: Photo by Peripitus
Figure 6.4.2 Non-use Value: Sumatran Tiger. Although wild tigers do not directly impact people living in the United States, many
Americans wish for the species to continue existing in their natural environment. This is an example of a non-use value. Source:
Photo by Nevit Dilman
Willingness to Pay/Accept
Economists use measures of value that are anthropocentric, or human centered. A rigorous body of theory about consumer choice
lies beneath those measures. Mathematical complexity can make that theory seem like unreliable trickery, but in truth, consumer
theory rests on only a very small number of fundamental assumptions:
People have preferences over things.
People are able to rank any two bundles of goods to identify which one they prefer.
People are rational in that they will choose the bundle they prefer (over bundles they do not prefer) if they can afford it.
Those uncontroversial axioms are actually enough to derive all the results economists use when working with valuation
methodology. However, the derivations are easier and sometimes more intuitive with a little more structure added to our
hypothetical consumer choice problem:
People face budget constraints (total expenditures can’t exceed their income).
6.4.2 [Link]
People make choices to make themselves as well off as possible (“maximize their utility”) within the rationing forced by their
budget constraints.
This framework yields two ways to think about the values of changes to the quality or quantity of environmental goods. Consider
first a situation where we are trying to determine the value of a project that yields an environmental improvement—say, for
example, water in the Chicago River will be cleaner. The social benefit of that project turns out to be what people are willing to pay
for it. The second measure of value is appropriate if we want to measure the value of environmental goods that will be lost or
degraded by a deleterious change—say, for example, climate change leading to the extinction of polar bears. In that context, the
value of the change is given by the amount of money you would have to pay people in order to make them willing to accept it.
“Willingness to pay” (WTP) is a budget-constrained measure of a change in welfare; a person cannot be willing to pay more money
for a change than they have income. In contrast, “willingness to accept” (WTA) is not a budget constrained measure of value—you
might have to increase a person’s income many times over in order to fully compensate them for the loss of an environmental
amenity they hold dear—and can theoretically approach infinity. Empirical studies tend to find that WTA value estimates are larger
than equivalent estimates of WTP.
Analysts usually choose whether to use WTA or WTP approaches as a function of the context of the analysis. The “right” measure
to use may depend on whether you want value estimates to inform a policy that would improve conditions relative to the current
legal status quo, or to understand the consequences of a change that would cause deterioration of some environmental good citizens
currently enjoy. Another factor in choosing a valuation method is that WTP is budget constrained while WTA is not. WTP
estimates of value tend to be lower in places where people have lower incomes. That variation captures a realistic pattern in the size
of willingness to pay for environmental improvements. However, equity problems clearly plague a study that concludes, for
example, that improvements in air quality are more valuable to society if they happen in rich areas rather than poor.
All valuation methodologies—WTP and WTA—are designed to estimate values of fairly small changes in the environment, and
those values are often setting-specific. Careful analysts can do benefit transfer studies in which they use the results of one valuation
study to inform value estimates in a different place. However, such applications must be carried out carefully. The value of a unit
change in a measure of environmental integrity is not an immutable constant, and the values of very large changes in either quantity
or quality of an environmental amenity usually cannot be estimated. A cautionary example is an influential but widely criticized
paper published in Nature by Robert Costanza and colleagues that carried out sweeping benefit transfer estimates of the total social
values of a number of Earth’s biomes (open oceans, forests, wetlands, etc.). The resulting estimates were too large to be correct
estimates of WTP because they exceeded the value of the whole world’s GDP, and too small to be correct estimates of WTA
because life on earth would cease to exist if oceans disappeared, so WTA for that change should be infinity (Costanza et al., 1997).
An Economist’s Environmental Valuation Toolkit: Direct, Revealed Preference, and Stated Preference
Methods
Early work on environmental valuation estimated the benefits of improved environmental quality using direct methods that exploit
easily obtained information about the monetary damage costs of pollution. These methods are still sometimes used (most often by
people who are not environmental economists) because of their simple intuitive appeal. An analyst can measure costs associated
with pollution; the benefits of environmental cleanup are then the reductions in those costs. Following are some examples:
Production damage measures: Pollution has a deleterious effect on many production processes. For example, air pollution
lowers corn yields, thus increasing the cost of producing a bushel of corn. An analyst could try to measure the benefits of
eliminating air pollution by calculating the increase in net social benefits that would flow from the corn market as a result of
higher yields.
Avoided cost measures: Environmental degradation often forces people to spend money on efforts to mitigate the harm caused
by that degradation. One benefit of reversing the degradation is not having to spend that money on mitigation—the avoided
cost. For example, hydrological disruption from impervious surfaces in urban areas forces cities to spend money on expensive
storm sewer infrastructure to try to reduce floods. A benefit of installing rain gardens and green roofs to manage stormwater
might be avoided storm sewer infrastructure costs.
Health cost measures: Pollution has adverse effects on human health. For example, toxic chemicals can cause cancer, and
ground level ozone causes asthma. Some measures of the damages caused by pollution simply count the financial costs of such
illnesses, including the costs of cancer treatment and lost wages from adults missing work during asthma attacks.
6.4.3 [Link]
These measures seem appealing, but are in fact deeply problematic. One of the most serious problems with direct measures is that
they often yield woefully incomplete estimates of the benefits of environmental cleanup. Consider the example of cancer above.
Suppose a woman gets cancer from drinking contaminated well water. By the time her illness is diagnosed, the cancer is so
advanced that doctors can do little to treat her, and she dies a few months later. The medical expenditures associated with this
illness are not very large; she and her family would surely have been willing to pay much more money to have eliminated the
toxins so she did not ever get sick. The direct health cost measure of the benefits of cleaning up the contaminated water is a serious
underestimate of the true benefit to society of that environmental improvement.
A second set of valuation tools called revealed preference methods work to estimate WTP for environmental amenities and quality
by exploiting data on actual behaviors and market choices that are related to the environmental good in question. People reveal
their WTP for environmental goods with their actions. Three examples of such methods are below.
Hedonic price analysis: We often cannot observe individuals taking direct action to change the quality of the environment to
which they are exposed in a given location because they simply cannot effect such change; no one person, for example, can
reduce the concentration of fine particles in the air near his house. We do, however, observe market data about the choices
people make about where to live. If two houses are otherwise identical but one house is situated in a place with much cleaner air
than the other, the benefit of breathing cleaner air will get capitalized in the value of that house. All else equal, neighborhoods
with better environments will have more expensive homes. Analysts can gather data on housing prices and house characteristics
(both environmental and nonenvironmental) and use a statistical analysis to estimate marginal WTP for elements of
environmental quality that vary among the houses in the data set. The hedonic price analysis approach has been used to value
amenities such as air quality, hazardous waste site cleanup, and open space.
Hedonic wage analysis: Some forms of pollution cause people to face higher risk of death in any given year. Thus, one
important goal of valuation is to estimate a dollar value of reduced mortality resulting from pollution cleanup. Except in the
movies, we rarely observe people choosing how much money they are willing to pay to save a specific person from certain
death. However, all of us make choices every day that affect our risk of death. One important choice is which job to accept.
Elementary school teachers face little job-related mortality risk. In contrast, coal miners, offshore oil rig workers, and deep sea
fishermen accept high rates of accidental death when they take their jobs. By analyzing data on wage rates and worker death
rates in a variety of different industries, we can estimate WTP to reduce the risk of death. Using such hedonic wage analysis,
economists have developed measures of the value of a statistical life(VSL), which can be applied to physical estimates of
reductions in pollution-related deaths to find the benefits of reduced mortality.
Travel cost analysis: Many natural amenities, such as forests, lakes, and parks are enjoyed by the public free of charge. While
there is no formal market for “hours of quality outdoor recreation,” people do incur costs associated with such recreation—gas
purchased to drive to the site, hotel expenses for overnight trips, and the opportunity cost of the time spent on the trip. If
environmental quality is valued, people will be willing to pay higher travel costs to visit recreation sites with higher levels of
environmental quality (e.g., cleaner water in the lake, more fish to catch, a better view from a mountain with low air pollution).
In travel cost analysis, researchers gather data on the environmental features of a set of recreation sites and the choices people
make about visiting those sites—which they choose to visit, and how often—and apply statistical analysis to those data to
estimate WTP for improved quality of natural amenities.
One of the greatest strengths of revealed preference valuation methods is that they use information about real behavior rather than
hypothetical choices. These approaches also yield estimates of WTP that are often more complete than the results of direct market
measure studies.
Revealed preference studies do, however, have weaknesses and limitations. First, they only give good estimates of WTP for
environmental goods if people have full and accurate information about environmental quality and associated risks. For example,
hedonic estimates of WTP to avoid living with polluted air will be biased downward if people in a city do not know how air
pollution varies among neighborhoods. Second, some revealed preference approaches are only valid if the relevant markets (labor
markets for a wage study, housing markets for a hedonic price study) are not plagued by market power and transaction costs that
prevent efficient equilibria from being reached. For example, if workers find it too daunting and costly to move from one region to
another, then coal miners may fail to earn the wage premium that would be associated with such a risky job in the absence of
relocation hurdles. Third, revealed preference approaches cannot be used to estimate values for levels of environmental quality that
are not observed in real-world data. If all the lakes in a region are terribly polluted, we cannot use a travel cost study of lake site
choice to identify WTP for very clean lakes. Fourth, revealed preference methods can capture only use values, not non-use values.
6.4.4 [Link]
The limitations of revealed preference valuation tools motivated environmental and natural resource economists to develop
valuation methods that do not require analysts to be able to observe real-world behavior related to the amenity being valued. These
stated preference methods are now highly refined, but the essential idea is simple. These studies design a survey that presents
people with information about hypothetical scenarios involving an environmental good, gather data on their responses to questions
about how much they would pay for something or whether they would choose one scenario over another, and then analyze the data
to estimate WTP for the good or WTA compensation for elimination or degradation of the good.
Contingent valuation: The methodology called contingent valuation (or CV) gained prominent attention when it was used by
economists to estimate the damage done to society by the oil spilled by Exxon’s Valdez oil tanker in Prince William Sound in
1989 (Carson et al., 2003). A CV survey gives a clear description of a single environmental amenity to be valued, such as a
wetland restoration, whale populations, or improved water quality in a local lake. The description includes details about how the
amenity would be created, and how the survey respondent would pay any money they claim to be willing to pay in support of
the amenity. Respondents are then asked a question to elicit their WTP. This value elicitation question can be open ended (“How
much would you be willing to pay in taxes to increase whale populations?) or closed ended (“Would you be willing to pay $30
to increase whale populations?). The resulting data set is analyzed to find the average WTP of people in the sample population.
Conjoint analysis: Conjoint analysis is also referred to as choice experiment survey analysis. It was developed first by analysts
in business marketing and psychology, and only later adopted by economists for environmental valuation. The main difference
between conjoint analysis and CV is that CV elicits WTP for an environmental amenity with a single fixed bundle of features,
or attributes. Conjoint analysis estimates separate values for each of a set of attributes of a composite environmental amenity.
For example, grasslands can vary in bird species diversity, wildflower coverage, and distance from human population centers. A
conjoint analysis of grassland ecosystems would construct a set of hypothetical grasslands with varied combinations of
attributes (including the cost to the respondent of a chosen grassland). The survey would present respondents with several
choice questions; in each choice, the respondent would be asked to pick which of several hypothetical grasslands they would
prefer. The resulting data would be analyzed to find how each attribute affects the likelihood that one grassland is preferred over
another. This would yield estimates of marginal values for each attribute; those values could then be used to find WTP for
composite grasslands with many different combinations of features.
Both CV and conjoint analysis methods can be designed to estimate WTP for improvement or WTA degradation depending on
which context is most appropriate for the problem at hand. These stated preference methods have two main strengths. First, they
can capture non-use values. Second, their hypothetical nature allows analysts to estimate WTP for improvements out of the range
of current experience (or WTA for degradation we have fortunately not yet experienced).
However, stated preferences approaches do have weaknesses and limitations. For example, many economists are uncomfortable
using value estimates derived from hypothetical choices, worrying whether consumers would make the same choices regarding
payment for public environmental goods if the payments were real. Scholars also worry about whether people give responses to
stated preference surveys that are deliberately skewed from their true WTP. Understatements of value could arise to protest a
government policy (“Why should I have to pay to clean up the environment when someone else made it dirty in the first place?”) or
out of a desire to free ride. Finally, the hypothetical nature of stated preference surveys can mean that some respondents are not
familiar with the thing being valued, and thus may have trouble giving meaningful responses to the questions. Stated preference
surveys must be designed to give respondents enough information without biasing their responses.
References
Carson, R. T., Mitchell, R. C., Hanemann, M., Kopp, R. J., Presser, S., and Ruud, P. A. (2003). Contingent valuation and lost
passive use: Damages from the Exxon Valdez oil spill. Environmental and Resource Economics, 25(3), 257-286. DOI:
10.1023/A:1024486702104.
Costanza, R., d'Arge, R., de Groot, R., Farber, S., Grasso, M., Hannon, B., Limburg, K., Naeem, S., O'Neill, R. V., Paruelo, J.,
Raskin, R. G., Sutton, P., & van den Bel, M. (1997). The value of the world's ecosystem services and natural capital. Nature, 387,
253–260. doi:10.1038/387253a0
Review Questions
Question:
Why might it be useful to estimate dollar values for features of the environment?
6.4.5 [Link]
Question:
What are the three types of valuation tools? List at least one strength and one weakness of each.
Glossary
avoided cost
A type of direct method that equates the value of an environmental improvement with a cost that can then be avoided.
benefit transfer
A method for estimating the value of a natural amenity by applying estimates from a complex study of a slightly different (but
similar) amenity to the case at hand.
conjoint analysis
A stated preference valuation tool that allows an analyst to estimate the marginal values of multiple attributes of an
environmental good.
contingent valuation
A stated preference valuation tool that describes a single environmental good and elicits survey responses from which the
analyst can estimate people’s value for that good.
direct methods
Valuation tools that use data on actual market transactions to estimate the value of a change in the environment.
ecosystem services
Resources and processes through which the environment gives benefits to humanity.
non-use values
Values people have for nature that do not stem from direct interaction.
revealed preference
Valuation tools that use behaviors such as job choice, housing choice, and recreational site choice to reveal information about
the values people have for features of the environment.
stated preference
Valuation tools that use survey responses to hypothetical questions rather than data on actual choices.
use values
Benefits associated with direct interaction with nature and the environment.
valuation
The process of estimating a dollar value for an amenity or a disamentity.
6.4.6 [Link]
A statistical concept that can be used as the value of reducing the number of deaths in a population by one.
willingness to accept
The amount of money you would have to pay someone to compensate them for a deleterious change.
willingness to pay
The amount of money someone is willing to pay for something good.
This page titled 6.4: Environmental Valuation is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto Cabezas
(GALILEO Open Learning Materials) .
6.4.7 [Link]
6.5: Evaluating Projects and Policies
Learning Objectives
After reading this module, students should be able to
know five important features of how economists think about costs
understand why discounting is both important and controversial, and be able to calculate the net present value of a project
or policy
know what cost-benefit analysis is, and be aware of some of its limitations
think about four criteria for evaluating a project that are not captured in a basic cost-benefit analysis
Introduction
Environmental valuation methods help analysts to evaluate the benefits society would gain from policies or cleanup and restoration
projects that improve environmental quality or better steward our natural resources. Another set of tools can yield information
about the costs of such actions (a brief description is below). But even if we have plausible estimates of the costs and benefits of
something, more work needs to be done to put all that information together and make some rational choices about public policy and
investments. This module discusses the challenges of policy evaluation when costs and benefits accrue over time, outlines the main
features of cost-benefit analysis, and presents several other criteria for policy evaluation.
Opportunity Cost
Not all costs involve actual outlays of money. An opportunity cost is the foregone benefit of something that we choose (or are
forced) not to do. The oppurtunity cost of a year of graduate school is the money you could have made if you had instead gotten a
full-time job right after college. Endangered species protection has many opportunity costs: timber in old-growth forests can’t be
cut and sold; critical habitat in urban areas can’t be developed into housing and sold to people who want to live in the area.
Opportunity costs do not appear on firms’ or governments’ accounting sheets and are thus often overlooked in estimates of the
costs of a policy. Studies of U.S. expenditures on endangered species’ recoveries have used only information about costs like direct
government expenditures because opportunity costs are so challenging to measure (e.g. Dawson and Shogren, 2001).
6.5.1 [Link]
Figure 6.5.1 A Redwood Forest in California. Forests can’t both be cut down and preserved for habitat. The dollar cost of lumber is
straightforward to quantify, but it is more difficult to quantify the value of ecosystems. Cutting down the forest therefore has an
opportunity cost that is hard to measure, and this can bias people and governments towards resource extraction. Source: Photo by
Michael Barera
Additionality
A careful analysis of the costs of a program includes only costs that are additional, that is, new additions to costs that would have
existed even in the absence of the program. For example, current regulations require developers to use temporary controls while
constructing a new building to prevent large amounts of sediment from being washed into local rivers and lakes. Suppose EPA
wants to estimate the costs of a new regulation that further requires new development to be designed such that stormwater doesn’t
run off the site after the building is finished. A proper analysis would not include the costs of the temporary stormwater controls in
the estimate of the cost of the new regulation, because those temporary controls would be required even in the absence of the new
regulation (Braden and Ando, 2011). The concept of additionality has been made famous in the context of benefit estimation by a
debate over whether programs that pay landowners not to deforest their lands have benefits that are additional; some of those lands
might not have been deforested even without the payments, or the landowners may receive conservation payments from multiple
sources for the same activity.
6.5.2 [Link]
Control for Associated Market Changes
A careful cost analysis must pay attention to market changes associated with cost increases. To illustrate, suppose the government
is thinking of passing a ban on agricultural use of methyl bromide. This ozone-depleting chemical is widely used as an agricultural
fumigant, and is particularly important in strawberry production and shipping. A ban on methyl bromide might, therefore, increase
the marginal cost of producing strawberries. A simple approach to estimating the cost of the proposed methyl bromide ban would
be to find out how many strawberries were sold before the ban and calculate the increase in the total cost of producing that many
strawberries. However, the increase in production costs will drive up the price of strawberries and lower the number of strawberries
sold in the marketplace. There is a cost to society with two parts: (a) deadweight loss associated with the net benefits of the
strawberries not sold, and (b) the increased cost of producing the strawberries that still are sold. That total social cost is lower,
however, than the estimate yielded by the simple approach outlined above because the simple approach includes increased
production costs for strawberries that are not sold. An accurate cost estimate must take into account market changes.
The concept of net benefits was introduced above; in the context of policy or project evaluation, net benefits are, quite simply, the
difference between the benefits and the costs of a policy in a given year. However, environmental policies typically have benefits
and costs that play out over a long period of time, and those flows are often not the same in every year. For example, wetland
restoration in agricultural areas has a large fixed cost at the beginning of the project when the wetland is constructed and planted.
Every year after that there is an opportunity cost associated with foregone farm income from the land in the wetland, but that
annual cost is probably lower than the fixed construction cost. The wetland will yield benefits to society by preventing the flow of
some nitrogen and phosphorus into nearby streams and by providing habitat for waterfowl and other animals. However, the wildlife
benefits will be low in the early years, increasing over time as the restored wetland vegetation grows and matures. It is not too
difficult to calculate the net benefits of the restoration project in each year, but a different methodology is needed to evaluate the net
benefits of the project over its lifetime.
Some analysts simply add up all the costs and benefits for the years that they accrue. However, that approach assumes implicitly
that we are indifferent between costs and benefits we experience now and those we experience in the future. That assumption is
invalid for two reasons. First, empirical evidence has shown that humans are impatient and prefer benefits today over benefits
tomorrow. One need only ask a child whether they want to eat a candy bar today or next week in order to see that behavior at work.
Second, the world is full of investment opportunities (both financial and physical). Money today is worth more than money
tomorrow because we could invest the money today and earn a rate of return. Thus, if there is a cost to environmental cleanup, we
would rather pay those costs in the future than pay them now.
Economists have developed a tool for comparing net benefits at different points in time called discounting. Discounting converts a
quantity of money received at some point in the future into a quantity that can be directly compared to money received today,
controlling for the time preference described above. To do this, an analyst assumes a discount rate r, where r ranges commonly
between zero and ten percent depending on the application. If we denote the net benefits t years from now as Vt(in the current year,
Vt
t=0), then we say the present discounted value of Vt is P DV (V ) = t Figure 6.4.2 shows how the present value of $10,000
t
(1+r)
declines with time, and how the rate of the decrease varies with the choice of discount rate r. If a project has costs and benefits
Vt
every year for T years, then the net present value of the entire project is given by P DV (V ) = ∑
t . t
(1+r)
Figure 6.5.2 The Impact of a Discount Rate on Present Value Estimates. Source: California Department of Transportation
6.5.3 [Link]
A particular cost or benefit is worth less in present value terms the farther into the future it accrues and the higher the value of the
discount rate. These fundamental features of discounting create controversy over the use of discounting because they make projects
to deal with long-term environmental problems seem unappealing. The most pressing example of such controversy swirls around
analysis of climate-change policy. Climate-change mitigation policies typically incur immediate economic costs (e.g. switching
from fossil fuels to more expensive forms of energy) to prevent environmental damages from climate change several decades in the
future. Discounting lowers the present value of the future improved environment while leaving the present value of current costs
largely unchanged.
Cost-benefit analysis is just that: analysis of the costs and benefits of a proposed policy or project. To carry out a cost-benefit
analysis, one carefully specifies the change to be evaluated, measures the costs and benefits of that change for all years that will be
affected by the change, finds the totals of the presented discounted values of those costs and benefits, and compares them. Some
studies look at the difference between the benefits and the costs (the net present value), while others look at the ratio of benefits to
costs. A “good” project is one with a net present value greater than zero and a benefit/cost ratio greater than one.
The result of a cost-benefit analysis depends on a large number of choices and assumptions. What discount rate is assumed? What
is the status quo counterfactual against which the policy is evaluated? How are the physical effects of the policy being modeled?
Which costs and benefits are included in the analysis—are non-use benefits left out? Good cost-benefit analyses should make all
their assumptions clear and transparent. Even better practice explores whether the results of the analysis are sensitive to
assumptions about things like the discount rate (a practice called sensitivity analysis). Scandal erupted in 2000 when a whistle-
blower revealed that the Army Corps of Engineers was pressuring its staff to alter assumptions to make sure a cost-benefit analysis
yielded a particular result (EDV&CBN, 2000). Transparency and sensitivity analysis can help to prevent such abuses.
Efficiency
A policy is efficient if it maximizes the net benefits society could get from an action of that kind. Many projects and policies can
pass a cost-benefit test but still not be efficient. Several levels of carbon dioxide emission reduction, for example, could have
benefits exceeding costs, but only one will have the largest difference between benefits and costs possible. Such efficiency will
occur when the marginal benefits of the policy are equal to its marginal costs. Sometimes a cost-benefit analysis will try to estimate
the total costs and benefits for several policies with different degrees of stringency to try to see if one is better than the others.
However, only information about the marginal benefit and marginal cost curves will ensure that the analyst has found the efficient
policy. Unfortunately, such information is often very hard to find or estimate.
Cost Effectiveness
As we saw in the Module 8.4, it can be particularly difficult to estimate the benefits of environmental policy, and benefit estimates
are necessary for finding efficient policies. Sometimes policy goals are just set through political processes—reducing sulfur dioxide
emissions by 10 million tons below 1980 levels in the Clean Air Act acid rain provisions, cutting carbon dioxide emissions by 5%
from 1990 levels in the Kyoto protocol—without being able to know whether those targets are efficient. However, we can still
evaluate whether a policy will be cost-effective and achieve its goal in the least expensive way possible. For example, for total
pollution reduction to be distributed cost-effectively between all the sources that contribute pollution to an area (e.g. a lake or an
urban airshed), it must be true that each of the sources is cleaning up such that they all face the same marginal costs of further
abatement. If one source had a high marginal cost and another’s marginal cost was very low, total cost could be reduced by
switching some of the cleanup from the first source to the second.
Incentives to Innovate
At any one point in time, the cost of pollution control or resource recovery depends on the current state of technology and
knowledge. For example, the cost of reducing carbon dioxide emissions from fossil fuels depends in part on how expensive solar
and wind power are, and the cost of wetland restoration depends on how quickly ecologists are able to get new wetland plants to be
established. Everyone in society benefits if those technologies improve and the marginal cost of any given level of environmental
stewardship declines. Thus, economists think a lot about which kinds of policies do the best job of giving people incentives to
develop cheaper ways to clean and steward the environment.
6.5.4 [Link]
Fairness
A project can have very high aggregate net benefits, but distribute the costs and benefits very unevenly within society. We may
have both ethical and practical reasons not to want a policy that is highly unfair. Some people have strong moral or philosophical
preferences for policies that are equitable. In addition, if the costs of a policy are borne disproportionately by a single group of
people or firms, that group is likely to fight against it in the political process. Simple cost-benefit analyses do not speak to issues of
equity. However, it is common for policy analyses to break total costs and benefits down among subgroups to see if uneven patterns
exist in their distribution. Studies can break down policy effects by income category to see if a policy helps or hurts people
disproportionately depending on whether they are wealthy or poor. Other analyses carry out regional analyses of policy effects. .
For example, climate-change mitigation policy increases costs disproportionately for poor households because of patterns in energy
consumption across income groups. Furthermore, the benefits and costs of such policy are not uniform across space in the U.S. The
benefits of reducing the severity of climate change will accrue largely to those areas that would be hurt most by global warming
(coastal states hit by sea level rise and more hurricanes, Western states hit by severe water shortages) while the costs will fall most
heavily on regions of the country with economies dependent on sales of oil and coal.
Some of our evaluative criteria are closely related to each other; a policy cannot be efficient if it is not cost-effective. However,
other criteria have nothing to do with each other; a policy can be efficient but not equitable, and vice versa. Cost-benefit analyses
provide crude litmus tests—we surely do not want to adopt policies that have costs exceeding their benefits. However, good policy
development and evaluation considers a broader array of criteria.
Review Questions
1. What are some common mistakes people make in evaluating the costs of a policy or project, and what should you do to avoid
them?
2. What is discounting, and how do we use it in calculating the costs and the benefits of a project that has effects over a long
period of time?
3. Why is discounting controversial?
4. How does cost-benefit analysis complement some of the other measures people use to evaluate a policy or project?
References
Braden, J. B. & A. W. Ando. 2011. Economic costs, benefits, and achievability of low-impact development based stormwater
regulations, in Economic Incentives for Stormwater Control, Hale W. Thurston, ed., Taylor & Francis, Boca Raton, FL.
Carson, R. T., Mitchell, R. C., Hanemann, M., Kopp, R. J., Presser, S., and Ruud, P. A. (2003).Contingent valuation and lost
passive use: Damages from the Exxon Valdez oil spill. Environmental and Resource Economics, 25(3), 257-286. DOI:
10.1023/A:1024486702104.
Dawson, D. & Shogren, J. F. (2001). An update on priorities and expenditures under the Endangered Species Act. Land Economics,
77(4), 527-532.
EDV&CBN (2000). Environmental groups protest alteration of U.S. Army Corps cost benefit analysis. Environmental Damage
Valuation and Cost Benefit News, 7(4), 1-3. [Link] .
Glossary
additionality
The extent to which a new action (policy, project etc.) adds to the benefits or costs associated with existing conditions.
cost-benefit analysis
Evaluation of how the overall benefits of a project compare to its costs.
cost effectiveness
The extent to which an outcome is achieved at the lowest cost possible.
counterfactual
The scenario against which a different scenario should be compared; in policy analysis, the way the world would have been in
the absence of the policy.
6.5.5 [Link]
discounting
The process of converting future values (costs or benefits) into an equivalent amount of money received today; controls for
human time preference.
opportunity cost
The cost of foregoing the next best choice when making a decision.
sensitivity analysis
Evaluation of how sensitive the results of an analysis are to changes in assumptions used in the analysis.
This page titled 6.5: Evaluating Projects and Policies is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto
Cabezas (GALILEO Open Learning Materials) .
6.5.6 [Link]
6.6: Solutions- Property Rights, Regulations, and Incentive Policies
Learning Objectives
After reading this module, students should be able to
know why having clearly defined property rights might improve environmental outcomes and be aware of the limitations of
that approach
define several different types of command and control regulations, and understand their comparative advantages
know what incentive polivies (taxes and tradeable permits) are, what they do and what their strengths and weaknesses are
Introduction
Governments have implemented many policies to solve problems with environmental quality and natural resource depletion. Every
policy is unique and deserves detailed individual analysis in the policymaking process—the devil is always in the details. However,
economists have developed a taxonomy of policy types. This taxonomy helps us to understand general principles about how
policies of different types are likely to perform and under which circumstances they are likely to work best. Policies are broadly
characterized as either command-and-control or incentive policies. Command and control includes several types of standards.
Incentive policies include taxes, tradable permits, and liability.
Property Rights
In 1960, Ronald Coase wrote the pioneering article "The Problem of Social Cost" in which he put forth ideas about externalities
that have come to be known as the Coase theorem (Coase, 1960). The basic idea of the Coase theorem is that if property rights over
a resource are well specified, and if the parties with an interest in that resource can bargain freely, then the parties will negotiate an
outcome that is efficient regardless of who has the rights over the resource. The initial allocation of rights will not affect the
efficiency of the outcome, but it will affect the distribution of wealth between the parties because the party with the property rights
can extract payment from the other parties as part of the agreement.
To bring this abstract idea to life, we will draw on the classic example employed by generations of economists to think about the
Coase theorem. Suppose a farmer and a rancher live next door to each other. There is land between them on which the farmer wants
to plant crops, but the rancher's cows keep eating the crops. The farmer would like to have no cows on the land, and the rancher
would like the farmer to stop planting crops so the cows could eat as much grass as they like. The efficient outcome is where the
marginal benefit of a cow to the rancher is just equal to the marginal cost to the farmer of that cow's grazing. If the farmer is given
property rights over the land, the rancher will have an incentive to pay the farmer to allow the efficient number of cows rather than
zero; if the rancher has the rights, then the farmer will have to pay the rancher to limit the herd to just the efficient size. Either way
they have incentives to negotiate to the efficient outcome because otherwise both of them could be made better off.
The Coase theorem is invoked by some scholars and policy analysts to argue that government policy is not needed to correct
problems of externalities; all you need is property rights, and private negotiations will take care of the rest. However, Coase himself
recognized in his writing that often the real world does not have the frictionless perfect negotiation on which the conclusions of the
theorem rest. For example, there are transaction costs in bargaining, and those transaction costs can be prohibitively large when
many people are involved, as in the case of air pollution from a factory. Furthermore, perfect bargaining requires perfect
information. People often are unaware of the threats posed to their health by air and water pollution, and thus do not know what
kind of bargaining would actually be in their own best interests.
Despite these limitations, there is a move afoot to use property right development to effect environmental improvement and
improve natural resource stewardship, particularly in developing countries. In parts of Africa, new systems have given villages
property rights over wildlife on their lands, yielding stronger incentives to manage wildlife well and demonstrably increasing
wildlife populations. In South America, land-tenure reform is promoted as a way to reduce deforestation.
6.6.1 [Link]
some facet of pollution control or natural resource use, and then a government agency enforces the rules. Here we discuss and
explore examples of a few kinds of such "command-and-control" regulations.
Ambient Standards
Some policies have targets for the quality of some element of the environment that results from human behavior and natural
processes. An ambient standard
establishes a level of environmental quality that must be met. The Clean Air Act directs the U.S. Environmental Protection Agency
(EPA) to establish National Ambient Air Quality Standards (NAAQSs) for a range of air pollutants such as ozone and fine
particles. The Clean Water Act directs state offices of the EPA to set ambient water quality standards for rivers and streams in their
boundaries. In practice, however, such standards are binding only on state regulators. State EPA offices are responsible for
developing plans to ensure that air and surface water bodies meet these ambient quality standards, but they cannot do the clean up
on their own. They need to use a different set of tools to induce private agents to actually reduce or clean up pollution such that the
ambient standards can be met.
Some ambient standards (such as the NAAQSs) have provoked criticism from economists for being uniform across space. Every
county in the country has to meet the same air quality goals, even though the efficient levels of air quality might vary from one
county to the next with variation in the marginal benefits and marginal costs of cleaning the air. However, uniform ambient
standards grant all people in the U.S. the same access to clean air—a goal that has powerful appeal on the grounds of equity.
Individual Standards
First, we discuss a kind of policy applied to individual people or companies called a technology standard. Pollution and resource
degradation result from a combination of human activity and the characteristics of the technology that humans employ in that
activity. Behavior can be difficult to monitor and control. Hence, lawmakers have often drafted rules to control our tools rather than
our behaviors. For example, automakers are required to install catalytic converters on new automobiles so that cars have lower
pollution rates, and people in some parts of the country must use low-flow showerheads and water-efficient toilets to try to reduce
water usage.
Technology standards have the great advantage of being easy to monitor and enforce; it is easy for a regulator to check what
pollution controls are in the design of a car. Under some circumstances technology standards can reduce pollution and the rate of
natural resource destruction, but they have several serious limitations. First, they provide no incentives for people to alter elements
of their behavior other than technology choice. Cars may have to have catalytic converters to reduce emissions per mile, but people
are given no reason to reduce the number of miles they drive. Indeed, these policies can sometimes have perverse effects on
behavior. Early generations of water-efficient toilets performed very poorly; they used fewer gallons of water per flush, but people
found themselves flushing multiple times in order to get waste down the pipes. Thus, these standards are neither always efficient
nor cost effective. Second, technology standards are the worst policy in the toolkit for promoting technological innovation. Firms
are actively forbidden from using any technology other than the one specified in the standards. Automakers might think of a better
and cheaper way to reduce air pollution from cars, but the standard says they have to use catalytic converters.
A second type of policy applied to individual agents is called a performance standard. Performance standards set strict limits on an
outcome of human activity. For example, in order to meet the NAAQSs, state EPA offices set emission standards for air pollution
sources in their states. Those standards limit the amount of pollution a factory or power plant can release into the air, though each
source can control its pollution in any way it sees fit. The limits on pollution are the same for all sources of a given type (e.g.,
power plant, cement factory, etc.). Performance standards are also used in natural resource regulation. For example, because
stormwater runoff causes flooding and harms aquatic habitat, the city of Chicago requires all new development to be designed
handle the first inch of rainfall in a storm onsite before runoff begins.
To enforce a performance standard the regulator must be able to observe the outcome of the agents' activities (e.g. measure the
pollution, estimate the runoff). If that is possible, these policies have some advantages over technology standards. Performance
standards do give people and firms some incentive to innovate and find cheaper ways to reduce pollution because they are free to
use any technology they like to meet the stated requirements. Performance standards are also more efficient because they give
people and firms incentives to change multiple things about their activity to reduce the total cost of pollution abatement; a power
plant can reduce sulfur dioxide emissions by some combination of installing scrubber technology, switching to low-sulfur coal, and
reducing total energy generation.
6.6.2 [Link]
Performance standards also have some drawbacks and limitations, however. It is difficult for a regulator to figure out the cost
effective allocation of total pollution reduction between sources and then set different performance standards for each source to
reach that cost effective allocation. Hence, performance standards tend to be uniform across individual pollution sources, and so
pollution reduction is not done in the cheapest way possible for the industry and society overall. This problem is particularly severe
where there is great variation among sources in their abatement costs, and thus the cost-effective allocation of cleanup among
sources is far from uniform.
Incentive Policies
Other approaches to environmental policy give firms and individuals incentives to change their behavior rather than mandating
specific changes. These incentive policies try to make use of market forces for what they do best—allocating resources cost-
effectively within an economy—while correcting the market failures associated with externalities, public goods, and common pool
resources.
Tax/Subsidy
Environmental taxes are based on a simple premise: if someone is not bearing the full social costs of their actions, then we should
charge them an externality tax per unit of harmful activity (e.g. ton of pollution, gallon of stormwater runoff) that is equal to the
marginal cost that is not borne by the individual. In this way, that person must internalize the externality, and will have the
incentive to choose a level of activity that is socially optimal. Thus, if we think the social marginal cost of ton of carbon dioxide
(because of its contribution to climate change) is $20, then we could charge a tax of $20 per ton of carbon dioxide emitted. The
easiest way to do this would be to have a tax on fossil fuels according to the amount of carbon dioxide that will be emitted when
they are burned.
If a price is placed on carbon dioxide, all agents would have an incentive to reduce their carbon dioxide emissions to the point
where the cost to them of reducing one more unit (their marginal abatement cost) is equal to the per unit tax. Therefore, several
good things happen. All carbon dioxide sources are abating to the same marginal abatement cost, so the total abatement is
accomplished in the most cost-effective way possible. Furthermore, total emissions in the economy overall will go down to the
socially efficient level. Firms and individuals have very broad incentives to change things to reduce carbon dioxide emissions—
reduce output and consumption, increase energy efficiency, switch to low carbon fuels—and strong incentives to figure out how to
innovate so those changes are less costly. Finally, the government could use the revenue it collects from the tax to correct any
inequities in the distribution of the program's cost among people in the economy or to reduce other taxes on things like income.
While taxes on externality-generating activities have many good features, they also have several drawbacks and limitations. First,
while an externality tax can yield the efficient outcome (where costs and benefits are balanced for the economy as a whole), that
only happens if policy makers know enough about the value of the externality to set the tax at the right level. If the tax is too low,
we will have too much of the harmful activity; if the tax is too high, the activity will be excessively suppressed.
Second, even if we are able to design a perfect externality tax in theory, such a policy can be difficult to enforce. The enforcement
agency needs to be able to measure the total quantity of the thing being taxed. In some cases that is easy—in the case of carbon
dioxide for example, the particular fixed link between carbon dioxide emissions and quantities of fossil fuels burned means that
through the easy task of measuring fossil fuel consumption we can measure the vast majority of carbon dioxide emissions.
However, many externality-causing activities or materials are difficult to measure in total. Nitrogen pollution flows into streams as
a result of fertilizer applications on suburban lawns, but it is impossible actually to measure the total flow of nitrogen from a single
lawn over the course of a year so that one could tax the homeowner for that flow.
Third, externality taxes face strong political opposition from companies and individuals who don't want to pay the tax. Even if the
government uses the tax revenues to do good things or to reduce other tax rates, the group that disproportionately pays the tax has
an incentive to lobby heavily against such a policy. This phenomenon is at least partly responsible for the fact that there are no
examples of pollution taxes in the U.S. Instead, U.S. policy makers have implemented mirror-image subsidy policies, giving
subsidies for activities that reduce negative externalities rather than taxing activities that cause those externalities. Environmental
policy in the case of U.S. agriculture is a prime example of this, with programs that pay farmers to take lands out of production or
to adopt environmentally friendly farming practices. A subsidy is equivalent to the mirror-image tax in most ways. However, a
subsidy tends to make the relevant industry more profitable (in contrast to a tax, which reduces profits), which in turn can stimulate
greater output and have a slight perverse effect on total pollution or environmental degradation; degradation per unit output might
go down, but total output goes up.
6.6.3 [Link]
Tradable Permits
Another major type of incentive policy is a tradable permits scheme. Tradable permits are actually very similar to externality taxes,
but they can have important differences. These policies are colloquially known as "cap and trade". If we know the efficient amount
of the activity to have (e.g., number of tons of pollution, amount of timber to be logged) the policy maker can set a cap on the total
amount of the activity equal to the efficient amount. Permits are created such that each permit grants the holder permission for one
unit of the activity. The government distributes these permits to the affected individuals or firms, and gives them permission to sell
(trade) them to one another. In order to be in compliance with the policy (and avoid punishment, such as heavy fines) all agents
must hold enough permits to cover their total activity for the time period. The government doesn't set a price for the activity in
question, but the permit market yields a price for the permits that gives all the market participants strong incentives to reduce their
externality-generating activities, to make cost-effective trades with other participants, and to innovate to find cheaper ways to be in
compliance. Tradable permit policies are similar to externality taxes in terms of efficiency, cost-effectiveness, and incentives to
innovate.
Tradable permit policies have been used in several environmental and natural resource policies. The U.S. used tradable permits
(where the annual cap declined to zero over a fixed number of years) in two separate policy applications to reduce the total cost to
society of (a) phasing out the use of lead in gasoline and (b) eliminating production of ozone-depleting chlorofluorocarbons. The
Clean Air Act amendments of 1990 put in place a nationwide tradable permit program for emissions of acid-rain precursor sulfur
dioxide from electric power plants. The European Union used a tradable permit market as part of its policy to reduce carbon
dioxide emissions under the Kyoto protocol. Individual tradable quotas for fish in fisheries of Alaska and New Zealand have been
used to rationalize fishing activity and keep total catches down to efficient and sustainable levels (see 8.3).
Tradable permits have been adopted more widely than externality taxes. Two factors may contribute to that difference. First,
tradable permit policies can have different distributional effects from taxes depending on how the permits are given out. If the
government auctions the permits to participants in a competitive marketplace, then the tradable permit scheme is the same as the
tax; the industry pays the government an amount equal to the number of permits multiplied by the permit price. However, policy
makers more commonly design policies where the permits are initially given for free to participants in the market, and then
participants sell the permits to each other. This eliminates the transfer of wealth from the regulated sector (the electric utilities, the
fishing boats, etc.) to the government, a feature that has been popular with industry. Second, taxes and tradable permits behave
differently in the face of uncertainty. A tax policy fixes the marginal cost to the industry, but might yield more or less of the
harmful activity than expected if market conditions fluctuate. A cap and trade program fixes the total amount of the harmful
activity, but can yield costs to industry that are wildly variable. Environmentalists have liked the outcome certainty of tradable
permits.
Liability
A third type of environmental policy was not designed by economists, but still functions to give agents incentives to take efficient
actions to reduce environmental degradation: liability. Liability provisions can make people or firms pay for the damages caused by
their actions. If the expected payment is equal to the total externality cost, then liability makes the agent internalize the externality
and take efficient precautions to avoid harming the environment.
Two kinds of liability exist in the U.S.: statutory and common law. Common law derives from a long tradition of legal history in
the U.S.—people have sued companies for damages from pollution under tort law under doctrines such as nuisance, negligence, or
trespass. This approach has been highly problematic for a number of reasons. For example, tort law places a high burden of proof
on the plaintiff to show that damages resulted directly from actions taken by the defendant. Plaintiffs have often struggled with that
burden because pollution problems are often caused by many sources, and the harm caused by pollution can display large lags in
space and time. If the defendant expects with high probability not to be held responsible by the courts, then liability does not
function effectively to make agents internalize the externality costs of their actions.
Frustration with common law has led to several strong statutory liability laws in the U.S. which make explicit provisions for
holding firms liable for damages from pollution with much more manageable burdens of proof. The Oil Pollution Act of 1990
holds companies like Exxon and British Petroleum strictly liable for the damages caused by oil spills from accidents such as the
Valdez grounding in Prince William Sound or the Deepwater Horizon explosion in the Gulf of Mexico. Under a rule of strict
liability, a party is liable for harm if the harm occurred as a result of their actions regardless of the presence (or absence) of
negligence or intent. The Comprehensive Environmental Response, Compensation, and Liability Act (CERCLA, or "Superfund")
holds companies strictly liable for damages from toxic waste "Superfund" sites.
6.6.4 [Link]
These laws have surely increased the extent to which oil and chemical companies take precautions to avoid spills and other releases
of hazardous materials into the environment. However, enforcement of these provisions is very costly. The legal proceedings for a
big case like Deepwater Horizon entail court, lawyer, and expert witness activity (and high fees) for many years. The transaction
costs are so burdensome to society that liability may not be a viable approach for all environmental problems.
Review Questions
1. What are some of the strengths and weaknesses of command and control regulation? When would these be the best policy tool
to use?
2. What are some of the strengths and weaknesses of incentive policies? When would these be the best policy tool to use?
3. Did Coase think government policy was not necessary to solve externality problems? Briefly explain.
4. How do liability laws function as incentive policies? What are some of their limitations?
References
Coase, R.H. 1960. The problem of social cost. Journal of Law and Economics, 3, 1-44.
Glossary
Ambient Standard
A minimum level of overall environmental quality that must be reached.
Coase Theorem
The idea that with property rights and frictionless negotiation, private agents will bargain to reach efficient outcomes even in the
face of externalities.
Eternality Tax
A tax on something that causes negative externalities.
Liability
A legal construct meaning that an agent is held responsible by the courts to pay when that agent does something that imposes costs
on other people in society.
Performance Standard
A regulation specifying something about the outcome of private behaviors.
Technology Standard
A regulation specifying what kind of technology agents must or must not use in their activities.
Tradable Permits
A policy in which the total amount of an activity is limited, but agents can trade the rights to engage in that activity (permits).
This page titled 6.6: Solutions- Property Rights, Regulations, and Incentive Policies is shared under a CC BY license and was authored, remixed,
and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
6.6.5 [Link]
CHAPTER OVERVIEW
This page titled 7: Modern Environmental Management is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto
Cabezas (GALILEO Open Learning Materials) .
1
7.1: Modern Environmental Management – Chapter Introduction
Introduction
In chapter four, the ways in which our current environmental policy evolved were presented and discussed. Although the National
Environmental Policy Act (NEPA) provided lofty goals for our environmental policy, and most importantly a legal basis for action,
the fact remains, then and today, that human actions produce very large quantities of waste, virtually all of it harmful to human and
ecosystem health if not managed properly. This chapter is about how we currently manage these wastes (Module 9.2), the laws and
regulations that define our system of waste management (Module 9.4), and how we determine the consequences, i.e. risks,
associated with chemicals released into the environment (Module 9.5). Of course, environmental policies will continue to evolve,
and although we may not know the exact pathway or form this will take, environmental policy of the future will most certainly
build upon the laws and regulations that are used today to manage human interactions with the environment. Thus, it is important to
develop an understanding of our current system, its legal and philosophical underpinnings, and the quantitative basis for setting
risk-based priorities.
An interesting example of how our current system of environmental management has adapted to modern, and global, problems is
the U.S. Supreme Court ruling, in April of 2007, in the case of Massachusetts vs. the Environmental Protection Agency that the
USEPA had misinterpreted the Clean Air Act in not classifying, and regulating, carbon dioxide, as a pollutant (the plaintiffs
actually involved twelve states and several cities). Up until that time several administrations had said that the Act did not give the
EPA legal authority to regulate CO2 (and by inference all greenhouse gases). At the time the Clean Air Act was passed (most
recently in 1990), "clean air" was thought to mean both visibly clean air, and also air free of pollutants exposure to which could
cause harm to humans – harm being defined as an adverse outcome over a course of time that might extend to a human lifetime.
And although there was concern about global climate change due to greenhouse gas emissions, the gases themselves were not
thought of as "pollutants" in the classical sense. This ruling set the stage for the EPA to regulate greenhouse gases through a series
of findings, hearings, rulings, and regulations in accord with terms set out in the Clean Air Act. This process is underway at the
present time.
In addition to its significance for potentially mitigating the problem of global climate change, this case illustrates more generally
how the environmental management system we have put in place today might adapt to problems of the future. Laws that are
forward-thinking, not overly proscriptive, and administratively flexible may well accommodate unforeseen problems and needs. Of
course, this does not preclude the passage of new laws or amendments, nor does it imply that all our laws on the environment will
adapt in this way to future problems.
This page titled 7.1: Modern Environmental Management – Chapter Introduction is shared under a CC BY license and was authored, remixed,
and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
7.1.1 [Link]
7.2: Systems of Waste Management
Learning Objectives
After reading this module, students should be able to
recognize various environmental regulations governing the management of solid and hazardous wastes, radioactive waste
and medical waste
understand the environmental concerns with the growing quantities and improper management of wastes being generated
recognize integrated waste management strategies that consist of prevention, minimization, recycling and reuse,
biological treatment, incineration, and landfill disposal
Introduction
Waste is an inevitable by-product of human life. Virtually every human activity generates some type of material side effect or by-
product. When the materials that constitute these by-products are not useful or have been degraded such that they no longer fulfill
their original or other obvious useful purpose, they are classified as a waste material.
Practically speaking, wastes are generated from a wide range of sources and are usually classified by their respective sources.
Common generative activities include those associated with residences, commercial businesses and enterprises, institutions,
construction and demolition activities, municipal services, and water/wastewater and air treatment plants, and municipal incinerator
facilities. Further, wastes are generated from numerous industrial processes, including industrial construction and demolition,
fabrication, manufacturing, refineries, chemical synthesis, and nuclear power/nuclear defense sources (often generating low- to
high-level radioactive wastes).
Population growth and urbanization (with increased industrial, commercial and institutional establishments) contribute to increased
waste production, as do the rapid economic growth and industrialization throughout the developing world. These social and
economic changes have led to an ever-expanding consumption of raw materials, processed goods, and services. While these trends
have, in many ways, improved the quality of life for hundreds of millions of people, it has not come without drastic costs to the
environment. Proper management of a range of wastes has become necessary in order to protect public health and the environment
as well as ensure sustained economic growth.
It is commonly believed that incineration and landfill disposal represent preferred options in dealing with waste products; however,
many wastes have the potential to be recycled or re-used for some purpose or in some manner. Some waste materials may be
reclaimed or re-generated and used again for their original or similar purpose, or they may be physically or chemically changed and
employed for alternative uses. As natural resources continue to be depleted, and as incineration and landfill disposal options
become more costly and unsustainable, numerous economic and social incentives are being promoted by government agencies to
prevent or reduce waste generation and develop new methods and technologies for recycling and reusing wastes. Such efforts can
have broader implications for energy conservation and the reduction of greenhouse gas emissions that contribute to global climate
change, while concurrently fostering sustainable waste management practices.
This section provides an overview of the existing regulatory framework mandating the management of wastes, environmental
concerns associated with waste generation and management, and various alternatives for the proper management of wastes. Recent
developments towards the development of sustainable waste management systems are also highlighted. It should be mentioned here
that although the content of this section reflects the regulatory framework and practices within the United States, similar
developments and actions have occurred in other developed countries and are increasingly being initiated in numerous developing
countries.
Regulatory Framework
During the course of the 20th century, especially following World War II, the United States experienced unprecedented economic
growth. Much of the growth was fueled by rapid and increasingly complex industrialization. With advances in manufacturing and
chemical applications also came increases in the volume, and in many cases the toxicity, of generated wastes. Furthermore, few if
any controls or regulations were in place with respect to the handling of toxic materials or the disposal of waste products.
Continued industrial activity led to several high-profile examples of detrimental consequences to the environment resulting from
these uncontrolled activities. Finally, several forms of intervention, both in the form of government regulation and citizen action,
occurred in the early 1970s.
7.2.1 [Link]
Ultimately, several regulations were promulgated on the state and federal levels to ensure the safety of public health and the
environment (see Module 9.4). With respect to waste materials, the Resource Conservation and Recovery Act (RCRA), enacted by
the United States Congress, first in 1976 and then amended in 1984, provides a comprehensive framework for the proper
management of hazardous and non-hazardous solid wastes in the United States. RCRA stipulates broad and general legal objectives
while mandating the United States Environmental Protection Agency (USEPA) to develop specific regulations to implement and
enforce the law. The RCRA regulations are contained in Title 40 of the Code of Federal Regulations (CFR), Parts 239 to 299.
States and local governments can either adopt the federal regulations, or they may develop and enforce more stringent regulations
than those specified in RCRA. Similar regulations have been developed or are being developed worldwide to manage wastes in a
similar manner in other countries.
The broad goals of RCRA include: (1) the protection of public health and the environment from the hazards of waste disposal; (2)
the conservation of energy and natural resources; (3) the reduction or elimination of waste; and (4) the assurance that wastes are
managed in an environmentally-sound manner (e.g. the remediation of waste which may have spilled, leaked, or been improperly
disposed). It should be noted here that the RCRA focuses only on active and future facilities and does not address abandoned or
historical sites. These types of environmentally impacted sites are managed under a different regulatory framework, known as the
Comprehensive Environmental Response, Compensation, and Liability Act (CERCLA), or more commonly known as "Superfund."
7.2.2 [Link]
Toxicity (as defined in 40 CFR 261.24), which refers to substances that can induce harmful or fatal effects when ingested or
absorbed, or inhaled.
7.2.3 [Link]
mercury; and toxic organics such as dioxins and furans. All medical wastes represent a small fraction of total waste stream,
estimated to constitute a maximum of approximately two percent. Medical wastes are commonly disposed of through incineration:
as with most wastes, the resulting volume is greatly reduced, and it assures the destruction and sterilization of infectious pathogens.
Disadvantages include the potential of air pollution risks from dioxins and furans as well as the necessary disposal of potentially
hazardous ash wastes. New options for disposal of medical wastes (including infectious wastes) are still being explored. Some
other technologies include irradiation, microwaving, autoclaving, mechanical alternatives, and chemical disinfection, among others.
7.2.4 [Link]
small compared to hazardous waste sites. For example, explosives and radioactive wastes are primarily located at Department of
Energy (DOE) sites because many of these facilities have been historically used for weapons research, fabrication, testing, and
training. Organic contaminants are largely found at oil refineries, or petroleum storage sites, and inorganic and pesticide
contamination usually is the result of a variety of industrial activities as well as agricultural activities. Yet, soil and groundwater
contamination are not the only direct adverse effects of improper waste management activities – recent studies have also shown
that greenhouse gas emissions from the wastes are significant, exacerbating global climate change.
A wide range of toxic chemicals, with an equally wide distribution of respective concentrations, is found in waste streams. These
compounds may be present in concentrations that alone may pose a threat to human health or may have a synergistic/cumulative
effect due to the presence of other compounds. Exposure to hazardous wastes has been linked to many types of cancer, chronic
illnesses, and abnormal reproductive outcomes such as birth defects, low birth weights, and spontaneous abortions. Many studies
have been performed on major toxic chemicals found at hazardous waste sites incorporating epidemiological or animal tests to
determine their toxic effects.
As an example, the effects of radioactive materials are classified as somatic or genetic. The somatic effects may be immediate or
occur over a long period of time. Immediate effects from large radiation doses often produce nausea and vomiting, and may be
followed by severe blood changes, hemorrhage, infection, and death. Delayed effects include leukemia, and many types of cancer
including bone, lung, and breast cancer. Genetic effects have been observed in which gene mutations or chromosome abnormalities
result in measurable harmful effects, such as decreases in life expectancy, increased susceptibility to sickness or disease, infertility,
or even death during embryonic stages of life. Because of these studies, occupational dosage limits have been recommended by the
National Council on Radiation Protection. Similar studies have been completed for a wide range of potentially hazardous materials.
These studies have, in turn, been used to determine safe exposure levels for numerous exposure scenarios, including those that
consider occupational safety and remediation standards for a variety of land use scenarios, including residential, commercial, and
industrial land uses.
Figure shows the hierarchy of management of wastes in order or preference, starting with prevention as the most favorable to
disposal as the least favorable option. Source: Drstuey via Wikimedia Commons
7.2.5 [Link]
Waste Prevention
The ideal waste management alternative is to prevent waste generation in the first place. Hence, waste prevention is a basic goal of
all the waste management strategies. Numerous technologies can be employed throughout the manufacturing, use, or post-use
portions of product life cycles to eliminate waste and, in turn, reduce or prevent pollution. Some representative strategies include
environmentally conscious manufacturing methods that incorporate less hazardous or harmful materials, the use of modern leakage
detection systems for material storage, innovative chemical neutralization techniques to reduce reactivity, or water saving
technologies that reduce the need for fresh water inputs.
Waste Minimization
In many cases, wastes cannot be outright eliminated from a variety of processes. However, numerous strategies can be
implemented to reduce or minimize waste generation. Waste minimization, or source reduction, refers to the collective strategies of
design and fabrication of products or services that minimize the amount of generated waste and/or reduce the toxicity of the
resultant waste. Often these efforts come about from identified trends or specific products that may be causing problems in the
waste stream and the subsequent steps taken to halt these problems. In industry, waste can be reduced by reusing materials, using
less hazardous substitute materials, or by modifying components of design and processing. Many benefits can be realized by waste
minimization or source reduction, including reduced use of natural resources and the reduction of toxicity of wastes.
Waste minimization strategies are extremely common in manufacturing applications; the savings of material use preserves
resources but also saves significant manufacturing related costs. Advancements in streamlined packaging reduces material use,
increased distribution efficiency reduces fuel consumption and resulting air emissions. Further, engineered building materials can
often be designed with specific favorable properties that, when accounted for in overall structural design, can greatly reduce the
overall mass and weight of material needed for a given structure. This reduces the need for excess material and reduces the waste
associated with component fabrication.
The dry cleaning industry provides an excellent example of product substitution to reduce toxic waste generation. For decades, dry
cleaners used tetrachloroethylene, or "perc" as a dry cleaning solvent. Although effective, tetrachloroethylene is a relatively toxic
compound. Additionally, it is easily introduced into the environment, where it is highly recalcitrant due to its physical properties.
Further, when its degradation occurs, the intermediate daughter products generated are more toxic to human health and the
environment.
Because of its toxicity and impact on the environment, the dry cleaning industry has adopted new practices and increasingly utilizes
less toxic replacement products, including petroleum-based compounds. Further, new emerging technologies are incorporating
carbon dioxide and other relatively harmless compounds. While these substitute products have in many cases been mandated by
government regulation, they have also been adopted in response to consumer demands and other market-based forces.
7.2.6 [Link]
paper manufacturing, or the processing of old aluminum cans into new aluminum products. In other cases, reclaimed materials
undergo little or no processing prior to their re-use. Some common examples include the use of tree waste as wood chips, or the use
of brick and other fixtures into new structural construction. In any case, the success of recycling depends on effective collection
and processing of recyclables, markets for reuse (e.g. manufacturing and/or applications that utilize recycled materials), and public
acceptance and promotion of recycled products and applications utilizing recycled materials.
Biological Treatment
Landfill disposal of wastes containing significant organic fractions is increasingly discouraged in many countries, including the
United States. Such disposal practices are even prohibited in several European countries. Since landfilling does not provide an
attractive management option, other techniques have been identified. One option is to treat waste so that biodegradable materials
are degraded and the remaining inorganic waste fraction (known as residuals) can be subsequently disposed or used for a beneficial
purpose.
Biodegradation of wastes can be accomplished by using aerobic composting, anaerobicdigestion, or mechanical biological
treatment (MBT) methods. If the organic fraction can be separated from inorganic material, aerobic composting or anaerobic
digestion can be used to degrade the waste and convert it into usable compost. For example, organic wastes such as food waste,
yard waste, and animal manure that consist of naturally degrading bacteria can be converted under controlled conditions into
compost, which can then be utilized as natural fertilizer. Aerobic composting is accomplished by placing selected proportions of
organic waste into piles, rows or vessels, either in open conditions or within closed buildings fitted with gas collection and
treatment systems. During the process, bulking agents such as wood chips are added to the waste material to enhance the aerobic
degradation of organic materials. Finally, the material is allowed to stabilize and mature during a curing process where pathogens
are concurrently destroyed. The end-products of the composting process include carbon dioxide, water, and the usable compost
material.
Compost material may be used in a variety of applications. In addition to its use as a soil amendment for plant cultivation, compost
can be used remediate soils, groundwater, and stormwater. Composting can be labor-intensive, and the quality of the compost is
heavily dependent on proper control of the composting process. Inadequate control of the operating conditions can result in
compost that is unsuitable for beneficial applications. Nevertheless, composting is becoming increasingly popular; composting
diverted 82 million tons of waste material away the landfill waste stream in 2009, increased from 15 million tons in 1980. This
diversion also prevented the release of approximately 178 million metric tons of carbon dioxide in 2009 – an amount equivalent to
the yearly carbon dioxide emissions of 33 million automobiles.
In some cases, aerobic processes are not feasible. As an alternative, anaerobic processes may be utilized. Anaerobic digestion
consists of degrading mixed or sorted organic wastes in vessels under anaerobic conditions. The anaerobic degradation process
produces a combination of methane and carbon dioxide (biogas) and residuals (biosolids). Biogas can be used for heating and
electricity production, while residuals can be used as fertilizers and soil amendments. Anaerobic digestion is a preferred
degradation for wet wastes as compared to the preference of composting for dry wastes. The advantage of anaerobic digestion is
biogas collection; this collection and subsequent beneficial utilization makes it a preferred alternative to landfill disposal of wastes.
Also, waste is degraded faster through anaerobic digestion as compared to landfill disposal.
Another waste treatment alternative, mechanical biological treatment (MBT), is not common in the United States. However, this
alternative is widely used in Europe. During implementation of this method, waste material is subjected to a combination of
mechanical and biological operations that reduce volume through the degradation of organic fractions in the waste. Mechanical
operations such as sorting, shredding, and crushing prepare the waste for subsequent biological treatment, consisting of either
aerobic composting or anaerobic digestion. Following the biological processes, the reduced waste mass may be subjected to
incineration.
Incineration
Waste degradation not only produces useful solid end-products (such as compost), degradation by-products can also be used as a
beneficial energy source. As discussed above, anaerobic digestion of waste can generate biogas, which can be captured and
incorporated into electricity generation. Alternatively, waste can be directly incinerated to produce energy. Incineration consists of
waste combustion at very high temperatures to produce electrical energy. The byproduct of incineration is ash, which requires
proper characterization prior to disposal, or in some cases, beneficial re-use. While public perception of incineration can be
negative, this is often based reactions to older, less efficient technologies. New incinerators are cleaner, more flexible and efficient,
and are an excellent means to convert waste to energy while reducing the volume of waste. Incineration can also offset fossil fuel
7.2.7 [Link]
use and reduce greenhouse gas (GHG) emissions (Bogner et al., 2007). It is widely used in developed countries due to landfill
space limitations. It is estimated that about 130 million tons of waste are annually combusted in more than 600 plants in 35
countries. Further, incineration is often used to effectively mitigate hazardous wastes such as chlorinated hydrocarbons, oils,
solvents, medical wastes, and pesticides.
Despite all these advantages, incineration is often viewed negatively because of the resulting air emissions, the creation of daughter
chemical compounds, and production of ash, which is commonly toxic. Currently, many 'next generation" systems are being
researched and developed, and the USEPA is developing new regulations to carefully monitor incinerator air emissions under the
Clean Air Act.
Landfill Disposal
Despite advances in reuse and recycling, landfill disposal remains the primary waste disposal method in the United States. As
previously mentioned, the rate of MSW generation continues to increase, but overall landfill capacity is decreasing. New
regulations concerning proper waste disposal and the use of innovative liner systems to minimize the potential of groundwater
contamination from leachate infiltration and migration have resulted in a substantial increase in the costs of landfill disposal. Also,
public opposition to landfills continues to grow, partially inspired by memories of historic uncontrolled dumping practices the
resulting undesirable side effects of uncontrolled vectors, contaminated groundwater, unmitigated odors, and subsequent
diminished property values.
Landfills can be designed and permitted to accept hazardous wastes in accordance with RCRA Subtitle C regulations, or they may
be designed and permitted to accept municipal solid waste in accordance with RCRA Subtitle D regulations. Regardless of their
waste designation, landfills are engineered structures consisting of bottom and side liner systems, leachate collection and removal
systems, final cover systems, gas collection and removal systems, and groundwater monitoring systems (Sharma and Reddy, 2004).
An extensive permitting process is required for siting, designing and operating landfills. Post-closure monitoring of landfills is also
typically required for at least 30 years. Because of their design, wastes within landfills are degraded anaerobically. During
degradation, biogas is produced and collected. The collection systems prevent uncontrolled subsurface gas migration and reduce
the potential for an explosive condition. The captured gas is often used in cogeneration facilities for heating or electricity
generation. Further, upon closure, many landfills undergo "land recycling" and redeveloped as golf courses, recreational parks, and
other beneficial uses.
Wastes commonly exist in a dry condition within landfills, and as a result, the rate of waste degradation is commonly very slow.
These slow degradation rates are coupled with slow rates of degradation-induced settlement, which can in turn complicate or
reduce the potential for beneficial land re-use at the surface. Recently, the concept of bioreactor landfills has emerged, which
involves recirculation of leachate and/or injection of selected liquids to increase the moisture in the waste, which in turn induces
rapid degradation. The increased rates of degradation increase the rate of biogas production, which increases the potential of
beneficial energy production from biogas capture and utilization.
Summary
Many wastes, such as high-level radioactive wastes, will remain dangerous for thousands of years, and even MSW can produce
dangerous leachate that could devastate an entire eco-system if allowed infiltrate into and migrate within groundwater. In order to
protect human health and the environment, environmental professionals must deal with problems associated with increased
generation of waste materials. The solution must focus on both reducing the sources of wastes as well as the safe disposal of
wastes. It is, therefore, extremely important to know the sources, classifications, chemical compositions, and physical
characteristics of wastes, and to understand the strategies for managing them.
Waste management practices vary not only from country to country, but they also vary based on the type and composition of waste.
Regardless of the geographical setting of the type of waste that needs to be managed, the governing principle in the development of
any waste management plan is resource conservation. Natural resource and energy conservation is achieved by managing materials
more efficiently. Reduction, reuse, and recycling are primary strategies for effective reduction of waste quantities. Further, proper
waste management decisions have increasing importance, as the consequences of these decisions have broader implications with
respect to greenhouse gas emissions and global climate change. As a result, several public and private partnership programs are
under development with the goal of waste reduction through the adoption of new and innovative waste management technologies.
Because waste is an inevitable by-product of civilization, the successful implementation of these initiatives will have a direct effect
on the enhanced quality of life for societies worldwide.
7.2.8 [Link]
Review Questions
1. How is hazardous waste defined according to the Resource Conservation and Recovery Act (RCRA)? In your opinion, is this
definition appropriate? Explain.
2. Explain specific characteristics of radioactive and medical wastes that make their management more problematic than MSW.
3. Compare and contrast environmental concerns with wastes in a rural versus urban setting.
4. What are the pros and cons of various waste management strategies? Do you agree or disagree with the general waste
management hierarchy?
5. Explain the advantages and disadvantages of biological treatment and incineration of wastes.
References
1. Bogner, J., Ahmed, M.A., Diaz, C. Faaij, A., Gao, Q., Hashimoto,S., et al. (2007). Waste Management, In B. Metz, O.R.
Davidson, P.R. Bosch, R. Dave, L.A. Meyer (Eds.), Climate Change 2007: Mitigation. Contribution of Working Group III to the
Fourth Assessment Report of the Intergovernmental Panel on Climate Change (pp. 585-618). Cambridge University Press,
Cambridge, United Kingdom and New York, NY, USA. Retrieved August 19, 2010 from [Link]/pdf/assessment-re...-
[Link]
2. Sharma, H.D. & Reddy, K.R. (2004). Geoenvironmental Engineering: Site Remediation, Waste Containment, and Emerging
Waste Management Technologies. Hoboken, NJ: John Wiley.
Glossary
aerobic
Living systems or processes that require, or are not destroyed by, the presence of oxygen.
anaerobic
A living system or process that occurs in, or is not destroyed by, the absence of oxygen.
bioaccumulation
The increase in concentration of a substance in an organism over time.
biological treatment
A treatment technology that uses bacteria to consume organic fraction of municipal solid waste/wastewater.
compost
The stable, decomposed organic material resulting from the composting process.
corrosivity
The ability to corrode metal. Corrosive wastes are wastes that are acidic and capable of corroding metal such as tanks,
containers, drums, and barrels.
digestion
The biochemical decomposition of organic matter of MSW, resulting in its partial gasification, liquefaction, and mineralization.
genetic effects
Effects from some agent, like radiation that are seen in the offspring of the individual who received the agent. The agent must
be encountered pre-conception.
ignitability
Ability to create fire under certain conditions. Ignitable wastes can create fires under these certain conditions.
7.2.9 [Link]
incineration
A thermal process of combusting MSW.
landfills
Designed, controlled and managed disposal sites for MSW spread in layers, compacted to the smallest practical volume, and
covered by material applied at the end of each operating day.
leachate
Wastewater that collects contaminants as it trickles through MSW disposed in a landfill.
medical waste
Any municipal solid waste generated in the diagnosis, treatment, or immunisation of human beings or animals.
mill tailings
Waste material from a conventional uranium recovery facility.
pollution prevention
The active process of identifying areas, processes, and activities which generate excessive waste for the purpose of substitution,
alteration, or elimination of the process to prevent waste generation in the first place.
radioactive waste
Any waste that emits energy as rays, waves, or streams of energetic particles.
reactivity
Materials susceptible to unstable conditions. Reactive wastes are unstable under normal conditions and can create explosions
and or toxic fumes, gases, and vapors when mixed with water.
recycling
Separation physical/mechanical process by which secondary raw materials (such as paper, metals, glass, and plastics.) are
obtained from MSW.
reuse
Using a component of MSW in its original form more than once.
solid waste
According to the Resource Conservation and Recovery Act (RCRA), solid waste is: garbage; refuse; sludge from a waste
treatment plant, water supply treatment plant, or air pollution control facility; and other discarded materials, including solid,
liquid, or contained gaseous material resulting from industrial, commercial, mining, and agricultural operations, and from
community activities.
somatic effects
Effects from some agent, like radiation that are seen in the individual who receives the agent.
7.2.10 [Link]
toxicity
The degree to which a chemical substance (or physical agent) elicits a deleterious or adverse effect upon the biological system
of an organism exposed to the substance over a designated time period.
waste to energy
Combustion of MSW to generate electrical energy or heat
waste minimization
Measures or techniques that reduce the amount of wastes generated during industrial production processes; the term is also
applied to recycling and other efforts to reduce the amount of waste going into the waste management system.
waste prevention
The design, manufacture, purchase or use of materials or products to reduce their amount or toxicity before they enter the
municipal solid waste stream. Because it is intended to reduce pollution and conserve resources, waste prevention should not
increase the net amount or toxicity of wastes generated throughout the life of a product.
This page titled 7.2: Systems of Waste Management is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto
Cabezas (GALILEO Open Learning Materials) .
7.2.11 [Link]
7.3: Case Study- Electronic Waste and Extended Producer Responsibility
Electronic waste, commonly known as e-waste, refers to discarded electronic products such as televisions, computers and computer
peripherals (e.g. monitors, keyboards, disk drives, and printers), telephones and cellular phones, audio and video equipment, video
cameras, fax and copy machines, video game consoles, and others (see Figure 7.3.1).
Figure 7.3.1 Electronic Waste. Photograph shows many computers piled up in a parking lot as waste. Source: Bluedisk via
Wikimedia Commons
In the United States, it is estimated that about 3 million tons of e-waste are generated each year. This waste quantity includes
approximately 27 million units of televisions, 205 million units of computer products, and 140 million units of cell phones. Less
than 15 to 20 percent of the e-waste is recycled or refurbished; the remaining percentage is commonly disposed of in landfills
and/or incinerated. It should be noted that e-waste constitutes less than 4 percent of total solid waste generated in the United States.
However, with tremendous growth in technological advancements in the electronics industry, many electronic products are
becoming obsolete quickly, thus increasing the production of e-waste at a very rapid rate. The quantities of e-waste generated are
also increasing rapidly in other countries such as India and China due to high demand for computers and cell phones.
In addition to the growing quantity of e-waste, the hazardous content of e-waste is a major environmental concern and poses risks
to the environment if these wastes are improperly managed once they have reached the end of their useful life. Many e-waste
components consist of toxic substances, including heavy metals such as lead, copper, zinc, cadmium, and mercury as well as
organic contaminants, such as flame retardants (polybrominated biphenyls and polybrominated diphenylethers). The release of
these substances into the environment and subsequent human exposure can lead to serious health and pollution issues. Concerns
have also been raised with regards to the release of toxic constituents of e-waste into the environment if landfilling and/or
incineration options are used to manage the e-waste.
Various regulatory and voluntary programs have been instituted to promote reuse, recycling and safe disposal of bulk e-waste.
Reuse and refurbishing has been promoted to reduce raw material use energy consumption, and water consumption associated with
the manufacture of new products. Recycling and recovery of elements such as lead, copper, gold, silver and platinum can yield
valuable resources which otherwise may cause pollution if improperly released into the environment. The recycling and recovery
operations have to be conducted with extreme care, as the exposure of e-waste components can result in adverse health impacts to
the workers performing these operations. For economic reasons, recycled e-waste is often exported to other countries for recovery
operations. However, lax regulatory environments in many of these countries can lead to unsafe practices or improper disposal of
bulk residual e-waste, which in turn can adversely affect vulnerable populations.
In the United States, there are no specific federal laws dealing with e-waste, but many states have recently developed e-waste
regulations that promote environmentally sound management. For example, the State of California passed the Electronic Waste
Recycling Act in 2003 to foster recycling, reuse, and environmentally sound disposal of residual bulk e-waste. Yet, in spite of
recent regulations and advances in reuse, recycling and proper disposal practices, additional sustainable strategies to manage e-
waste are urgently needed.
One sustainable strategy used to manage e-waste is extended producer responsibility (EPR), also known as product stewardship.
This concept holds manufacturers liable for the entire life-cycle costs associated with the electronic products, including disposal
7.3.1 [Link]
costs, and encourages the use of environmental-friendly manufacturing processes and products. Manufacturers can pursue EPR in
multiple ways, including reuse/refurbishing, buy-back, recycling, and energy production or beneficial reuse applications. Life-cycle
assessment and life-cycle cost methodologies may be used to compare the environmental impacts of these different waste
management options. Incentives or financial support is also provided by some government and/or regulatory agencies to promote
EPR. The use of non-toxic and easily recyclable materials in product fabrication is a major component of any EPR strategy. A
growing number of companies (e.g. Dell, Sony, HP) are embracing EPR with various initiatives towards achieving sustainable e-
waste management.
EPR is a preferred strategy because the manufacturer bears a financial and legal responsibility for their products; hence, they have
an incentive to incorporate green design and manufacturing practices that incorporate easily recyclable and less toxic material
components while producing electronics with longer product lives. One obvious disadvantage of EPR is the higher manufacturing
cost, which leads to increased cost of electronics to consumers.
There is no specific federal law requiring EPR for electronics, but the United States Environmental Protection Agency (USEPA)
undertook several initiatives to promote EPR to achieve the following goals: (1) foster environmentally conscious design and
manufacturing, (2) increase purchasing and use of more environmentally sustainable electronics, and (3) increase safe,
environmentally sound reuse and recycling of used electronics. To achieve these goals, USEPA has been engaged in various
activities, including the promotion of environmental considerations in product design, the development of evaluation tools for
environmental attributes of electronic products, the encouragement of recycling (or e-cycling), and the support of programs to
reduce e-waste, among others. More than 20 states in the United States and various organizations worldwide have already
developed laws and/or policies requiring EPR in some form when dealing with electronic products. For instance, the New York
State Wireless Recycling Act emphasizes that authorized retailers and service providers should be compelled to participate in take-
back programs, thus allowing increased recycling and reuse of e-waste. Similarly, Maine is the first U.S. state to adopt a household
e-waste law with EPR.
In Illinois, Electronic Products Recycling & Reuse Act requires the electronic manufacturers to participate in the management of
discarded and unwanted electronic products from residences. The Illinois EPA has also compiled e-waste collection site locations
where the residents can give away their discarded electronic products at no charge. Furthermore, USEPA compiled a list of local
programs and manufacturers/retailers that can help consumers to properly donate or recycle e-waste.
Overall, the growing quantities and environmental hazards associated with electronic waste are of major concern to waste
management professionals worldwide. Current management strategies, including recycling and refurbishing, have not been
successful. As a result, EPR regulations are rapidly evolving throughout the world to promote sustainable management of e-waste.
However, neither a consistent framework nor assessment tools to evaluate EPR have been fully developed.
This page titled 7.3: Case Study- Electronic Waste and Extended Producer Responsibility is shared under a CC BY license and was authored,
remixed, and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
7.3.2 [Link]
7.4: Government and Laws on the Environment
Learning Objectives
After reading this section, students should be able to
understand the purpose of government regulations set for the protection of human health and the environment
distinguish the current environmental laws and regulations for various types of pollutants present in different media or
phases of the environment
discern the need for future environmental laws as related to the sustainability of industrial activity and the economy
Introduction
In the United States, the laws and regulations pertaining to the protection of the environment have been enacted by the U.S.
Congress. The U.S. Environmental Protection Agency (EPA) is authorized to enforce the environmental laws and to implement the
environmental regulations. The United States environmental laws cover various phases of the environment such as water, air, and
hazardous waste, where most of the regulations have been based on the risk assessment of the pollutants. The major environmental
laws and regulations are briefly listed in the Table 7.4.1.
Table 7.4.1 This Summary of Major Environmental Laws lists major environmental laws enacted from the 1950s onward.
Environmental Issue Description Acronym Year Enacted
7.4.1 [Link]
Water
Clean Water Act
To protect the surface waters of the United States such as lakes, rivers, streams, shorelines and estuaries, the Federal Water
Pollution Control Act was established in 1956. The amendment to the Federal Water Pollution Control Act (FWPCA) of 1972
focused on surface water quality goals, effluent limits based on available technology, and a national discharge permit system. The
FWPCA (1972) introduced effluent limits for chemical substances in surface waters in conjunction with a National Pollutant
Discharge Elimination System (NPDES) allowing for enforceable control over permits obtained by industry for discharge of
effluents containing pollutants into natural water systems. The Clean Water Act (CWA) of 1977 placed emphasis on the control of
waterborne toxic substances released into natural surface waters. The CWA introduced a Priority List of Pollutants which includes
127 toxic chemical substances including synthetic organic compounds and heavy metals. In accordance with the CWA, the EPA
must establish effluent limitations for chemical substances on the List of Priority Pollutants for discharge by industrial facilities and
municipal wastewater treatment plants.
The CWA aims to provide a system of national effluent standards for each industry, a set of water quality standards, an enforceable
discharge permit program, provisions for special wastes such as toxic chemicals and oil spills, and a construction program for
publicly owned treatment works (POTWs). Municipal wastewater treatment plants are examples of POTWs. The NPDES permits
are issued according to the effluent limitations required by the Federal Water Pollution Control Act and the Clean Water Act.
Because of higher costs associated with treatment of industrial effluents before discharge into natural waters which requires an
NPDES permit, many industries discharge to a municipal sewer and have their wastes treated at the POTW following pretreatment
regulations overseen by the POTW. In addition, the CWA provides permits for stormwater and other non-point source (see
definition in Module "Sustainable Stormwater Management") pollution to prevent stormwater runoff over contaminated land and
paved areas from polluting surface waters such as rivers and lakes. Stormwater pollution prevention plans and stormwater
treatment facilities have to be implemented to avoid contamination of clean water.
7.4.2 [Link]
Nervous
Benzene 0.005 Zero Cancer Arsenic 0.010 Zero system,
cancer
Liver,
Liver, Chromium kidney,
Atrazine 0.003 0.003 0.1 0.1
kidney, lung (total) circulatory
system
Central
Pentachloro
0.001 Zero Cancer Cyanide 0.2 0.2 nervous
phenol
system
Polychlorina
ted Blue baby
0.0005 Zero Cancer Nitrate 10 10
biphenyls syndrome
(PCBs)
Kidney,
Benzo(a)pyr central
0.0002 Zero Cancer Mercury 0.002 0.002
ene nervous
system
7.4.3 [Link]
EPA for each utility, where allowances may be traded between utilities or within a company. Companies with emissions less than
the EPA allowance may trade their excess allowance with companies with allowance deficits, preventing severe hardships to those
utilities that that are dependent on high-sulfur coal. The CAA has also set provisions for reduction of NOx emissions. A market-
based approach is employed by the 1990 CAA amendments to eliminate ozone-destroying chemical substances (such as
chlorofluorocarbons) that deplete the ozone layer using a phasing-out schedule by terminating the production of these chemicals in
accordance with the Montreal Protocol (1989). The recycling of chlorofluorocarbons (CFCs) and the labeling of ozone-friendly
substitute chemicals are mandated by the CAA.
Hazardous Wastes
Hazardous wastes are wastes that pose a health and safety risk to humans and to the environment. The EPA designates hazardous
wastes as wastes which contain components that have one of the four general characteristics of reactivity, corrosivity, ignitability
and toxicity, in addition to other EPA classifications of hazardous wastes. The laws and regulations governing the management of
hazardous wastes and materials may be divided into two categories: present and future hazardous materials and wastes are
regulated under the Resource Conservation and Recovery Act (RCRA), while past and usually abandoned hazardous waste sites are
managed under the Comprehensive Environmental Response, Compensation and Liability Act (CERCLA).
7.4.4 [Link]
The EPA uses the National Priority List (NPL) to identify contaminated sites that present a risk to public health or the environment
and that may be eligible for Superfund money. A numeric ranking system known as the Hazard Ranking System (HRS) has been
established to determine the eligibility of contaminated sites for Superfund money, where sites with high HRS scores are most
likely to be added to the NPL. The National Contingency Plan (NCP) provides guidance for the initial assessment and the HRS
ranking of contaminated sites. After the initial assessment of a contaminated site, a remedial investigation is carried out where the
NCP provides for a detailed evaluation of the risks associated with that site. A remedial investigation results in a work plan, which
leads to the selection of an appropriate remedy referred to as a feasibility study. The feasibility study assesses several remedial
alternatives, resulting in Record of Decision (ROD) as the basis for the design of the selected alternative. The degree of cleanup is
specified by the NCP in accordance with several criteria such as the degree of hazard to the public health and the environment,
where the degree of cleanup varies for different contaminated sites.
A separate addition to the provisions of CERCLA is Title III of SARA known as the Emergency Planning and Community Right-
to-Know Act (EPCRA). The State Emergency Response Commission must be notified by a regulated facility that has extremely
hazardous substances exceeding the EPA specified Threshold Planning Quantities. The community is responsible for establishing
Local Emergency Planning Committees to develop a chemical emergency response plan which provides information on the
regulated facilities, emergency response procedures, training and evacuation plans. The awareness of a community about the
specific chemicals present in the community is an integral part of the Community's Right-to-Know, in addition to public
information about potential hazards from those chemicals. The EPCRA also stipulates that each year those facilities that release
chemicals above specified threshold levels should submit a Toxics Release Inventory (TRI) according to EPA specifications. The
TRI includes information on both accidental and routine releases in addition to off-site transfers of waste. The availability of the
TRI data to the public has led to serious consideration by industry to control their previously unregulated and uncontrolled
emissions due to the heightened public concern about the presence and the releases of chemicals in their community.
Figure 7.4.1 Exxon Valdez Oil Spill. Heavy sheens of oil covering large areas of the Prince William Sound, Alaska a few days after
the Exxon Valdez oil spill. Source: U.S. National Oceanic and Atmospheric Administration via Wikimedia Commons
The prevention of oil spills in navigable waters and shorelines of the United States is stipulated through the OPA statute. The OPA
encompasses oil spill prevention, preparedness, and response performance of industry and the federal government. Incentives are
provided to owners and operators for oil spill prevention, enforced by the EPA through the oil spill liability and penalty provisions
of the OPA. Oil companies in the United States engage in oil exploration, both offshore and onshore, resulting in accidental
releases of crude petroleum into the environment from wells, drilling rigs, offshore platforms and oil tankers. With the exception of
the 2010 BP Deepwater Horizon oil spill in the Gulf of Mexico, the number and amount of oil spills have decreased over the past
twenty years in the United States despite the increasing demand for oil. This decline has been attributed to the OPA after the Exxon
Valdez oil spill incident. The Exxon Valdez oil spill was the largest ever in United States waters until the 2010 BP Deepwater
Horizon oil spill (see Figure 7.4.2). BP has been held to be responsible for the Deepwater Horizon oil spill, and has been made
accountable for all cleanup costs and other damages by the federal government.
7.4.5 [Link]
Figure 7.4.2 BP Deepwater Horizon Oil Spill The Deep Horizon oil spill in the Gulf of Mexico as seen from space. Source:
NASA/GSFC, MODIS Rapid Response AND [Link] AND FT2, via Wikimedia Commons
7.4.6 [Link]
Occupational Safety and Health Act (OSHA)
The Occupational Safety and Hazard Act (OSHA) of 1970 and its amendment of 1990 aim to ensure safe and healthful working
conditions for workers through enforcement of standards developed under OSHA, and to provide for research, training and
education in the field of occupational safety and health. The standards for occupational health and safety are established by the
Occupational Safety and Health Administration and its state partners, which are enforced through inspections of industry and
providing guidance on better operating practices. The National Institute for Occupational Safety and Health (NIOSH) was
established to recommend occupational safety and health standards based on extensive scientific testing, which are afterwards
enforced by OSHA. Those industries which have followed OSHA standards have experienced a decline in overall injury and illness
rates, where the costs due to worker injuries, illnesses and compensation associated with occupational safety are a major loss for
industry. The OSHA standards for worker health and safety are recommended to be used in conjunction with various industrial
pollution prevention programs.
Summary
Environmental laws and regulations serve the purpose of limiting the amount of pollution in the environment from anthropogenic
sources due to industrial and other economic activities. Environmental regulations are specific to different phases of the
environment such as water and air. Government regulations help industry to curtail the environmental impact of pollution, leading
to the protection of human health and the environment. Future environmental laws and policy should convey and work in tandem
with the efforts of the public and industry for a more sustainable economy and society.
Resources
1) For more information on environmental engineering, read Chapter 1 of:
Davis, M.L. & Cornwell, D.A. (2008). Introduction to Environmental Engineering (4th ed.). New York: McGraw-Hill.
2) For more information about managing environmental resources, read:
LaGrega, M.D., Buckingham, P.L., Evans, J.C., & Environmental Resources Management (2001). Hazardous Waste Management
(2nd ed.). New York: McGraw-Hill.
3) For more information on the U.S. Environmental Protection Agency's laws and regulations, visit:
[Link]
This page titled 7.4: Government and Laws on the Environment is shared under a CC BY license and was authored, remixed, and/or curated by
Heriberto Cabezas (GALILEO Open Learning Materials) .
7.4.7 [Link]
7.5: Risk Assessment Methodology for Conventional and Alternative Sustainability
Options
Learning Objectives
After reading this module, students should be able to
understand the content and the goals of four-step of risk assessment process
know how to estimate dose received via each exposure pathway
know how to integrate exposure and toxicity information to characterize health risks
understand how to quantitatively estimate cumulative cancer and noncancer risks
understand how to identify/evaluate uncertainties in risk assessment
Introduction
Risk assessment is a scientific process used by federal agencies and risk management decision-makers to make informed decisions
about actions that may be taken to protect human health by ascertaining potential human health risks or health hazard associated
with exposure to chemicals in the environment.
Figure shows emissions billowing from a smoke stack into the atmosphere. Risk assessment helps federal agencies and risk
management decision-makers arrive at informed decisions about actions to take to protect human health from hazards such as air
pollution, pictured here. Source: Alfred Palmer via Wikipedia
Some of the real-world examples of risk assessment includes: establishment of national ambient air quality and drinking water
standards for protection of public health (e.g. ozone, particulate matter in outdoor air; chromium, chloroform or benzene in water);
establishment of clean-up levels for hazardous waste site remediation; development of fish consumption advisories for pregnant
women and general population (e.g. PCBs, mercury); assessment of risks and benefits of different alternative fuels for sound
energy policy development (e.g. oxygenated gasoline, biodiesel); and estimation of health risks associated with pesticide residues
in food. The estimated risk is a function of exposure and toxicity, as described in detail in NAS (1983) and EPA (1989). The
regulatory risk assessment follows a four-step paradigm using qualitative and/or quantitative approaches. In quantitative risk
assessment using either deterministic or probabilistic approaches, the risk estimates pertaining to an exposure scenario is
particularly useful when comparing a number of exposure or risk reduction measures among one another as an optimization
protocol to determine the best economically viable option for protection of public health and the environment. With environmental
sustainability and life-cycle analysis in the forefront of green technological innovation, energy, and economic savings in the 21st
Century, risk assessment will pay a pivotal role in discerning the option(s) with the most benefit in health protection and, thus, will
be an integral part of any environmentally sustainability analysis. Such comparative risk assessment can be performed for
traditional approaches vs. environmentally sustainable approaches. They can also be performed among different environmentally
sustainable options for an environmental pollution problem such hazardous waste site remediation and redevelopment, air quality
management in urban areas, pest management practices, agricultural health and safety, alternative energy sources for transportation
sources and among others.
The four steps of risk assessment are i) hazard identification ii) toxicity (or dose-response) assessment; iii) exposure assessment;
and iv) risk characterization, which are described below in detail. The emphasis is given in documenting the resources necessary to
successfully perform each step.
7.5.1 [Link]
Hazard Identification
In the hazard identification step, a scientific weight of evidence analysis is performed to determine whether a particular substance
or chemical is or is not causally linked to any particular health effect at environmentally relevant concentrations. Hazard
identification is performed to determine whether, and to what degree, toxic effects in one setting will occur in other settings. For
example, is a chemical that is shown to cause carcinogenicity in animal test species (e.g. rat, mouse) likely to be a carcinogen in
exposed humans? In order to assess the weight of evidence for adverse health effects, risk analysts follow the following steps (EPA,
1993): (1) Compile and analyze all the available toxicology data on the substance of interest; (2) Weigh the evidence that the
substance causes a toxic effect (cancer of non-cancer health end-points); and (3) Assess whether adverse health effect (or toxicity)
could occur from human exposure in a real-life setting.
In the first task of hazard identification, risk analyst examines the toxicity literature using the following analytical tools in the order
of importance:
Epidemiological studies
Controlled human exposure chamber experiments
In-vivo animal bioassays
In-vitro cell and tissue culture bioassays
Quantitative Structure –Activity Relationship Analysis (QSAR)
Among these, in-vivo animal bioassays are, by far, the most utilized source of information for hazard identification for chemicals
and, on rare instances, for chemical mixtures (e.g. diesel). When available, well-conducted epidemiological studies are regarded as
the most valuable source of human health hazard identification information since they provide direct human evidence for potential
health effects. Epidemiology is the study of the occurrence and distribution of a disease or physiological condition in human
populations and of the factors that influence this distribution (Lilienfeld and Lilienfeld, 1980). The advantages of epidemiological
studies for hazard identification are (EPA, 1989; EPA, 1993): animal-to-human extrapolation is not necessary, real exposure
conditions, and a wide range of subjects with different genetic and life-style patterns. However, epidemiological studies have a
number of shortcomings, which limit their usefulness in hazard identification. Some of these disadvantages include difficulty in
recruiting and maintaining a control group; having no control over some of the non-statistical variables related to exposures,
lifestyles, co-exposure to other chemicals, etc.; absence of actual exposure measurements along with memory bias for retrospective
studies; lengthy latency periods for chronic health effects such as cancer; and poor sensitivity and inability to determine cause-
effect relationships conclusively.
Animal bioassays remedy some of the weaknesses of epidemiological studies by allowing for greater control over the experiment
and are deemed to be reliable measurement of toxic effects, although they require "high dose in animals-to low dose in humans"
extrapolation. The selection of design parameters of animal bioassays is critically important in observing or missing an actual
hazard. These parameters include: animal species selected for the experiment (rat, mouse); strain of the test species; age/sex of the
test species; magnitude of exposure concentrations/doses applied or administered; number of dose levels studied; duration of
exposure; controls selected; and route of exposure. Animal studies are characterized as acute (a single dose or exposures of short
duration), chronic (exposures for full lifetimes of test species – about two years in rats/mice) and sub-chronic (usually 90 days)
based on the exposure duration. In the hazard identification step, the following measures of toxicity are commonly compiled:
LD50/LC50/EC50: The dose or concentration in a toxicity study at which causing 50 percent mortality in test species was
observed. The EC50 is the effective concentration causing adverse effects or impairment in 50% of the test species.
NOAEL (No Observable Adverse Effect Level): The highest dose or concentration in a toxicity study at which no adverse
effect was observed.
LOAEL (Lowest Observable Adverse Effect Level): The lowest dose or concentration in a toxicity study at which an adverse
effect was observed.
MTD (Maximum Tolerated Dose): The largest dose a test animal can receive for most of its lifetime without demonstrating
adverse health effects other than carcinogenicity.
Risk scientists rely on a number of reputable sources to gather, compile, and analyze hazard identification information to be able to
perform weight of evidence analysis and to conclude whether a chemical may cause a health effect in humans. Some of these
sources are:
Hazardous Substances Data Bank (HSDB) maintained by the National Library of Medicine: This scientifically peer-reviewed
data bank provides human and animal toxicity data for about 5,000 chemicals and can be accessed via [Link]/cgi-
7.5.2 [Link]
bin/sis/htmlgen?HSDB
ChemicIDplus Advanced database maintained by the National Library of Medicine: This database allows users to search the
NLM ChemIDplus database of over 370,000 chemicals. Compound identifiers such as Chemical Name, CAS Registry Number,
Molecular Formula, Classification Code, Locator Code, and Structure or Substructure can be entered to display toxicity data via
[Link]/chemidplus/
National Toxicology Program (NTP) of the Department of Health and Human Services:
Report of Carcinogens (RoC): The RoC is an informational scientific and public health document that identifies and
discusses agents, substances, mixtures, or exposure circumstances that may pose a hazard to human health by virtue of their
carcinogenicity. The RoC is published biennially and serves as a meaningful and useful compilation of data on: a) the
carcinogenicity (ability to cause cancer), genotoxicity (ability to damage genes), and biologic mechanisms (modes of action
in the body) of the listed substance in humans and/or animals; b) the potential for human exposure to these substances; and
c) Federal regulations to limit exposures. The link to the most recent version of the RoC can be accessed via
[Link]/?objectid=B...F8CBFACF0FC7EF
NTP Toxicity Testing Study Results and Research Areas: NTP tests chemicals for their toxicity in human and animal
systems. The results of these toxicity testing studies along with current research areas can be obtained at:
[Link]/[Link]?o...CC2DE0A230C920
National Institute of Occupational and Safety Health (NIOSH) Hazard Identification Databases: The following NIOSH website
houses a multitude of databases and information for chemicals and their hazards under a single umbrella, including NIOSH's
"Pocket Guide to Chemical Hazards": [Link]/niosh/[Link]
Agency for Toxic Substances and Disease Registry (ATSDR) Toxicological Profiles and Public Health Statements: ATSDR
produces "toxicological profiles" for hazardous substances found at National Priorities List (NPL) Superfund sites. About 300
toxicological profiles have so far been published or are under development. The chemical-specific toxicological profiles can be
accessed via [Link]/toxprofiles/[Link]
World Health Organization (WHO) International Programme of Chemical Safety (IPCS): ICPS publishes "Environmental
Health Criteria" (EHC) for chemical substances, which provide critical reviews on the effects of chemicals or combinations of
chemicals and physical and biological agents on human health and the environment. The IPCS site can be accessed via
[Link]
Material Safety Data Sheets (MSDS): MSDS are invaluable resource to obtain compositional data for products and mixtures.
7.5.3 [Link]
based on the assumption that if the critical toxic effect is prevented, then all toxic effects are prevented (EPA, 1989). The
derivation of toxicity value, RfD/RfC, for noncarcinogens assumes that they are threshold chemicals, meaning there is a
threshold below which no adverse effects are observed in test species. This dose level (i.e. NOAEL) in animals is simply
adjusted by a number of factors (UFs and MF) to determine the safe dose level in humans (i.e. RfD) as shown by the following
equation:
N OAEL
Rf D =
U F1 × U F2 × U F3 . . . ×M F
Cancer Slope Factor (CSF) for oral/dermal pathway or Unit Risk Factor (URF) for inhalation pathway – Carcinogens:
Unlike the noncarcinogens, carcinogens assumed to be non-threshold chemicals based on the Environmental Protection Agency
(EPA) assumption that a small number of molecular events can evoke changes in a single cell that can lead to uncontrolled
cellular proliferation and eventually to cancer. In deriving slope factors, firstly, an appropriate dose-response data set is selected.
In this exercise, whenever available, human data of high quality are preferable to animal data. However, if only animal data are
available, dose-response data from species that responds most similarly to humans with respect to metabolism, physiology, and
pharmacokinetics is preferred. When no clear choice is possible, the most sensitive species is chosen. Secondly, a model to the
available data set is applied and extrapolation from the relatively high doses administered to test species in animal bioassay (or
the exposures recorded in epidemiologic studies) to the lower environmental exposure levels expected for humans is performed
using the model. Although various models have been developed for this purpose (e.g. probit, logit, Weibull), the linearized
multistage model has commonly been used by the EPA. After the data are fit to the appropriate model, the upper 95th percent
confidence limit of the slope of the resulting dose-response curve is calculated, which is known as the Cancer Slope Factor
(CSF). It represents an upper 95th percent confidence limit on the probability of a response per unit intake of a chemical over a
lifetime (i.e. dose). Thus, its units are (mg/kg-day) -1. This indicates that there is only a five percent chance that the probability
of a response could be greater than the estimated value of CSF. Because the dose-response curve generally is linear only in the
low-dose region, the slope factor estimate only holds true for low doses. Toxicity values for carcinogenic effects also can be
expressed in terms of risk per unit concentration of the chemical, which are called unit risk factors (URFs). They are calculated
by dividing the CSF by adult body weight (70 kg) and multiplying by the adult inhalation rate (20 m3/day), for risk associated
with unit concentration in air (EPA, 1989).
A number of regulatory agencies responsible for environmental and public health protection have devoted resources in developing
and documenting toxicity values for noncarcinogens (RfDs/RfCs) and carcinogens (CSFs/URFs). The following hierarchy of
sources is recommended by the EPA in evaluating chemical toxicity for Superfund sites (EPA, 2003):
Integrated Risk Information System (IRIS) and cited references, which is the prime source for the chemical-specific toxicity
value information and can be accessed via: [Link]
The Provisional Peer Reviewed Toxicity Values (PPRTV) and cited references developed for the EPA Office of Solid Waste and
Emergency Response (OSWER) Office of Superfund Remediation and Technology Innovation (OSRTI) programs (not publicly
available).
Other toxicity values, which includes the following sources of toxicity values that are commonly consulted by the EPA
Superfund Program when a relevant toxicity value is not available from either IRIS or the PPRTV database:
California Environmental Protection Agency (Cal EPA) Toxicity Criteria Database, available at:
[Link]
The Agency for Toxic Substances and Disease Registry (ATSDR) Minimal Risk Levels (MRLs, addressing noncancer
effects only). MRL is an estimate of the daily human exposure to a hazardous substance that is likely to be without
appreciable risk of adverse noncancer health effects over a specified duration of exposure. These substance-specific
estimates, which are intended to serve as screening levels, are used by ATSDR health assessors and other responders to
identify contaminants and potential health effects that may be of concern at hazardous waste sites. To date, 137 inhalation
MRLs, 226 oral MRLs and 8 external radiation MRLs have been derived and can be found at:
[Link]
The EPA Superfund Health Effects Assessment Summary Tables (HEAST) database and cited references; and
Additional sources of toxicity values.
There are a number of other valuable sources for toxicity values (RfDs/RfCs for non-carcinogens and URFs/CSFs for carcinogens),
which can be compiled via the following sources:
7.5.4 [Link]
EPA Region 9 tabulated "Preliminary Remediation Goals (PRGs)" or "Regional Screening Levels (RSL)" for Chemical
Contaminants at Superfund Sites," which also lists toxicity values (oral/inhalation RfD and oral/inhalation CSF) used in the
medium-specific PRG/RSL calculation for each chemical. This table can be accessed via
[Link]/region09/waste/sf...prg/[Link]
The Hot Spot Guidelines published by California EPA for Air Toxics Program includes technical background documentation
for toxicity criteria/values for chemicals (i.e. Cancer Potency Factors (CPFs), which is equivalent to EPA's CSFs and Chronic
Recommended Exposure Limits (RELs), which are similar to USEPA's RfCs). The most recent version of REL Table is
located at: [Link]/air/[Link]. The Technical Support Document for CPFs that contains cancer unit risks and
potency factors for 121 of the 201 carcinogenic substances or groups of substances can be accessed via
[Link]
The Department of Energy's Oak Ridge National Laboratory (ORNL) maintains Risk Assessment Information System
(RAIS) website, which contains useful information for risk assessment, including chemical-specific toxicity values. The RAIS
information can be accessed via [Link]
Toxicology Excellence in Risk Assessment (TERA), a non-profit organization, manages and distributes a free Internet
database of human health risk values and cancer classifications for over 600 chemicals of environmental concern from multiple
organizations worldwide. This database, Integrated Toxicity Estimates for Risk (ITER), can be accessed via:
[Link] or via NLM's TOXNET database at: [Link].
The dermal RfDs and CSFs can be derived from oral RfDs and CSFs, adjusted for chemical-specific gastrointestinal absorption
efficiency, based on the recommended methodology in EPA's Guidance for Dermal Risk Assessment (EPA, 2004a).
Exposure Assessment
In the third step of risk assessment, the magnitude of exposure is determined by measuring or estimating the amount of an agent to
which humans are exposed (i.e. exposure concentration) and the magnitude of dose (or intake) is estimated by taking the
magnitude, frequency, duration, and route of exposure into account. Exposure assessments may consider past, present, and future
exposures.
Figure shows the human health effects of environmental pollution from pollution source to receptor. Source: Mikael Häggström via
Wikimedia Commons
While estimates of past or current exposure concentration/dose can be based on measurements or models of existing conditions,
estimates of future exposure concentration/dose can be based on models of future conditions. In the case of inhalation exposures,
personal or area monitoring to sample for contaminants in the air can be employed. The sampling data can be augmented with
modeling efforts using default and/or site-specific input parameters. The model application can begin with simple screening level
dispersion models and/or can utilize higher-level 2-D or 3-D models depending on the complexity of the environmental pollution
problem in hand.
In any exposure assessment, the risk scientists ask a number of questions to hypothesize the exposure scenario pertaining to
environmental pollution affecting the population or a sub-group. Some of these are:
What is the source of pollution at the site? (e.g. underground storage tank leak, emissions from an industrial plant, surface water
run-off from agricultural fields)
7.5.5 [Link]
Which environmental compartments are likely to be contaminated? (i.e. air, water, sediment, soil, plants, animals, fish)
What are the chemicals of concern (COC) originating from the pollution source?
What are the fate and transport properties of these chemicals that may inform the aging of the pollution in the environment over
time and resultant chemical signature in each environmental medium?
Who is exposed? (e.g. children, elderly, asthmatics, general population, workers)
How many people are exposed?
Where are people exposed? (e.g. home, workplace, outside environment, retirement communities, schools)
How are people exposed? (i.e. exposure pathway – inhalation, dermal contact or ingestion)
How often are people exposed? (i.e. exposure frequency)
How long are people exposed? (i.e. exposure duration)
Answers to these questions frame the problem at hand. In the next step, a number of exposure parameters are integrated into an
estimate of daily dose received by an exposed individual via each exposure route (ingestion, dermal contact or skin absorption, and
inhalation). The magnitude of human exposures, in general, is dependent on COC concentration in soil, exposure parameters
describing human physiology (e.g. soil ingestion rate, body weight), and population-specific parameters describing exposure
behavior (exposure frequency, duration). When evaluating subchronic or chronic exposures to noncarcinogenic chemicals, dose is
averaged over the period of exposure, termed "Average Daily Dose" (ADD). However, for carcinogens, dose is averaged over an
entire lifetime (i.e. 70 years), thus referred to as "Lifetime Average Daily Dose" (LADD). Both ADD and LADD represent
normalized exposure rate in the units of mg of chemical per kg body weight per day (mg/kg-day). The ADD for noncarcinogenic
COCs and LADD for carcinogenic COCs are estimated for four most commonly studied exposure pathways in EPA risk
assessments, particularly for hazardous waste sites, as shown below (Erdal, 2007):
C1 ×I R0 ×EF ×ED×C F
Soil Ingestion: L(ADD) o =
BW ×AT
BW ×AT
1
Cs ×I Ri ×EF ×ED×
1
Cs ×I Ri ×EF ×ED×
Cs: Exposure Concentration (i.e., 95th Upper Confidence Limit on the Mean) of COC in soil (mg/kg) – (chemical-specific; can be
estimated using EPA 2004b)
IRo: Ingestion rate of soil (mg/d)
IRi: Inhalation rate (m3/d)
SA: Skin surface area (cm2)
AF: Soil-to-skin adherence factor (mg/cm2)
ABS: Dermal absorption fraction (unitless – chemical-specific)
EV: Event frequency (events/d)
EF: Exposure frequency (d/y)
ED: Exposure duration (y)
PEF: Particulate emission factor (m3/kg) – 1.36 x 109 m3/kg per (EPA 2002a)
VF: Soil-to-air volatilization factor (m3/kg – chemical-specific)
BW: Body weight (kg)
AT: Averaging time (days) – (ED*365 d/y for noncarcinogens; 70 y*365 d/y for carcinogens)
CF: Conversion factor – 10-6 kg/mg
In deterministic risk assessment, ADD and LADD estimates are performed for a reasonable maximum exposure scenario (RME)
and a central tendency exposure scenario (CTE), resulting in a range. EPA's reliance on the concept of RME for estimating risks is
based on a conservative but plausible exposure scenario (which is defined to be the 90th to 95th percentile exposure, signifying that
fewer than five percent to 10 percent of the population would be expected to experience higher risk levels), and has been
7.5.6 [Link]
scientifically challenged over the years. For example, Burmaster and Harris (1993) showed that the use of EPA recommended
default exposure parameter values resulted in exposure and risk estimates well in excess of the 99th percentile due to multiplication
of three upper-bound values (i.e. 95th percentiles) for IR, EF, and ED. The authors argued that this leads to hazardous waste site
cleanup decisions based on health risks that virtually no one in the surrounding population would be expected to experience. They
advised the EPA to endorse and promote the use of probabilistic methods (e.g. Monte-Carlo simulation) as a way to supplement or
replace current risk assessment methods, in order to overcome the problem of "compounded conservatism" and enable calculation
of risks using a more statistically defensible estimate of the RME. In probabilistic risk assessment, the input parameters are
characterized by their unique probability distribution. The EPA's Exposure Factors Program provides information on development
of exposure parameter distributions in support of probabilistic distributions and can be accessed via:
[Link]
The values of the exposure parameters corresponding to RME or CTE scenarios are often compiled from EPA's Exposure Factors
Handbook (EFH):
General Exposure Factors Handbook (EPA, 1997) provides exposure assessors with data needed on standard factors to calculate
human exposure to toxic chemicals as part of risk assessments. These factors include: drinking water consumption, soil
ingestion, inhalation rates, dermal factors including skin area and soil adherence factors, consumption of fruits and vegetables,
fish, meats, dairy products, homegrown foods, breast milk intake, human activity factors, consumer product use, and residential
characteristics. Recommended values are for the general population and also for various segments of the population who may
have characteristics different from the general population. The most recent version of the EFH can be accessed via
[Link]
The Children-Specific EFH (EPA, 2002b) provides a summary of the available and up-to-date statistical data on various factors
assessing child exposures, which can be accessed via [Link]
Risk Characterization
In the last step, a hazard quotient (HQ) as an indicator of risks associated with health effects other than cancer and excess cancer
risk (ECR) as the incremental probability of an exposed person developing cancer over a lifetime, are calculated by integrating
toxicity and exposure information, as shown below. If HQ > 1, there may be concern for potential adverse systemic health effects in
the exposed individuals. If HQ ≤ 1, there may be no concern. It should be noted that HQs are scaling factors and they are not
statistically based. The EPA's acceptable criterion for carcinogenic risks is based on public policy as described in the National
Contingency Plan (NCP) and is the exposure concentration that represent an ECR in the range of 10-4 – 10-6, i.e. 1 in 10,000 to 1 in
1,000,000 excess cancer cases (EPA, 1990).
Noncancer Risk: HazardQuotient(HQ)= ADD
Rf D
C OCNC =1
(H Qo + H Qd + H Qi )
n n
Cumulative Excess Cancer Risk: ∑ C OCc =1
EC R = ∑C OC
c
=1
(EC Ro + EC Rd + EC Ri )
Here, o, d and i subscripts express oral (ingestion), dermal contact and inhalation pathways.
As discussed above, the HQ, HI, and ECR estimates are performed for RME and CTE scenarios separately in the case of
deterministic risk assessment. Although EPA published the probabilistic risk assessment guidelines in 2001 (EPA, 2001), its
application has so far been limited. Proper evaluation of uncertainties, which are associated with compounded conservatism and
potential underestimation of quantitative risk estimates (e.g. due to the presence of COCs without established toxicity values), is
intrinsic to any risk-based scientific assessment. In general, uncertainties and limitations are associated with sampling and analysis,
chemical fate and transport, exposure parameters, exposure modeling, and human dose-response or toxicity assessment (derivation
of CSFs/RfDs, extrapolation from high animal doses to low human doses), and site-specific uncertainties.
7.5.7 [Link]
Conclusion
The improvement in the scientific quality and validity of health risk estimates depends on advancements in our understanding of
human exposure to, and toxic effects associated with, chemicals present in environmental and occupational settings. For example,
life-cycle of and health risks associated with pharmaceuticals in the environment is poorly understood due to lack of environmental
concentration and human exposure data despite extensive toxicological data on drugs. There are many other examples for which
either data on exposure or toxicity or both have not yet been developed, preventing quantitative assessment of health risks and
development of policies that protect the environment and public health at the same time. Therefore, it is important to continue to
develop research data to refine future risk assessments for informed regulatory decision-making in environmental sustainability and
to ensure that costs associated with different technological and/or engineering alternatives are scientifically justified and public
health-protective. One area that, particularly, requires advancement is the assessment of health risks of chemical mixtures. Current
risk assessment approaches consider one chemical at a time. However, chemicals are present in mixtures in the environment.
Furthermore, physical, chemical and biological transformations in the environment and interactions among chemicals in the
environment may change the toxic potential of the mixture over time. Thus, risk assessment is an evolving scientific discipline that
has many uncertainties in all of the four steps. These uncertainties should be thoroughly documented and discussed and the risk
assessment results should be interpreted within the context of these uncertainties.
Review Questions
1. What are the human health hazards of vinyl chloride?
2. What are the human toxicity values (RfD, CSF) of vinyl chloride and how are these values estimated?
3. How do you calculate the dose received by children and adults via ingestion of vinyl chloride-contaminated drinking water
under the RME scenario? Please document and explain your assumptions along with your references for the exposure
parameters for each receptor of concern
4. How do you calculate RME cancer and noncancer risks to children and adults for the above exposure scenario?
5. What does excess cancer risk of three cases out of ten thousand exposed (3x10-4) signify?
6. If drinking water were also contaminated with benzene, how would you estimate cumulative cancer and noncancer risks
associated with exposure to drinking water contaminated with vinyl chloride and benzene for children and adults under the
RME scenario?
References
Burmaster, D.E. & Harris, R.H. (1993). The magnitude of compounding conservatisms in Superfund risk assessments. Risk
Analysis, 13, 131-34.
EPA (U.S. Environmental Protection Agency). (1989). Risk assessment guidance for Superfund, volume I: Human health
evaluation manual (Part A) (Interim Final) (EPA/540/1-89002). Office of Emergency and Remedial Response, Washington, DC
EPA (U.S. Environmental Protection Agency). (1990, March 8). National contingency plan. Federal Register, 55, 8848.
Washington, DC.
EPA (U.S. Environmental Protection Agency). (1993, September). SI 400: Introduction to risk assessment and risk management for
hazardous air pollutants. Air Pollution Training Institute. Environmental Research Center, Research Triangle Park, NC.
EPA (U.S. Environmental Protection Agency). (1997, August). Exposure factors handbook, volume I – General factors
(EPA/600/P-95/002Fa). National Center for Environmental Assessment, Office of Research and Development, Washington, DC.
EPA (U.S. Environmental Protection Agency). (2001, December). Risk assessment guidance for Superfund, volume III – Part A:
Process for conducting probabilistic risk assessment (EPA540-R-02-002). Office of Emergency and Remedial Response,
Washington, DC.
EPA (U.S. Environmental Protection Agency). (2002a, December). Supplemental guidance for developing soil screening levels for
Superfund sites (OSWER 9355.4-24). Office of Emergency and Remedial Response, Washington, DC.
EPA (U.S. Environmental Protection Agency). (2002b, September). Child-specific exposure factors handbook (EPA-600-P-00-
002B). Interim Report. National Center for Environmental Assessment, Office of Research and Development, Washington, DC.
EPA (U.S. Environmental Protection Agency). (2003, December 5). Human health toxicity values in Superfund risk assessments
(OSWER Directive 9285.7-53). Memorandum from Michael B. Cook, Director of Office of Superfund Remediation and
7.5.8 [Link]
Technology Innovation to Superfund National Policy Managers, Regions 1 – 10. Office of Emergency and Remedial Response,
Washington, DC.
EPA (U.S. Environmental Protection Agency). (2004a, July). Risk assessment guidance for Superfund, volume I: Human health
evaluation manual (Part E, supplemental guidance for dermal risk assessment) (EPA/540/R/99/005). Final. Office of Superfund
Remediation and Technology Innovation, Washington, DC.
EPA (U.S. Environmental Protection Agency). (2004b, April). ProUCL Version 3.00.02 user's guide and software
(EPA/600/R04/079). Prepared by Anita Singh and Robert Maichle with Lockheed Martin Environmental Services and Ashok K.
Singh of University of Nevada for the U.S. Environmental Protection Agency.
Erdal, S. (2007). Case study: Multi-pathway risk assessment for adults and children living near a hazardous waste site. In M.G.
Robson & W.A. Toscano (Eds.), Environmental Health Risk Assessment for Public Health(pp. 523-530). Association of Schools of
Public Health.
Lilienfeld A.M. & Lilienfeld, D.E. (1980). The French influence on the development of epidemiology. Henry E Sigerist Suppl Bull
Hist Med, 4, 28-42.
NAS (National Academy of Sciences). (1983). Risk assessment in the federal government: Managing the process. National
Academy Press, Washington, DC.
Glossary
bioassay
An assay for determining the potency (or concentration) of a substance that causes a biological change in experimental animals.
carcinogenicity
Defines the ability or tendency to produce cancer.
epidemiology
The study of the distribution and determinants of health-related states or events in specified populations.
monte-carlo method
A repeated random sampling from the distribution of values for each of the parameters in a generic (exposure or dose) equation
to derive an estimate of the distribution of (doses or risks in) the population.
This page titled 7.5: Risk Assessment Methodology for Conventional and Alternative Sustainability Options is shared under a CC BY license and
was authored, remixed, and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
7.5.9 [Link]
CHAPTER OVERVIEW
This page titled 8: Sustainable Energy Systems is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto Cabezas
(GALILEO Open Learning Materials) .
1
8.1: Sustainable Energy Systems - Chapter Introduction
Learning Objectives
After reading this module, students should be able to
outline the history of human energy use
understand the challenges to continued reliance on fossil energy
understand the motivations and time scale for transitions in energy use
Sustainability Challenges
Eighty five percent of world energy is supplied by combustion of fossil fuels. The use of these fuels (coal since the middle ages for
heating; and coal, oil and gas since the Industrial Revolution for mechanical energy) grew naturally from their high energy density,
abundance and low cost. For approximately 200 years following the Industrial Revolution, these energy sources fueled enormous
advances in quality of life and economic growth. Beginning in the mid-20th Century, however, fundamental challenges began to
emerge suggesting that the happy state of fossil energy use could not last forever.
Environmental Pollution
The first sustainability challenge to be addressed was environmental pollution, long noticed in industrial regions but often ignored.
Developed countries passed legislation limiting the pollutants that could be emitted, and gradually over a period of more than two
decades air and water quality improved until many of the most visible and harmful effects were no longer evident.
8.1.1 [Link]
Figure 8.1.1 Crude Oil Reserves. The global distribution of crude oil resources. 1 Includes 172.7 billion barrels of bitumen in oil
sands in Alberta, Canada. 2 Excludes countries that were part of the former U.S.S.R. See " Union of Soviet Socialist Republics
(U.S.S.R.)" in Glossary. 3 Includes only countries that were part of the former [Link]: U.S. Energy Information
Administration, Annual Review, 2009, p. 312 (Aug. 2010)
8.1.2 [Link]
Figure 8.1.2 Temperature, Sea Level, and Snow Cover 1850-2000. Three graphs show trends in average surface temperature,
average sea level and northern hemisphere snow cover from 1850-2000. Source: Climate Change 2007: Synthesis Report:
Contribution of Working Groups I, II and III to the Fourth Assessment Report of the Intergovernmental Panel on Climate Change,
Cambridge University Press, figure 1.1, page 31
There can be no doubt of the rising trends, and there are disturbing signs of systematic change in other indicators as well (Arndt, et
al., 2010). The short-term extension of these trends can be estimated by extrapolation. Prediction beyond thirty or so years requires
developing scenarios based on assumptions about the population, social behavior, economy, energy use and technology advances
that will take place during this time. Because trends in these quantities are frequently punctuated by unexpected developments such
as the recession of 2008 or the Fukushima nuclear disaster of 2011, the pace of carbon emissions, global warming and climate
change over a century or more cannot be accurately predicted. To compensate for this uncertainty, predictions are normally based
on a range of scenarios with aggressive and conservative assumptions about the degrees of population and economic growth,
energy use patterns and technology advances. Although the hundred year predictions of such models differ in magnitude, the
common theme is clear: continued reliance on fossil fuel combustion for 85 percent of global energy will accelerate global
warming and increase the threat of climate change.
The present reliance on fossil fuels developed over time scales of decades to centuries. Figure 8.1.3 shows the pattern of fuel use in
the United States since 1775.
Figure 8.1.3 Primary Energy Consumption by Source, 1775-2009. Graph shows the pattern of fuel use in the United States since
1775. Source: U.S. Energy Information Administration, Annual Review, 2009, p. xx (Aug. 2010)
Wood was dominant for a century until the 1880s, when more plentiful, higher energy density and less expensive coal became king.
It dominated until the 1950s when oil for transportation became the leading fuel, with natural gas for heating a close second. Coal
is now in its second growth phase, spurred by the popularity of electricity as an energy carrier in the second half of the 20th
Century. These long time scales are built into the energy system. Uses such as oil and its gasoline derivative for personal
transportation in cars or the widespread use of electricity take time to establish themselves, and once established provide social and
infrastructural inertia against change.
8.1.3 [Link]
The historical changes to the energy system have been driven by several factors, including price and supply challenges of wood, the
easy availability and drop-in replaceability of coal for wood, the discovery of abundant supplies of oil that enabled widespread use
of the internal combustion engine, and the discovery of abundant natural gas that is cleaner and more transportable in pipelines than
coal. These drivers of change are based on economics, convenience or new functionality; the resulting changes in our energy
system provided new value to our energy mix.
The energy motivations we face now are of a different character. Instead of adding value, the motivation is to avert "doomsday"
scenarios of diminishing value: increasing environmental degradation, fuel shortages, insecure supplies and climate change. The
alternatives to fossil fuel are more expensive and harder to implement, not cheaper and easier than the status quo. The historical
motivations for change leading to greater value and functionality are reversed. We now face the prospect that changing the energy
system to reduce our dependence on fossil fuels will increase the cost and reduce the convenience of energy.
Summary
Continued use of fossil fuels that now supply 85 percent of our energy needs leads to challenges of environmental degradation,
diminishing energy resources, insecure energy supply, and accelerated global warming. Changing to alternate sources of energy
requires decades, to develop new technologies and, once developed, to replace the existing energy infrastructure. Unlike the
historical change to fossil fuel that provided increased supply, convenience and functionality, the transition to alternative energy
sources is likely to be more expensive and less convenient. In this chapter you will learn about the environmental challenges of
energy use, strategies for mitigating greenhouse gas emissions and climate change, electricity as a clean, efficient and versatile
energy carrier, the new challenges that electricity faces in capacity, reliability and communication, the challenge of transitioning
from traditional fossil to nuclear and renewable fuels for electricity production. You will also learn about the promise of biofuels
from cellulose and algae as alternatives to oil, heating buildings and water with solar thermal and geothermal energy, and the
efficiency advantages of combining heat and power in a single generation system. Lastly, you will learn about the benefits,
challenges and outlook for electric vehicles, and the sustainable energy practices that will reduce the negative impact of energy
production and use on the environment and human health.
Review Questions
1. Fossil fuels have become a mainstay of global energy supply over the last 150 years. Why is the use of fossil fuels so
widespread?
2. Fossil fuels present four challenges for long-term sustainability. What are they, and how do they compare in the severity of their
impact and cost of their mitigation strategies?
3. The dominant global energy supply has changed from wood to coal to oil since the 1700s. How long did each of these energy
transitions take to occur, and how long might a transition to alternate energy supplies require?
References
1. Arndt, D. S., Baringer, M. O., & Johnson, M. R. (eds.). (2010). State of the Climate in 2009. Bull. Amer. Meteor. Soc., 91, S1–
S224, [Link]/bams-state-...imate/[Link]
2. Hirsch, R.L., Bezdek, R., & Wendling, R. (2006). Peaking of World Oil Production and Its Mitigation. AIChE Journal, 52, 2 –
8. doi: 10.1002/aic.10747
3. Owen, N.A., Inderwildi, O.R., & King, D.A. (2010). The status of conventional world oil reserves – Hype or cause for concern?
Energy Policy,38, 4743 – 4749. doi: 10.1016/[Link].2010.02.026
Glossary
fossil fuels
Oil, gas and coal produced by chemical transformation of land plants (coal) and marine animals (oil and gas) trapped in the
earth's crust under high pressure and temperature and without access to oxygen. The formation of fossil fuels can take.
industrial revolution
The transition from simple tools and animal power for producing products to complex machinery powered by the combustion of
fuels. The Industrial Revolution began in England in the mid-18th Century initially centered around the development of the
steam engine powered by coal.
8.1.4 [Link]
The combustion of fuel inside or "internal" to the cylinder and moving piston which produces motion; gasoline engines are a
common example. In contrast, steam engines are external combustion engines where combustion and steam generation are
outside the cylinder containing the moving piston. The internal combustion engine is lighter and more portable than the steam
engine, enabling modern transportation in cars, diesel powered trains, ships and airplanes.
This page titled 8.1: Sustainable Energy Systems - Chapter Introduction is shared under a CC BY license and was authored, remixed, and/or
curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
8.1.5 [Link]
8.2: Environmental Challenges in Energy, Carbon Dioxide, Air, Water and Land Use
Learning Objectives
After reading this module, students should be able to
outline environmental impacts of energy use
evaluate the different energy sources based on their environmental impact
understand the global capacity for each non-renewable energy source
Introduction
Energy to illuminate, heat and cool our homes, businesses and institutions, manufacture products, and drive our transportation
systems comes from a variety of sources that are originate from our planet and solar system. This provides a social and economic
benefit to society. The earth’s core provides geothermal energy. The gravitational pull of moon and sun create tides. The sun makes
power in multiple ways. By itself, the sun generates direct solar power. The sun’s radiation in combination with the hydrologic
cycle can make wind power and hydroelectric power. Through photosynthesis, plants grow making wood and biomass that decay
after they die into organic matter. Over the course of thousands of years, this decay results in fossil fuels that have concentrated or
stored energy. To learn more about measuring different kinds of energy, known as emergy, see Chapter 11. Each of these types of
energy can be defined as renewable or non-renewable fuels and they each have some environmental and health cost.
Fossil fuel reserves are not distributed equally around the planet, nor are consumption and demand. We will see in this chapter that
fuel distribution is critical to the sustainability of fossil fuel resources for a given geographic area. Access to renewable resources
and their viability is greatly dependent on geography and climate. Making energy requires an input of energy so it is important to
look at the net energy generated – the difference of the energy produced less the energy invested.
8.2.1 [Link]
Figure 8.2.1 Environmental Impacts of Nonrenewable and Renewable Electricity Sources. Source: C. Klein-Banai using data from
U.S. Energy Information Administration and U.S. Environmental Protection Agency
Figure illustrates that the United States imported over half of the crude oil and refined petroleum products that it consumed during
2009. Source: U.S. Energy Information Administration, Petroleum Supply Annual, 2009, preliminary data
The holder of oil reserves in the oil market is the Organization of Petroleum Exporting Countries, (OPEC) (see Figure 8.2.3). As of
January 2009, there were 12 member countries in OPEC: Algeria, Angola, Ecuador, Iran, Iraq, Kuwait, Libya, Nigeria, Qatar,
Saudi Arabia, the United Arab Emirates, and Venezuela. OPEC attempts to influence the amount of oil available to the world by
assigning a production quota to each member except Iraq, for which no quota is presently set. Overall compliance with these quotas
8.2.2 [Link]
is mixed since the individual countries make the actual production decisions. All of these countries have a national oil company but
also allow international oil companies to operate within their borders. They can restrict the amounts of production by those oil
companies. Therefore, the OPEC countries have a large influence on how much of world demand is met by OPEC and non-OPEC
supply. A recent example of this is the price increases that occurred during the year 2011 after multiple popular uprisings in Arab
countries, including Libya.
Figure 8.2.3 Proven Oil Reserves Holders. Pie chart shows proven oil reserves holders. Source: C. Klein-Banai using data from
BP Statistical Review of World Energy (2010)
This pressure has lead the United States to developing policies that would reduce reliance on foreign oil such as developing
additional domestic sources and obtaining it from non-Middle Eastern countries such as Canada, Mexico, Venezuela, and Nigeria.
However, since fossil fuel reserves create jobs and provide dividends to investors, a lot is at stake in a nation that has these
reserves. Depending on whether that oil wealth is shared with the country’s inhabitants or retained by the oil companies and
dictatorships, as in Nigeria prior to the 1990s, a nation with fossil fuel reserves may benefit or come out even worse.
Figure 8.2.4 Fuel Type and Carbon Emissions. The two charts show the relationship between fuel type and carbon emissions for
U.S. energy consumption in 2010. Source: U.S. Energy Information Administration
8.2.3 [Link]
Figure 8.2.5 Historic U.S. Coal Production. Graph shows U.S. Coal Production from 1950-2010. Source: U.S. Energy Information
Administration
Coal is plentiful and inexpensive, when looking only at the market cost relative to the cost of other sources of electricity, but its
extraction, transportation, and use produces a multitude of environmental impacts that the market cost does not truly represent.
Coal emits sulfur dioxide, nitrogen oxide, and mercury, which have been linked to acid rain, smog, and health issues. Burning of
coal emits higher amounts of carbon dioxide per unit of energy than the use of oil or natural gas. Coal accounted for 35 percent of
the total United States emissions of carbon dioxide released into the Earth’s atmosphere in 2010 (see Figure 8.2.4). Ash generated
from combustion contributes to water contamination. Some coal mining has a negative impact on ecosystems and water quality,
and alters landscapes and scenic views. There are also significant health effects and risks to coal miners and those living in the
vicinity of coal mines.
Traditional underground mining is risky to mine workers due to the risk of entrapment or death. Over the last 15 years, the U.S.
Mine Safety and Health Administration has published the number of mine worker fatalities and it has varied from 18-48 per year
(see Figure 8.2.6).
Figure 8.2.6 U.S. Coal Mining Related Fatalities. Graph shows U.S. coal mining related fatalities from 1995-2010. Source: C.
Klein-Banai using data from the U.S. Department of Labor, Mine Safety and Health Administration
Twenty-nine miners died on April 6, 2010 in an explosion at the Upper Big Branch coal mine in West Virginia, contributing to the
uptick in deaths between 2009 and 2010. In other countries, with less safety regulations, accidents occur more frequently. In May
2011, for example, three people died and 11 were trapped in a coalmine in Mexico for several days. There is also risk of getting
black lung disease (pneumoconiosis) This is a disease of the lungs caused by the inhalation of coal dust over a long period of time.
It causes coughing and shortness of breath. If exposure is stopped the outcome is good. However, the complicated form may cause
shortness of breath that gets increasingly worse.
Mountain Top Mining (MTM), while less hazardous to workers, has particularly detrimental effects on land resources. MTM is a
surface mining practice involving the removal of mountaintops to expose coal seams, and disposing of the associated mining waste
8.2.4 [Link]
in adjacent valleys – "valley fills." The process of MTM is described in more detail by the U.S. Environmental Protection Agency
(U.S. EPA).
Figure 8.2.7 Mountaintop Removal Coal Mining in Martin County, Kentucky. Photograph shows mountaintop coal removal
mining in Martin County, Kentucky. Source: Flashdark.
The following are some examples of the impact of MTM:
an increase of minerals in the water that negatively impact fish and macroinvertebrates, leading to less diverse and more
pollutant-tolerant species
streams are sometimes covered up by silt from mining
the re-growth of trees and woody plants on regraded land may be slowed due to compacted soils
affects the diversity of bird and amphibian species in the area since the ecosystem changes from wooded areas to other
there may be social, economic and heritage issues created by the loss of wooded land that may have been important to traditions
and economies of the area
A study by Epstein, et al. (2011) assigned a monetary value (full cost accounting) for the life cycle of coal in the United States,
accounting for many environmental and health impacts of coal. The authors found the cost to be about $0.178/kWh of electricity
generated from coal ($345.4 billion in 2008), doubling or tripling the price of coal-generated electricity. This study accounted for
all of the impacts discussed above and more.
8.2.5 [Link]
Carbon dioxide (CO2) is a greenhouse gas and a source of climate change.
Sulfur dioxide (SO2) causes acid rain, which damages plants and animals that live in water, and it increases or causes
respiratory illnesses and heart diseases, particularly in vulnerable populations like children and the elderly.
Nitrous oxides (NOx) and Volatile Organic Carbons (VOCs) contribute to ozone at ground level, which is an irritatant and
causes damage to the lungs.
Particulate Matter (PM) produces hazy conditions in cities and scenic areas, and combines with ozone to contribute to asthma
and chronic bronchitis, especially in children and the elderly. Very small, or “fine PM,” is also thought to penetrate the
respiratory system more deeply and cause emphysema and lung cancer.
Lead can have severe health impacts, especially for children.
Air toxins are known or probable carcinogens.
There are other domestic sources of liquid fossil fuel that are being considered as conventional resources and are being depleted.
These include soil sands/tar sands – deposits of moist sand and clay with 1-2 percent bitumen (thick and heavy petroleum rich in
carbon and poor in hydrogen). These are removed by strip mining (see section above on coal). Another source is oil shale in United
States west which is sedimentary rock filled with organic matter that can be processed to produce liquid petroleum. Also, mined by
strip mines or subsurface mines, oil shale can be burned directly like coal or baked in the presence of hydrogen to extract liquid
petroleum. However, the net energy values are low and they are expensive to extract and process. Both of these resources have
severe environmental impacts due to strip mining, carbon dioxide, methane and other air pollutants similar to other fossil fuels.
Figure 8.2.8 U.S. Natural Gas Supply, 1990-2035. Graph shows U.S. historic and projected natural gas production from various
sources. Source: U.S. Energy Information Administration
8.2.6 [Link]
Natural gas is a preferred energy source when considering its environmental impacts. Specifically, when burned, much less carbon
dioxide (CO2), nitrogen oxides, and sulfur dioxide are omitted than from the combustion of coal or oil (see Figure 8.2.1). It also
does not produce ash or toxic emissions.
Figure 8.2.9 Hydraulic Fracturing Process Graphic illustrates the process of hydraulic fracturing. Source: Al Granberg,
ProPublica. This graphic may not be relicensed for sale except by the copyright holder (ProPublica).
8.2.7 [Link]
environmental challenge for nuclear power is the wastes including uranium mill tailings, spent (used) reactor fuel, and other
radioactive wastes. These materials have long radioactive half-lives and thus remain a threat to human health for thousands of
years. The U.S. Nuclear Regulatory Commission regulates the operation of nuclear power plants and the handling, transportation,
storage, and disposal of radioactive materials to protect human health and the environment.
By volume, uranium mill tailings are the largest waste and they contain the radioactive element radium, which decays to produce
radon, a radioactive gas. This waste is placed near the processing facility or mill where they come from, and are covered with a
barrier of a material such as clay to prevent radon from escaping into the atmosphere and then a layer of soil, rocks, or other
materials to prevent erosion of the sealing barrier.
High-level radioactive waste consists of used nuclear reactor fuel. This fuel is in a solid form consisting of small fuel pellets in long
metal tubes and must be stored and handled with multiple containment, first cooled by water and later in special outdoor concrete
or steel containers that are cooled by air. There is no long-term storage facility for this fuel in the United States.
There are many other regulatory precautions governing permitting, construction, operation, and decommissioning of nuclear power
plants due to risks from an uncontrolled nuclear reaction. The potential for contamination of air, water and food is high should an
uncontrolled reaction occur. Even when planning for worst-case scenarios, there are always risks of unexpected events. For
example, the March 2011 earthquake and subsequent tsunami that hit Japan resulted in reactor meltdowns at the Fukushima Daiichi
Nuclear Power Station causing massive damage to the surrounding area.
Fukushima Daiichi Nuclear Power Station
March 11, 2011: Magnitude 9.0 earthquake 231 miles northeast of Tokyo. Less than 1 hour later a 14m tsunami hit
50 power station employees worked around the clock to try to stabilize the situation
United States’ nuclear reactors have containment vessels that are designed to withstand extreme weather events and earthquakes.
However, in the aftermath of the Japan incident, they are reviewing their facilities, policies, and procedures.
Figure 8.2.10 U.S. Energy Consumption by Energy Source, 2009. Renewable energy makes up 8% of U.S. energy consumption.
Source: U.S. Energy Information Administration
Hydropower
Hydropower (hydro-electric) is considered a clean and renewable source of energy since it does not directly produce emissions of
air pollutants and the source of power is regenerated. However, hydropower dams, reservoirs, and the operation of generators can
have environmental impacts. Figure 8.2.11 shows the Hoover Power Plant located on the Colorado River. Hydropower provides 35
percent of the United States’ renewable energy consumption (see Figure 8.2.10). In 2003 capacity was at 96,000 MW and it was
estimated that 30,000 MW capacity is undeveloped.
8.2.8 [Link]
Figure 8.2.10 Hoover Power Plant. View of Hoover Power Plant on the Colorado River as seen from above. Source: U.S.
Department of the Interior
Migration of fish to their upstream spawning areas can be obstructed by a dam that is used to create a reservoir or to divert water to
a run-of-river hydropower plant. A reservoir and operation of the dam can affect the natural water habitat due to changes in water
temperatures, chemistry, flow characteristics, and silt loads, all of which can lead to significant changes in the ecology and physical
characteristics of the river upstream and downstream. Construction of reservoirs may cause natural areas, farms, and archeological
sites to be covered and force populations to relocate. Hydro turbines kill and injure some of the fish that pass through the turbine
although there are ways to reduce that effect. In areas where salmon must travel upstream to spawn, such as along the Columbia
River in Washington and Oregon, the dams get in the way. This problem can be partially alleviated by using “fish ladders” that help
the salmon get up the dams.
Carbon dioxide and methane may also form in reservoirs where water is more stagnant and be emitted to the atmosphere. The exact
amount of greenhouse gases produced from hydropower plant reservoirs is uncertain. If the reservoirs are located in tropical and
temperate regions, including the United States, those emissions may be equal to or greater than the greenhouse effect of the carbon
dioxide emissions from an equivalent amount of electricity generated with fossil fuels (EIA, 2011).
Biomass
Biomass is derived from plants. Examples include lumber mill sawdust, paper mill sludge, yard waste, or oat hulls from an oatmeal
processing plant. A major challenge of biomass is determining if it is really a more sustainable option. It often takes energy to make
energy and biomass is one example where the processing to make it may not be offset by the energy it produces. For example,
biomass combustion may increase or decrease emission of air pollutants depending on the type of biomass and the types of fuels or
energy sources that it replaces. Biomass reduces the demand for fossil fuels, but when the plants that are the sources of biomass are
grown, a nearly equivalent amount of CO2 is captured through photosynthesis, thus it recycles the carbon. If these materials are
grown and harvested in a sustainable way there can be no net increase in CO2 emissions. Each type of biomass must be evaluated
for its full life-cycle impact in order to determine if it is really advancing sustainability and reducing environmental impacts.
8.2.9 [Link]
Figure 8.2.12 Woodchips. Photograph shows a pile of woodchips, which are a type of biomass. Source: Ulrichulrich
8.2.10 [Link]
Biofuels typically replace petroleum and are used to power vehicles. Although ethanol has higher octane and ethanol-gasoline
mixtures burn cleaner than pure gasoline, they also are more volatile and thus have higher "evaporative emissions" from fuel tanks
and dispensing equipment. These emissions contribute to the formation of harmful, ground level ozone and smog. Gasoline
requires extra processing to reduce evaporative emissions before it is blended with ethanol.
Biodiesel can be made from used vegetable oil and has been produced on a very local basis. Compared to petroleum diesel,
biodiesel combustion produces less sulfur oxides, particulate matter, carbon monoxide, and unburned and other hydrocarbons, but
more nitrogen oxide.
Figure 8.2.13 Installing a Geothermal Pipe. System Drilling to install geothermal ground source pipe system. Source: Office of
Sustainability, UIC
Geothermal power plants do not burn fuel to generate electricity so their emission levels are very low. They release less than one
percent of the carbon dioxide emissions of a fossil fuel plant. Geothermal plants use scrubber systems to clean the air of hydrogen
sulfide that is naturally found in the steam and hot water. They emit 97 percent less acid rain-causing sulfur compounds than are
emitted by fossil fuel plants. After the steam and water from a geothermal reservoir have been used, they are injected back into the
earth.
Geothermal ground source systems utilize a heat-exchange system that runs in the subsurface about 20 feet (5 meters) below the
surface where the ground is at a constant temperature. The system uses the earth as a heat source (in the winter) or a heat sink (in
the summer). This reduces the energy consumption requires to generate heat from gas, steam, hot water, and chiller and
conventional electric air-conditioning systems. See more in chapter 10.
Solar Energy
Solar power has minimal impact on the environment, depending on where it is placed. In 2009, one percent of the renewable
energy generated in the United States was from solar power (1646 MW) out of the eight percent of the total electricity generation
that was from renewable sources. The manufacturing of photovoltaic (PV) cells generates some hazardous waste from the
chemicals and solvents used in processing. Often solar arrays are placed on roofs of buildings or over parking lots or integrated into
construction in other ways. However, large systems may be placed on land and particularly in deserts where those fragile
ecosystems could be damaged if care is not taken. Some solar thermal systems use potentially hazardous fluids (to transfer heat)
that require proper handling and disposal. Concentrated solar systems may need to be cleaned regularly with water, which is also
needed for cooling the turbine-generator. Using water from underground wells may affect the ecosystem in some arid locations.
8.2.11 [Link]
Figure 8.2.14 Rooftop Solar Installations. Rooftop solar installation on Douglas Hall at the University of Illinois at Chicago has no
effect on land resources, while producing electricity with zero emissions. Source: Office of Sustainability, UIC
Wind
Wind is a renewable energy source that is clean and has very few environmental challenges. Wind turbines are becoming a more
prominent sight across the United States, even in regions that are considered to have less wind potential. Wind turbines (often
called windmills) do not release emissions that pollute the air or water (with rare exceptions), and they do not require water for
cooling. The U.S. wind industry had 40,181 MW of wind power capacity installed at the end of 2010, with 5,116 MW installed in
2010 alone, providing more than 20 percent of installed wind power around the globe. According to the American Wind Energy
Association, over 35 percent of all new electrical generating capacity in the United States since 2006 was due to wind, surpassed
only by natural gas.
Figure 8.2.15 Twin Groves Wind Farm, Illinois. Wind power is becoming a more popular source of energy in the United States.
Source: Office of Sustainability, UIC
Since a wind turbine has a small physical footprint relative to the amount of electricity it produces, many wind farms are located on
crop, pasture, and forest land. They contribute to economic sustainability by providing extra income to farmers and ranchers,
allowing them to stay in business and keep their property from being developed for other uses. For example, energy can be
produced by installing wind turbines in the Appalachian mountains of the United States instead of engaging in mountain top
removal for coal mining. Off shore wind turbines on lakes or the ocean may have smaller environmental impacts than turbines on
land.
Wind turbines do have a few environmental challenges. There are aesthetic concerns to some people when they see them on the
landscape. A few wind turbines have caught on fire, and some have leaked lubricating fluids, though this is relatively rare. Some
people do not like the sound that wind turbine blades make. Listen to one here and see what you think.
8.2.12 [Link]
Turbines have been found to cause bird and bat deaths particularly if they are located along their migratory path. This is of
particular concern if these are threatened or endangered species. There are ways to mitigate that impact and it is currently being
researched.
There are some small impacts from the construction of wind projects or farms, such as the construction of service roads, the
production of the turbines themselves, and the concrete for the foundations. However, overall life cycle analysis has found that
turbines make much more energy than the amount used to make and install them.
Summary
We derive our energy from a multitude of resources that have varying environmental challenges related to air and water pollution,
land use, carbon dioxide emissions, resource extraction and supply, as well as related safety and health issues. A diversity of
resources can help maintain political and economic independence for the United States. Renewable energy sources have lower
environmental impact and can provide local energy resources. Each resource needs to be evaluated within the sustainability
paradigm. In the near future, we can expect the interim use of more difficult and environmentally-challenging extraction methods
to provide fossil fuels until the growth and development of renewable and clean energy sources will be able to meet our energy
demands.
Review Questions
1. Describe three major environmental challenges for fossil fuels in general or one in particular.
2. What are the compelling reasons to continue using coal in spite of its challenges?
3. Rate the following electricity sources for their contribution to climate change from most to least: biomass, coal, solar, wind,
nuclear, natural gas, oil, geothermal, hydroelectric, MSW. Is there any compelling reason not to use any of the carbon neutral
(no net carbon emissions) sources?
4. Describe the environmental and social concerns with regard to biofuels.
Resources
To learn more about global energy issues, visit the International Energy Agency website.
To learn more about United States and international energy issues, visit the U.S. Energy Information Administration website.
To learn more about the U.S. Nuclear Regulatory Commission, please click here.
Learn about your clean energy options here.
References
1. American Wind Energy Association. (2011). Industry Statistics. Retrieved September 6, 2011 from
[Link]/learnabout/indus...stats/[Link]
2. Epstein, P.R., Buonocare, J.J, Eckerle, K., Hendryx, M., Stout III, B.M., Heinberg, R., et al. (2011). Full cost accounting for the
life cycle of coal. Annals of the New York Academy of Sciences, 1219, 73-98. Retrieved May 17, 2011 from
[Link]/downloads/CoalExtern...[Link]
3. U.S. Energy Information Administration. (2011). Hydropower generators produce clean electricity, but hydropower does have
environmental impacts. Retrieved September 6, 2011 from [Link]
page=hydropower_environment
4. Wood, J.H., Long, G.R, & Morehouse, D.F. (2004). Long-term world oil supply scenarios:The future is neither as bleak or rosy
as some assert. Energy Information Administration. Retrieved May 17, 2011
from[Link]
Glossary
biodiesel
A fuel usually made from soybean, canola, or other vegetable oils; animal fats; and recycled grease and oils. It can serve as a
substitute for conventional diesel or distillate fuel.
biofuels
8.2.13 [Link]
Liquid fuels and blending components produced from biomass materials, used primarily in combination with transportation
fuels, such as gasoline.
biomass
Organic, non-fossil material of biological origin that is renewable because it can be quickly re-grown, taking up the carbon that
is released when it is burned.
geothermal energy
Hot water or steam extracted from geothermal reservoirs in the earth's crust. Water or steam extracted from geothermal
reservoirs can be used for geothermal heat pumps, water heating, or electricity generation. Geothermal heat or cooling may also
come from ground source heat exchange taking advantage of the constant temperature in the ground below the surface.
geothermal plant
A power plant in which the prime mover is a steam turbine. The turbine is driven either by steam produced from hot water or by
natural steam that heat source is found in rock.
non-renewable fuels
Fuels that will be used up, irreplaceable.
photovoltaic cells
An electronic device consisting of layers of semiconductor materials that are produced to form adjacent layers of materials with
different electronic characteristics and electrical contacts and being capable of converting incident light directly into electricity
(direct current).
radioactive half-lives
The amount of time necessary to decrease the radioactivity of radioactive material to one-half the original level.
renewable fuels
Fuels that are never exhausted or can be replaced.
This page titled 8.2: Environmental Challenges in Energy, Carbon Dioxide, Air, Water and Land Use is shared under a CC BY license and was
authored, remixed, and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
8.2.14 [Link]
8.3: Case Study- Greenhouse Gases and Climate Change
Introduction
If increased greenhouse gas emissions from human activity are causing climate change, then how do we reduce those emissions?
Whether dictated by an international, national, or local regulation or a voluntary agreement, plans are needed to move to a low-
carbon economy. In the absence of federal regulation, cities, states, government institutions, and colleges and universities, have all
taken climate action initiatives. This case study provides two examples of climate action plans – one for a city (Chicago) and one
for an institution (the University of Illinois at Chicago).
Figure illustrates the emissions calculated for Chicago through 2005. Source: City of Chicago, Chicago Climate Action Plan
A good climate action plan includes reporting of greenhouse gas emissions, as far back as there is data, preferably to 1990. Figure
8.3.1 depicts the emissions calculated for Chicago through 2005. From that point there is an estimate (the dotted line) of a further
increase before the reductions become evident and the goals portrayed can be obtained. The plan was released in September 2008
and provides a roadmap of five strategies with 35 actions to reduce greenhouse gas emissions (GHG) and adapt to climate change.
The strategies are shown in Table 8.3.1. Figure 8.3.2 identifies the proportion of emissions reductions from the various strategies.
Sources of the CCAP Emission Reductions by Strategy
Figure 8.3.2 Graph shows the sources of the Chicago CAP emission reductions by strategy. Source: C. Klein-Banai using data
from City of Chicago, Chicago Climate Action Plan.
In 2010 CCAP put out a progress report wherein progress is measured by the many small steps that are being taken to implement
the plan. It is not translated exactly to emissions reductions but reports on progress for each step such as the number of residential
units that have been retrofitted for energy efficiency, the number of appliances traded in, the increase in the number of rides on
public transit, and the amount of water conserved daily.
8.3.1 [Link]
University Climate Action Plan
Several factors caused a major Chicago university to develop a climate action plan. As part of the American College and University
Presidents’ Climate Commitment (ACUPCC), nearly 670 presidents have signed a commitment to inventory their greenhouse
gases, publicly report it, and to develop a climate action plan. Part of the Chicago CAP is to engage businesses and organizations
within the city in climate action planning. In order to be a better steward of the environment, the University of Illinois at Chicago
(UIC) developed a climate action plan. The goals are similar to Chicago’s: a 40 percent GHG emissions reduction by 2030 and at
least 80 percent by 2050, using a 2004 baseline. The strategies align with those of the city in which the campus resides (see Table
8.3.1). UIC’s greenhouse gas reports are also made publically available on the ACUPCC reporting site. Figure 8.3.3 displays
UIC’s calculated emissions inventory (in red) and then the predicted increases for growth if activities continue in a “business as
usual (BAU)” approach. The triangular wedges below represent emissions reductions through a variety of strategies, similar to
those of the wedge approach that Professors Sokolow and Pacala proposed. Those strategies are displayed in Table 8.3.1, alongside
Chicago’s for comparative purposes.
Figure 8.3.3 UIC’s Projected Emissions Reductions. Projected emissions reductions from 2004 to 2030. Where BAU stands for
Business as Usual, what would happen if no action were taken? Source: UIC Climate Action Plan, figure 6.
Improve power plant efficiency Purchase electricity from a renewable electricity provider
8.3.2 [Link]
Increase distributed generation
Capture storm water on site Sustainable food purchase and use of biodegradable packaging
Collecting and converting vegetable oil
Develop a user-friendly property management system
Expand the waste minimization program
Recycle construction debris
Purchasing policies
Preparation (Adaptation) Improved Grounds Operations
Engage businesses
8.3.3 [Link]
Childcare center
Public Engagement
Table 8.3.1 Alignment of the Chicago and UIC Climate Action Plans Source: C. Klein-Banai using data from Chicago
Climate Action Plant and UIC Climate Action Plan
Conclusion
There is no one approach that will effectively reduce greenhouse gas emissions. Climate action plans are helpful tools to represent
strategies to reduce emissions. Governmental entities such as nations, states, and cities can develop plans, as can institutions and
businesses. It is important that there be an alignment of plans when they intersect, such as a city and a university that resides within
it.
This page titled 8.3: Case Study- Greenhouse Gases and Climate Change is shared under a CC BY license and was authored, remixed, and/or
curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
8.3.4 [Link]
8.4: Energy Sources and Carriers
Electricity
Introduction
Over the past century and a half electricity has emerged as a popular and versatile energy carrier. Communication was an early
widespread use for electricity following the introduction of the telegraph in the 1840s. In the 1870s and 1880s electric motors and
lights joined the telegraph as practical electrical devices, and in the 1890s electricity distribution systems, the forerunners of today's
electricity grid, began to appear. The telegraph became wireless with the invention of radio, demonstrated in the laboratory in the
1880s and for transatlantic communication in 1901. Today, electricity is exploited not only for its diverse end uses such as lighting,
motion, refrigeration, communication and computation, but also as a primary carrier of energy. Electricity is one of two backbones
of the modern energy system (liquid transportation fuels are the other), carrying high density energy over short and long distances
for diverse uses. In 2009, electricity consumed the largest share of the United States’ primary energy, 38 percent, with
transportation a close second at 37 percent (EIA Annual Energy Review, 2009). These two sectors also accounted for the largest
shares of U.S. carbon emissions, 38 percent for electricity and 33 percent for transportation (EIA Annual Energy Review, 2009).
Figure 8.4.1 shows the growth of electricity as an energy carrier since 1949 and the growing range of its uses.
United States Electricity Net Generation Since 1949 and Uses
Figure 8.4.1 United States Electricity Net Generation Since 1949 and Uses The growth of United States electricity generation since
1949 and some of its uses. Source: G. Crabtree using data from EIA Annual Energy Review 2009, Table 8.2a, p 230.; Felix O, U.S.
CPSC, Joe Mabel,Marcin Wichary, Samboy, Andrew, Jan Ainali, Lovelac7
Figure Electricity Energy Chain shows the electricity energy chain from generation to use. By far most electricity is generated by
combustion of fossil fuels to turn steam or gas turbines. This is the least efficient step in the energy chain, converting only 36
percent of the chemical energy in the fuel to electric energy, when averaged over the present gas and coal generation mix. It also
produces all the carbon emissions of the electricity chain. Beyond production, electricity is a remarkably clean and efficient carrier.
Conversion from rotary motion of the turbine and generator to electricity, the delivery of electricity through the power grid, and the
conversion to motion in motors for use in industry, transportation and refrigeration can be more than 90 percent efficient. None of
these steps produces greenhouse gas emissions. It is the post-production versatility, cleanliness, and efficiency of electricity that
make it a prime energy carrier for the future. Electricity generation, based on relatively plentiful domestic coal and gas, is free of
immediate fuel security concerns. The advent of electric cars promises to increase electricity demand and reduce dependency on
foreign oil, while the growth of renewable wind and solar generation reduces carbon emissions. The primary sustainability
challenges for electricity as an energy carrier are at the production step: efficiency and emission of carbon dioxide and toxins.
Electric Energy Chain
Figure 8.4.2 Electricity Energy Chain Graph shows the electricity energy chain from generation to use. Source: G. Crabtree
Figure 8.4.3 Superconducting Underground Cables The superconducting wires on the right carry the same current as the
conventional copper wires on the left. Superconducting cable wound from these wires carries up to five times the current of
conventional copper cables. Source: Courtesy, American Superconductor Corporation
The reliability of the electricity grid presents a second challenge. The United States’ grid has grown continuously from origins in
the early 20th Century; much of its infrastructure is based on technology and design philosophy dating from the 1950s and 1960s,
when the major challenge was extending electrification to new rural and urban areas. Outside urban areas, the grid is mainly above
ground, exposing it to weather and temperature extremes that cause most power outages. The response to outages is frustratingly
slow and traditional – utilities are often first alerted to outages by telephoned customer complaints, and response requires sending
8.4.1 [Link]
crews to identify and repair damage, much the same as we did 50 years ago. The United States’ grid reliability is significantly
lower than for newer grids in Europe and Japan, where the typical customer experiences ten to 20 times less outage time than in the
United States. Reliability is especially important in the digital age, when an interruption of even a fraction of a cycle can shut down
a digitally controlled data center or fabrication line, requiring hours or days to restart.
Reliability issues can be addressed by implementing a smart grid with two-way communication between utility companies and
customers that continuously monitors power delivery, the operational state of the delivery system, and implements demand
response measures adjusting power delivered to individual customers in accordance with a previously established unique customer
protocol. Such a system requires installing digital sensors that monitor power flows in the delivery system, digital decision and
control technology and digital communication capability like that already standard for communication via the Internet. For
customers with on-site solar generation capability, the smart grid would monitor and control selling excess power from the
customer to the utility.
Figure 8.4.4 illustrates the two-way communication features of the smart grid. The conventional grid in the upper panel sends
power one way, from the generating station to the customer, recording how much power leaves the generator and arrives at the
customer. In the smart grid, the power flow is continuously monitored, not only at the generator and the customer, but also at each
connection point in between. Information on the real time power flow is sent over the Internet or another special network to the
utility and to the customer, allowing real time decisions on adding generation to meet changes in load, opening circuit breakers to
reroute power in case of an outage, reducing power delivered to the customer during peak periods to avoid outages (often called
"demand response"), and tracking reverse power flows for customers with their own solar or other generation capacity. The
conventional power grid was designed in the middle of the last century to meet the simple need of delivering power in one
direction. Incorporating modern Internet-style communications and control features could bring the electricity grid to a
qualitatively new level of capability and performance required to accommodate local generation and deliver higher reliability.
Smart Grid
Figure 8.4.4 Smart Grid The addition of real-time monitoring and communicating capability like that used on the Internet would
add 'smart' operation of the electricity grid. Source: National Institute of Standards and Technology
Smart components incorporated throughout the grid would be able to detect overload currents and open breakers to interrupt them
quickly and automatically to avoid unnecessary damage and triggering a domino effect cascade of outages over wide areas as
happened in the Northeast Blackout of 2003. For maximum effectiveness, such smart systems require fast automatic response on
millisecond time scales commensurate with the cycle time of the grid. Even simple digital communication meets this requirement,
but many of the grid components themselves cannot respond so quickly. Conventional mechanical circuit breakers, for example,
take many seconds to open and much longer to close. Such long times increase the risk of dangerous overload currents damaging
the grid or propagating cascades. Along with digital communications, new breaker technology, such as that based on fast, self-
healing superconducting fault current limiters, is needed to bring power grid operation into the modern era.
Renewable Variability
The grid faces major challenges to accommodate the variability of wind and solar electricity. Without significant storage capacity,
the grid must precisely balance generation to demand in real time. At present, the variability of demand controls the balancing
process: demand varies by as much as a factor of two from night to day as people go through their daily routines. This predictable
variability is accommodated by switching reserve generation sources in and out in response to demand variations. With renewable
generation, variation can be up to 70 percent for solar electricity due to passing clouds and 100 percent for wind due to calm days,
much larger than the variability of demand. At the present level of 1.6 percent wind and solar penetration, the relatively small
variation in generation can be accommodated by switching in and out conventional resources to make up for wind and solar
fluctuations. At the 20 percent penetration required by state Renewable Portfolio Standards, accommodating the variation in
generation requires a significant increase in the conventional reserve capacity. At high penetration levels, each addition of wind or
8.4.2 [Link]
solar capacity requires a nearly equal addition of conventional capacity to provide generation when the renewables are quiescent.
This double installation to insure reliability increases the cost of renewable electricity and reduces its effectiveness in lowering
greenhouse gas emissions.
A major complication of renewable variation is its unpredictability. Unlike demand variability, which is reliably high in the
afternoon and low at night, renewable generation depends on weather and does not follow any pattern. Anticipating weather-driven
wind and solar generation variability requires more sophisticated forecasts with higher accuracy and greater confidence levels than
are now available. Because today's forecasts often miss the actual performance target, additional conventional reserves must be
held at the ready to cover the risk of inaccuracies, adding another increase to the cost of renewable electricity.
Storage of renewable electricity offers a viable route to meeting the variable generation challenge. Grid electricity storage
encompasses many more options than portable electricity storage required for electric cars. Unlike vehicle storage, grid storage can
occupy a large footprint with little or no restriction on weight or volume. Grid storage can be housed in a controlled environment,
eliminating large temperature and humidity variations that affect performance. Grid storage must have much higher capacity than
vehicle storage, of order 150 MWh for a wind farm versus 20-50 kWh for a vehicle. Because of these differences, the research
strategy for grid and vehicle energy storage is very different. To date, much more attention has been paid to meeting vehicle
electricity storage requirements than grid storage requirements.
There are many options for grid storage. Pumped hydroelectric storage, illustrated in Figure 8.4.5, is an established technology
appropriate for regions with high and low elevation water resources. Compressed Air Energy Storage (CAES) is a compressed air
equivalent of pumped hydro that uses excess electricity to pump air under pressure into underground geologic formations for later
release to drive generators. This option has been demonstrated in Huntorf, Germany and in Mcintosh, Alabama. High temperature
sodium-sulfur batteries operating at 300 °C have high energy density, projected long cycle life, and high round trip efficiency; they
are the most mature of the battery technologies suggested for the grid. Flow batteries are an attractive and relatively unexplored
option, where energy is stored in the high charge state of a liquid electrolyte and removed by electrochemical conversion to a low
charge state. Each flow battery requires an electrolyte with a high and low charge state and chemical reaction that takes one into the
other. There are many such electrolytes and chemical reactions, of which only a few have been explored, leaving a host of
promising opportunities for the future. The energy storage capacity depends only on the size of the storage tank, which can be
designed fully independently of the power capacity that depends on the size of the electrochemical reactor. Sodium sulfur and flow
batteries store electric charge and can be used at any place in the electricity grid. In contrast, thermal storage applies only to
concentrating solar power technologies, where mirrors focus solar radiation to heat a working fluid that drives a conventional
turbine and generator. In these systems, heat energy can be stored as a molten salt in a highly insulated enclosure for hours or days,
allowing solar electricity to be generated on demand after sunset or on cloudy days. All of these options are promising and require
research and development to explore innovations, performance and cost limits.
Pumped Hydroelectric Storage
Figure 8.4.5 Pumped Hydroelectric Storage Upper storage reservoir for pumped hydroelectric storage, an established technology
for storing large amounts of grid electricity. Source: Ongrys via Wikimedia Commons
Figure 8.4.6 Renewable Resource Location vs. Demand Location Wind and solar electricity resources are located far from
population centers, requiring a dramatic improvement in long-distance electricity transmission – an "interstate highway system for
electricity." Source: Integrating Renewable Electricity on the Grid, Report of the Panel on Pubic Affairs, American Physical
Society (2010).
Summary
Electricity and liquid petroleum are the two primary energy carriers in the United States, and in the world. Once produced,
electricity is clean and versatile making it an appealing energy carrier for the future. The challenges facing the electricity grid are
8.4.3 [Link]
capacity, reliability, and accommodating renewable sources such as solar and wind whose output is variable and whose location is
remote from population centers. Electricity storage and long distance transmission are needed to accommodate these renewable
resources.
Figure 8.4.7 Electricity Generation by Source Chart shows U.S. electricity generation by source. Source: U.S. Energy Information
Administration, Annual Review, 2009, p. 228 (Aug. 2010)
Growth of Fuels Used to Produce Electricity in the United States
Figure 8.4.8 Growth of Fuels Used to Produce Electricity in the United States Graph shows the growth of fuels used to produce
electricity in the United States from 1950 to 2009. Source: U.S. Energy Information Administration, Annual Energy Review 2009,
p. 238 (Aug. 2010)
Figure 8.4.9 Global Carbon Cycle, 1990s The global carbon cycle for the 1990s, showing the main annual fluxes in GtC yr–1: pre-
industrial ‘natural’ fluxes in black and ‘anthropogenic’ fluxes in red. Source: Climate Change 2007: The Physical Science Basis:
Contribution of Working Group I to the Fourth Assessment Report of the Intergovernmental Panel on Climate Change, Cambridge
University Press, figure 7.3
Beyond a trend from coal to gas for electricity generation, there is a need to deal with the carbon emissions from the fossil
production of electricity. Figure Global Carbon Cycle, 1990s shows the size of these emissions compared to natural fluxes between
ocean and atmosphere and from vegetation and land use. The anthropogenic fluxes are small by comparison, yet have a large effect
on the concentration of carbon dioxide in the atmosphere. The reason is the step-wise dynamics of the carbon cycle. The ultimate
storage repository for carbon emissions is the deep ocean, with abundant capacity to absorb the relatively small flux from fossil
fuel combustion. Transfer to the deep ocean, however, occurs in three steps: first to the atmosphere, then to the shallow ocean, and
finally to the deep ocean. The bottleneck is the slow transfer of carbon dioxide from the shallow ocean to the deep ocean, governed
by the great ocean conveyor belt or thermohaline circulation illustrated in Figure Great Ocean Conveyor Belt. The great ocean
conveyor belt takes 400 – 1000 years to complete one cycle. While carbon dioxide waits to be transported to the deep ocean, it
saturates the shallow ocean and "backs up" in the atmosphere causing global warming and threatening climate change. If carbon
8.4.4 [Link]
emissions are to be captured and stored (or "sequestered") they must be trapped for thousands of years while the atmosphere adjusts
to past and future carbon emissions (Lenton, 2006).
Great Ocean Conveyor Belt
Figure 8.4.10 Great Ocean Conveyor Belt The great ocean conveyor belt (or thermohaline current) sends warm surface currents
from the Pacific to Atlantic oceans and cold deep currents in the opposite direction. The conveyor belt is responsible for
transporting dissolved carbon dioxide from the relatively small reservoir of the shallow ocean to much larger reservoir of the deep
ocean. It takes 400 - 1000 years to complete one cycle. Source: Argonne National Laboratory
Sequestration of carbon dioxide in underground geologic formations is one process that, in principle, has the capacity to handle
fossil fuel carbon emissions (Olajire, 2010); chemical reaction of carbon dioxide to a stable solid form is another (Stephens &
Keith, 2008). For sequestration, there are fundamental challenges that must be understood and resolved before the process can be
implemented on a wide scale.
The chemical reactions and migration routes through the porous rocks in which carbon dioxide is stored underground are largely
unknown. Depending on the rock environment, stable solid compounds could form that would effectively remove the sequestered
carbon dioxide from the environment. Alternatively, it could remain as carbon dioxide or transform to a mobile species and migrate
long distances, finally finding an escape route to the atmosphere where it could resume its contribution to greenhouse warming or
cause new environmental damage. The requirement on long term sequestration is severe: a leak rate of 1 percent means that all the
carbon dioxide sequestered in the first year escapes in a century, a blink of the eye on the timescale of climate change.
Summary
Coal (45 percent) and gas (23 percent) are the two primary fossil fuels for electricity production in the United States. Coal
combustion produces nearly twice the carbon emissions of gas combustion. Increasing public opinion and regulatory pressure to
lower carbon emissions are shifting electricity generation toward gas and away from coal. The domestic supply of gas is increasing
rapidly due to shale gas released by hydraulic fracturing, a technology with significant potential for harmful environmental impact.
Reducing the greenhouse gas impact of electricity production requires capturing and sequestering the carbon dioxide emitted from
power plants. Storing carbon dioxide in underground geologic formations faces challenges of chemical transformation, migration,
and longevity.
Nuclear Energy
From a sustainability perspective, nuclear electricity presents an interesting dilemma. On the one hand, nuclear electricity produces
no carbon emissions, a major sustainable advantage in a world facing human induced global warming and potential climate change.
On the other hand, nuclear electricity produces spent fuel that must be stored out of the environment for tens or hundreds of
thousands of years, it produces bomb-grade plutonium and uranium that could be diverted by terrorists or others to destroy cities
and poison the environment, and it threatens the natural and built environment through accidental leaks of long lived radiation.
Thoughtful scientists, policy makers and citizens must weigh the benefit of this source of carbon free electricity against the
environmental risk of storing spent fuel for thousands or hundreds of thousands of years, the societal risk of nuclear proliferation,
and the impact of accidental releases of radiation from operating reactors. There are very few examples of humans having the
power to permanently change the dynamics of the earth. Global warming and climate change from carbon emissions is one
example, and radiation from the explosion of a sufficient number of nuclear weapons is another. Nuclear electricity touches both of
these opportunities, on the positive side for reducing carbon emissions and on the negative side for the risk of nuclear proliferation.
Figure 8.4.11 Nuclear Share of United States Electricity Generation The percentage of electricity generated by nuclear power in
the United States, 1957-2009. Source: U.S. Energy Information Agency, Annual Energy Review 2009, p. 276 (Aug. 2010)
8.4.5 [Link]
The outcome of this debate (Ferguson, Marburger, & Farmer, 2010) will determine whether the world experiences a nuclear
renaissance that has been in the making for several years (Grimes & Nuttall, 2010). The global discussion has been strongly
impacted by the unlikely nuclear accident in Fukushima, Japan in March 2011. The Fukushima nuclear disaster was caused by an
earthquake and tsunami that disabled the cooling system for a nuclear energy complex consisting of operating nuclear reactors and
storage pools for underwater storage of spent nuclear fuel ultimately causing a partial meltdown of some of the reactor cores and
release of significant radiation. This event, 25 years after Chernobyl, reminds us that safety and public confidence are especially
important in nuclear energy; without them expansion of nuclear energy will not happen.
Operating and Decommissioned Nuclear Power Plants in the United States
Figure 8.4.12 Operating and Decommissioned Nuclear Power Plants in the United States Graph shows the number of operating
versus decommissioned nuclear power plants in the United States. Source: U.S. Energy Information Agency, Annual Energy
Review 2009, p. 274 (Aug. 2010)
There are two basic routes for handling the spent fuel of nuclear reactors: once through and reprocessing (World Nuclear
Association; Kazimi, Moniz, & Forsberg, 2010). Once through stores spent fuel following a single pass through the reactor, first in
pools at the reactor site while it cools radioactively and thermally, then in a long-term geologic storage site, where it must remain
for hundreds of thousands of years. Reprocessing separates the useable fraction of spent fuel and recycles it through the reactor,
using a greater fraction of its energy content for electricity production, and sends the remaining high-level waste to permanent
geologic storage. The primary motivation for recycling is greater use of fuel resources, extracting ~ 25 percent more energy than
the once through cycle. A secondary motivation for recycling is a significant reduction of the permanent geologic storage space (by
a factor of ~ 5 or more) and time (from hundreds of thousands of years to thousands of years). While these advantages seem natural
and appealing from a sustainability perspective, they are complicated by the risk of theft of nuclear material from the reprocessing
cycle for use in illicit weapons production or other non-sustainable ends. At present, France, the United Kingdom, Russia, Japan
and China engage in some form of reprocessing; the United States, Sweden and Finland do not reprocess.
Summary
Nuclear electricity offers the sustainable benefit of low carbon electricity at the cost of storing spent fuel out of the environment for
up to hundreds of thousands of years. Nuclear energy developed in only 11 years, unusually quickly for a major energy technology,
and slowed equally quickly due to public concerns about safety following Three Mile Island and Chernobyl. The Fukushima
reactor accident in March 2011 has raised further serious concerns about safety; its impact on public opinion could dramatically
affect the future course of nuclear electricity. Reprocessing spent fuel offers the advantages of higher energy efficiency and reduced
spent fuel storage requirements with the disadvantage of higher risk of weapons proliferation through diversion of the reprocessed
fuel stream.
8.4.6 [Link]
of transportation from uncertain and volatile foreign oil to domestic electricity and biofuels, and production of electricity from low
carbon sources.
Figure 8.4.13 Forms of Renewable Energy Provided by the Sun The sun is the ultimate source for many forms of renewable
energy: wind and running water that can be used for power generation without heat or combustion, and photosynthesis of green
plants (biomass) for combustion to provide heat and power generation and for conversion to biofuels (upper panels). Solar energy
can be directly captured for water and space heating in buildings, after concentration by mirrors in large plants for utility-scale
power generation by conventional turbines, and without concentration in photovoltaic cells that produce power without heat or
combustion (lower panels). Source: G. Crabtree using images from Linuxerist, Mor plus, Richard Dorrell, Hernantron, BSMPS,
Cachogaray, and Andy F.
As relative newcomers to energy production, renewable energy typically operates at lower efficiency than its conventional
counterparts. For example, the best commercial solar photovoltaic modules operate at about 20 percent efficiency, compared to
nearly 60 percent efficiency for the best combined cycle natural gas turbines. Photovoltaic modules in the laboratory operate above
40 percent efficiency but are too expensive for general use, showing that there is ample headroom for performance improvements
and cost reductions. Wind turbines are closer to their theoretical limit of 59 percent (known as Betz's law) often achieving 35 – 40
percent efficiency. Biomass is notoriously inefficient, typically converting less than one percent of incident sunlight to energy
stored in the chemical bonds of its roots, stalks and leaves. Breeding and genetic modification may improve this poor energy
efficiency, though hundreds of millions of years of evolution since the appearance of multicelled organisms have not produced a
significant advance. Geothermal energy is already in the form of heat and temperature gradients, so that standard techniques of
thermal engineering can be applied to improve efficiency. Wave and tidal energy, though demonstrated in several working plants,
are at early stages of development and their technological development remains largely unexplored.
Figure 8.4.14 Renewable Energy Share of Global Final Energy Consumption, 2008 The contribution of fossil, nuclear and
renewable energy to global final energy consumption in 2008. Source: REN21. 2010. Renewables 2010 Global Status Report
(Paris: REN21 Secretariat), p. 15
8.4.7 [Link]
The potential of renewable energy resources varies dramatically. Solar energy is by far the most plentiful, delivered to the surface
of the earth at a rate of 120,000 Terawatts (TW), compared to the global human use of 15 TW. To put this in perspective, covering
100x100 km2 of desert with 10 percent efficient solar cells would produce 0.29 TW of power, about 12 percent of the global human
demand for electricity. To supply all of the earth’s electricity needs (2.4 TW in 2007) would require 7.5 such squares, an area about
the size of Panama (0.05 percent of the earth’s total land area). The world’s conventional oil reserves are estimated at three trillion
barrels, including all the oil that has already been recovered and that remain for future recovery. The solar energy equivalent of
these oil reserves is delivered to the earth by the sun in 1.5 days.
The global potential for producing electricity and transportation fuels from solar, wind and biomass is limited by geographical
availability of land suitable for generating each kind of energy (described as the geographical potential), the technical efficiency of
the conversion process (reducing the geographical potential to the technical potential), and the economic cost of construction and
operation of the conversion technology (reducing the technical potential to the economic potential). The degree to which the global
potential of renewable resources is actually developed depends on many unknown factors such as the future extent of economic and
technological advancement in the developing and developed worlds, the degree of globalization through business, intellectual and
social links among countries and regions, and the relative importance of environmental and social agendas compared to economic
and material objectives. Scenarios evaluating the development of renewable energy resources under various assumptions about the
world’s economic, technological and social trajectories show that solar energy has 20-50 times the potential of wind or biomass for
producing electricity, and that each separately has sufficient potential to provide the world’s electricity needs in 2050 (de Vries,
2007).
The geographical distribution of useable renewable energy is quite uneven. Sunlight, often thought to be relatively evenly
distributed, is concentrated in deserts where cloud cover is rare. Winds are up to 50 percent stronger and steadier offshore than on
land. Hydroelectric potential is concentrated in mountainous regions with high rainfall and snowmelt. Biomass requires available
land that does not compete with food production, and adequate sun and rain to support growth. Figure 8.4.15 shows the
geographical distribution of renewable electricity opportunities that are likely to be economically attractive in 2050 under an
aggressive world development scenario.
Map shows areas where one or more of the wind, solar, and biomass options of renewable electricity
is estimated to be able to produce electricity in 2050 at costs below 10 b kWh.
Figure 8.4.15 Renewable Electricity Opportunities Map shows areas where one or more of the wind, solar, and biomass options of
renewable electricity is estimated to be able to produce electricity in 2050 at costs below 10 b kWh. Source: de Vries, B.J.M., van
Vuuren, D.P., & Hoogwijk, M.M. (2007). A hyper-text must be included to the Homepage of the journal from which you are
licensing at [Link] Permission for reuse must be obtained from Elsevier.
Figure shows the average wind speeds in the United States at 80 meters. Also see offshore wind resource maps. Source: U.S.
Department of Energy, National Renewable Energy Laboratory and AWS Truepower LLC
Barriers to Deployment
Renewable energy faces several barriers to its widespread deployment. Cost is one of the most serious, illustrated in Figure 8.4.17.
Although the cost of renewables has declined significantly in recent years, most are still higher in cost than traditional fossil
alternatives. Fossil energy technologies have a longer experience in streamlining manufacturing, incorporating new materials,
8.4.8 [Link]
taking advantage of economies of scale and understanding the underlying physical and chemical phenomena of the energy
conversion process. As Figure 8.4.17 shows, the lowest cost electricity is generated by natural gas and coal, with hydro and wind
among the renewable challengers. Cost, however, is not an isolated metric; it must be compared with the alternatives. One of the
uncertainties of the present business environment is the ultimate cost of carbon emissions. If governments put a price on carbon
emission to compensate the social cost of global warming and the threat of climate change, the relative cost of renewables will
become more appealing even if their absolute cost does not change. This policy uncertainty in the eventual cost of carbon-based
power generation is a major factor in the future economic appeal of renewable energy.
Cost Estimates of Electricity in 2020 by Fossil, Nuclear and Renewable Generation
Figure 8.4.17 Production Cost of Electricity - 2020 Projection Estimates of the cost of electricity in 2020 by fossil, nuclear and
renewable generation. Source: European Commission, Strategic Energy Technologies Information System
A second barrier to widespread deployment of renewable energy is public opinion. In the consumer market, sales directly sample
public opinion and the connection between deployment and public acceptance is immediate. Renewable energy is not a choice that
individual consumers make. Instead, energy choices are made by government policy makers at city, state and federal levels, who
balance concerns for the common good, for “fairness” to stakeholders, and for economic cost. Nevertheless, public acceptance is a
major factor in balancing these concerns: a strongly favored or disfavored energy option will be reflected in government decisions
through representatives elected by or responding to the public. Figure 8.4.18 shows the public acceptance of renewable and fossil
electricity options. The range of acceptance goes from strongly positive for solar to strongly negative for nuclear. The disparity in
the public acceptance and economic cost of these two energy alternatives is striking: solar is at once the most expensive alternative
and the most acceptable to the public.
The importance of public opinion is illustrated by the Fukushima nuclear disaster of 2011. The earthquake and tsunami that
ultimately caused meltdown of fuel in several reactors of the Fukushima complex and release of radiation in a populated area
caused many of the public in many countries to question the safety of reactors and of the nuclear electricity enterprise generally.
The response was rapid, with some countries registering public consensus for drastic action such as shutting down nuclear
electricity when the licenses for the presently operating reactors expire. Although its ultimate resolution is uncertain, the sudden
and serious impact of the Fukushima event on public opinion shows the key role that social acceptance plays in determining our
energy trajectory.
European Union Citizens’ Acceptance of Renewable and Fossil Electricity Generation Technologies
Figure shows the European Union citizens’ public acceptance of renewable and fossil electricity generation technologies. Source:
European Commission, Eurobarameter on Energy Technologies: Knowledge-Perception-Measures, p. 33
Summary
Strong interest in renewable energy arose in the 1970s as a response to the shortage and high price of imported oil, which disrupted
the orderly operation of the economies and societies of many developed countries. Today there are new motivations, including the
realization that growing greenhouse gas emission accelerates global warming and threatens climate change, the growing
dependence of many countries on foreign oil, and the economic drain of foreign oil payments that slow economic growth and job
creation. There are three ultimate sources of all renewable and fossil energies: sunlight, the heat in the earth’s core and crust, and
the gravitational pull of the moon and sun on the oceans. Renewable energies are relatively recently developed and typically
operate at lower efficiencies than mature fossil technologies. Like early fossil technologies, however, renewables can be expected
to improve their efficiency and lower their cost over time, promoting their economic competitiveness and widespread deployment.
The future deployment of renewable energies depends on many factors, including the availability of suitable land, the technological
cost of conversion to electricity or other uses, the costs of competing energy technologies, and the future need for energy. Scenario
analyses indicate that renewable energies are likely to be technically and economically capable of supplying the world’s electricity
needs in 2050. In addition to cost, public acceptance is a key factor in the widespread deployment of renewable energy.
8.4.9 [Link]
The rise in the price of oil in the last decade makes dependence on imported energy for transportation an economic as well as an
energy issue. The United States, for example, now spends upwards of $350 billion annually on imported oil, a drain of economic
resources that could be used to stimulate growth, create jobs, build infrastructure and promote social advances at home.
From a sustainability perspective, oil presents several challenges. First is the length of time over which the world's finite oil
reserves can continue to supply rising demand. Second is the impact on global warming and climate change that carbon emissions
from oil combustion will have, and third is the challenge of finding a sustainable replacement for oil for transportation. The first
challenge, how much oil is left and when its production will peak, was discussed in Module Sustainable Energy Systems - Chapter
Introduction. The bottom line is that, as Yogi Berra famously said, making predictions is difficult, especially about the future.
Although we know the general course of initial rise and ultimate fall that global oil production must take, we do not know with
confidence the time scale over which it will play out.
The uncertainty of the timing of the peak in global oil production encourages us to find other issues and motivations for dealing
with an inevitably unsustainable supply. A prime motivation is energy security, the threat that oil supplies could be interrupted by
any of several events including weather, natural disaster, terrorism and geopolitics. Much of the world feels these threats are good
reasons for concerted effort to find replacements for oil as our primary transportation fuel. A second motivation is the
environmental damage and accumulation of greenhouse gases in the atmosphere due to transportation emissions. Unlike electricity
generation, transportation emissions arise from millions of tiny sources, e.g. the tailpipes of cars and trucks and the exhaust of
trains and airplanes. The challenge of capturing and sequestering carbon dioxide from these distributed and moving sources is
dramatically greater than from the large fixed sources of power plants. A more achievable objective may be replacing oil as a
transportation fuel with biofuel that recycles naturally each year from tailpipes of cars to biofuel crops that do not compete with
food crops. Other options include replacing liquid fuels with electricity produced domestically, or increasing the efficiency of
vehicles by reducing their weight, regeneratively capturing braking energy, and improving engine efficiency. Each of these options
has promise and each must overcome challenges.
Changes in the energy system are inevitably slow, because of the time needed to develop new technologies and the operational
inertia of phasing out the infrastructure of an existing technology to make room for a successor. The transportation system exhibits
this operational inertia, governed by the turnover time for the fleet of vehicles, about 15 years. Although that time scale is long
compared to economic cycles, the profit horizon of corporations and the political horizon of elected officials, it is important to
begin now to identify and develop sustainable alternatives to oil as a transportation fuel. The timescale from innovation of new
approaches and materials to market deployment is typically 20 years or more, well matched to the operational inertia of the
transportation system. The challenge is to initiate innovative research and development for alternative transportation systems and
sustain it continuously until the alternatives are established.
Summary
Oil for transportation and electricity generation are the two biggest users of primary energy and producers of carbon emissions in
the United States. Transportation is almost completely dependent on oil and internal combustion engines for its energy. The
concentration of oil in a few regions of the world creates a transportation energy security issue. Unlike electricity generation
emissions, carbon emissions from transportation are difficult to capture because their sources, the tailpipes of vehicles, are many
and moving. The challenges of oil energy security and capturing the carbon emissions of vehicles motivate the search for an oil
replacement, such as biofuels, electricity or greater energy efficiency of vehicles.
8.4.10 [Link]
petroleum crude and in a few centuries for coal and natural gas. Peak Oil refers to the peak in oil production that must occur as
petroleum crude runs out. As shown in Figure 8.4.19, the main discoveries of crude oil occurred prior to 1980.
Peak Oil – The Growing Gap
Figure 8.4.19 Peak Oil – The Growing Gap Petroleum crude oil discoveries versus refined oil production. Source: Rep. Roscoe
Bartlett, Maryland
Since oil is getting harder and harder to find, we now have to obtain it from less accessible places such as far under the ocean,
which has led to hard-to-repair accidents such as the Deepwater Horizon oil spill in May, 2010. An additional effect is the higher
cost of refining the petroleum since it comes from more remote locations or in less desirable forms such as thick, rocky “tar sand”
or “oil sand” found in Canada or Venezuela. Overall, the use of petroleum crude cannot exceed the amount of petroleum that has
been discovered, and assuming that no major oil discoveries lie ahead, the production of oil from crude must start to decrease.
Some analysts think that this peak has already happened.
An additional aspect of oil scarcity is energy independence. The United States currently imports about two thirds of its petroleum,
making it dependent on the beneficence of countries that possess large amounts of oil. These countries are shown in Figure 8.4.20,
a world map rescaled with the area of each country proportional to its oil reserves. Middle Eastern countries are among those with
the highest oil reserves. With its economy and standard of living so based on imported petroleum crude it is easy to see why the
United States is deeply involved in Middle East politics. It should be noted that Figure 8.4.19 corresponds to the entire world and
even currently oil-rich countries such as Saudi Arabia will soon experience peak oil.
World According to Oil
Figure 8.4.19 The World According to Oil World map redrawn with country area proportional to oil resources. Source: Rep.
Roscoe Bartlett, Maryland
A second major motivation to move away from petroleum crude is global climate change. While the correlation of carbon dioxide
(CO2) concentration in the atmosphere to average global temperature is presently being debated, the rise of CO2 in our atmosphere
that has come from burning fossil fuel since the industrial revolution is from about 280 ppm to about 390 ppm at present, and
cannot be denied. Energy sources such as wind, solar, nuclear, and biomass are needed that minimize or eliminate the release of
atmospheric CO2. Biomass is included in this list since the carbon that makes up plant fiber is taken from the atmosphere in the
process of photosynthesis. Burning fuel derived from biomass releases the CO2 back into the atmosphere, where it can again be
incorporated into plant mass. The Energy Independence and Security Act (EISA) of 2007 defines an advanced biofuel as one that
lowers lifecycle greenhouse gas emissions (emissions from all processes involved in obtaining, refining, and finally burning the
fuel) by 60% relative to the baseline of 2005 petroleum crude.
Figure 8.4.20 Gas/Ethanol Fuel Pump A fuel pump in Brazil offering either ethanol alcohol (A) or gasoline (G). Source: Natecull
There, if the cost of alcohol (as it is known colloquially) is less than 70% than the cost of gasoline, tanks are filled with ethanol. If
the cost of alcohol is more than 70% of the cost of gasoline, people fill up with gasoline since there is about a 30% penalty in gas
mileage with ethanol. This comes about simply because the chemical structure of ethanol has less energy per volume (about 76,000
Btu/gallon or 5,100 kcal/liter) than gasoline (115 Btu/gallon or 7,600 kcal/liter) or diesel (132,000 Btu/gallon or 8,800 kcal/liter).
Cane ethanol qualifies, per EISA 2007, as an advanced biofuel.
In the United States, for a cost of about twice that of cane-derived ethanol, corn starch is saccharified and fermented into ethanol.
Ethanol is used predominantly as a high octane, oxygenated blend at 10% to improve the combustion in gasoline engines. The
distribution of ethanol as E85 flex fuel (85% ethanol and 15% gasoline) has faltered probably because the price, even with a 50
cents/gallon federal subsidy, does not make up for the 25 – 30% decrease in gas mileage (see Figure 8.4.21).
Mileage Comparisons
Figure 8.4.21 Mileage Comparisons Mileage comparison of gasoline and E85 flex fuel. Source: U.S. Department of Energy,
Energy Efficiency and Renewable Energy. Image created at [Link]
First generation biodiesel is made via the base catalyzed transesterification of plant oils such as soy and palm. The main
disadvantage with plant oil-based biofuels is the high cost of the plant oil, owing to the relatively little oil that can be produced per
acre of farmland compared to other biofuel sources. The problem with transesterification is that it produces a fuel relatively high in
8.4.11 [Link]
oxygen, which a) causes the biodiesel to become cloudy (partially freeze) at relatively high temperature, and makes the biodiesel b)
less stable, and c) less energy dense than petroleum-derived diesel.
Cane ethanol qualifies as an advanced biofuel, as its production lowers greenhouse gas emissions more than 60% relative to the
2005 petroleum baseline (per EISA 2007). Corn ethanol is far from this energy efficiency. However, ethanol made from
lignocellulose – the non-food part of plants - comes close, at a 50% reduction. This brings us to the second generation of biofuels.
Figure 8.4.22 Second Generation Biofuels Cellulosic ethanol and green diesel. Source: John Regalbuto
Collectively termed “lignocellulose,” this material consists of three main components; cellulose, hemicellulose, and lignin.
Chemical or biological pretreatments are required to separate the whole biomasss into these fractions. Hemicellulose and cellulose,
with the appropriate enzymes or inorganic acids, can be deconstructed into simple sugars and the sugars fermented into ethanol, or
with some newer strains of microbes, into butanol. Butanol has only 10% less energy density than gasoline. The lignin fraction of
biomass is the most resistant to deconstruction by biological or chemical means and is often burned for heat or power recovery.
At the same time attention turned toward cellulosic ethanol, petroleum refining companies set about to improve biodiesel. A
petroleum refining process called hydrotreating was used to upgrade plant oil. In this process, the oil is reacted with hydrogen in
the presence of inorganic catalysts, and the plant oil is converted into a much higher quality, oxygen-free “green diesel” and jet
fuel. This type of biofuel is in fact a “drop in replacement” to petroleum-derived diesel and jet fuel and passes all of the stringent
regulations demanded by the automobile and defense industries. It has been tested in a number of commercial and military aircraft.
Routes to Advanced Biofuels
Figure 8.4.23 Routes to Advanced Biofuels Various routes to drop-in replacement hydrocarbon biofuels. Source: John Regalbuto
The various routes to drop-in replacement hydrocarbon biofuels are shown in Figure Routes to Advanced Biofuels. On the left side
of the figure, feedstocks are ordered relative to their abundance and cost. The most abundant and, therefore, cheapest feedstock is
lignocellulose from sources such as agricultural residue, forest waste, and energy crops such as switch grass and short rotation
poplar trees. Of lesser abundance and higher expense are the sugars and starches – corn and sugar cane. The least abundant and
most expensive biofuels, lipid-based feedstocks from plant oil or animal fat, are shown at the bottom. Efforts are underway to mass
produce oil-laden algae. The oils harvested from algae are relatively easy to convert to hydrocarbon biofuels, by using processing
similar to hydrotreating. The main set of problems associated with algae lie in its mass production. Algal feedstocks are easy to
convert to hydrocarbons but algae itself is difficult to mass produce, whereas lignocellulose is very abundant but more difficult to
convert into hydrocarbons.
Two of the routes to hydrocarbon biofuels compete directly with fermentation of sugars to ethanol. The same sugars can be treated
with inorganic catalysts, via the blue liquid phase processing routes seen in the center of Figure 8.4.23, or with microbial routes to
yield hydrocarbons as the fermentation product (pink routes). Microbes are examples of biocatalysts; enzymes within the microbe
act in basically the same way that inorganic catalysts act in inorganic solutions. The field of research in which enzymes are
engineered to alter biological reaction pathways is called synthetic biology.
A flow sheet of an inorganic catalytic set of processes to hydrocarbon biofuels, from a leading biofuel startup company (Virent
Energy Systems of Madison, Wisconsin) is shown in Figure 8.4.24. Both of the biocatalytic and the inorganic catalytic processes
involve an intrinsic separation of the hydrocarbon product from water, which eliminates the energy intensive distillation step
needed for alcohol fuels. For the microbial route the added benefit of this self-separation is that the microbes are not poisoned by
the accumulation of product as occurs in fermentation to alcohol.
8.4.12 [Link]
Inorganic Catalytic Routes to Advanced Biofuels
Figure 8.4.24 Inorganic Catalytic Routes to Advanced Biofuels A flow sheet of an inorganic catalytic set of processes to
hydrocarbon biofuels, from a leading biofuel startup company, Virent Energy Systems. Source: Virent Energy Systems, figure 1
Two other main routes to hydrocarbon biofuels are seen in the upper section of Figure 8.4.23: gasification and pyrolysis. An
advantage of both of these routes is that they process whole biomass, including the energy-rich lignin fraction of it. Gasification
produces a mixture of carbon monoxide and hydrogen called synthesis gas, which can be converted to hydrocarbon fuels by a
number of currently commercialized catalytic routes including Fischer-Tropsch synthesis and methanol-to-gasoline. The challenge
with biomass is to make these processes economically viable at small scale. The second process is pyrolysis, which yields a crude-
like intermediate called pyrolysis oil or bio-oil. This intermediate must be further treated to remove oxygen; once this is done it can
be inserted into an existing petroleum refinery for further processing.
Summary
The motivations for hydrocarbon biofuels are energy independence and a reduction in greenhouse gas emissions. The first
renewable biofuels were biodiesel and bioethanol. With inorganic catalysis and synthetic biology, these have been supplanted with
drop-in replacement gasoline, diesel, and jet fuels. These can be made in the United States in a number of ways from presently
available, sustainably produced lignocellulosic feedstocks such as corn stover, wood chips, and switchgrass, and in the future, from
mass-produced algae. It is too early to tell which production method will prevail, if in fact one does. Some processes might end up
being particularly advantageous for a particular feedstock such as wood or switchgrass. What we do know is that something has to
be done; our supply of inexpensive, easily accessible oil is running out. Biofuels will be a big part of the country’s long-term
energy independence. A great deal of scientific and engineering research is currently underway; it’s an exciting time for biofuels.
Figure 8.4.26 Estimated Cooling Costs Comparison Estimated cooling costs of geothermal systems compared with conventional
systems. Source: Sohail Murad
8.4.13 [Link]
Types of Geothermal Systems
There are two major types of geothermal systems: in ground and pond systems. In ground geothermal systems can be vertical and
horizontal as shown in Figure 8.4.27. The excavation cost of vertical systems is generally higher and they require more land area
for installation, which is generally not an option in urban locations. Other than excavation costs, vertical and horizontal GHPs have
similar efficiencies since the ground temperature below the frost line is essentially constant.
daigram of a vertical closed loop system
Figure 8.4.27 In Ground Geothermal Systems Examples of horizontal and vertical ground systems. Source: U.S. Department of
Energy, Energy Efficiency and Renewable Energy
Pond geothermal systems are generally preferable if there is water available in the vicinity at almost constant temperature year
round. These systems are especially suited to industrial units (e.g. oil refineries) with water treatment facilities to treat processed
water before it is discharged. The temperature of treated water from these facilities is essentially constant throughout the year and
is an ideal location for a pond system. Pond geothermal systems are constructed with either open loops or closed loops (see Figure
8.4.28). Open loop systems actually remove water from the pond, while the close loop systems only remove energy in the form of
heat from the pond water. Of course, in open pond system this water is again returned to the pond, albeit at a lower temperature
when used for heating.
Diagram of an Open Pond Geothermal System
Figure 8.4.28 Pond Geothermal Systems Examples of closed and open loop pond systems. Source: U.S. Department of Energy,
Energy Efficiency and Renewable Energy
Figure 8.4.29 Return of Investment in Geothermal System Return of additional capital investment in a typical geothermal system.
Source: Murad, S., & Al-Hallaj, S. from Feasibility Study For a Hybrid Fuel Cell/Geothermal System, Final Report, HNTB
Corporation, August 2009.
8.4.14 [Link]
only a few hours. This then reduces the load on the geothermal system during peak hours when electricity cost is generally the
highest.
PCM Storage Tanks reduce the overall cost of the geothermal heat pump system significantly since it does not have to be designed
to address peak heating/cooling needs. In addition, it also shifts energy loads from peak hours to non-peak hours. Figure 8.4.30
shows temperature variations for a typical summer day in July 2010 in Chicago. The high temperature of 90 degree lasted only for
a short period of about 4 hours, and then returned to below 85 degrees rapidly. These relatively short temperature peaks can be
easily managed by PCMs.
Temperature Variation
Figure 8.4.30 Temperature Variation Temperature variation during a typical July day in Chicago. Source: Sohail Murad produced
figure using data from Great Lakes Environmental Research Laboratory
In conclusion, geothermal heat pumps are a very attractive, cost efficient sustainable energy source for both heating and cooling
with a minimal carbon print. It is a well-developed technology that can be easily incorporated into both residential and commercial
buildings at either the design stage or by retrofitting buildings.
References
1. Crabtree, G. & Misewich, J. (Co-Chairs). (2010). Integrating Renewable Electricity on the Grid, American Physical Society.
American Physical Society, Washington D.C. Retrieved August 12, 2011 from
[Link]
2. Owen, D. (2009). Green Metropolis: Why Living Smaller, Living Closer, And Driving Less Are the Keys to Sustainability. New
York: Riverhead Books.
3. U.S. Energy Information Administration. (2010). Annual Energy Review 2009. Retrieved August 12, 2011 from
[Link]
4. Lenton, T.M. (2006). Climate change to the end of the millennium. Climatic Change, 76, 7-29. doi: 10.1007/s10584-005-9022-
1
5. Olajire, A. (2010). CO2 capture and separation technologies for end-of-pipe applications: A review, Energy 35, pp. 2610-2628.
doi: 10.1016/[Link].2010.02.030
6. Osborn, S.G., Vengosh, A., Warner, N.R., & Jackson, R.B. (2011). Methane contamination of drinking water accompanying
gas-well drilling and hydraulic fracturing. PNAS, 108, pp. 8172-1876. doi: 10.1073/pnas.1100682108
7. Stephens, J.C. & Keith, D.W. (2008). Assessing geochemical carbon management. Climatic Change, 90, 217-242. doi:
10.1007/s10584-008-9440-y
8. U.S. Energy Information Administration. (2010). Annual Energy Review 2009. Retrieved August 12, 2011 from
[Link]
9. U.S. Energy Information Administration. (2011). Annual Energy Outlook 2011. Retrieved September 2, 2011 from
[Link]/forecasts/aeo/pdf/0383(2011).pdf
10. Ferguson, C.D.,Marburger, L.E. & Farmer, J.D. (2010) A US nuclear future? Nature, 467, 391-393. doi: 10.1038/467391a
11. Grimes, R.J. & Nuttall, W.J. (2010). Generating the Option of a two-stage nuclear renaissance. Science, 329, 799-803. doi:
10.1126/science.1188928
12. Kazimi, M., Moniz, E.J., & Forsberg, C. (2010) The Future of the Nuclear Fuel Cycle. MIT Energy Initiative. Retrieved May
30, 2011 from [Link]
13. World Nuclear Association (2011). Processing of Used Nuclear Fuel. Retrieved May 30, 2011 from [Link]
[Link]/info/[Link].
14. Attari, S.Z., DeKay, M.L., Davidson, C.I., & de Bruin, W.B. (2010). Public perceptions of energy consumption and savings.
PNAS, 107, 16054.
15. Black & Veatch (2007, October). Twenty percent wind energy penetration in the United States: A technical analysis of the
energy resource. Walnut Creek, CA: Black & Veatch Corp. Retrieved December 9, 2011 from
[Link]/Black_Veatch_20_Percent_Report.pdf
16. Clean Energy Progress Report, IEA (2011), [Link]/publications/free...p?PUBS_ID=2384
17. de Vries, B.J.M., van Vuuren, D.P., & Hoogwijk, M.M. (2007). Renewable energy sources: Their global potential for the first-
half of the 21st century at a global level: An integrated approach. Energy Policy, 35, 2590.
18. Jacobson, M.Z. & Delucchi, M.A. (2009). A plan for a sustainable future. Scientific American, 301, 58.
19. Kaldellis, J.K. & Zafirakis, D. (2011). The wind energy (r)evolution: A short review of a long history. Renewable Energy, 36,
1887-1901.
8.4.15 [Link]
20. Lu, X., McElroy, M.B., & Kiviluoma, J. (2009). Global potential for wind-generated electricity. PNAS, 106, 10933.
21. Renewables 2010 Global Status Report, REN21, Renewable Energy Policy Network for the 21st Century,
[Link]/REN21Activities...4/[Link]
22. Schwartz, M., Heimiller, D., Haymes, S., & Musial, W. (2010, June). Assessment of offshore wind energy resources for the
United States (NREL-TP-500-45889). Golden, CO: National Renewable Energy Laboratory. Retrieved December 9, 2011 from
[Link]/docs/fy10osti/[Link]
23. Trench, C.J. (n.d.). Oil Market Basics. [Link] Information Administration. Retrieved September 12, 2011 from
[Link]/pub/oil_gas/pe...cs/[Link]
Review Questions
1. Electricity is the fastest growing energy carrier in the world, trailed by liquid fuels for transportation. Why is electricity more
appealing than liquid fuels?
2. A primary challenge for the electricity grid is capacity to handle the "urban power bottleneck" in cities and suburbs. How can
superconducting cables address urban capacity issues?
3. Renewable wind and solar electricity is plentiful in the United States, but they are located remotely from high population
centers and their output is variable in time. How can these two issues be addressed?
4. The United States’ electricity supply is provided primarily by coal, natural gas, nuclear, and hydropower. How safe are these
fuel supplies from interruption by international disasters, weather events or geopolitical tension?
5. Natural gas reserves from shale are increasing rapidly due to increased use of hydrofracturing technology ("fracking"). The
increased domestic resource of shale gas has the potential to provide greater energy security at the expense of greater
environmental impact. What are the long-term costs, benefits, and outlook for tapping into domestic shale gas reserves?
6. Anthropogenic carbon emissions are small compared to natural exchange between ocean and atmosphere and fluxes from
vegetation and land use. Why do anthropogenic emissions have such a large effect on the concentration of carbon dioxide in the
atmosphere?
7. One proposal for mitigating carbon emissions is capturing and storing them in underground geologic formations (sequestration).
What scientific, technological and policy challenges must be overcome before sequestration can be deployed widely?
8. Nuclear electricity came on the scene remarkably quickly following the end of World War II, and its development stagnated
quickly following the Three Mile Island and Chernobyl accidents. The Fukushima disaster of 2011 adds a third cautionary note.
What conditions must be fulfilled if the world is to experience an expansion of nuclear electricity, often called a nuclear
renaissance?
9. Nuclear fuel can be used once and committed to storage or reprocessed after its initial use to recover unused nuclear fuel for re-
use. What are the arguments for and against reprocessing?
10. Storage of spent nuclear fuel for tens to hundreds of thousands of years is a major sustainability challenge for nuclear
electricity. Further development of the Yucca Mountain storage facility has been halted. What are some of the alternatives for
storing spent nuclear fuel going forward?
11. What events in the 1970s and late 1990s motivated the modern interest in renewable energy?
12. Renewable energy is often divided into solar, wind, hydropower, biomass, geothermal, wave and tide. What are the ultimate
sources of each of these renewable energies? What is the ultimate source of fossil fuel and why is it not classified as renewable?
13. Renewable energy has the technical potential to supply global electricity needs in 2050. What factors determine whether
renewable energy will actually be deployed to meet this need? How can unknowns, such as the rate of technological and
economic advances, the economic, intellectual and social connections among countries, and the relative importance of
environmental and social agendas be taken into account in determining the course of deployment of renewable energy?
14. Public acceptance is a key factor in the growth of renewable energy options. What is the public acceptance of various energy
options, and how might these change over the next few decades?
15. The almost exclusive dependence of the transportation system on liquid fuels makes oil an essential commodity for the orderly
operation of many societies. What are some alternatives to oil as a transportation fuel?
16. There are many reasons to reduce consumption of oil, including an ultimately finite supply, the high cost and lost economic
stimulus of payments to foreign producers, the threat of interruption of supply due to weather, natural disaster, terrorism or
geopolitical decisions, and the threat of climate change due to greenhouse gas emissions. Which of these reasons are the most
important? Will their relative importance change with time?
17. The transportation system changes slowly, governed by the lifetime of the fleet of vehicles. Compare the time required for
change in the transportation system with the timescale of economic cycles, the profit horizon of business, the political horizon
8.4.16 [Link]
of elected officials and the time required to develop new transportation technologies such as electric cars or biofuels. What
challenges do these time scales present for changing the transportation system?
18. What are the potential advantages of hydrocarbon biofuels over alcohol biofuels?
19. How could biofuels be used with other alternate energy forms to help the United States become energy independent?
20. On what principle does a geothermal heat pump work?
21. What makes it more cost efficient than electrical heating or conventional furnaces?
22. Are geothermal heat pumps suitable for moderate climates (e.g. Miami, FL)? Are conventional electrical or gas furnaces the
only choices in these areas?
Glossary
Energy Carrier
A medium, such as electricity, gasoline or hydrogen, that can move energy from one place to another, usually from the point of
production (e.g. an electrical generator or petroleum refinery) to the point of use (e.g. an electric light or motor or a gasoline
engine).
Biocatalysis
Catalysis conducted by enzymes – catalysis within the body, for example.
Energy Density
The amount of energy contained in a given volume (say a gas tank). The higher the energy density of a fuel, the farther the car will
go on a tank of the fuel.
Fermentation
The conversion of sugars into alcohols or hydrocarbons by microbes.
Fischer-Tropsch synthesis
The inorganic catalytic reaction between CO and H2 (synthesis gas), which produces diesel and jet fuel.
Gasification
The conversion of biomass at very high temperature (1000 – 1200°C) in an oxygen atmosphere that results in a “synthesis gas”
intermediate – a mixture of carbon monoxide (CO) and hydrogen (H2).
Hydrotreating
Reaction in the presence of hydrogen.
Infrastructure Compatible
Compatible with existing oil pipelines, storage tanks, petroleum refineries, and internal combustion engines.
Inorganic Catalysis
Solid, inorganic materials such as platinum nanoparticles deposited onto activated carbon, which accelerate the rate of chemical
reactions without being consumed in the process.
Lignocellulose
The non-food portion of plants such as the stalks and leaves of corn plants (corn stover).
Peak Oil
The peak in world oil production that must come about as oil consumption surpasses the discovery of new oil
Pyrolysis
The conversion of biomass at moderately high temperature (500 – 800°C) in an inert atmosphere that results in a “bio-oil”
intermediate.
Synthetic Biology
The field of biology in which microbes are engineered to control metabolic pathways.
Transesterification
8.4.17 [Link]
The base catalyzed reaction of plant oil with methanol with breaks the oil into long fatty acid chains, which can be used as a low
quality diesel fuel.
Geothermal Energy
Energy from the earth.
Heat Pump
A device that allows heat to be removed at a lower temperature and supplied at a higher temperature, for example an air
conditioner.
8.4: Energy Sources and Carriers is shared under a not declared license and was authored, remixed, and/or curated by LibreTexts.
8.4.18 [Link]
8.5: Energy Uses
Electrical and Plug-in Hybrids
Since the early 20th Century, oil and the internal combustion engine have dominated transportation. The fortunes of oil and
vehicles have been intertwined, with oil racing to meet the energy demands of the ever growing power and number of personal
vehicles, vehicles driving farther in response to growing interstate highway opportunities for long distance personal travel and
freight shipping, and greater personal mobility producing living patterns in far-flung suburbs that require oil and cars to function. In
recent and future years, the greatest transportation growth will be in developing countries where the need and the market for
transportation is growing rapidly. China has an emerging middle class that is larger than the entire population of the United States,
a sign that developing countries will soon direct or strongly influence the emergence of new technologies designed to serve their
needs. Beyond deploying new technologies, developing countries have a potentially large second advantage: they need not follow
the same development path through outdated intermediate technologies taken by the developed world. Leapfrogging directly to the
most advanced technologies avoids legacy infrastructures and long turnover times, allowing innovation and deployment on an
accelerated scale.
The internal combustion engine and the vehicles it powers have made enormous engineering strides in the past half century,
increasing efficiency, durability, comfort and adding such now-standard features as air conditioning, cruise control, hands-free cell
phone use, and global positioning systems. Simultaneously, the automobile industry has become global, dramatically increasing
competition, consumer choice and marketing reach. The most recent trend in transportation is dramatic swings in the price of oil,
the lifeblood of traditional vehicles powered with internal combustion engines.
Figure 8.5.1 Electric Transportation Transportation is electrified by replacing the gasoline engine with an electric motor, powered
by electricity from a battery on board the car (upper panel) or electricity from a fuel cell and hydrogen storage system on board the
car (lower panel). For maximum effectiveness, both routes require renewable production of electricity or hydrogen. Source: George
Crabtree using images from Rondol, skinnylawyer, Tinu Bao, U.S. Department of Energy, Office of Science
8.5.1 [Link]
large fleet of electric cars in the United States would require significant additional electricity, as much as 130 GW if the entire
passenger and light truck fleet were converted to electricity, or 30 percent of average United States electricity usage in 2008.
The energy usage of electric cars is about a factor of four less than for gasoline cars, consistent with the higher efficiency of electric
motors over internal combustion engines. Although gasoline cars vary significantly in their energy efficiency, a "typical" middle of
the road value for a five-passenger car is 80kWh/100km. A typical electric car (such as the Think Ox from Norway, the Chevy Volt
operating in its electric mode, or the Nissan Leaf) uses ~ 20 kWh/100km. While the energy cost of electric cars at the point of use
is significantly less, one must consider the cost at the point of production, the electricity generating plant. If the vehicle's electricity
comes from coal with a conversion efficiency of 33 percent, the primary energy cost is 60 kWh/100km, approaching but still
smaller than that of the gasoline car. If electricity is generated by combined cycle natural gas turbines with 60 percent efficiency,
the primary energy cost is 33 kWh/100km, less than half the primary energy cost for gasoline cars. These comparisons are
presented in Table 8.5.1.
Table 8.5.1 Comparisons of Energy UseComparison of energy use for gasoline driven and battery driven cars, for the cases of inefficient coal
generation (33%) and efficient combined cycle natural gas generation (60%) of electricity. Source: George Crabtree.
Battery Electric Nissan Leaf, Chevy Volt
Gasoline Engine 5 Passenger Car
(battery mode), Think Ox
Energy at Point of Use 80 kWh/100 km 20 kWh/100 km
Table 8.5.2 Comparisons of Carbon EmissionsComparison of carbon emissions from gasoline driven and battery driven cars, for the cases of high
emission coal generation (2.1 lb CO2/kWh), lower emission natural gas (1.3 lbCO2/kWh) and very low emission nuclear, hydro, wind or solar
electricity. Source: George Crabtree.
Battery Electric Nissan Leaf, Chevy Volt
Gasoline Engine 5 Passenger Car
(batter mode), Think Ox
CO2 Emissions at point of use 41 lbs ~0
The carbon footprint of electric cars requires a similar calculation. For coal-fired electricity producing 2.1 lb CO2/kWh, driving
100km produces 42 lbs (19 kgs) of carbon dioxide; for gas-fired electricity producing 1.3 lb CO2/kWh, 100km of driving produces
26 lbs (11.7 kgs) of carbon dioxide. If electricity is produced by nuclear or renewable energy such as wind, solar or hydroelectric,
no carbon dioxide is produced. For a "typical" gasoline car, 100km of driving produces 41 lbs (18.5 kgs) of carbon dioxide. Thus
the carbon footprint of a "typical" electric car is, at worst equal, to that of a gasoline car and, at best, zero. Table 8.5.3 summarizes
the carbon footprint comparisons.
8.5.2 [Link]
The battery in hybrid cars has two functions: it drives the electric motor and also collects electrical energy from regenerative
braking, converted from kinetic energy at the wheels by small generators. Regenerative braking is effective in start-stop driving,
increasing efficiency up to 20 percent. Unlike gasoline engines, electric motors use no energy while standing still; hybrids therefore
shut off the gasoline engine when the car comes to a stop to save the idling energy. Gasoline engines are notoriously inefficient at
low speeds (hence the need for low gear ratios), so the electric motor accelerates the hybrid to ~15 mph (24 kph) before the
gasoline engine restarts. Shutting the gasoline engine off while stopped increases efficiency as much as 17 percent.
The energy saving features of hybrids typically lower their energy requirements from 80 kWh/100km to 50-60 kWh/100km, a
significant savings. It is important to note, however, that despite a supplementary electric motor drive system, all of a hybrid's
energy comes from gasoline and none from the electricity grid.
The plug-in hybrid differs from conventional hybrids in tapping both gasoline and the electricity grid for its energy. Most plug-in
hybrids are designed to run on electricity first and on gasoline second; the gasoline engine kicks in only when the battery runs out.
The plug-in hybrid is thus an electric car with a supplemental gasoline engine, the opposite of the conventional hybrid cars
described above. The value of the plug-in hybrid is that it solves the "driving range anxiety" of the consumer: there are no worries
about getting home safely from a trip that turns out to be longer than expected. The disadvantage of the plug-in hybrid is the
additional supplemental gasoline engine technology, which adds cost and complexity to the automobile.
Summary
Electricity offers an attractive alternative to oil as a transportation fuel: it is domestically produced, uses energy more efficiently,
and, depending on the mode of electricity generation, can emit much less carbon. Electric vehicles can be powered by fuel cells
producing electricity from hydrogen, or from batteries charged from the electricity grid. The hydrogen option presents greater
technological challenges of fuel cell cost and durability and high capacity on-board hydrogen storage. The battery option is ready
for implementation in the nearer term but requires higher energy density batteries for extended driving range, and a fast charging or
battery swapping alternative to long battery charging times.
8.5.3 [Link]
industrial facilities with efficiencies of 80 percent and higher. Meaning, for every 100 units of fuel energy into the boiler/furnace,
we get about 80 units of useful thermal energy.
Commercial and industrial facilities that utilize the conventional energy system found in the United States (electricity supplied
from the electric grid and thermal energy produced on-site through the use of a boiler/furnace) will often times experience overall
fuel efficiencies of between 40 to 55 percent (actual efficiency depends on the facilities heat to power ratio).
Combined Heat and Power (CHP) is a form of distributed generation. It is an integrated system located at or near the
building/facility that generates utility grade electricity which satisfies at least a portion of the electrical load of the facility, and
captures and recycles the waste heat from the electric generating equipment to provide useful thermal energy to the facility.
Conventional CHP (also referred to as topping cycle CHP) utilizes a single dedicated fuel source to sequentially produce useful
electric and thermal power. Figure 8.5.2 provides a diagram of a typical topping cycle CHP system. A variety of fossil fuels,
renewable fuels, and waste products are utilized as input fuel to power a prime mover that generates mechanical shaft power
(exception is fuel cells). Prime movers might include reciprocating engines, gas turbines, steam turbines or fuel cells. The
mechanical shaft power is converted into utility grade electricity through a highly efficient generator. Since the CHP system is
located at or near the building/facility, the heat lost through the prime mover can be recycled through a heat exchanger and provide
heating, cooling (absorption chillers), and/or dehumidification (desiccants) to meet the thermal load of the building. These systems
can reach fuel use efficiencies of as high as 75 to 85 percent (versus the conventional energy system at approximately 40 to 55
percent).
Conventional (Topping Cycle) CHP
Figure 8.5.2 Conventional (Topping Cycle) CHP Diagram illustrates a typical topping cycle of CHP systems. Source: John Cuttica
In our example of 100 units of fuel into the CHP system, only 30 to 35 units of electricity are generated, but another 40 to 50 units
of the fuels energy can be recovered and utilized to produce thermal power. What this tells us is that for conventional CHP systems
to reach the high efficiency level, there must be a use for the recovered thermal energy. Thus a key factor for conventional CHP
systems is the coincidence of electric and thermal loads in the building. This is shown in Figure 8.5.3. The “Y” axis represents the
cost of generating electricity with a CHP system utilizing a 32 percent efficient reciprocating engine. The “X” axis represents the
cost of natural gas utilized to operate the CHP system and also the value of the natural gas being displaced if the recycled heat from
the engine can be utilized. The lines in the chart show various levels of recoverable heat available from the engine. If no heat is
recovered (no use for the thermal energy), the cost of generating electricity with the CHP system is $0.08/kWhr. When the full
amount of heat from the engine is recovered (full use of the thermal energy), the cost of generating electricity with the CHP system
then drops to $0.03/kWhr.
Importance of Waste Heat Recovery
Figure 8.5.3 Importance of Waste Heat Recovery Graph shows the importance of waste heat recovery in CHP systems. Source:
John Cuttica
Since the high efficiency of a CHP system is dependent on the effective use of the recoverable heat, CHP systems are often times
sized to meet the thermal load of the application and the amount of electricity produced is the by-product. The electricity is used to
off set the electricity otherwise purchased from the local electric utility. When the CHP system does not produce enough electricity
to satisfy the load, the utility supplies the difference from the grid. When the CHP system (sized from the thermal requirements)
produces more electricity than the load requires, the excess electricity can be sold to the local utility (normally at the avoided cost
of power to the utility).
There are three general modes of operation for CHP on-site generators relative to the electric utility grid:
Stand Alone (totally isolated from the grid)
Isolated from the grid with utility back-up (when needed)
Parallel operation with the grid
The preferred mode of operation is parallel with the grid. Both the on-site CHP system and the utility grid power the facility
simultaneously. With a proper sizing and configuration of the CHP system, the parallel mode of operation provides the most
flexibility. Should the grid go down, the CHP system can keep operating (e.g. during the 2003 Northeast Blackout and the 2005
Hurricane Katrina), and should the CHP system go down, the utility grid can supply power to the load. Overall reliability of power
to the load is increased.
The basic components of a conventional (topping cycle) CHP system are:
Prime Mover that generates mechanical shaft energy
8.5.4 [Link]
Reciprocating engine
Turbines (gas, micro, steam)
Fuel Cell (fuel cells ustilize an electrochemical process rather than a mechanical shaft process)
Generator converts the mechanical shaft energy into electrical energy
Synchronous generator (provides most flexibility and independence from the grid)
Induction generator (grid goes down - the CHP system stops operating)
Inverter (used mainly on fuel cells - converts DC power to utility grade AC power)
Waste Heat Recovery is one or more heat exchangers that capture and recycle the heat from the prime mover
Thermal Utilization equipment converts the recycled heat into useful heating, cooling (absorption chillers) and/or
dehumidification (deisiccant dehumidifiers)
Operating Control Systems insure the CHP components function properly together
Figure shows the cost of CO2 reduction technologies. Source: Oak Ridge National Laboratory (2008), p. 13, and McKinsey &
Company, “Reducing U.S. Greenhouse Gas Emissions: How Much at What Cost?,” December, 2007
CHP Applications
Today there are more than 3,500 CHP installations in the United States, totaling more than 85,000 MW of electric generation. That
represents approximately 9 percent of the total electric generation capacity in the United States. The 85,000 MW of installed CHP
reduces energy consumption by 1.9 Quads (1015 Btus) annually and eliminates approximately 248 million metric tons (MMT) of
CO2 annually.
CHP systems are generally more attractive for applications that have one or more of the following characteristics:
Good coincidence between electric and thermal loads
Maximum cost differential between electricity cost from the local utility and the cost of the fuel utilized in the CHP system
(referred to as spark spread)
Long operating hours (normally more than 3,000 hours annually)
Need for good power quality and reliability
The following are just a few of the type applications where CHP makes sense:
Hospitals
Colleges and Universities
High Schools
Fitness Centers
Office Buildings
Hotels
Data Centers
Prisons
Pulp and Paper Mills
Chemical Manufacturing Plants
8.5.5 [Link]
Metal Fabrication Facilities
Glass Manufacturers
Ethanol Plants
Food Processing Plants
Waste Water Treatment Facilities
Livestock Farms
CHP Benefits
CHP is not the only solution to our energy problems. In fact, CHP is not the most cost effective solution in all applications or in all
areas of the country. There are many variables that determine the viability of CHP installations. However, when the technical and
financial requirements of the application are met, a well designed, installed and operated CHP system provides benefits for the
facility owner (end user), the electric utility, and society in general. The high efficiency attained by the CHP system provides the
end user with lower overall energy costs, improved electric reliability, improved electric power quality, and improved energy
security. In areas where the electric utility distribution grid is in need of expansion and/or upgrades, CHP systems can provide the
electric utility with a means of deferring costly modifications to the grid. Although the electricity generated on-site by the end user
displaces the electricity purchased from the local electric utility and is seen as lost revenue by many utilities, energy efficiency and
lower utility costs are in the best interest of the utility customer and should be considered as a reasonable customer option by
forward-looking customer oriented utilities. Finally, society in general benefits from the high efficiencies realized by CHP systems.
The high efficiencies translate to less air pollutants (lower greenhouse gas and NOx emissions) than produced from central station
electric power plants.
Figure 8.5.5 Waste Heat to Power (Bottoming Cycle) CHP Diagram illustrates a waste heat to power (bottoming cycle) CHP
system. Source: John Cuttica
Summary
Combined Heat and Power (CHP) represents a proven and effective near-term alternative energy option that can enhance energy
efficiency, ensure environmental quality, and promote economic growth. The concept of generating electricity on-site allows one to
capture and recycle the waste heat from the prime mover providing fuel use efficiencies as high as 75 to 85 percent. Like other
forms of alternative energy, CHP should be considered and included in any portfolio of energy options.
References
1. Crabtree, G.W., Dresselhaus, M.S., & Buchanan, M.V. (2004). The Hydrogen Economy, Physics Today, 57, 39-45. Retrieved
September 2, 2011 from [Link]/depositi/t...en_economy.pdf
2. Crabtree, G.W. & Dresselhaus, M.S. (2008). The Hydrogen Fuel Alternative. MRS Bulletin,33, 421-428. Retrieved September
2, 2011 from [Link]/~w...ergy/[Link]
3. Doucette, R.T. & McCulloch, M.D. (2011). Modeling the CO2 emissions from battery electric vehicles given the power
generation mixes of different countries. Energy Policy, 39, 803-811. doi: 10.1016/[Link].2010.10.054
4. Eberle, U. & Helmolt, R.V. (2010). Sustainable transportation based on electric vehicle concepts: a brief overview. Energy and
Environmental Science, 3, 689-699. doi: 10.1039/C001674H
5. Oak Ridge National Laboratory. (2008). Combined Heat and Power, Effective Energy Solutions for a Sustainable Future.
Retrieved September 26, 2011 from [Link]/industry...port_12-[Link]
Questions
1. Transportation relies almost exclusively for its fuel on oil, whose price fluctuates significantly in response to global geopolitics
and whose long-term availability is limited. What are the motivations for each of the stakeholders, including citizens,
8.5.6 [Link]
companies and governments, to find alternatives to oil as a transportation fuel?
2. Electricity can replace oil as a transportation fuel in two ways: by on board production in a hydrogen fuel cell, and by on board
storage in a battery. What research and development, infrastructure and production challenges must be overcome for each of
these electrification options to be widely deployed?
3. Electric- and gasoline-driven cars each use energy and emit carbon dioxide. Which is more sustainable?
4. How do gasoline-driven, battery-driven and hybrid cars (like the Prius) compare for (i) energy efficiency, (ii) carbon emissions,
and (iii) reducing dependence on foreign oil?
5. What drives the system efficiency in a conventional CHP system?
6. To ensure high system efficiency, how would you size a conventional CHP system?
7. What is the preferred method of operating a CHP system that provides the most flexibility with the utility grid?
8. Why are CHP systems considered one of the most cost-effective CO2 abatement practices?
9. Name at least three application characteristics that make CHP an attractive choice.
Glossary
Hybrid Vehicle
A car that contains two drive systems, one based on the internal combustion engine and one on the electric motor. Conventional
hybrids, such as the Toyota Prius, use the electric motor only when high power is needed: starting from a stop, passing, and going
uphill. The electricity to run the motor is generated on board by an alternator powered by the internal combustion engine and by
regenerative breaking. Plug-in hybrids such as the Chevy Volt, in contrast, use the electric motor as the main drive for the car,
relying on the gasoline engine only when the battery is low or empty.
Internal Combustion Engine
The engine that converts the chemical energy of gasoline into the mechanical energy of motion, by exploding small amounts of fuel
in the confined space of fixed cylinder containing a moving piston. A precise amount of fuel must be metered in, and a spark
created at a precise moment in the piston's journey to produce the maximum explosive force to drive the piston. The internal
combustion engine is an engineering marvel (the word engineering celebrates it) perfected over more than a century. In contrast,
the electric motor is much simpler, more efficient and less expensive for the same power output.
Point of Production
The first (or at least an early) step in the energy chain, where the energy that ultimately will perform a function at the point of use is
put into its working form. For gasoline-driven cars, this is the refinery where gasoline is produced from crude oil, for battery-driven
cars this is the power generation plant were electricity is produced. Gasoline is then delivered to the pump and finally to the car,
where it is converted (the point of use) to mechanical motion by the engine. Similarly, electricity is delivered to the battery of an
electric car by the grid, and converted by the electric motor of the car (the point of use) to mechanical motion.
Point of Use
The last step in the energy chain, where energy accomplishes its intended function. For vehicles, this is the conversion of chemical
energy in gasoline cars or electric energy in battery cars to motion of the wheels that moves the car along the road.
Absorption Chiller
Utilizes heat instead of mechanical energy to provide cooling. A thermal compressor (fueled by the waste heat from the CHP
system) is used in place of an electrically powered mechanical compressor in the refrigeration process.
Avoided Cost of Power
The marginal cost for a utility to produce one more unit of power.
Combined Heat and Power (CHP)
An integrated system, located at or near the building or facility, that generates utility grade electricity which satisfies at least a
portion of the electrical load of the facility and captures/ recycles the waste heat from the electric generating equipment to provide
useful thermal energy to the facility.
Conventional CHP (Topping Cycle CHP)
8.5.7 [Link]
Utilizes a single dedicated fuel source to sequentially produce useful electric and thermal power.
Desiccant Dehumidification
Process that removes moisture (latent load) from a building air stream by passing the air over a desiccant wheel (normally a silica
gel). The recovered heat from a CHP system is utilized to regenerate the desiccant by driving the moisture off the desiccant wheel
to the outside.
Fuel Cell
An exothermic electrochemical reaction that combines hydrogen and oxygen ions through an electrolyte material to generate
electricity (DC) and heat.
Gas Turbine
An internal-combustion engine consisting essentially of an air compressor, combustion chamber, and turbine wheel that is turned
by the expanding products of combustion.
Induction Generator
Converts the mechanical shaft power from the CHP prime mover to utility grade Alternating Current Power. An induction
generator can only operate when connected to an external reactive power source (normally provided by the utility grid).
Inverter
Converts Direct Current electric power into utility grade Alternating Current electric power. Normally used with fuel cell systems.
Organic Rankine Cycle (ORC)
Uses an organic, high molecular mass fluid with a liquid-vapor phase change or boiling point occurring at a lower temperature than
the water-steam phase change. The fluid allows rankine cycle heat recovery from lower temperature sources where the heat is
converted into useful work, which can then be converted into electricity.
Prime Mover
The term utilized to denote the CHP system equipment that converts input fuel into mechanical shaft power (reciprocating engine,
gas turbine, steam turbine, micro-turbine).
Reciprocating Engine
A heat engine that uses one or more reciprocating pistons to convert pressure into mechanical rotating shaft power.
Steam Turbine
Utilizes the Rankine Cycle to extract heat from steam and transform the heat into mechanical shaft power by expanding the steam
from high pressure to low pressure through the turbine blades.
Synchronous Generator
Converts the mechanical shaft power from the CHP prime mover to utility grade Alternating Current Power. A synchronous
generator is self-exciting (contains its own source of reactive power) and can operate independent of, or isolated from, the utility
grid.
Waste Heat to Power (Bottoming Cycle CHP)
Captures the waste heat generated by an industrial or commercial process, utilizing the waste heat as the free fuel source for
generating electricity.
Energy Density
The energy contained in a volume or mass divided by the volume or mass it occupies. High energy density materials pack a large
energy into a small space or mass; low energy density materials require more space or mass to store the same amount of energy.
The electrical energy of batteries is at the low end of the energy density scale, the chemical energy of gasoline is at the high end,
approximately a factor of 30-50 larger than batteries.
8.5: Energy Uses is shared under a not declared license and was authored, remixed, and/or curated by LibreTexts.
8.5.8 [Link]
8.6: Applications of Phase Change Materials for Sustainable Energy
The growing demand for sustainable energy from consumers and industry is constantly changing. The highest demand of energy
consumption during a single day brings a continuous and unsolved problem: how to maintain a consistent desired temperature in a
sustainable way. Periods of extreme cold or warm weather are the triggering factors for increasing the demand on heating or
cooling. Working hours, industry processes, building construction, operating policies, and type and volume of energy production
facilities are some of the main reasons for peak demand crises. Better power generation management and significant economic
benefit can be achieved if some of the peak load could be shifted to the off peak load period. This can be achieved by thermal
storage for space heating and cooling purposes.
Thermal energy can be stored as a change in the internal energy of certain materials as sensible heat, latent heat or both. The most
commonly used method of thermal energy storage is the sensible heat method, although phase change materials (PCM), which
effectively store and release latent heat energy, have been studied for more than 30 years. Latent heat storage can be more efficient
than sensible heat storage because it requires a smaller temperature difference between the storage and releasing functions. Phase
change materials are an important and underused option for developing new energy storage devices, which are as important as
developing new sources of renewable energy. The use of phase change material in developing and constructing sustainable energy
systems is crucial to the efficiency of these systems because of PCM’s ability to harness heat and cooling energies in an effective
and sustainable way.
liquid, for example, has kinetic energy of the motion of atoms that is not present in the solid, so its energy is higher. The higher
energy of the liquid compared to the solid is the latent heat. When the solid is fully transformed to liquid, added energy reverts to
going into sensible heat and raising the temperature of the liquid.
Temperature Profile of a PCM
Figure shows the temperature profile of a PCM. In the region where latent heat is effective, the temperature keeps either constant or
in a narrow range. The phase of the material turns from one to another and both phases appears in the medium. Source: Said Al-
Hallaj & Riza Kizilel
A PCM is a substance with a high latent heat (also called the heat of fusion if the phase change is from solid to liquid) which is
capable of storing and releasing large amounts of energy at a certain temperature. A PCM stores heat in the form of latent heat of
fusion which is about 100 times more than the sensible heat. For example, latent heat of fusion of water is about 334kJ/kg whereas
sensible heat at 25° Celsius (77°F) is about 4.18kJ/kg. PCM will then release thermal energy at a freezing point during
solidification process (Figure 8.6.2). Two widely used PCMs by many of us are water and wax. Think how water requires
significant amount of energy when it changes from solid phase to liquid phase at 0°C (32°F) or how wax extends the burning time
of a candle. Moreover, the cycle of the melting and solidification can be repeated many times.
Phase Change of a PCM
Figure represents the phase change of a PCM when the heat is applied or removed. Source: Said Al-Hallaj & Riza Kizilel
There are large numbers of PCMs that melt and solidify at a wide range of temperatures, making them attractive in a number of
applications in the development of the energy storage systems. Materials that have been studied during the last 40 years include
hydrated salts, paraffin waxes, fatty acids and eutectics of organic and non-organic compounds (Figure 8.6.3). Therefore, the
selection of a PCM with a suitable phase transition temperature should be part of the design of a thermal storage system. It should
be good at heat transfer and have high latent heat of transition. The melting temperature should lie in the range of the operation, be
chemically stable, low in cost, non-corrosive and nontoxic.
Energy Storage Systems
Figure shows materials commonly studied for use in PCMs due to their ability to melt and solidify at a wide range of temperatures.
Source: Said Al-Hallaj & Riza Kizilel
Even though the list of the PCMs is quite long, only a limited number of the chemicals are possible candidates for energy
applications due to the various limitations of the processes. Paraffins and hydrated salts are the two most promising PCMs.
8.6.1 [Link]
Generally, paraffins have lower fusion energy than salt hydrates but do not have the reversibility issue, i.e paraffin is only in
physical changes and keeps its composition when heat is released or gained whereas hydrated salt is in chemical change when heat
is released or gained. Therefore, a major problem with salt hydrates is incongruent melting, which reduces the reversibility of the
phase change process. This also results in a reduction of the heat storage capacity of the salt hydrate. On the other hand, paraffins
also have a major drawback compared to salt hydrates. The low thermal conductivity creates a major drawback which decreases the
rates of heat stored and released during the melting and crystallization processes and hence results in limited applications. The
thermal conductivity of paraffin used as PCM is slightly above 0.20 W/mK (compare with ice; kice=∼2 W/mK). Several methods
such as finned tubes with different configurations and metal matrices filled with PCM have been investigated to enhance the heat
transfer rate of PCM. Novel composite materials of PCM, which have superior properties, have also been proposed for various
applications. For example, when PCM is embedded inside a graphite matrix, the heat conductivity can be considerably increased
without much reduction in energy storage.
Applications of PCMs
The three applications of PCMs listed below (solar energy, buildings, and vehicles) are only a small portion of the many areas
where they can be used (catering, telecom shelters, electronics, etc.). The applications of PCMs in these areas have been widely
studied in order to minimize the greenhouse effect and to minimize the need for foreign gasoline which costs U.S. economy
millions of dollars every year.
Increasing concerns of the impact of fossil fuels on the environment and their increasing cost has led to studies on thermal energy
storage for the space heating and cooling of buildings. Extreme cold or warm weather increases the demand on heating or cooling.
If the thermal energy of heat or coolness is stored and then provided during the day or night, part of the peak loads can be shifted to
off-peak hours. Therefore, an effective energy management and economic benefit can be achieved.
Solar energy is recognized as one of the most promising alternative energy resource options. However, it is intermittent by nature:
there is no sun at night. The reliability of solar energy can be increased by storing it when in excess of the load and using the stored
energy whenever needed.
The minimization of heat loss or gain through walls, ceilings, and floors has been studied for a long time and PCM applications
have been considered for more than 30 years to minimize these losses/gains, and thus reduce the cost of electricity or natural gas
use in buildings. Studies on viability of PCMs in vehicle applications are also growing widely. Denaturation of food during
transport brings a major problem which is being partially solved by refrigerated trucks. However, this causes not only more
expensive foods, but also irreversible environmental effects on living organisms.
systems using PCMs have been good candidates for thermal energy storage and have been applied since 1980s. At first, the water
heaters were supported by filling the bottom of the heaters with PCMs, which was a first step in storing energy in heating systems.
However, the quantity of the available energy in the storage system was limited by low thermal conductivity of the PCM.
Improvements on thermal storage systems and developments in the incorporation of PCMs that utilize the solar energy have been
extensively studied since then.
Solar Heater
Figure shows solar heating system with and without PCM. Source: Said Al-Hallaj & Riza Kizilel
Later studies have mainly concentrated on increasing thermal conductivity using composite materials. Adding PCM modules at the
top of the water tank gives the system a higher storage density and compensate for heat loss in the top layer because of the latent
heat of PCM. The configuration of the PCM storage unit can result in advantageous control of the water temperature rise and drop
during both day and night time. Therefore, thermally stratified water tanks are widely used for short term thermal energy storage.
Application of these tanks significantly increases not only the energy density with the number of PCM modules, but also the
cooling period and keeps the water temperature higher compared to the ones without PCMs. Besides, solar water heating systems
operate within a wide range of temperatures from ambient temperatures to 80°C (176°F). A PCM has much larger heat storage
capacity relative to water over a narrow temperature range, close to its melting temperature.
8.6.2 [Link]
A major component of total household energy consumption is cooking. Solar energy offers an economical option for cooking in
households, especially in third world countries. A solar cooker is a device which uses the energy of sunlight to heat food or drink to
cook or sterilize it (Figure 8.6.5). It uses no fuel, costs nothing to operate, and reduces air pollution. A solar cooker’s reflective
surface concentrates the light into a small cooking area and turns the light into heat. It is important to trap the heat in the cooker
because heat may be easily lost by convection and radiation. The feasibility of using a phase change material as the storage medium
in solar cookers have been examined since 1995. A box-type solar cooker with stearic acid based PCM has been designed and
fabricated by Buddhi and Sahoo (1997), showing that it is possible to cook food even in the evening with a solar cooker. The rate of
heat transfer from the PCM to the cooking pot during the discharging mode of the PCM is quite slow and more time is required for
cooking food in the evening. Fins that are welded at the inner wall of the PCM container were used to enhance the rate of heat
transfer between the PCM and the inner wall of the PCM container. Since the PCM surrounds the cooking vessel, the rate of heat
transfer between the PCM and the food is higher and the cooking time is shorter. It is remarkable that if food is loaded into the solar
cooker before 3:30 p.m. during the winter season, it could be cooked. However, the melting temperature of the PCM should be
selected carefully. The more the input solar radiation, the larger quantity of heat there is in a PCM. Few examples for PCMs for
solar cooker applications are acetamide (melting point of 82 °C), acetanilide (melting point of 118 °C), erythritol (melting point of
118 °C) and magnesium nitrate hexahydrate (melting point of 89–90 °C).
Solar Cooker
Figure 8.6.5 Solar Cooker. Photograph shows solar heating system. Source: Atlascuisinesolaire via Wikimedia Commons.
Building Applications
PCMs can be used for temperature regulation, heat or cold storage with high storage density, and thermal comfort in buildings that
require a narrow range of temperature (Figure 8.6.6). Therefore, if the solar energy is stored effectively, it can be utilized for night
cold. The use of PCMs brings an opportunity to meet the demand for heating. It helps to store the energy which is available during
daytime and to keep the temperature of the building in the comfort level.
Typical Application of PCM in Buildings
Figure illustrates a typical application of PCM in buildings. Heat storage and delivery occur over a fairly narrow temperature range.
Wallboards containing PCM have a large heat transfer area that supports large heat transfer between the wall and the space. Source:
Said Al-Hallaj & Riza Kizilel
Energy storage in the walls or other components of the building may be enhanced by encapsulating PCM within the surfaces of the
building. The latent heat capacity of the PCM is used to capture solar energy or man-made heat or cold directly and decrease the
temperature swings in the building. It also maintains the temperature closer to the desired temperature throughout the day.
Researchers have proposed macro or micro level encapsulated PCM in concrete, gypsum wallboard, ceiling and floor in order to
achieve a reasonably constant temperature range.
Today, it is possible to improve the thermal comfort and reduce the energy consumption of buildings without substantial increase in
the weight of the construction materials by the application of micro and macro encapsulated PCM. The maximum and minimum
peak temperatures can be reduced by the use of small quantities of PCM, either mixed with the construction material or attached as
a thin layer to the walls and roofs of a building. In addition, the energy consumption can also be reduced by absorbing part of the
incident solar energy and delaying/reducing the external heat load.
The absorption of heat gains and the release of heat at night by a paraffin wax-based PCMs encapsulated within a co-polymer and
sandwiched between two metal sheets (PCM board) have been used in some building materials. The PCM boards on a wall reduce
the interior wall surface temperature during the charging process, whereas the PCM wall surface temperature is higher than the
other walls during the heat releasing process. The heat flux density of a PCM wall in the melting zone is almost twice as large as
that of an ordinary wall. Also, the heat-insulation performance of a PCM wall is better than that of an ordinary wall during the
charging process, while during the heat discharging process, the PCM wall releases more heat energy.
Unlike structural insulated panels, which exhibit fairly uniform thermal characteristics, a PCM’s attributes vary depending upon
environmental factors. The structural insulated panel works at all times, resisting thermal flow from hot temperatures to colder
temperatures. The thermal flux is directly proportional to the temperature difference across the structural insulated panel insulation.
The usefulness of PCM is seen when the in-wall temperatures are such that it causes the PCM to change state. It can be inferred
that the greater the temperature difference between day and night, the better the PCM works to reduce heat flux. The use of a phase
change material structural insulated panel wall would be excellent for geographic areas where there is typically a large temperature
swing, warm during the day and cool at night.
8.6.3 [Link]
Vehicle Applications
Studies on viability of PCM in vehicle applications are growing widely. For example, PCMs are studied with regard to refrigerated
trucks, which are designed to carry perishable freight at specific temperatures. Refrigerated trucks are regulated by small
refrigeration units that are placed outside the vehicle in order to keep the inside of the truck trailer at a constant temperature and
relative humidity. They operate by burning gas, hence the cost of shipment is highly affected by the changes of temperature in the
trailer. The use of PCM has helped in lowering peak heat transfer rates and total heat flows into a refrigerated trailer. Ahmed,
Meade, and Medina (2010) modified the conventional method of insulation of the refrigerated truck trailer by using paraffin-based
PCMs in the standard trailer walls as a heat transfer reduction technology. An average reduction in peak heat transfer rate of 29.1
percent was observed when all walls (south, east, north, west, and top) were considered, whereas the peak heat transfer rate was
reduced in the range of 11.3 - 43.8 percent for individual walls. Overall average daily heat flow reductions into the refrigerated
compartment of 16.3 percent were observed. These results could potentially translate into energy savings, pollution abatement from
diesel-burning refrigeration units, refrigeration equipment size reduction, and extended equipment operational life.
Vehicles are mainly powered by gasoline (i.e gas or petrol). Liquified petroleum gases and diesel are other types of fluids used in
vehicles. Lately, hybrid vehicles became popular among consumers as they significantly reduce the toxic exhaust gases if the
vehicles run in electric mode. Li-ion batteries have been used in electronic devices for a long time (cell-phones, laptops, and
portable devices). Many scientists, especially in the United States, have been working on the possibility of using Li-ion batteries for
transportation applications in order to double the fuel efficiency and reduce emissions of hybrid vehicles. Li-ion battery modules
can be connected in order to meet the nominal voltage of the vehicle to run the vehicle in the electric mode. However this brings a
huge problem which keeps away the uses of Li-ion batteries in many applications: as a result of exothermic electrochemical
reactions, Li-ion batteries release energy during discharge. The generated energy should be transferred from the body of the battery
to environment. If the rate of the transfer is not sufficient, some of the gelled phase materials turn into gas phase and increase the
internal pressure of the cell. Therefore the energy should be released from the cell as soon as possible or the temperature of the cell
should not lead to an increase. Sveum, Kizilel, Khader, and Al-Hallaj (2007) have shown that Li-ion batteries with thermal
management using PCM eliminate the need for additional cooling systems and improve available power (Figure 8.6.7). The
researchers maintained battery packs at an optimum temperature with proper thermal management and the PCM was capable of
removing large quantities of heat due to its high latent heat of fusion.
Application with PCM Technology
Figure 8.6.7 Application with PCM Technology. A pack of Li-ion batteries kept at a narrow temperature range with a proper use of
passive thermal management system. Source: AllCell’s PCM Technology©
Summary
There is a great interest in saving energy and in the use of renewable energies. PCMs provide an underused option for developing
new energy storage devices in order to minimize greenhouse effects. They operate at constant temperature; as heat is added to the
material, the temperature remains stable, but the heat drives the change to a higher energy phase. A PCM stores heat in the form of
latent heat of fusion which is about 100 times more than the sensible heat. Hydrated salts, paraffin waxes, fatty acids and eutectics
of organic and non-organic compounds are the major types of PCMs that melt at a wide range of temperatures. The specific melting
point of the PCM determines the design of thermal storage system.
In this module, applications of PCM in solar energy, buildings, and vehicles were reviewed. Solar heaters have been popular since
1960s and PCMs have been used to store the precious energy from sun since 1980s. They have been used extensively in solar
cookers, especially in the third world countries in order to decrease the thermal related costs. The cookers do not use fuel and hence
reduce air pollution.
PCM can be used for temperature regulation in order to minimize the heat loss or gain through building walls. They have been used
to capture solar heat and decrease the temperature fluctuations in buildings. Moreover, since a small amount of PCM is sufficient in
order to store solar energy, thermal comfort is achieved without substantial increase in the weight of the construction materials.
Application of PCMs in transportation is growing widely. Today, refrigerated trucks are regulated by refrigeration units, but the use
of PCMs is a viable option to prevent the denaturation of food during transportation. The transfer rate of heat can be reduced
significantly with PCMs. Moreover, PCM makes Li-ion batteries, which have high energy density, viable for high-power
applications. The generated energy during discharge or drive mode can be transferred from the body of the battery to environment
with the help of PCMs. Battery packs can be maintained at an optimum temperature with proper thermal management and the PCM
has been shown to be capable of removing large quantities of heat due to its high latent heat of fusion.
8.6.4 [Link]
Even though there is a lot of on-going research on effective and efficient applications of PCMs in a variety of areas (e.g. solar
cookers, buildings, vehicles), PCMs have yet to become a widely used technology for sustainable energy. The advantages of PCMs
are hardly known by many people and, therefore, the applications of PCMs and their benefits should be offered to consumers. The
sun is out there, continuously transferring its energy for free, but we need to do more to harness that sustainable energy for our own
needs.
References
1. Ahmed, M., Meade, O., & Medina, M. A. (2010, March). Reducing heat transfer across the insulated walls of refrigerated truck
trailers by the application of phase change materials. Energy Conversion and Management, 51, 383-392. doi:
10.1016/[Link].2009.09.003
2. Buddhi, D. & Sahoo, L. K. (1997, March). Solar cooker with latent heat storage: Design and experimental testing. Energy
Conversion and Management, 38, 493-498. doi: 10.1016/S0196-8904(96)00066-0
3. Sveum, P., Kizilel, R., Khader, M., & Al-Hallaj, S. (2007, September). IIT Plug-in Conversion Project with the City of Chicago.
Paper presented at the Vehicle Power and Propulsion Conference, Arlington, TX. doi: 10.1109/VPPC.2007.4544174
Review Questions
1. Explain briefly how phase change materials work.
2. What is the main disadvantage of the paraffin wax as a phase change material?
3. Name three different areas in the sustainable energy field in which PCMs are a key element in balancing heating and cooling.
Glossary
Ambient Temperature
The temperature of the surrounding environment.
Denaturation
A process in which proteins or nucleic acids lose their tertiary structure and secondary structure by application of heat.
Diesel
Any liquid fuel used in diesel engines.
Eutectics
A combination of two or more compounds of either organic, inorganic or both which may have a different melting point to their
individual and separate compounds.
Finned Tube
Tube with an extending part on a surface to facilitate cooling
Gas Phase
One of the three classical states of matter.
Gasoline
A toxic translucent, petroleum-derived liquid that is primarily used as a fuel in internal combustion engines. The term "gasoline" is
often shortened in colloquial usage to gas. Under normal ambient conditions its material state is liquid, unlike liquefied petroleum
gas or "natural gas."
Graphite Matrix
Composite material with graphite being a metal (see metal matrices).
Heat of Fusion
The amount of heat required to convert a unit mass of a solid at its melting point into a liquid without an increase in temperature.
Hydrated Salt
A solid compound containing water molecules combined in a definite ratio as an integral part of a crystal.
Latent Heat
8.6.5 [Link]
The heat which flows to or from a material without a change to temperature.
Li-ion Battery
A type of rechargeable battery in which lithium ions move from the negative electrode to the positive electrode during discharge
and from the positive electrode to negative electrode during charge.
Liquified Petroleum Gas
A flammable mixture of hydrocarbon gases used as a fuel in heating appliances and vehicles.
Metal Matrices
Composite material with at least two constituent parts, one being a metal.
Nominal Voltage
Voltage of a fully charged cell or battery when delivering maximum amount of energy that can be withdrawn from a battery at a
specific discharge rate.
Phase Change Material
A material that stores heat in the form of latent heat of fusion.
Paraffin
A white, odorless, tasteless, waxy solid to store heat with a specific heat capacity of 2.14–2.9 J g−1 K−1 and a heat of fusion of 200–
300 J g−1.
Sensible Heat
The heat energy stored in a substance as a result of an increase in its temperature.
Solar Energy
The sun’s radiation that reaches the earth.
8.6: Applications of Phase Change Materials for Sustainable Energy is shared under a not declared license and was authored, remixed, and/or
curated by LibreTexts.
8.6.6 [Link]
CHAPTER OVERVIEW
This page titled 9: Problem-Solving, Metrics, and Tools for Sustainability is shared under a CC BY license and was authored, remixed, and/or
curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
1
9.1: Problem-Solving, Metrics, and Tools for Sustainability - Chapter Introduction
“What gets measured gets done” is an oft-quoted saying (attributed to many individuals) that attempts to capture the essential role
of quantification in order to understand a system, solve a problem, advance a cause, or establish a policy. Throughout this text a
wide variety of measurements are put forth, cited, and discussed in connection with particular concepts including climate change,
economics, social well-being, engineering efficiency, and consumer habits. This chapter is devoted to a special collection of
methods, measurements, tools, indicators, and indices that are used to assess the comparative sustainability among potential and
often competing options, designs, or decisions, and to measure progress toward achieving the goals of sustainability over time.
The chapter begins in the Module Life Cycle Assessment with a brief discussion of industrial ecology, an emerging science that
focuses on understanding material and energy flows to and through different kinds of human-created systems. This kind of
understanding is essential for framing problems that need to be solved in a holistic way. Industrial ecologists study such topics as
recycling and reuse of materials, energy efficiency, organizational structures, supply chains, the social impacts of decisions, and the
economics of product development. It has been termed “the science of sustainability” (Graedel, 2000).
One of the principal tools of industrial ecology which is discussed in this chapter is life cycle assessment (LCA), a comprehensive
set of procedures for quantifying the impacts associated with the energy and resources needed to make and deliver a product or
service. LCA’s are carried out for two main reasons: (a) to analyze all the steps in a product chain and see which use the greatest
amount of energy and materials or produce the most waste, and (b) to enable comparisons among alternative products or supply
chains and to see which one create the least environmental impact. Inherent in the concept of LCA is the notion of tradeoffs – the
recognition that in a finite world choosing one product, pathway, or way of living has consequences for environmental and social
well-being. Of course choices must be made, but the goal of quantifying the implications of our actions as holistically as possible is
to avoid consequences that are “unintended.”
Although life cycle assessment grew out of the needs of industry to better design products and understand the implications of their
decisions, the systemic manner of framing problems upon which LCA is based has permeated a wide variety of fields, stimulating
what might be termed “life cycle thinking” in each of them. The Subcollection Derivative Life Cycle Concepts in this chapter
contains modules devoted to presentations of a number of ways of expressing the impacts of humans on the environment. These are
derived from life cycle principles and are drawn from the fields of ecology, thermodynamics, and environmental science. They
include “footprinting” and several sustainability indicators, all of which quantify human impacts in terms of resource consumption
and waste production over an extended geographic range and/or over timeframes that go beyond the immediate. A case study on
the UN Millennium Development Goals Indicator presents a comprehensive approach for assessing not only environmental
sustainability, but also hunger and poverty, education, gender equity, infant mortality, maternal health, disease, and global
partnerships – all elements of sustainable development made clear in the Brundtland report. Finally, this chapter concludes with a
module about sustainability and business.
This page titled 9.1: Problem-Solving, Metrics, and Tools for Sustainability - Chapter Introduction is shared under a CC BY license and was
authored, remixed, and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
9.1.1 [Link]
9.2: Life Cycle Assessment
Introduction
“What gets measured gets done” is an oft-quoted saying (attributed to many individuals) that attempts to capture the essential role
of quantification in order to understand a system, solve a problem, advance a cause, or establish a policy. Throughout this text a
wide variety of measurements are put forth, cited, and discussed in connection with particular concepts including climate change,
economics, social well-being, engineering efficiency, and consumer habits. This chapter is devoted to a special collection of
methods, measurements, tools, indicators, and indices that are used to assess the comparative sustainability among potential and
often competing options, designs, or decisions, and to measure progress toward achieving the goals of sustainability over time.
The chapter begins in the Module 11.2 with a brief discussion of industrial ecology, an emerging science that focuses on
understanding material and energy flows to and through different kinds of human-created systems. This kind of understanding is
essential for framing problems that need to be solved in a holistic way. Industrial ecologists study such topics as recycling and
reuse of materials, energy efficiency, organizational structures, supply chains, the social impacts of decisions, and the economics of
product development. It has been termed “the science of sustainability” (Graedel, 2000).
One of the principal tools of industrial ecology which is discussed in this chapter is life cycle assessment (LCA), a comprehensive
set of procedures for quantifying the impacts associated with the energy and resources needed to make and deliver a product or
service. LCA’s are carried out for two main reasons: (a) to analyze all the steps in a product chain and see which use the greatest
amount of energy and materials or produce the most waste, and (b) to enable comparisons among alternative products or supply
chains and to see which one create the least environmental impact. Inherent in the concept of LCA is the notion of tradeoffs – the
recognition that in a finite world choosing one product, pathway, or way of living has consequences for environmental and social
well-being. Of course choices must be made, but the goal of quantifying the implications of our actions as holistically as possible is
to avoid consequences that are “unintended.”
Although life cycle assessment grew out of the needs of industry to better design products and understand the implications of their
decisions, the systemic manner of framing problems upon which LCA is based has permeated a wide variety of fields, stimulating
what might be termed “life cycle thinking” in each of them. The Subcollection Derivative Life Cycle Concepts in this chapter
contains modules devoted to presentations of a number of ways of expressing the impacts of humans on the environment. These are
derived from life cycle principles and are drawn from the fields of ecology, thermodynamics, and environmental science. They
include "footprinting” and several sustainability indicators, all of which quantify human impacts in terms of resource consumption
and waste production over an extended geographic range and/or over timeframes that go beyond the immediate. Module 11.3.6
presents a comprehensive approach for assessing not only environmental sustainability, but also hunger and poverty, education,
gender equity, infant mortality, maternal health, disease, and global partnerships – all elements of sustainable development made
clear in the Brundtland report. Finally, this chapter concludes with a module about sustainability and business.
9.2.1 [Link]
capabilities to create artificial fertilizer and to convert biomass into agricultural products (see Module [Link]). And, after all, such
a venture is both “domestic” and “natural” – attributes that incline many, initially at least, to be favorably disposed. However upon
closer examination this direction is not quite as unequivocally positive as we might have thought. Yes it is possible to convert grain
into ethanol and plant oils into diesel fuel, but the great majority of these resources have historically been used to feed Americans
and the animals that they consume (and not just Americans; the United States is the world’s largest exporter of agricultural
products). As demand has increased, the prices for many agricultural products have risen, meaning that some fraction of the world’s
poor can no longer afford as much food. More marginal lands (which are better used for other crops, grazing, or other uses) have
been brought under cultivation for fermentable grains, and there have been parallel “indirect” consequences globally – as the world
price of agricultural commodities has risen, other countries have begun diverting land from existing uses to crops as well.
Furthermore, agricultural runoff from artificial fertilizers has contributed to over 400 regional episodes of hypoxia in estuaries
around the world, including the U.S. Gulf Coast and Chesapeake Bay.
In response to such problems, U.S. Congress passed the Energy Independence and Security Act in 2007, which limits the amount of
grain that can be converted into biofuels in favor of using agriculturally-derived cellulose, the chief constituent of the cell walls of
plants. This has given rise to a large scientific and technological research and development program to devise economical ways to
process cellulosic materials into ethanol, and parallel efforts to investigate new cellulosic cropping systems that include, for
example, native grasses. Thus, the seemingly simple decision to grow our biofuels industry in response to a political objective has
had unintended political, financial, dietary, social, land use, environmental quality, and technological consequences.
With hindsight, the multiple impacts of biofuels have become clear, and there is always the hope that we can learn from examples
like this. But we might also ask if there is a way to foresee all or at least some of these impacts in advance, and adjust our designs,
processes, and policies to take them into account and make more informed decisions, not just for biofuels but also for complex
societal problems of a similar nature. This approach is the realm of the field of industrial ecology, and the basis for the tool of life
cycle assessment (LCA), a methodology that has been designed to perform holistic analyses of complex systems.
Industrial Ecology
Many systems designed by humans focus on maximizing profitability for the firm, business or corporation. In most cases this
means increasing production to meet demand for the products or services being delivered. An unfortunate byproduct of this is the
creation of large amounts of waste, many of which have significant impacts if they enter the environment. Figure 9.2.1 is a general-
purpose diagram of a typical manufacturing process, showing the inputs of materials and energy, the manufacturing of products,
and the generation of wastes (the contents of the “manufacturing box” are generic and not meant to depict any particular industry—
it could be a mine, a factory, a power plant, a city, or even a university). What many find surprising is the large disparity between
the amounts of waste produced and the quantity of product delivered. Table 9.2.1 provides such information, in the form of waste-
to-product ratios, for a few common industries.
Human-Designed Industry
Paper 10/1
Chemicals 0.1-100/1
9.2.2 [Link]
Modern Agriculture ~4/1
In 1989, Robert Frosch & Nicholas Gallopoulos, who worked in the General Motors Research Laboratory, published an important
analysis of this problem in Scientific American (Frosch and Gallopoulos, 1989). Their paper was entitled “Strategies for
Manufacturing”; in it they posed a critical question: Why is it that human-designed manufacturing systems are so wasteful, but
systems in nature produce little, if any, waste? Although there had been many studies on ways to minimize or prevent wastes, this
was the first to seek a systemic understanding of what was fundamentally different about human systems in distinction to natural
systems. The paper is widely credited with spawning the new field of Industrial Ecology, an applied science that studies material
and energy flows through industrial systems. Industrial Ecology is concerned with such things as closing material loops (recycling
and reuse), process and energy efficiency, organizational behavior, system costs, and social impacts of goods and services. A
principle tool of Industrial Ecology is life cycle assessment.
9.2.3 [Link]
Well-to-wheel: a special type of LCA involving the application of fuel cycles to transportation vehicles.
Embodied energy: A cradle-to-gate analysis of the life cycle energy of a product, inclusive of the latent energy in the materials,
the energy used during material acquisition, and the energy used in manufacturing intermediate and final products. Embodied
energy is sometimes referred to as “emergy”, or the cumulative energy demand (CED) of a product or service.
LCA Methodology
Over time the methodology for conducting Life Cycle Analyses (LCAs) has been refined and standardized; it is generally described
as taking place in four steps: scoping, inventory, impact assessment, and interpretation. The first three of these are consecutive,
while the interpretation step is an ongoing process that takes place throughout the methodology. Figure 9.2.2 illustrates these in a
general way.
General Framework for Life Cycle Assessment
Figure 9.2.2 General Framework for Life Cycle Assessment The four steps of life cycle assessment and their relationship to one
another. Source: Mr3641 via Wikipedia
Scoping
Scoping is arguably the most important step for conducting an LCA. It is here that the rationale for carrying out the assessment is
made explicit, where the boundaries of the system are defined, where the data quantity, quality, and sources are specified, and
where any assumptions that underlie the LCA are stated. This is critically important both for the quality of the resultant analysis,
and for comparison among LCAs for competing or alternative products.
Inventory Analysis
The inventory analysis step involves the collection of information on the use of energy and various materials used to make a
product or service at each part of the manufacturing process. If it is true that scoping is the most important step in an LCA then the
inventory is probably the most tedious since it involves locating, acquiring, and evaluating the quality of data and specifying the
sources of uncertainties that may have arisen. For products that have been produced for a long time and for which manufacturing
processes are well known, such as making steel, concrete, paper, most plastics, and many machines, data are readily available. But
for newer products that are either under development or under patent protection, data are often considered proprietary and are
generally not shared in open sources. Uncertainty can arise because of missing or poorly documented data, errors in measurement,
or natural variations caused by external factors (e.g., weather patterns can cause considerable variation in the outputs of agricultural
systems or the ways that consumers use products and services can cause variability in the emission of pollutants and the disposition
of the product at end of life). Often the manufacturing chain of a process involves many steps resulting in a detailed inventory
analysis. Figure 9.2.3, for example, shows the manufacturing flow for a bar of soap (this diagram is for making bar soap using
saponification—the hydrolysis of triglycerides using animal fats and lye). The inventory requires material and energy inputs and
outputs for each of these steps, although it may turn out that some steps contribute little to the ultimate impact analysis. For
example, the inventory associated with capital equipment for a manufacturing process, i.e. machines that are replaced at lengthy
intervals such that their impacts in the short term are minimal, are often omitted from the analysis.
There are two additional aspects of LCA that should also be addressed during inventory analysis: the functional unit of comparison,
and the allocation of inventory quantities among co-products or services. The functional unit is the basis for comparing two or
more products, processes, or services that assure equality of the function delivered. This may seem like a straightforward task. For
example, for the soap produced by the process of Figure 9.2.3, one might choose “one bar of soap” as a functional unit of
comparison. But then how would a LCA comparison be made with, say, liquid hand soap or a body wash product (which combines
the functionality of soap and shampoo)? Perhaps “number of washings” would be a better choice, or maybe concentration of
surfactant made available per average use (in the latter case an “average dose” would need to be defined). Furthermore, soaps have
other additives and attributes such as scents, lotions, colors, and even the functionality of the shape – factors that may not affect
cleaning effectiveness but certainly do have an impact on consumer preferences, and hence quantity sold. Since it is quite likely
that essentially all soaps purchased by consumers will eventually be washed down the drain, such marketability factors may indeed
have an environmental impact.
Inventory data are virtually always sought for a total supply-manufacturing-consumer-use chain rather than individual products,
thus when that same chain produces multiple products it is necessary to allocate the materials, energy, and wastes among them.
Again, referring to Figure 9.2.3, there are potentially several co-products produced: tallow and other animal products, forest
products, cardboard and paper, and salable scrap. There are generally three ways to allocate materials and energy among co-
products: mass, volume, and economic value. Mass and volume allocations are the most straightforward, but may not capture
9.2.4 [Link]
market forces that are important in bringing materials into the environment. Allocation via economic valuation usually reflects the
value of the energy and any “value added” to the raw materials, but may miss the impacts of the materials themselves. In addition,
market values may fluctuate over time. In the final analysis the important aspect of any allocation procedure is that it be fully
documented.
Detailed System Flow Diagram for Bar Soap
Figure 9.2.3 Detailed System Flow Diagram for Bar Soap The manufacturing flow for a bar of soap (this diagram is for making bar
soap using saponification—the hydrolysis of triglycerides using animal fats and lye). Source: (U.S. Environmental Protection
Agency, 2006)
Impact Assessment
The life cycle impact assessment (LCIA) takes the inventory data on material resources used, energy consumed, and wastes emitted
by the system and estimates potential impacts on the environment. At first glance, given that an inventory may include thousands of
substances, it may seem that the number of potential impacts is bewilderingly large, but the problem is made more tractable
through the application of a system of impact classifications within which various inventory quantities can be grouped as having
similar consequences on human health or the environment. Sometimes inventoried quantities in a common impact category
originate in different parts of the life cycle and often possess very different chemical/biological/physical characteristics. The LCIA
groups emissions based on their common impacts rather than on their chemical or physical properties, choosing a reference
material for which health impacts are well known, as a basic unit of comparison. A key aspect is the conversion of impacts of
various substances into the reference unit. This is done using characterization factors, some of which are well-known, such as
global warming potential and ozone depletion potential, and LC50 (the concentration of a substance at which fifty percent of an
exposed population is killed), and others are still under development. Table 9.2.2 presents several impact categories that are
frequently used in the LCIA along with their references. The categories listed in Table 9.2.2 are not exhaustive – new types of
impact categories, such as land use and social impacts – and continue to be developed.
Table 9.2.2 Common Impact Categories and Their References: Several impact categories that are frequently used in the LCIA along with their
references. Source: An example will help to illustrate the type of information that results from T. Theis adapted from (U.S. Environmental
Protection Agency, 2006)
Human Health (cancer) Kg Benzene eq/unit
Ecotoxicity, Aquatic, Terrestrial Toxicity Kg 2,4 D eq/unit LC50 eq from exposure modeling
Acidification Kg H+/unit
An example will help to illustrate the type of information that results from life cycle inventory and impact assessments. In this case,
a system that produces a biologically-derived plastic, polylactide, is examined. (PLA). PLA has been proposed as a more
sustainable alternative to plastics produced from petroleum because it is made from plant materials, in this case corn, yet has
properties that are similar to plastics made from petroleum. Figure 9.2.4 shows a schematic of the system, which is a cradle-to-gate
assessment. As with any plastic, PLA can be turned into a variety of final products and each will have different cradle-to-grave
LCA characteristics. The production of PLA involves growing corn, harvesting and processing the grain, and polymerizing the
lactic acid molecules produced from fermentation. At each step a variety of chemicals and energy are used or produced. It is these
production materials that contribute to the impact analysis. Inventory quantities were allocated among major bio-products on a
mass basis.
Processing Diagram for Making Polylactide (PLA)
Figure 9.2.4 Processing Diagram for Making Polylactide (PLA) The production of PLA involves growing corn, harvesting and
processing the grain, and polymerizing the lactic acid molecules produced from fermentation. At each step a variety of chemicals
and energy are used or produced. It is these production materials that contribute to the impact analysis. Source: Landis, A.E. (2007)
Among the inventory data acquired in this case is life cycle fossil fuel used by the system, mostly to power farming equipment
(“Agriculture”), wet-mill corn (“CWM”), heat fermentation vats (“Ferment”), and Polymerization (“Polym”). The transport of
9.2.5 [Link]
intermediate products from sources to the processing center is also included. Figure 9.2.5 shows the fossil fuel used to make PLA
compared with fossil fuel used for making several petroleum-based plastics. Figure 9.2.6 shows the global warming potential
impact analysis. As might have been expected, the fossil fuels used to make PLA are slightly less than for the petroleum-derived
plastics on an equal mass basis (the functional unit is one kilogram of plastic). The PLA inventory also shows the sources of fossil
fuel used for each step along the manufacturing chain, with the fermentation step being the most intense user. What may not be
obvious is that the total greenhouse gases (GHG) emitted from the process, on an equivalent carbon dioxide (CO2) basis, are
generally higher for the biopolymer in comparison with the petroleum polymers in spite of the lower fossil fuel usage. When the
data are examined closely this is due to the agricultural step, which consumes generates relatively little fossil fuel, but is
responsible for a disproportionate amount of emissions of GHGs, mostly in the form of nitrous oxide, a powerful greenhouse gas
(310 times the global warming potential of CO2) that is a by-product of fertilizer application to fields. This example also illustrates
counter-intuitive results that LCAs often generate, a principal reason why it is important to conduct them.
Fossil Fuel Used to Make PLA vs. Petroleum-Based Products
Figure 9.2.5 Fossil Fuel Use to Make PLA vs. Petroleum-Based Plastics The amount of fossil fuels used when making PLA is
slightly less in comparison to making several petroleum-based products. (note: PS-GPPS – General Purpose Polystyrene; HDPE –
High Density Polyethylene; PET – Polyethylene Terephthalate; LDPE – Low Density Polyethylene; PP – Polypropylene). Source:
Landis, A.E. using data from: PLA-L, PLA-L2, (Landis, A.E., 2007); PLA-V, (Vink, et al., 2003); PLA-B, (Bohlmann, 2004);
PLA-P, (Patel, et al., 2006).
Figure 9.2.6 Global Warming Potential Impact Analysis Global warming impact for PLA compared with several other
petroleum- derived plastics. Source: Landis, A.E. (2007)
Interpretation of LCA
The interpretation step of LCA occurs throughout the analysis. As noted above, issues related to the rationale for conducting the
LCA, defining the system and setting its boundaries, identifying data needs, sources, and quality, and choosing functional units,
allocation procedures, and appropriate impact categories must all be addressed as the LCA unfolds. There are essentially two
formal reasons for conducting an LCA: (a) identification of “hot spots” where material and/or energy use and waste emissions, both
quantity and type, are greatest so that efforts can be focused on improving the product chain; and (b) comparison of results between
and among other LCAs in order to gain insight into the preferable product, service, process, or pathway. In both cases, there are
cautions that apply to the interpretation of results.
Assumptions
Typically a variety of assumption must be made in order to carry out the LCA. Sometimes these are minor, for example, exclusion
of elements of the study that clearly have no appreciable impact on the results, and sometimes more critical, for example choosing
one set of system boundaries over another. These must be explicitly stated, and final results should be interpreted in light of
assumptions made.
Incommunesurability
Sometimes LCA impact categories, such as those shown in Table 9.2.1, overlap in the sense that the same pollutant may contribute
to more than one category. For instance, if a given assessment comes up with high scores for both aquatic toxicity and human
toxicity from, say, pesticide use then one might be justified in using both of these categories to draw conclusions and make choices
based on LCA results. However, more typically elevated scores are found for categories that are not directly comparable. For
instance, the extraction, refining, and use of petroleum generate a high score for global warming (due to GHG release), while the
product chain for the biofuel ethanol has a high score for eutrophication (due to nitrogen release during the farming stage). Which
problem is worse – climate change or coastal hypoxia? Society may well choose a course of action that favors one direction over
another, but in this case the main value of the LCA is to identify the tradeoffs and inform us of the consequences, not tell us which
course is “correct.”
9.2.6 [Link]
Risk Evaluation and Regulation
One of the inherent limits to LCA is its use for assessing risk. Risk assessment and management, as described in the Modules 4.1
and 9.1, is a formal process that quantifies risks for a known population in a specific location exposed to a specific chemical for a
defined period of time. It generates risk values in terms of the probability of a known consequence due to a sequence of events that
are directly comparable, and upon which decisions on water, land, and air quality standards and their violation can be and are made.
LCA is a method for evaluating the impacts of wastes on human health and the environment from the point of view of the
product/service chain rather than a particular population. It can be used to identify the sources of contamination and general
impacts on the environment – a sort of “where to look” guide for regulation, but its direct use in the environmental regulatory
process has been, to date, rather limited. One application for LCA that has been suggested for regulatory use is for assessing the
impacts of biofuel mandates on land use practices, in the United States and other regions, however no regulatory standards for land
use have yet been proposed.
The Greenhouse Gases, Regulated Emissions, and Energy Use in Transportation Model (GREET)
GREET is a spreadsheet-based database developed by Argonne National Laboratory that links energy use to emissions on a life
cycle basis. Early versions were limited to greenhouse gases, but as the model has been refined many other types of contaminants
have been added. Although it has been widely used for comparing transportation and fuel options (hence its title), GREET has been
used for many other applications that have a significant energy component, including agriculture, material and product
development, and strategies for recycling.
SimaPro
SimaPro was developed by PRé Consultants in the Netherlands. It is a process-based tool for analyzing products and systems for
their energy usage and environmental impacts over their life cycle. It contains a number of databases for simulating processes,
performing inventories, assembling products and systems, analyzing results, and assessing life cycle impacts, and features modules
for performing uncertainty and sensitivity analyses.
Tool for the Reduction and Assessment of Chemical and Other Environmental Impacts
TRACI is a tool for performing life cycle impact analyses developed by the U.S. Environmental Protection Agency. It uses
inventory data as input information to perform a “mid-point” impact analysis using categories such as those shown in Table 9.2.1.
A mid-point analysis assesses impact based upon results at a common point in the risk chain, for example, global warming
potential, because subsequent end-point impact assessments require several assumptions and value choices that often differ from
case to case. The values for the various impact categories given in Table 9.2.1 are mid-point references.
9.2.7 [Link]
Of course, at this level of aggregation much information is lost, especially on how the system actually functions. For example, the
“energy” sector of the economy includes electricity generated, but doesn’t distinguish among nuclear, fossil, or renewable sources.
And if one is concerned with the functional reasons for a particular result, EIO-LCA will be of limited use. Often the “bottom-up”
and EIO-LCA approaches are combined (a “hybrid” approach).
Conclusions
The life cycle approach is a useful way to come to an understanding of the material and energy needed to make a product or deliver
a service, see where wastes are generated, and estimate the subsequent impacts that these wastes may have on the environment. It is
a good way to improve a product chain, articulate tradeoffs, and make comparisons among alternative processes and products. In
these contexts LCA facilitates decision making by managers, designers, and other stakeholders. Most importantly, LCA is a way of
framing policy options in a comprehensive and systematic way.
References
1. Bohlmann, G. M. (2004). Biodegradable packaging life cycle assessment. Environmental Progress, 23(4), 49-78. doi:
10.1002/ep.10053
2. Frosch, R. & Gallopoulos, N. (1989). Strategies for Manufacturing. Scientific American, 261(3), 144-152.
3. Landis, A. E. (2007). Environmental and Economic Impacts of Biobased Production. Unpublished doctoral dissertation,
University of Illinois at Chicago.
4. Patel, M., Crank, M., Dornburg, V., Hermann, B., Roes, L., Huesling, B., et al. (2006, September). Medium and long term
opportunities and risks of the biotechnological production of bulk chemicals from renewable resources – The potential of white
biotechnology: The BREW Project. Utrecht University, Netherlands: European Commission’s GROWTH Programme (DG
Research).
5. U.S. Environmental Protection Agency. (2006). Life cycle assessment: Principles and practice. (EPA Publication No.
EPA/600/R-06/060). Systems Analysis Branch, National Risk Management Research Laboratory. Cincinnati, Ohio.
[Link]/nrmrl/lcaccess/pdfs/[Link].
6. Vink, E. T. H., Rábagno, K.R., Glassner, D.A., & Gruber, P.R. (2003). Applications of life cycle assessment to NatureWorks
polylactide (PLA) production. Polymer Degradation and Stability 80(3), 403-419. doi: 10.1016/S0141-3910(02)00372-5
Review Questions
1. Using the information in 9.2.1, fill in numerical values, per unit of product, for the diagram in Figure 9.2.1. One diagram for
each industrial sector.
2. What are some of the reasons to use Life Cycle Assessments?
3. What are the basic stages of a product or service chain that serve as the basis for a life cycle assessment?
4. What are the steps involved in performing a life cycle assessment?
5. Name several characteristic scopes of life cycle assessments.
6. What is “embodied energy”?
7. Name several impact assessment categories and the reference units typically used to express them.
8. Name several life cycle impact analysis tools and their major characteristics.
9. What are some of the limitations of life cycle assessments?
10. Locate and read a completed Life Cycle Assessment online. Consider whether widespread adoption by society would result in
measureable lowering of environmental impacts? If so what kind? What might the obstacles be? Are there any tradeoffs
associated with adoption, i.e. some impacts may be reduced, but others might get worse?)
Glossary
allocation
For a chain which produces multiple products or services, the partitioning of inventory quantities among these co-products or co-
services.
economic input ouput life cycle assessment (EIO-LCA)
An aggregated approach to LCA in which the environmental impacts of a product or service are determined through an analysis of
the complete economy.
functional unit
9.2.8 [Link]
The basis for comparing two or more products, processes, or services that assures equality of the function delivered.
industrial ecology
An applied science that is concerned with material and energy flows through industrial systems.
life cycle assessment (LCA)
A method for quantifying the materials and energy needed to make or deliver a product or service that assesses the wastes produced
and potential environmental impacts across all or a part of the product chain.
life cycle impact assessment (LCIA)
The stage of an LCA in which the environmental impacts associated with the manufacture and delivery use and disposal of a
product are calculated.
life cycle inventory (LCI)
The stage of an LCA in which information on the use of energy and various materials used to make a product or service at each
part of the manufacturing process is collected.
scoping
The stage of a LCA in which the rationale for carrying out the assessment is made explicit, where the boundaries of the system are
defined, where the data quantity, quality, and sources are specified, and where any assumptions that underlie the LCA are stated.
This page titled 9.2: Life Cycle Assessment is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto Cabezas
(GALILEO Open Learning Materials) .
9.2.9 [Link]
9.3: Derivative Life Cycle Concepts
Sustainability Metrics and Rating Systems
Introduction
The ideal method to measure sustainability would reflect the three-legged stool paradigm – environmental protection, social equity, and
economic benefit. The metrics must make the connection between what the indicators measure and actual sustainability. A useful indicator
will reflect changes over time that show whether a system is becoming more or less sustainable, and generally substitutes for something else
or represents several measures (Sahely, 2005). The challenge of studying sustainability as an objective science is that the work is value-
loaded and socially charged. If we are aware of the purpose of the analysis we can use a multidisciplinary approach to the problem definition
and the research methodology (Lele and Norgaard, 1996).
Information Pyramid
Figure 9.3.1 Information Pyramid The Information Pyramid shows ways of handling data when studying sustainability. Source: C. Klein-
Banai.
In general, three approaches to sustainability measurement and reporting are commonly utilized: accounts that use quantitative data and
convert them to a common unit such as money, area or energy; narrative assessments that include text, maps, graphics and tabular data; and
indicator-based systems that may include the information that a narrative assessment has but they are organized around indicators or
measurable parts of a system. Indicator-based systems are generally found to perform better and are easily measurable and comparable since
they are more objective than narrative systems, or use only individual data points (Dalal-Clayton, 2002). Decision-makers and stakeholders
need to participate in the development of indicators to be sure that their values and concerns are addressed. However, the system does need to
be technically and scientifically based.
In the next few modules we will briefly discuss existing sustainability metrics that are generally based within certain disciplines such as
ecology, economics, and physics, and how they may reflect other disciplines (see Table 9.3.1). Most of these metrics are described in greater
details in the following modules: The IPAT Equation, Biodiversity, Species Loss, and Ecosystem Function, Tragedy of the Commons,
Environmental Valuation, Evaluating Projects and Policies, and Life Cycle Assessment.
Table 9.3.1 Common Sustainability MetricsTable lists common sustainability metrics. Source: C. Klein-Banai
9.3.1 [Link]
The maximum amount of resource
Carrying capacity: Maximum extraction while not depleting the
Ecosystem
sustainable yield (MSY) & IPAT resource from one harvest to the
next
Ecological Measures
Ecological measures of sustainability are used for natural systems. These measures include resilience and several constructs that are
derivatives from carrying capacity. Resilience is the time needed for a system that provides desirable ecosystem goods and services to go
back to a defined dynamic regime after disturbance. Resilience stresses the changing nature of ecosystems, rather than seeing them as static
and providing a continuous and constant amount of natural resources. Carrying capacity estimates society’s total use of the resource stocks
and flows provided by an ecosystem relative to the remaining resources needed by the ecosystem for stability and regeneration. Maximum
sustainable yield (MSY) is an outgrowth of carrying capacity and the goal is to reach the maximum amount of resource extraction while not
depleting the resource from one harvest to the next. Sustainability, in this context, can be understood as the point when the rate of resource
extraction or harvest (MSY) equals the amount produced by the ecosystem. Previously discussed methods are types of measures of
sustainability such as IPAT (see Module The IPAT Equation) which accounts for the effect of society on the amount of resources used when
looking at carrying capacity. This type of measure looks at whether the impact of a human society is increasing or decreasing over time and
can be used to compare impacts between societies of difference sizes or affluence levels.
Footprinting (see Module Footprinting: Carbon, Ecological and Water) is often used as a measure of sustainability that can be understood
intuitively and is, therefore, useful when talking to the general public. The ecological footprint, which also represents the carrying capacity of
the earth, is defined as “the total area of productive land and water ecosystems required to produce the resources that the population
consumes and assimilate the wastes that the population produces, wherever on Earth that land and water may be located” (Rees and
Wackernagel, 1996). This results in an evaluation of the demand and supply of natural capital of a given population (individual to planet) or a
product/service.
Life-cycle assessment (LCA), a structured methodology that can be utilized to evaluate the environmental impacts of products, processes,
projects, or services throughout their life cycles from cradle to grave (see Module Life Cycle Assessment) may be considered an ecological
metric. A greenhouse gas emissions inventory is an example of this methodology (see Case Study: Greenhouse Gases and Climate Change).
Economic Measures
Economic measures place a monetary value on sustainability. Economists use the following measures of sustainability: ecosystem valuation,
contingent valuation, and net national product, which are discussed in Chapter . Standard economic methods can be used to evaluate
environmental projects.
Indices that are used on a national and international level by organizations like the United Nations may be used to examine the economic and
social welfare of a region. The Index of Sustainable Economic Welfare (ISEW) and other related frameworks that account for sustainable
development have been conceived to provide an alternative to the Gross Domestic Product, which does not capture human welfare in its
calculations. This system weights personal expenditures within a population with an index of income inequality and a set of factors are then
added or subtracted to this monetary value. Monetary analysis of sustainability does not value the variety of sustainability issues especially
those that cannot be measured as a product or service in today’s markets (Gasparatos, et al., 2008).
Physical Measures
Physical measures of sustainability use thermodynamic concepts in their calculations. Two physical approaches to measuring sustainability
are exergy and emergy. These concepts are derived from the second law of thermodynamics which states that a closed system with constant
9.3.2 [Link]
mass and no energy inputs tends toward higher entropy or disorder. For instance, a piece of wood that is the product of many years of
complex tree growth releases energy (light and heat in the flame) when burned, and becomes carbon ash, smoke, gases, and water vapor. This
means that as properties within a system such as mass, energy, and chemical concentrations degrade (decompose) over time or burn, they also
make available useful energy (exergy) for work. Ecosystems and human economies function under this second law, but they can use external
energy (the sun) to maintain or increase energy supplies.
Emergy is the amount of energy of one kind (solar) that has been used directly or indirectly (through a transformation process) to make a
service or a product of one type and it is expressed in units of (solar energy) emjoule. It can be thought of as a measure of all the entropy that
has been produced over the whole process of creating a given product or service (Brown and Ulgiati, 2002). An example is the process of
fossil fuel creation: solar energy was used by plants to grow and is stored in the complex molecular structures that held the plants together,
when those plants died they decomposed and were buried over time under the changing earth, and the energy was concentrated into fossil
fuels. Emergy, thus, allows us to account for all the environmental support needed by human and eco-systems or inputs.
Measures of energy inputs are transformed to emergy by use of a factor that represents the amount of environmental work needed to produce
a product or provide a service. The emergy flows within a system include renewable resources (sunlight, rain, wind, agricultural production,
timber harvest, etc.), non-renewable production (fossil fuels, metals, minerals, soils), and imports/exports. A sustainable system would have a
net positive (or zero) emergy flow across its boundary (Mayer, et al., 2004). Emergy evaluations have been used, for instance, to
quantitatively demonstrate that renewable energy plants had higher sustainability compared to thermal plants (Brown and Ugliati, 2002).
Exergy can be defined as the maximum work that can be extracted from a system as it moves to thermodynamic equilibrium with a reference
state, as in the example of burned wood above. It has been used to study efficiency of chemical and thermal processes. This represents an
entropy-free form of energy that is a measure of its usefulness, quality or potential to make change. Exergy accounting provides insights into
the metabolism of a system and its effect on the environment using a common denominator. It can address energy utilization, be used for
design and analysis of energy systems and to quantify waste and energy losses reflecting resource use. Exergy can account for an economic
component, labor input, and impact of emissions on human health (Gasparatos, et al., 2008; Odum, 1996).
Comparison of Measures
So far we highlighted three categories for measures of sustainability – ecological, economic, and physical – and provided a few examples.
Sustainability measures is an evolving field of study and the metrics are innumerable. Ecological measures include indicators that try to
measure the sustainability of the ecosystem as a whole. Economic metrics use monetary measures and try to put a price on the environment
and its services. They are valued based on currency, which is an anthropocentric value, meaning it is significant only to humans. They
account only for human welfare to the extent that it depends on nature to survive. They do not account for the effect on an ecosystem as a
whole, including plants and animals. Physical metrics are closely tied to thermodynamics and energy, and are generally expressed in units of
energy.
Sustainability indicators are needed to improve our understanding of the nature of human demands on ecosystems and the extent to which
these can be modified. Society uses resources for physical and social infrastructure and continually increases its needs due to population
growth which is made possible by changing the way we grow and produce food, thus manipulating the food web. Some of these economic
metrics are closely tied to social sustainability metrics as well and try to account for the social welfare of a population. Overall, while
physical tools can capture certain environmental and economic issues, too, they do not address economic issues from the same perspective as
conventional economic analysis. Moreover, they do not capture most social issues.
Economic markets do not usually directly value goods and services that ecological systems provide to human economies and societies. These
ecosystem services include the uptake of carbon dioxide by plants and trees, purification of water by microorganisms, enrichment of soil
through degradation of plant and animal materials, and rainfall that provides irrigation (see Constanza, et al., 1997). Also economists do not
agree on the degree of substitutability between natural and man-made capital. This concept of substitutability means that natural capital such
as 100 year old (“old forest”) trees used to build homes and furniture can be replaced by replanting fast-growing trees and provide the same
value (Pearce, 1993). Technology also transforms the use of resources for instance by making them more readily available and more
economic. An example of this is the use of “fracking” to produce natural gas from sources that were difficult to extract from a decade ago
(see Module 10.2).
9.3.3 [Link]
A group of indicators can then be evaluated using a composite indicator/index (CI) or rating. CIs stand at the top of an information pyramid
where primary data lies at the base, followed by analyzed data, then indicators, and topped by indices. A composite indicator is formed by the
compilation of various individual indicators into one index based on an underlying model. (Nardo, et al., 2005). An example is the
Leadership in Energy and Environmental Design (LEED) which is a green building certification system developed by the U.S. Green
Building Council (USGBC). It accounts for a large variety of building attributes that contribute to a building being considered “sustainable”
such as building materials, location, landscape, energy usage, access to alternative transportation and so on. The final result is a numerical
rating for the building that is then associated with a certain certification level (Certified, Silver, Gold, Platinum). This kind of system is most
widely accepted and valued when a peer-review is conducted to determine what weights should be given to each attribute. When the USGBC
decided to update its rating system because it did not accurately reflect the values of its members, it underwent review through its various
committees.
Sometimes, when you have a lot of different measures that use different units you do not want to aggregate them together into one number. In
this case, a multi-criteria assessment (MCA) can be used where constituent indicators are not aggregated into a single index. Multi-criteria
analysis (similar name, different context) can be used as a tool to establish weights for several criteria, without requiring that all data be
converted into the same units (Hermann, 2007). There are several multi-criteria evaluation methods that can be used for this. These methods
may either be data-driven (bottom-up) when high-quality data is available or theory driven (top-down) when data is available for only one of
the aspects. A broader review of this can be found in Gasparatos, et al. (2008).
Many industry sectors are developing frameworks or rating systems that provide ways to report and measure sustainability. Two examples are
discussed here.
The Global Reporting Initiative (GRI) provides a system for organizations to publish their sustainability performance. Its purpose is to
provide transparency and accountability for stakeholders and to be comparable among organizations. It is developed in an international,
multi-stakeholder process and it is continuously improved. An organization determines which indicators from among those proposed it will
report. However, no overall index or scores are reported. There is also usually a narrative portion to the report (Global Reporting Initiative).
The indicators are broken down in environmental, economic and social performance indicators. Each area has core indicators with some
additional indicators that may be used based on the organization’s choice.
The American Association of Higher Education (AASHE), is the lead organization in North America for sustainability in colleges and
universities. One of their major projects has been the development of the Sustainability Tracking, Assessment and Rating System (STARS).
This is a voluntary, self-reporting framework that is to be used to measure relative progress of universities and colleges as they work toward
sustainability. STARS was developed using a collaborative process that involved input from many institutions. In 2008, a pilot study of 66
institutions was conducted to test the viability of the system and STARS version 1.0 was released in January 2010 with many schools
reporting by January 2011. The credits are given in three categories of equal weight – education and research; operations; planning,
administration and engagement. Each credit is given a weight based on the extent to which the credit contributes to improved environmental,
financial and social impacts, and whether there are educational benefits associated with the achievement of this credit and the breadth of that
impact. The result is a composite indicator, with transparent individual scoring. Schools participating in STARS will use an on-line reporting
tool which makes the results publicly available. Depending on the total points achieved, a level of achievement is be assigned. The STARS
rating will be good for three years but a school may choose to update annually. See 11.3.3 for an example of this reporting.
9.3.4 [Link]
Operating profit Water Consumption No. of non-profit projects
Recycling
Hazardous waste
An analytical tool, called COMPLIMENT, was developed to provide detailed information on the overall environmental impact of a business
(Hermann, 2007). This tool integrates the concepts of life cycle assessment, multi-criteria analysis and environmental performance indicators.
The combinations of environmental performance indicators used depend on the organization and reflect the relevant parts of the production
train. The method includes setting system boundaries, data collection, calculation of potential environmental impacts and their normalization,
aggregation of impacts using a multi-criteria analysis, the weights per impact category are multiplied by normalization potential impacts and
the results can be added up for each perspective. The system boundary strives to be cradle-to-grave (from extraction of resources to disposal)
although it may be a cradle-to-gate (from extraction of resources to completion of production) analysis.
Adoption of any single group of tools means that a certain perspective will be more highly represented in the sustainability assessment. “The
need to address the multitude of environmental, social, economic issues, together with intergenerational and intragenerational equity
concerns” (Gasparatos, et al., 2008, p. 306) produces problems that none of the disciplinary approaches can solve separately. Combining the
outputs of biophysical and monetary tools will result in a more comprehensive sustainability perspective. The result is that the choice of
metrics and tools must be made based on the context and characteristics that are desired by the analysts (Gasparatos, et al., 2008). Using a
composite indicator or a set of individual indicators presented together can overcome the problem of using a single metric to measure
sustainability.
Existing indicator-based sustainability assessments vary in the number of subsystems or assessment areas, the number of levels between
subsystem and indicator, and whether they result in an index (compound indicator) of the state of the system and subsystems. These would
include the ecosystem or environment, people or economy and society, and possibly institutions. The more subsystems assigned, the lower
the weight given to the environmental portion. As more indicator systems are developed they become increasingly complex, yet there is a
demand for a simple presentation that does not erase the complexity. A single indicator with true significance is not achievable, but by
combining indicators into indices the results are more meaningful.
Figure 9.3.2 Visualizing Results of Sustainability Assessments Hypothetical graphical representation of the Environmental Dimension of the
GRI for universities. The red numbers indicate the percentage of points achieved within each sub-category within the category of
environment. Source: D. Fredman adapted from Lozano (2006).
Another example of visualizing sustainability is seen in a framework developed for universities to use (Troschinetz et al., 2007). Again,
multidimensional sustainability indicators, each having an economic, environmental and social component are used. The categories are listed
in Figure 9.3.3. Each indicator was examined using a sustainability indicator triangle where each corner is delineated as economic,
environmental or social and the indicators are placed within to the triangle to reflect how well each measures those aspects.
Sustainability Indicator Triangle
Figure 9.3.3 Sustainability Indicator Triangle The thirteen sustainability indicators are placed according to how well each measures a
dimension of sustainability, i.e. environmental, societal and economic. Source: C. Klein-Banai adapted from Troschinetz, et al. (2007).
Conclusion
Measuring sustainability is difficult because of the interdisciplinary nature and complexity of the issues that this concern represents. Methods
have been developed out of the different disciplines that are based in the ecological, economic, physical and social sciences. When
approaching a measure of sustainability it is important to understand what you will use the results of that measure for, what the major
concerns you want to address are, and the limits of the system you choose. Often it is more meaningful to measure progress of the entity you
are examining – is it more sustainable than it was before? It is difficult to compare similar entities (countries, companies, institutions, even
9.3.5 [Link]
products) due to the complexity and variability in the data. Using visualization to represent the data is a helpful way to show the state of
sustainability rather than trying to express it in one number or in a table of numbers.
Ecological Footprint
Concept
The Merriam-Webster Dictionary defines footprint as:
1. an impression of the foot on a surface;
2. the area on a surface covered by something
Similarly, the ecological footprint (EF) represents the area of land on earth that provides for resources consumed and that assimilates the
waste produced by a given entity or region. It is a composite index (see Module 10.2) that represents the amount of biologically productive
land and water area required to support the demands of the population in that entity or region The EF is beneficial because it provides a single
value (equal to land area required) that reflects resource use patterns (Costanza, 2000). The use of EF in combination with a social and
economic impact assessment can provide a measure of sustainability’s triple bottom line (Dawe, et al., 2004). It can help find some of the
“hidden” environmental costs of consumption that are not captured by techniques such as cost-benefit analysis and environmental impact
(Venetoulis, 2001). Using the ecological footprint, an assessment can be made of from where the largest impact comes (Flint, 2001).
Next, we will discuss the how an EF is calculated.
Methodology
The ecological footprint methodology was developed by William Rees and Mathis Wackernagel (1996), and consists of two methodologies:
1. Compound calculation is typically used for calculations involving large regions and nations and is shown in Figure Compound
Calculation Steps for Ecological Footprint Analysis. First, it involves a consumption analysis of over 60 biotic resources including meat,
dairy produce, fruit, vegetables, pulses, grains, tobacco, coffee, and wood products. That consumption is then divided by biotic
productivity (global average) for the type of land (arable, pasture, forest, or sea areas) and the result represents the area needed to sustain
that activity. The second part of the calculation includes energy generated and energy embodied in traded goods. This is expressed in the
area of forested land needed to sequester CO2 emissions from both types of energy. Finally, equivalence factors are used to weight the six
ecological categories based on their productivity (arable, pasture, forest, sea, energy, built-up land). The results are reported as global
hectares (gha) where each unit is equal to one hectare of biologically productive land based on the world's average productivity. We derive
sub-national footprints based on apportioning the total national footprint between sub-national populations. The advantage of this method
is that it captures many indirect of effects of consumption so the overall footprint is more accurate.
2. Component-based calculation resembles life-cycle analysis in that it examines individual products and services for their cradle-to-grave
resource use and waste, and results in a factor for a certain unit or activity. The footprint is typically broken down into categories that
include energy, transportation, water, materials and waste, built-up land, and food. This method is better for individuals or institutions
9.3.6 [Link]
since it accounts for specific consumption within that entity. However, it probably under-counts as not all activities and products could
practically be measured or included. It also may double-count since there may be overlap between products and services.
Compound Calculation Steps for Ecological Footprint Analysis
Figure shows the compound calculation steps for ecological footprint analysis. Source: C. Klein-Banai.
Figure 9.3.5 Ecological Footprints of Select Nations Graph shows the ecological footprints of select nations. The bars show average EF in
global hectares per person for each nation. Each color on the bar represents the different types of land. Source: © 2010 WWF ([Link]).
Some rights reserved. Living Planet Report, 2010 , figure under CC BY-SA 3.0 License
The Living Planet Report prepared by the World Wildlife Fund, the Institute of Zoology in London, and Wackernagel’s Global Footprint
Network reports on the footprints of various nations. Figure Ecological Footprints of Select Nations displays the footprint of several nations
as shown in the report. The bars show average EF in global hectares per person for each nation. Each color on the bar represents the different
types of land. Here we see that the United Arab Emirates has the largest footprint of 10.2 gha per person, with the majority of its footprint due
to carbon (same as energy land described above). Whereas Latvia has the lowest footprint displayed at 6.0 gha per person, with the majority
of its footprint due to forestland.
United States’ Ecological Footprint
Figure shows the United States’ Ecological Footprint compared to the global average. Source: © 2010 WWF ([Link]). Some rights
reserved. Living Planet Report, 2010, figure under CC BY-SA 3.0 License
Figure 9.3.6 shows the national footprint in 2007 of the United States as 7.99 gha per person both with a bar display and with specific metrics
on the right that show the exact footprint and the United States’ ranking among all nations in the report (e.g. carbon is 5.57 gha and ranks 3rd
largest overall). The bar to the left expresses the world average. The United States’ footprint of 7.99 gha stands in contrast to the earth's
global biocapacity of 1.8 gha per person. Globally, the total population’s footprint was 18 billion gha, or 2.7 gha per person. However, the
earth’s biocapacity was only 11.9 billion gha, or 1.8 gha per person. This represents an ecological demand of 50 percent more than the earth
can manage. In other words, it would take 1.5 years for the Earth to regenerate the renewable resources that people used in 2007 and absorb
CO2 waste. Thus, earth’s population used the equivalent of 1.5 planets in 2007 to support their lives.
Carbon Footprint
Since climate change (see Chapter Climate and Global Change) is one of the major focuses of the sustainability movement, measurement of
greenhouse gases or carbon footprint is a key metric when addressing this problem. A greenhouse gas emissions (GHG) inventory is a type of
carbon footprint. Such an inventory evaluates the emissions generated from the direct and indirect activities of the entity as expressed in
carbon dioxide equivalents (see below). Since you cannot manage what you cannot measure, GHG reductions cannot occur without
establishing baseline metrics. There is increasing demand for regulatory and voluntary reporting of GHG emissions such as Executive Order
13514, requiring federal agencies to reduce GHG emissions, the EPA’s Mandatory GHG Reporting Rule for industry, the Securities and
Exchange Commission’s climate change disclosure guidance, American College and University Presidents’ Climate Commitment
(ACUPCC) for universities, ICLEI for local governments, the California Climate Action Registry, and numerous corporate sustainability
reporting initiatives.
9.3.7 [Link]
Scoping the Inventory
The first step in measuring carbon footprints is conducting an inventory is to determine the scope of the inventory. The World Business
Council for Sustainable Development (WBCSD) and the World Resource Institute (WRI) defined a set of accounting standards that form the
Greenhouse Gas Protocol (GHG Protocol). This protocol is the most widely used international accounting tool to understand, quantify, and
manage greenhouse gas emissions. Almost every GHG standard and program in the world uses this framework as well as hundreds of GHG
inventories prepared by individual companies and institutions. In North America, the most widely used protocol was developed by The
Climate Registry.
The GHG Protocol also offers developing countries an internationally accepted management tool to help their businesses to compete in the
global marketplace and their governments to make informed decisions about climate change. In general, tools are either sector-specific (e.g.
aluminum, cement, etc.) or cross-sector tools for application to many different sectors (e.g. stationary combustion or mobile combustion).
Scopes of a Greenhouse Gas Emissions Inventory
Figure shows the three scopes of a greenhouse gas emissions inventory. Source: New Zealand Business Council for Sustainable
Development, The challenges of greenhouse gas emissions: The “why” and “how” of accounting and reporting for GHG emissions (2002,
August), figure 3, p. 10.
The WRI protocol addresses the scope by which reporting entities can set boundaries (see Figure 9.3.4). These standards are based on the
source of emissions in order to prevent counting emissions or credits twice. The three scopes are described below:
Scope 1: Includes GHG emissions from direct sources owned or controlled by the institution – production of electricity, heat or steam,
transportation or materials, products, waste, and fugitive emissions. Fugitive emissions are due to intentional or unintentional release of
GHGs including leakage of refrigerants from air conditioning equipment and methane releases from farm animals.
Scope 2: Includes GHG emissions from imports (purchases) of electricity, heat or steam – generally those associated with the generation
that energy.
Scope 3: Includes all other indirect sources of GHG emissions that may result from the activities of the institution but occur from sources
owned or controlled by another company, such as business travel; outsourced activities and contracts; emissions from waste generated by
the institution when the GHG emissions occur at a facility controlled by another company, e.g. methane emissions from landfilled waste;
and the commuting habits of community members.
Depending on the purpose of the inventory the scope may vary. For instance, the EPA mandatory reporting requirements for large carbon
dioxide sources require reporting of only Scope 1 emissions from stationary sources. However, many voluntary reporting systems require
accounting for all three scopes, such as the ACUPCC reporting. Numerous calculator tools have been developed, some publicly available and
some proprietary. For instance many universities use a tool called the Campus Carbon Calculator developed by Clean Air-Cool Planet, which
is endorsed by the ACUPCC. Numerous northeastern universities collaborated to develop the Campus Carbon Calculator and the calculator
has been used at more than 200 campuses in North America. It utilizes an electronic Microsoft Excel workbook that calculates estimated
GHG emissions from the data collected.
Methodology
GHG emissions calculations are generally calculated for the time period of one year. Figure 9.3.5 shows the steps for reporting GHG
emissions. It is necessary to determine what the baseline year is for calculation. This is the year that is generally used to compare future
increases or decreases in emissions, when setting a GHG reduction goal. The Kyoto Protocol proposes accounting for greenhouse gas
emissions from a baseline year of 1990. Sometimes calculations may be made for the current year or back to the earliest year that data is
available.
Steps for Preparing a GHG Emissions Report
Figure shows the required steps to take when preparing a GHG emissions report. Source: C. Klein-Banai
Next, the institutional or geographic boundaries need to be defined. Also, the gases that are being reported should be defined. There are six
greenhouse gases defined by the Kyoto Protocol. Some greenhouse gases, such as carbon dioxide, occur naturally and are emitted to the
atmosphere through natural and anthropogenic processes. Other greenhouse gases (e.g. fluorinated gases) are created and emitted solely
through human activities. The principal greenhouse gases that enter the atmosphere because of human activities are:
Carbon Dioxide (CO2): Carbon dioxide is released to the atmosphere through the combustion of fossil fuels (oil, natural gas, and coal),
solid waste, trees and wood products, and also as a result of non-combustion reactions (e.g. manufacture of cement). Carbon dioxide is
sequestered when plants absorb it as part of the biological carbon cycle.
Methane (CH4): Methane is emitted during the production and transport of coal, natural gas, and oil. Methane emissions also come from
farm animals and other agricultural practices and the degradation of organic waste in municipal solid waste landfills.
Nitrous Oxide (N2O): Nitrous oxide is emitted during agricultural and industrial activities, and combustion of fossil fuels and solid waste.
Fluorinated Gases: Hydrofluorocarbons, perfluorocarbons, and sulfur hexafluoride are synthetic, powerful greenhouse gases that are
emitted from a variety of industrial processes. Fluorinated gases are sometimes used as substitutes for ozone-depleting substances (i.e.
Chlorofluorocarbons (CFCs), hydrochlorofluorocarbon (HCFCs), and halons). CFCs and HCFCs are gases comprised of chloride,
fluoride, hydrogen, and carbon. Halons are elemental gases that include chlorine, bromine, and fluorine. These gases are typically emitted
9.3.8 [Link]
in smaller quantities, but because they are potent greenhouse gases, they are sometimes referred to as High Global Warming Potential
gases (“High GWP gases”)
Each gas, based on its atmospheric chemistry, captures different amounts of reflected heat thus contributing differently to the greenhouse
effect, which is known as its global warming potential. Carbon dioxide, the least capture efficient of these gases, acts as the reference gas
with a global warming potential of 1. Table 9.3.3 shows the global warming potential for the various GHGs.
Table 9.3.6 Global Warming Potentials Source: C. Klein-Banai created table using data from Climate Change 2007: The Physical Science Basis:
Contribution of Working Group I to the Fourth Assessment Report of the Intergovernmental Panel on Climate Change, Cambridge University Press, section
2.10.2
Gas GWP
CO2 1
CH4 21
N2O 310
HFC-23 11,700
HFC-32 650
HFC-125 2,800
HFC-134a 1,300
HFC-143a 3,800
HFC-152a 140
HFC-227ea 2,900
HFC-236fa 6,300
HFC-4310mee 1,300
CF4 6,500
C2F6 9,200
C4F10 7,000
C6F14 7,400
SF6 23,900
GHG emissions cannot be easily measured since they come from both mobile and stationary sources. Therefore, emissions must be
calculated. Emissions are usually calculated using the formula:
A × Fg = E (9.3.1)
where A is the quantification of an activity in units that can be combined with emission factor of greenhouse gas g (Fg) to obtain the resulting
emissions for that gas (Eg).
Examples of activity units include mmbtu (million British Thermal Units) of natural gas, gallons of heating oil, kilowatt hours of electricity,
and miles traveled. Total GHG emissions can be expressed as the sum of the emissions for each gas multiplied by its global warming
potential (GWP). GHG emissions are usually reported in metric tons of carbon dioxide equivalents (metric tons CO2-e):
′
GH G = ∑ Eg GW Pg (9.3.2)
Eg is usually estimated from the quantity of fuel burned using national and regional average emissions factors, such as those provided by the
US Department of Energy’s Energy Information Administration.
Emission factors can be based on government documents and software from the U.S. Department of Transportation, the U.S. Environmental
Protection Agency (EPA), and the U.S. Department of Energy, or from specific characteristics of the fuel used – such as higher heating value
and carbon content. Scope 3 emissions that are based on waste, materials, and commuting are more complex to calculate. Various calculators
use different inputs to do this and the procedures are less standardized. See Case Study: Comparing Greenhouse Gas Emissions, Ecological
Footprint and Sustainability Rating of a University for an example of these kinds of calculations.
9.3.9 [Link]
Greenhouse gas emissions inventories are based on standardized practice and include the steps of scoping, calculating, and reporting. They
are not based on actual measurements of emissions, rather on calculations based on consumption of greenhouse gas generating materials such
as fossil fuels for provision of energy and transportation or emissions from waste disposal. They can be conducted for buildings, institutions,
cities, regions, and nations.
Water Footprint
The water footprint of production is the volume of freshwater used by people to produce goods, measured over the full supply chain, as well
as the water used in households and industry, specified geographically and temporally. This is slightly different from the hydroprint described
above which simply compares the consumption of water by a geographic entity to the water that falls within its watershed. If you look at the
hydrologic cycle (see module Water Cycle and Fresh Water Supply), water moves through the environment in various ways. The water
footprint considers the source of the water as three components:
Green water footprint: The volume of rainwater that evaporates during the production of goods; for agricultural products, this is the
rainwater stored in soil that evaporates from crop fields.
Blue water footprint: The volume of freshwater withdrawn from surface or groundwater sources that is used by people and not returned;
in agricultural products this is mainly accounted for by evaporation of irrigation water from fields, if freshwater is being drawn.
Grey water footprint: the volume of water required to dilute pollutants released in production processes to such an extent that the quality
of the ambient water remains above agreed water quality standards.
The water footprint of an individual is based on the direct and indirect water use of a consumer. Direct water use is from consumption at
home for drinking, washing, and watering. Indirect water use results from the freshwater that is used to produce goods and services purchased
9.3.10 [Link]
by the consumer. Similarly, the water footprint of a business or institution is calculated from the direct and indirect water consumption.
Water Footprint of Production of Select Countries
Figure 9.3.6 Water Footprint of Production of Select Countries Graph shows the water footprint of production of select countries. Source: ©
2010 WWF ([Link]). Some rights reserved. Living Planet Report, 2010, figure under CC BY-SA 3.0 License
Figure 9.3.6 shows the water footprint of production for several countries as a whole. In this report, due to lack of data, one unit of return
flow is assumed to pollute one unit of freshwater. Given the negligible volume of water that evaporates during domestic and industrial
processes, as opposed to agriculture, only the grey water footprint for households and industry was included. This figure does not account for
imports and exports it is only based on the country in which the activities occurred not where they were consumed.
In contrast, the water footprint of a nation accounts for all the freshwater used to produce the goods and services consumed by the inhabitants
of the country. Traditionally, water demand (i.e. total water withdrawal for the various sectors of economy) is used to demonstrate water
demand for production within a nation. The internal water footprint is the volume of water used from domestic water resources; the external
water footprint is the volume of water used in other countries to produce goods and services imported and consumed by the inhabitants of the
country. The average water footprint for the United States was calculated to be 2480m3/cap/yr, while China has an average footprint of
700m3/cap/yr. The global average water footprint is 1240m3/cap/yr. As for ecological footprints there are several major factors that
determine the water footprint of a country including the volume of consumption (related to the gross national income); consumption pattern
(e.g. high versus low meat consumption); climate (growth conditions); and agricultural practice (water use efficiency) (Hoekstra &
Chapagain, 2007).
Using average water consumption for each stage of growth and processing of tea or coffee, the “virtual” water content of a cup can be
calculated (Table 9.3.8). Much of the water used is from rainfall that might otherwise not be “utilized” to grow a crop and the revenue from
the product contributes to the economy of that country. At the same time, the result is that many countries are “importing” water to support
the products they consume.
Table 9.3.8 Virtual Water Content of a Cup of Tea or CoffeeTable shows the virtual water content for a cup of tea or coffee. Source: C. Klein-Banai created
table using data from Chapagain and Hoekstra (2007). alt="Virtual Water Content of a Cup of Tea or Coffee" longdesc="Table shows the virtual water
content for a cup of tea or coffee."
Drink Preparation Vital Water Content (mL/cup)
To learn more about other countries’ water footprints, visit this interactive graph. To calculate your own water footprint, visit the Water
Footprint Calculator.
Summary
Footprinting tools can be useful ways to present and compare environmental impact. They are useful because they can combine impacts from
various activities into one measure. However, they have limitations. For instance, in a carbon footprint or greenhouse gas emissions
inventory, many of the “conventional” environmental impacts such as hazardous waste, wastewater, water consumption, stormwater, and
toxic emissions are not accounted for, nor are the impacts of resource consumption such as paper, food, and water generally measured.
Perhaps most importantly, certain low-carbon fuel sources (e.g. nuclear power) that have other environmental impacts (e.g. nuclear waste) are
neglected. Finally, the scope of the emissions inventory does not include upstream emissions from the manufacture or transport of energy or
materials. This suggests that there is a need to go beyond just GHG emissions analyses when evaluating sustainability and include all forms
of energy and their consequences.
The ecological footprint can be misleading when it is looked at in isolation, for instance with an urban area, the resources needed to support it
will not be provided by the actual geographic area since food must be “imported” and carbon offset by natural growth that does not “fit” in a
city. However, cities have many other efficiencies and advantages that are not recognized in an ecological footprint. When looked at on a
national level it can represent the inequities that exist between countries.
It is interesting to contrast the water and ecological footprints, as well. The water footprint explicitly considers the actual location of the water
use, whereas the ecological footprint does not consider the place of land use. Therefore it measures the volumes of water use at the various
locations where the water is appropriated, while the ecological footprint is calculated based on a global average land requirement per
consumption category. When the connection is made between place of consumption and locations of resource use, the consumer's
responsibility for the impacts of production at distant locations is made evident.
9.3.11 [Link]
Case Study: Comparing Greenhouse Gas Emissions, Ecological Footprint and Sustainability Rating of a
University
How do different measures of sustainability compare when looking at one institution? This case study compares these different measures for
the University of Illinois at Chicago (UIC). Located just southwest of downtown Chicago, UIC has 13 colleges serving 27,000 students and
12,000 employees, with over 100 buildings on 240 acres (97 hectares) of land. The activities of the faculty, staff and students and the
buildings and grounds have an impact on the sustainability of the institution. This case study will look at the results of the greenhouse gas
emission inventory, ecology footprint, and sustainability rating.
Greenhouse Gas Emissions Inventory
Figure 9.3.7 Greenhouse Gas Emissions Inventory UIC's Greenhouse gas emissions profile for FY2004-2010, using the regional mix for
purchased electricity. Source: C. Klein-Banai
Figure 9.3.7 displays UIC's GHG emissions profile for seven years. The emissions were calculated using the Campus Carbon Calculator
developed by the not-for-profit organization, Clean Air-Cool Planet. While this tool has a number of limitations it has been used by many of
the over 670 colleges and universities who are signatory to the American College and University Presidents Climate Commitment
(ACUPCC) to simplify the emissions inventory profile. The tool is also recommended by the ACUPCC as a standard method of emissions
calculation for United States universities. It is based on the World Resources Institute (WRI) and World Business Council for Sustainable
Development (WBSCD) Greenhouse Gas (GHG) Protocol Initiative that developed GHG emissions inventory standards. UIC's emissions
were calculated using the regional average electricity sources for the electric grid servicing the Chicago area. However, until August of 2009,
UIC purchased electricity from Commonwealth Edison which has a much lower greenhouse gas emissions factor due to the high percentage
of nuclear power in the Chicago region.
UIC operates two combined heat and power plants. However, the university has increasingly lowered its production of electricity from
natural gas by purchasing more electricity through block purchases (for defined amounts of electricity for a certain period of time) due to the
relatively low cost of purchasing electricity as compared to self-generating. This strategy has increased UIC's emissions as the regional mix
has a fair amount of coal-powered plants providing electricity. Neverthless, a downward trend in emissions is beginning in spite of the
increased electricity purchases between 2009 and 2010. This may be due to overall reduction in energy consumption on campus, which is
reducing the GHG emissions.
Figure 9.3.8 illustrates the relative contribution to UIC's 2010 emissions profile, with 77 percent of emissions coming from buildings (power
plants, purchased electricity, and other on-campus stationary, i.e. natural gas supply to the buildings), 20 percent due to transportation
(campus fleet , commuting to campus, and air travel), and less than one percent for emissions due to waste sent to the landfill (which
generates methane).
Breakdown of UIC's Greenhouse Gas Emissions
Figure shows the breakdown of UIC's greenhouse gas emissions inventory for the fiscal year 2010, in metric tons of carbon dioxide
equivalents (mt CO2-e). Total emissions: 354,758 mt CO2-e. Source: C. Klein-Banai
UIC's total emissions for fiscal year 2010 were 354,758 mt CO2-e, which amounts to 13.14 mt CO2-e per full-time equivalent student
enrolled. Table 9.3.9 compares UIC's emissions to those of the city of Chicago, state of Illinois, and the United States.
Table 9.3.9 Comparison of GHG EmissionsSources: C. Klein-Banai created table using data from UIC Climate Action Plan, Chicago Climate Action Plan,
U.S. EPA.
US 6,633.20 2009
An Ecological Footprint Analysis (EFA) was conducted using data from fiscal year 2008, including much of the same data used for the GHG
emissions inventory. In addition, water, food, recycling, and built-up land data were used to calculate the number of global hectares required
to provide the resources and absorb the waste and GHG emissions produced to support UIC's activities. The results are displayed in Table
9.3.10. The total footprint was 97,601 global hectares, on a per capita basis this is equivalent to 2.66 gha/person. This is in contrast to about
8.00 gha/person nationally in the United States, although one must use caution in making comparisons because the scope and methodology of
the analysis differ.
Table 9.3.10 UIC's Ecological Footprint Using FY2008 DataComposite Indicator: Sustainability Tracking, Assessment and Rating System. Source: C. Klein-
Banai.
Category Global Hectares Percent
9.3.12 [Link]
Energy 70,916 72.7%
The STARS system (see module Sustainability Metrics and Rating Systems) was used to rate UIC. The university received 39.1 points, for a
Bronze rating. The points break down into the categories shown in Table 9.3.11. There are three main categories of points – Education &
Research; Operations; and Planning, Administration & Engagement. Within each of the categories there are sub-categories such as
Curriculum, Climate, and Coordination & Planning. Within those sub-categories there are specific strategies that address them, with varying
amounts of points that depend on the assessed weight of each strategy. Each category's individual percentage score is weighted equally to the
others. In addition, four innovation strategies are available for which an institution can receive one point. These points are not attributed to a
particular category.
Table 9.3.11 STARS Points Received by UIC by CategorySource: C. Klein-Banai with data from STARS
Points Received Possible % Per Category Weight
Planning, Administration,
54.91% 33.33/100
and Engagement
Investment 0 16.75
Innovation 0 4.00
This reporting system shows that UIC's strengths lie in the areas outside of operations, which are what is measured with an EFA or GHG
emissions inventory. Most points were gained for Planning, Administration & Engagement. This rating system can be used to identify
9.3.13 [Link]
specific areas that can be targeted for advancing sustainability initiatives in a much broader realm than the other two metric allow. This case
study demonstrates the different types of information and sustainability tracking that can be done using different types of measures of
sustainability. Whether you use one measure or several depends on the purpose and scope of the sustainability reporting.
Food Miles
Introduction
Efforts to explore the impacts of the items on our dinner tables led to the broad concept of food miles (the distance food travels from
production to consumption) as being a quick and convenient way to compare products. With increasing globalization, our plates have
progressively included food items from other continents. Previously it would have been too expensive to transport these products. However,
changes to agricultural practices, transportation infrastructure, and distribution methods now mean that people in the United States can start
the day with coffee from Brazil, have a pasta lunch topped with Italian cheeses, snack on chocolate from Côte d'Ivoire, and end with a dinner
of Mediterranean bluefin tuna and Thai rice. However, the globalization that has led to increased availability of these products comes with
associated costs, such as the emission of greenhouse gases and other pollutants, increased traffic congestion, lack of support for local
economies, less fresh food, and decreased food security. Therefore, the concept of measuring food miles was meant to provide an easy
comparison of the relative impacts of our food choices.
Many individuals, groups, and businesses today measure or calculate food miles. But, when Andrea Paxton, a U.K.-based environmental
activist, coined the term in the 1990s the concept of food miles was intended to encompass more than simply a distance. The point was to
broaden awareness that our food choices have consequences that are often not apparent. Consumers frequently do not know the histories
behind their food purchases, and markets often cannot supply the information because of the many production processes and distribution
methods used.
While the distance food travels does determine some of the environmental, social, and economic impacts, there can be other hidden
consequences not so easily measured. Exploration of the utility of food miles in the general sense of knowing the impacts of our purchasing
decisions has resulted in a broadening awareness of the complexity of globalization. Although consumers can use the easy-to-compare
numbers representing food miles, that metric cannot reflect all of the impacts of food purchasing decisions.
Figure 9.3.9 Food Miles for Fruit Salad The various ingredients in this simple fruit salad travel different distances to Illinois' supermarkets.
Source. D. Ruez adapated from TUBS, akarlovic, Fir0002, and Abrahami
Most of our food from supermarkets is marked with a country or state of origin. That alone is usually enough to get an estimate of the
distance, especially if the location can be narrowed down by finding out the part of the country or state that most commonly produces the
product. If the fruit salad in Figure 9.3.10 is being made in Chicago, Illinois, and the apples are from the state of Washington, the likely
origin is in the center part of the state. The travel distance is approximately 2,000 miles (3,219 km). Bananas from Costa Rica traveled about
2,400 miles (3,862 km) to Chicago, and there are honey producers only 160 miles (257 km) from Chicago. A simple average of the miles the
ingredients traveled would not take into account that the fruit salad probably would not contain equal amounts of the three items. If the recipe
called for 2 pounds (.9 kg) of apples, 2 pounds (.9 kg) of bananas, and a ¼ pound (.1 kg) of honey, the miles would be weighted toward the
distances traveled by the fruit: 2080 food miles per pound of fruit salad (or 3,347 km/kg of fruit salad).
Benefits
The benefits of using food miles in evaluating food choices match the three main categories that represent sustainability: environmental,
social, and economic. All methods of transporting food over long distances, and most methods used to transport over short distances, involve
fossil fuels. Burning of fossil fuels creates greenhouse gases, which contribute to climate change. Using fossil fuels also results in the
emission of other gases and particulates that degrade air quality. Longer transportation distances intensify traffic congestion, resulting in lost
productivity, and increase the need for more extensive infrastructure (such as more highways) that negatively impact the environment by
increasing the amount of impervious cover and by requiring more natural resources. Increased roadways also encourage sprawl, leading to
more inefficient development patterns. Finally, traffic congestion and air pollution from driving contribute to an estimated 900,000 fatalities
per year worldwide.
Use of food miles is often tied to locavore movements, which emphasizes consumption of locally-grown food products. Local food is usually
fresher, with harvesting waiting until produce is ripe, and has less processing and fewer preservatives. Many people think locally-grown food
tastes better, but others chose to be a locavore because it strengthens local cultural identity or because the safety of the food is being
9.3.14 [Link]
controlled by people who also consume the products themselves. Eating local foods also promotes food security because availability and
price of imported foods is more dependent on fluctuating fuel costs and sociopolitical conflicts elsewhere.
The production of food in developing countries, and the subsequent exporting of those products, has several types of impacts. The
environmental burden of soil degradation, water depletion, and others, are imposed on developing countries, while more prosperous countries
enjoy the benefits. This can be especially problematic because some developing countries do not have the policies to require, or the resources
to implement, more environmentally-friendly food production practices. In particular, the low prices paid to food producers in developing
countries do not include sufficient funds, or requirements, for practices to preserve or restore ecosystem quality. Moreover, developing
countries disproportionately suffer malnutrition, yet the success of large-scale transport of food encourages cultivation of products to be
exported instead of planting nutritious foods to be self-sustaining.
Some businesses are embracing the basic concepts of food miles because transporting food over shorter distances uses less fuel, and is
therefore cheaper. Additionally, food that covers longer distances usually requires more packaging, which adds to the cost. By focusing on
local foods, local economies are supported. This has led to clearer labeling of food products, giving consumers the ability to make more
informed decisions about their purchases.
Criticism
Although the concept of food miles is useful, it has been heavily criticized for being too simplistic. For example, all miles are not created
equally. The consumption of fuel varies by the mode of transportation and the amount being moved. If you compare the consumption
required to move one pound of a product, ocean freighters are the most efficient of the common methods, followed by trains, trucks, and
finally planes. When a combination of transportation methods is used, making a comparison with food miles becomes even more complex.
This is especially a problem because most of us drive a personal vehicle to get our groceries. That means that it may be more efficient (from a
total fuel consumption perspective) to drive 1 mile (1.6 km) to a local supermarket who imports beef from Australia, than to drive 40 miles
(64 km) to visit a market selling locally-produced beef.
Food miles also do not take into consideration the variables of production before the products are transported. Growing outdoors requires
different amounts of energy input than greenhouses. A commonly cited example is that of tomatoes; heating greenhouses to grow tomatoes in
the United Kingdom consumes much more energy than growing tomatoes in warm Spain and importing them. Use of chemical fertilizers and
pesticides affect environmental quality and production levels differently than organic farming. Because soil quality varies, locally-grown
foods, in some cases, require more of the chemical additives. Some areas may have newer equipment, better climate, increased access to
water, or other factors that determine the overall efficiency of food production. Growing rice in deserts or oranges in the Arctic would have
more environmental impacts than the transportation of those products from locales where they can be grown more efficiently (from both an
environmental and economic perspective). Understanding these production variables is critical because several recent studies have suggested
that as much as 80% of greenhouse-gas generation from food consumption comes from the production phase.
There are also benefits to globalization and increased transport of food. There is now more widespread access to a broader range of food
products. This can lead to increased appreciation for other cultures and greater international cooperation. Long-distance transport of food
products can also provide jobs to developing nations by giving them access to larger, more prosperous markets internationally. Jobs and
economic incentives from food production are some of the few widespread opportunities for developing countries, and these may lead to
growth in other economic areas.
Criticism of the use of food miles can be unfairly disapproving of products that travel long distances. However, simple calculations of food
miles have also been said to underrepresent the importance of travel distances. Most food is transported with some packaging, and that
packaging also requires energy input for its production and transport. Because products that move shorter distances usually have less
packaging, the difference in calculated food miles may underestimate the actual environmental impact. Local foods also require less energy
and resource consumption because of reduced need for transportation infrastructure, chemical additives and preservatives, and refrigeration.
The impacts during the production phase also vary between types of foods, which can also result in underestimates of the impacts. Production
of meats, especially red meats, requires large amounts of land to generate the crops needed for animal feed. Because not all energy is passed
from feed to the animal, using meats for our food is inefficient from an energy perspective. It takes over 8 pounds of grain to feed a cow
enough to generate 1 pound of weight gain. That grain must be grown on land that can long longer produce food directly for human
consumption. The amount of land required to produce animal feed is known as ghost acres. Ghost acres also extend to the areas required to
provide the fuel, water, and other resources needed for animal feed, and for the overall support of animals1. While some other meats such as
pork, poultry, and especially fish, use proportionally less feed, there are other concerns about the environmental impacts of diets with large
amounts of meat.
Confined animal feeding operations (CAFOs), high-density animal farms, have become the primary source of livestock for meat in the U.S.,
Europe, and many other countries. The technological innovations employed in these operations have increased the speed and volume of meat
production, but have raised health concerns. Antibiotics and hormones used increasingly on animals in CAFOs may be passed on to humans
during consumption, even though there is currently no way of knowing a safe level of those substances in our diets. The overuse of
antibiotics in CAFOs also results in antibiotic-resistant pathogens. In addition to the impacts from the ghost acres, there are other ecological
9.3.15 [Link]
impacts such as pollution from massive amounts of concentrated manure. Although the distance meat is transported has an environmental
impact, the other concerns are more significant.
Confined Animal Feeling Operations (CAFOs)
Figure 9.3.10 Confined Animal Feeding Operations (CAFOs)This image shows a CAFO for cattle. CAFOs have raised health concerns for
human consumption of the meat produced in them. Source: eutrophication & hypoxia
Implementation
The ongoing investigations of food miles have affected businesses, groups, and individuals in various ways. As mentioned above, paying
closer attention to the distance food travels can be a good business strategy because fuel costs money. Centralization of processing and
distribution centers in the United States has resulted in a relative frequent occurrence of shipping produce thousands of miles only to end up
in supermarkets back near its origin. In many cases the initial savings from building fewer centralized facilities is exceeded in the long-term
by the continual shipping costs. As a result, some retailers are encouraging outside scrutiny of their habits, because it can result in increased
profits. At the other extreme, the rise in food miles in some cases is driven entirely by money. Fish caught offshore Norway and the United
Kingdom, for example, is sent to China for cheaper processing before being returned to European markets.
An awareness of the impact of food miles has led to many groups advocating local foods. Local farmers' markets have appeared and
expanded around the United States and elsewhere, providing increased access to fresh foods. Community-supported agriculture programs
create share-holders out of consumers, making them more personally invested in the success of local economies, while farmers gain some
financial security. Campaigns by community groups are influencing retailers and restaurants by scrutinizing the food-purchasing decisions.
The reciprocal is also true as retailers and restaurants advertise their sustainability efforts. See Resources for examples of local farmers'
markets in Illinois.
Illinois Farmer's Market
Figure 9.3.11 Illinois Farmer's Market Colorful produce at an Illinois farmer's market. Source: Maycomb Paynes
Yet, there are challenges to the implementation of food miles as a concept. Suppliers, such as individual farmers, might opt for the reliable
annual purchase from a mass-distributor. Consumers might make decisions solely on the sticker price, not knowing the other impacts, or
consumers might know the impacts but choose the immediate economic incentive. Some of these challenges can be addressed by education.
This can include efforts such as eco-labeling – labels, often by third parties, that independently attest to the environmental claims of products.
This can influence some consumers, but larger buyers like school systems and restaurant chains may require other incentives to change
purchasing practices. The source of these incentives, or alternatively, regulations, might come from government agencies, especially those
with desires to support local economies. However, there is no consensus regarding who should be evaluating and monitoring food miles.
Summary
The criticisms of food miles are valid, and work is continually being done incorporate the many factors that more completely show the
environmental impacts of transporting food. This can be a time consuming process, and the many variables are usually not readily available
to consumers. A frozen pizza might contain many types of ingredients from various areas that are transported to individual processing plants
before being assembled in another location and forwarded to distribution centers before being shipped to stores. Even if this process is
eventually simplified, eating decisions should not be made solely on the basis of food miles, which cannot account for the variations in
transportation and production methods or the social and economic impacts.
This does not mean that food miles are never a useful tool. When comparing similar products (e.g., onions to onions) with other similar
externalities (e.g., production and transportation methods), food miles provide a convenient way for consumers to begin to make informed
decisions about their purchases. Even though food transportation is a relatively small portion of the overall impact of our food consumption,
changes to any phase of the process can have a positive additive effect and make a real contribution to environmental health. Moreover, most
of the benefits for using food miles can likewise apply to many of our non-food purchases, with allowances for some of the same drawbacks.
Additionally, the discussion could be expanded to include other kinds of decisions, such as where to live in relation to location of job, and
where to take a vacation. In general, the concept of food miles reflects the need to understand how hidden influences generate environmental,
social, and economic impacts.
9.3.16 [Link]
Because of the diversity of observational scales and topics, not all EPIs are useful in all scenarios. However, all EPIs should indicate whether
the state of the environment is changed positively or negatively, and they should provide a measure of that change. An EPI is also more
meaningful if it can quantify the results to facilitate comparison between different types of activities. But before an EPI is selected, targets
and baselines must be clearly articulated. Vague targets are difficult to evaluate, and the results may be uninformative. The EPI selected must
use indicators that are definitively linked to the targets, are reliable and repeatable, and can be generated in a cost and time efficient manner.
To evaluate an activity, an EPI needs to include information from up to four types of indicators: inputs, outputs, outcomes, and impacts.
Inputs are the natural resources or ecosystem services being used. Outputs are the goods or services that result from that activity. While
outputs can often be quantified, outcomes typically cannot be and instead represent environmental, social, and economic dimensions of well-
being. In some cases it is useful to think of outcomes as why an output was sought; however, outcomes can also be unanticipated or unwanted
effects of an output. Impacts refer to the longer-term and more extensive results of the outcomes and outputs, and can include the interaction
of the latter two indicators.
For example, coal can be an input for an electricity-generating plant because we need the output (electricity) to turn on lights in our homes.
Two outcomes would include the ability to read at night because of the electricity and the visible air pollution from the power plant smoke
stacks. An impact of being able to read more can be a better-educated person, while an impact of the greenhouse gas emissions from burning
coal is increased potential for global climate change. This is a simplistic example which does not include the majority of relevant indicators
(inputs, outputs, outcomes, and impacts) for a complete and more meaningful analysis.
We can then evaluate each of the indicators. Is the input (coal) an appropriate choice? Is there enough for the practice of burning it to
continue? Are there problems, such as political instability that could interrupt continued access? Does the output (electricity) sufficiently
address the problem (in this case, energy for turning on lights)? Is the output produced and delivered in a timely manner? Is it provided to the
appropriate consumers and in a quantity that is large enough? Does the output create the desired outcome (being able to read at night)? Does
it also result in unwanted outcomes (air pollution)? Do the outcomes result in long-term impacts (such as life-long learning or decade-long
climate change) that are widespread?
Note that outcomes and impacts can be either positive or negative. The strength of an EPI lies in its ability to look at the bigger picture and
include multiple variables – particularly with regard to the impacts. However, whether an impact is considered meaningful depends on the
values and perspectives of the individuals and groups involved. Judgment plays a role because of the difficulty in comparing completely
different impacts. How do you compare life-long learning and climate change in the above example about the use of coal?
Uses
Monitoring the impacts of both short-term and long-term activities with EPIs allows decision makers to make changes that result in
performance with lesser environmental impacts. In some cases, changes can be made to ongoing projects, or the results of an EPI can be used
for publicity if the performance data indicate the activity is environmentally-sound. In other cases, the EPI establishes a performance
benchmark against which other projects are measured, or the results are used in the strategic planning phase while projects are in
development. In this way, past successes and failures can both be incorporated into future plans.
Use of EPIs requires production of multiple data points. A single application of an EPI does not mean much until placed into a larger context.
For example, an EPI might evaluate the impact of your city's recycling efforts (see Figure 9.3.12), but that result can be difficult to interpret
without additional data that can be presented in multiple ways:
Absolute values: Is the impact greater or less than that of other cities? How does the total cost of the recycling program compare?
Normalized values: How does the per person impact compare to another city, country, business, etc.? What is the amount of aluminum
recycled per dollar spent on recycling?
Trends: Is your city improving, or is the progress your city sees in recycling better than that that of other cities? This could be asked of
either absolute or normalized data: Is the total amount of aluminum recycled in your city increasing? Is the per-person amount of
aluminum recycled in your city increasing?
Municipal Solid Waste Recycling Rates
Figure 9.3.12 Municipal Solid Waste Recycling Rates Municipal solid waste recycling rates in the United States from 1960-2007. Source:
EPA
9.3.17 [Link]
influencing biodiversity. The better-known groups of organisms, such as birds and mammals, are also monitored for a direct count of
biodiversity, but vertebrates are a tiny proportion of life and cannot accurately reflect changes in all species.
Endangered Animals
Figure 9.3.13 Endangered Animals Illustration shows the number of endangered animals in each country of the world. Source: World Atlas
of Biodiversity
Wood is harvested for timber and fuel, but forests are also cleared for agricultural fields and housing developments. Such deforestation
frequently leads to rapid soil erosion and extinctions. Cutting of forests also results in changes to the water cycle, altering precipitation
patterns and rates, and nutrient cycles, such as the release of carbon dioxide into the atmosphere. At the same time as deforestation takes its
toll in places, trees are being planted elsewhere. Developed countries are increasing their forested areas, but this is commonly being done at
the expense of developing countries, which are exporting their wood (see Figure 9.3.13). Forestry indicators in EPIs include the annual
change in forested areas, but can be broken down into the types of forests because each has different environmental impacts. Another
indicator is the use of non-sustainable wood resources. Tree farms and some harvesting methods provide renewable supplies of wood, while
clear-cutting tropical forests does not. Irresponsible wood harvesting produces negative results for ecosystem health.
Deforestation in Haiti
Figure 9.3.14 Deforestation at the Haiti/Dominican Republic Border Satellite photograph show deforestation of Haiti (on the left) at the
border with the Dominican Republic (on the right). Deforestation on the Haitian side of the border is much more severe. Source: NASA
Air, water, and land pollution directly, and adversely, impacts human and ecosystem health. It also has economic consequences from the
damage of natural resources and human structures. In many cases the level of pollutants can be measured either in the environment or at the
point of emissions. Additional indicators include whether pollution monitoring even occurs, to what extent legal maximum levels are
enforced, and whether regulations are in place to clean up the damage. Visit the EPA's MyEnvironment application to learn more about
environmental issues in your area.
Greenhouse gas emissions and ozone depletion are results of air pollution, but are frequently placed in a separate category because they have
global impacts regardless of the source of the problem. Levels of greenhouse gases and ozone-depleting substances in the atmosphere can be
measured directly, or their impacts can be measured by looking at temperature change and the size of the ozone hole. However, those
methods are rarely part of EPIs because they do not assign a particular source. Instead, EPIs include the actual emissions by a particular
process or area.
Examples of EPIs
There are dozens of EPIs that can be used to evaluate sustainability. Below are two examples of multi-component methods that allow
comparisons at a national level, which is necessary for promoting many types of systemic change.
9.3.18 [Link]
Arab Emirates, all initiated major internal reviews that resulted in the initiation of efforts to improve environmental sustainability. Because
ESI data are presented not only as an overall average but also as 21 independent indicators, countries can focus their efforts where most
improvement could be made. Countries dissatisfied with their rankings have also begun to make more of their environmental data accessible.
Initial rankings by ESI score had missing or estimated data in many cases, but by making more data available, more accurate overall
assessments are possible. For example, the Global Environmental Monitoring System Water Program, an important source of water quality
information, had data contributions increase from less than 40 countries to over 100 as a result of the ESI.
Several similar ranking methodologies have emerged from the ESI. They vary in the number and type of variables included and indicators
produced. Some also calculate an overall average by weighting some indicators more than others. However, they all share the same 0-100
scale and have individual indicators that allow targeted improvement of the overall scores.
Comparisons
There are no perfect measures of sustainability, and different indicators can sometimes give conflicting results. In particular this happens
when perspectives on the most important components of sustainability, and the methods to address them, differ. Therefore, it is often useful to
examine the main characteristics of several Environmental Performance Indicators to find the one most appropriate for a particular study. As
an example, ESIs, EMPIs, and ecological footprinting (discussed in a previous section) are compared below.
Ecological footprinting (EF) has units that are the easiest to understand – area of land. Both EF and EMPI employ only a single type of unit,
allowing for use of absolute variables and permitting quantitative comparisons. However, EF does not use multiple indicators to allow for
focused attention on impacts. EMPI can also be used as scaled values (such as the proportion of emergy from renewable sources), in the same
manner as ESI. However, ESI combines multiple units of measurements, which can provide a more holistic perspective, but at the same time
leads to concerns about combining those data.
Of the three, ESI and EMPI take into account wastefulness and recycling, and only ESI includes the effects of all emissions. But while ESI
includes the most variables, it is the most complex to calculate; the simplest to calculate is EF.
Because ESI includes social and economic indicators, it can only compare nations (or in some cases, states or other levels of governments).
EF and EMPI are effective at comparing countries, but can also be used at scales from global down to individual products.
All three of the EPIs compared here can be useful, but each has their limitations. Additionally, there are scenarios where none of these are
useful. Specific environmental education projects, for example, would require different types of performance indicators.
9.3.19 [Link]
The underlying principle of the MDGs is that the world has sufficient knowledge and resources to implement sustainable practices to improve
the life of everyone. Annual progress reports suggest that this principle may be realistic only for some targets. The targets within
environmental sustainability are the implementation of national policies of sustainable development, increased access to safe drinking water,
and improvements to urban housing. There are success stories: efforts to increase availability of clean water have resulted in improvements
faster than expected. However, not all indicators are showing improvement. Impacts of climate change are accelerating, and risks of physical
and economic harm from natural disasters are increasing. Moreover, these impacts and risks are concentrated in poorer countries – those least
able to handle the threats. Overall, results are mixed.
The worldwide rate of deforestation is still high but has slowed. Large-scale efforts to plant trees in China, India, and Viet Nam have resulted
in combined annual increases of about 4 million hectares of forests since 2000. Unfortunately, that is about the same rate of forest loss in
South America and Africa each. Globally, the net loss of forest from 2000 to 2010 was 5.2 million hectares per year, down by a third from the
1990s.
The world will likely meet the target of halving the proportion of people without access to clean water, with the majority of progress made in
rural areas. By 2008 most regions exceeded or nearly met the target levels. The exceptions were Oceania and sub-Saharan Africa, which had
only 50% and 60% respectively of their populations with access to improved water sources. Those regions will almost certainly miss the
target. They, and most other developing regions, will also miss the target of halving the proportion of the population lacking improved
sanitation facilities. In fact, the total number of people without such access is expected to grow until at least 2015.
From 1990 to 2007, emissions of carbon dioxide rose in developed regions by 11%; in developing regions, which have much higher rates of
population growth, emissions increased by 110%. While most indicators have shown either progress or minimal additional harm, carbon
dioxide emissions stand out as one of the significant failures in achieving global sustainability.
Efforts to preserve biodiversity have made only minimal progress. One target was to have 10% of each of the world's terrestrial ecosystem
types protected by 2010; only half were. The proportion of key biodiversity areas protected stagnated in the 2000s, after showing faster
improvements in the 1970s-1990s. As a result, the number of birds and mammals expected to go extinct in the near future has increased.
The environmental sustainability target for urban housing was meant to significantly improve the lives of 100 million slum-dwellers by 2020.
This target differed from most others not only by using a later date of 2020, but by lacking a specified proportion of the population. Setting a
target as an absolute value for the entire globe obscures the progress in individual countries, so this criterion may be revisited. From 1990 to
2010, the proportion of slum-dwellers decreased from 46% to 33%. However, during the same time, the number of people living in slums
increased from 657 million to 828 million. Over 200 million slum-dwellers achieved access to clean water and improved sanitation facilities,
so the target was met. However, it is widely acknowledged that the target was set too low.
Even as we continue to strive toward the MDGs 2015 target date, it is also necessary to think beyond them. Changing demographics will
drive shifts in the global economy and the use of resources. Increased effects of climate change will result in greater volatility, while
technological developments can open new opportunities. In light of these changes, evaluation of the MDGs must assess their utility after
2015. Should the general framework stay in place, should it be modified with new approaches, or should it be replaced with something
fundamentally different?
Resources
Local Harvest [Link]
Illinois Stewardship Alliance [Link]
Slow Food Chicago [Link]
Illinois Farmers Markets, Illinois Department of Agriculture [Link]
Environmental Sustainability Index ([Link]
EPA MyEnvironment ([Link]
References
1. Brown, M.T. & Ulgiati, S. (2002). Emergy evaluations and environmental loading of electricity production systems. Journal of Cleaner
Production, 10, 321-334. doi:10.1016/S0959-6526(01)00043-9
2. Constanza, R., d'Arge, R., de Groot, R., Farber, S., Grasso, R., Hannon, B., et al. (1997). The value of the world's ecosystem services and
natural capital. Nature, 387, 253–260.
3. Dalal-Clayton, B. & Bass, S. (2002). Sustainable development strategies: A resource book. London: Earthscan Publications Ltd.
4. Gasparatos, A., El-Haram, M., & Horner, M. (2008). A critical review of reductionist approaches for assessing the progress towards
sustainability. Environmental Impact Assessment Review, 28, 286-311. doi:10.1016/[Link].2007.09.002
5. Hammond, A., Adriaanse, A., Rodenburg, E., Bryant, D., & Woodward, R. (1995). Environmental indicators: a systematic approach to
measuring and reporting on environmental policy performance in the context of sustainable development. World Resources Institute.
Retrieved March 7, 2011, from [Link]
9.3.20 [Link]
6. Hermann, B.G., Kroeze, C., & Jawjit, W. (2007). Assessing environmental performance by combining life cycle assessment, multi-criteria
analysis and environmental performance indicators. Journal of Cleaner Production, 15, 1878-1796. doi:10.1016/[Link].2006.04.004
7. Krajnc, D. & Glavic, P. (2005). A model for integrated assessment of sustainable development. Resources Conservation & Recycling, 43,
189-203.
8. Lele, S. & Norgaad, R.B. (1996, April). Sustainability and the scientist's burden. Conservation Biology, 10, 354-365. doi 10.1046/j.1523-
1739.1996.10020354.x
9. Lozano, R. (2006). A tool for graphical assessment of sustainability in universities. Journal of Cleaner Production, 14, 963-972. doi:
10.1016/[Link].2005.11.041
10. Mayer, A.L., Thurston, H.W. & Pawlowski, C.W. (2004). The multidisciplinary influence of common sustainability indices. Ecological
Environment, 2, 419-426. doi:10.1890/1540-9295(2004)002[0419:TMIOCS][Link];2
11. Nardo, M., Saisana, M., Saltelli, A, Tarantola S., Hoffman, A. & Giovannini, E. (2005). Handbook on constructing composite indicators:
Methodology and user guide (STD/DOC (2005)3). Paris: OECD Statistics Directorate. Retrieved February 26, 2011, from
[Link]/dataoecd/37/42/[Link]
12. Odum, H.T. (1996). Environmental Accounting, Emergy, and Environmental Decision Making. New York: John Wiley & Sons, Inc.
13. Pearce, D.W. & G.D. Atkinson (1993). Capital theory and the measurement of sustainable development: an indicator of “weak”
sustainability. Ecological economics, 8, 103-108.
14. Rees, W.E. & Wackernagel, M. (1996). Urban ecological footprints: why cities cannot be sustainable and why they are a key to
sustainability. Enviromental Impact Assessment Review, 16, 223-248.
15. Reporting Framework Overview. (n.d.). Global Reporting Initiative. Retrieved August 21, 2011, from
[Link]/Repor...eworkOverview/
16. Sahely, H.R., Kennedy, C.A. & Adams, B.J. (2005). Developing sustainability criteria for urban infrastructure systems. Canadian Journal
of Civil Engineering,32, 72-85. doi: 10.1139/l04-072
17. Troschinetz, A.M., Mihelcic, J.R., & Bradof, K.L. (2007). Developing Sustainability Indicators for a University Campus. International
Journal of Engineering Education, 23, 231-241.
18. Chambers, N., Simmons, C. & Wackernagel, M. (2000). Sharing Nature’s Interest: Ecological Footprints as an Indicator of Sustainability.
London: Earthscan Publications Ltd.
19. Chapagain, A.K. & Hoekstra, A.Y. (2007). The water footprint of coffee and tea consumption in the Netherlands. Ecological Economics,
64, 109-118.
20. Constanza, R. (2000). The dynamics of the ecological footprint concept. Ecological Economics,32, 341-345.
21. Dawe, G.F.M., Vetter, A. & Martin. S. (2004). An overview of ecological footprinting and other tools and their application to the
development of sustainability process. International Journal of Sustainability in Higher Education, 4, 340-371.
22. Flint, K. (2001). Institutional ecological footprint analysis: A case study of the University of Newcastle, Australia. International Journal of
Sustainability in Higher Education, 2, 48-62.
23. Hoekstra,Y. & Chapagain, A. K. (2007). Water footprints of nations: Water use by people as a function of their consumption pattern.
Water Resour Manage, 21, 35-48.
24. Klein-Banai, C. (2007). Greenhouse gas inventory for the University of Illinois at Chicago. UIC GHG Inventory,Retrieved November 21,
2009 from [Link]
25. Rees, W.E. and Wackernagal, M. (1996). Urban ecological footprints and why cities cannot be sustainable – and why they are a key to
sustainability. Environmental Impact Assessment Review, 16, 223-248.
26. Venetoulis, J. (2001). Assessing the ecological impact of a university: The ecological footprint for the University of Redlands.
International Journal of Sustainability in Higher Education, 2, 180-196.
27. Wackernagel, M., Monfreda, C., Moran, D., Wermer, P., Goldfinger, S., Deumling, D., & Murray, M. (2005, May 25). National footprint
and biocapacity accounts 2005: The underlying calculation method. Global Footprint Network. Retrieved March 2, 2010 from
[Link]/[Link]?id=5.
28. World Bank (2011). About Us. Retrieved September 20, 2011 from [Link]/3QT2P1GNH0.
29. Klein-Banai, C, Theis, T.L., Brecheisen, T.A. & Banai, A. (2010). A Greenhouse Gas Inventory as a Measure of Sustainability for an
Urban Public Research University, Environmental Practice, 12, 25-47.
30. Klein-Banai, C & Theis, T.L. (2011). An urban university's ecological footprint and the effect of climate change, Ecological Indicators,
11, 857–860.
31. UIC Office of Sustainability. (2011). State of Sustainability University of Illinois at Chicago Biennial Report. Retrieved May 30, 2011
from [Link]
32. Greenpeace (2011). Greenwashing. Greenpeace. Retrieved December 17, 2011 from [Link]/
Review Questions
1. What is the difference between data and an index?
2. What is the major challenge in measuring sustainability?
9.3.21 [Link]
3. Give three general categories of indicators that are used for measuring sustainability and provide one example of each.
4. Why is it important to have experts provide input to rating systems?
5. Choose a calculator from the box and calculate your own footprint. How does it compare to the national or global average? What can you
do to reduce your footprint?
6. Discuss what kind of inequities the various footprints represent between nations and the types of inequities.
7. How might the “food print” of a vegetarian differ from a carnivore?
8. What are some of the problems with comparing food miles for a cheeseburger to those for a vegetarian salad?
9. Why might food producers in isolated but prosperous areas (like Hawaii or New Zealand) argue against the use of food miles?
10. Do you think increased reliance on food miles is good or bad for rural areas in developing countries? Explain your decision.
11. What is the difference between energy and emergy?
12. In what way(s) is ESI a better method of assessing sustainability than EF and EMPI?
13. The ESI creates indicators in five areas. In which of the areas do you think the indicators are the least reliable?
14. Why do EPIs require multiple data points to be useful?
Glossary
Carrying Capacity
The maximum population that a given environment can sustain.
Ecosystem
All living organisms and non-living things that exist and interact in a certain area at the same time.
Ecosystem Goods and Services
An essential service an ecosystem provides that supports life and makes economic activity possible. For example, ecosystems clean air and
water naturally, recycle nutrients and waste generated by human economic activity.
Emergy
The amount of energy of one kind (solar) that has been used directly or indirectly (through a transformation process) to make a service or a
product as one type and it is expressed in units of (solar) emjoule.
Emjoule
The unit of emergy or emergy joule. Using emergy, sunlight, fuel, electricity, and human service can be put on a common basis by expressing
each of them in the emjoules of solar energy that is required to produce them. If solar emergy is the baseline, then the results are solar
emjoules (abbreviated seJ). Sometimes other baselines such as coal emjoules or electrical emjoules have been used but in most cases emergy
data are given in solar emjoules.
Entropy
The degree of disorder in a substance, system or process as in the second law of thermodynamics that states that the make-up of energy tends
to change from a more-ordered state to a less-ordered state, whereby increasing entropy.
Exergy
The maximum work that can be extracted from a system as it moves to thermodynamic equilibrium with a reference state.
Indicator
A variable equal to an operational representation of an attribute of a system.
Indicator-Based Systems
Systems that use quantitative measures of economic progress, social welfare, or environmental activity that can be interpreted to explain the
state of that system. Examples of these are gross domestic product, greenhouse gas emissions, and the unemployment rate.
Maximum Sustainable Yield (MSY)
An outgrowth of carrying capacity and the goal is to reach the maximum amount of resource extraction while not depleting the resource from
one harvest to the next.
Narrative Assessments
Descriptive documentation of a program, plan, or project.
Quantitative Data
Information that can be quantified numerically such as tons of waste, gallons of gasoline, and gallons of wastewater.
9.3.22 [Link]
Resilience
The ability of an ecological community to change in response to disturbance and the degree or time needed for that system that provides
desirable to go back to its original state.
Anthropogenic
Relating to or resulting from the influence that humans have on the natural world.
Cradle-to-Grave
From creation to disposal; throughout the life cycle.
Ecological Footprint (EF)
Represents the area of land on earth that provides for resources consumed and assimilates the waste produced by a given entity.
Global Warming Potential (GWP)
Each gas, based on its atmospheric chemistry, captures different amounts of reflected heat thus contributing differently to the greenhouse
effect contributing to its GWP. Carbon dioxide, the least capture efficient of these gases, acts as the reference gas with a global warming
potential of 1.
Gross Domestic Product
The sum of gross value added by all resident producers in the economy plus any product taxes and minus any subsidies not included in the
value of the products. It is calculated without making deductions for depreciation of fabricated assets or for depletion and degradation of
natural resources.
Sequestered
Removed from the atmosphere
Triple Bottom Line
Accounting for ecological and social performance in addition to financial performance)
Emergy (EMbodied energy)
The unit of energy into which any resource, product, or process can be converted to simplify comparisons between diverse items.
Emergy Performance Index (EMPI)
Value produced by converting all materials and processes to amounts of energy in order to evaluate renewability and sustainability.
Environmental Sustainability Index (ESI)
A composite value produced by including ecological, social, economic, and policy data.
Environmental Performace Indicators (EPI)
Any of the ways in which environmental outcomes and/or impacts can be assessed.
Impacts
Long-term and more widespread results of an activity.
Inputs
The specific resources or services used by an activity.
Outcomes
The short-term results of an activity.
Outputs
The goods and services being created by an activity, and the manner and degree in which they are delivered.
This page titled 9.3: Derivative Life Cycle Concepts is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto Cabezas
(GALILEO Open Learning Materials) .
9.3.23 [Link]
9.4: Sustainability and Business
Introduction
Throughout this text the integrative nature of environmental, social, and economic sustainability has been stressed. In this chapter,
various ways of framing the sustainability paradigm and measuring progress toward its achievement have been presented. This
section focuses more directly on businesses, and how they attempt to incorporate sustainability into their decisions and plans. The
business sector, continually seeking ways to create competitive advantages, has become acutely aware of the general value of
adjusting various business models to accommodate consumers’ desires for sustainable products and services. Still, the broad
definition of sustainable development provided by the Brundtland report is a difficult one to make operational. The World Business
Council for Sustainable Development has adapted Brundtland to a view more understandable to business interests, focusing on
living within the “interest” of natural systems and being cautious about drawing down the “principal” (i.e. degrading natural
ecosystems), but there remain substantial differences on precisely how to measure progress toward the goals of the sustainability
paradigm.
It is a common practice for businesses to refer to the triple bottom line, a reference to the value of a business going beyond dollar
profitability to include social and environmental costs and benefits as well. Indeed, many of the tools and indices outlined in
Module 11.2 and Module 11.3.1 are widely used by businesses to measure progress toward corporate goals. However, there is no
agreed upon way of using these tools, and many businesses have developed their own methods for assessing progress. This has,
inevitably perhaps, led to claims and counter-claims by various parties about the “sustainability” of their products or services. Such
claims usually find their way into corporate brochures and advertising so that, often without substantive backing or very subjective
analysis, the impression of significant corporate sustainability is created, practices known generally as greenwashing.
Greenwashing is a concern because these kinds of advertising messages can mislead consumers about the “the environmental
practices of a company or the environmental benefits of a product or service” (Greenpeace, 2011). Nevertheless, businesses must
ultimately generate profits to remain viable, and increasingly they are being held to account for their impacts on all aspects of
business operations, however difficult it may be to assign value to decisions made under conditions of considerable uncertainty.
The intergenerational mandate of Brundtland and the nature of modern environmental problems facing society ask that business
plans extend far beyond the usual five to ten year range.
9.4.1 [Link]
develop workable models that could be used to control and optimize material and energy flows, ensure product quality, manage
environmental impacts, and minimize costs (these functions are collectively referred to as the supply chain). An expanded use of
LCA incorporates the complete product chain, examining consumer uses, benefits and costs, and the post-consumer disposition of
the product. This has led to product conceptualization and development, and in some cases regulatory reform, that incorporate
business practices and plans built upon the concept of eco-efficiency, and extended product/producer responsibility (EPR). Eco-
efficiency is an evolutionary business model in which more goods and services are created with less use of resources, and fewer
emissions of waste and pollution. Extended product/producer responsibility involves the creation of financial incentives, and legal
disincentives, to encourage manufacturers to make more environmentally friendly products that incorporate end-of-life costs into
product design and business plans. For example one business model that is conducive to EPR is a “lease-and-take-back” model in
which products must eventually come back to the manufacturer or retailer, who then must reckon with the best way to minimize
end-of-life costs. Remanufacturing, recycling, and reuse of materials are the intended results of EPR, but ordinary disposal,
including landfilling or incineration, can also be an option.
Figure 9.4.1, illustrates in a general way how the LCA framework can be structured for understanding how product development
can benefit from the various material and information transfers and feedback loops along the product chain. Such a figure illustrates
the complexities involved in creating, marketing, and discerning the impacts of a product or service, and raises the general concept
of what is often referred to as product stewardship, an approach in which products are conceived, designed, manufactured, and
marketed within a systems thinking context. It is a way of framing environmental problems that recognizes the three parts of the
sustainability paradigm, and incorporates the concepts of sustainable manufacturing, marketing, utility-to-society, impacts of the
use of the product, and end-of-life disposition of the product.
LCA Framework Applied to Product Development
Figure 9.4.1 LCA Framework Applied to Product Development. Generalized view of the life cycle product chain, showing major
material and information transfers and feedback loops. Source: Tom Theis
Creating Uniformity
The problem of lack of uniformity in measuring, assessing, and valuing business actions taken at least in part for the sake of
sustainability might be dealt with more effectively through the development of uniform standards and metrics that are applied by an
agreed upon authority who uses transparent methodologies and reporting techniques so that other companies, and consumers, can
make more objective judgments about comparative performances. From what has been presented in this section this may appear to
be a near-impossible task. Yet attempts in this direction are being made, for example by the aforementioned World Business
Council for Sustainable Development, the Organization for Economic Cooperation and Development, and the United Nations
Millennium Development Goals. One of the more popular approaches for measuring and ranking corporate sustainability has been
developed by the Dow Jones Corporation (DJC), through its Sustainability Index (DJSI). It may seem ironic that such a bastion of
the free market economy has put together a system for measuring and assessing corporate sustainability, yet the size and general
acceptability of DJC by corporations and investors work in favor of the establishment of an objective and transparent index. The
DJSI itself was created in 1999 in response to the need, articulated from many sectors including consumers, for a way to assess
progress toward sustainable corporate responsibility. The index contains three general evaluative sectors – economic, social, and
environmental – that reflect the Brundtland definition. Each sector is composed, in turn, of specific categories as follows:
Economic
Codes of Conduct/Compliance/Corruption and Bribery
Corporate Governance
Risk and Crisis Management
Industry-specific Criteria
Social
Corporate Citizenship and Philanthropy
Labor Practice Indicators
Human Capital Development
Social Reporting
Talent Attraction and Retention
Industry-specific Criteria
Environmental
9.4.2 [Link]
Environmental Performance (Eco-efficiency)
Environmental Reporting
Industry-specific Criteria
Each of these categories is composed of quantitative measures and assigned a specific, and constant, weighting. From the data
gathered, a “best-in-class” (i.e. industry class) ranking is published annually. The index has engendered considerable corporate
competition such that mere attainment of the previous year’s statistics, for a given company, usually results in a drop in rank. Of
course one can argue with the choice of categories, or the data that are gathered and the way categories are parameterized, or with
the weighting scheme used, but the important aspects of DJSI (and other sustainability rankings) is its comprehensiveness,
uniformity, and transparency.
Summary
In the final analysis, no economy can move in the direction of sustainability without the active participation of the business sector.
In other words, significant progress cannot be achieved through government or individual actions alone. As noted above, this
creates difficulties and conflicts for businesses. As they continue to work together in the future, businesses and sustainability
experts face many questions such as: What are the best measures of sustainability and how should businesses develop and plan for
delivering more sustainable products and services? Is reliance on eco-efficiency enough to reduce the impacts of increasing
consumption? Should businesses play a more significant role in educating consumers on the factors that affect sustainable
development? How can businesses adapt to uncertainties that lie beyond the near term? What is the role of government in
overseeing or regulating business activities that contribute to sustainability?
References
1. Greenpeace (2011). Greenwashing. Greenpeace. Retrieved December 17, 2011 from [Link]/
Review Questions
1. Find a product chain for the manufacture of a major consumer item such as a flat screen television, a computer, or an
automobile and cast the stages of the chain in life cycle form as shown in Figure LCA Framework Applied to Product
Development. As part of your answer, define the various information transfers and feedback loops involved.
2. Consider the various types of costs in the total cost accounting framework. In proceeding from Type I to Type V, give reasons
why uncertainties usually increase at each level?
3. What are the main attributes of a sound index for measuring progress toward sustainability of products and services?
Glossary
Eco-efficiency
An evolutionary business model in which more goods and services are created with less use of resources, and fewer emissions of
waste and pollution.
End-of-Life Costs
Those costs that arise through activities associated with the disposition of a product at the end of its useful life. These include costs
associated with disposal, recycling, reuse, and remanufacturing.
Extended Product/Producer Responsibility
The creation of financial incentives, and legal disincentives, to encourage manufacturers to make more environmentally friendly
products that incorporate end-of-life costs into product design and business plans.
Greenwashing
Claims made by businesses about the superior contributions of their products and services to sustainability without substantive
backing or via a very subjective analysis.
Product Chain
Those stages in the conception, design, manufacture, marketing, use, and end-of-life that define the impacts of a product or service
on society.
Product Stewardship
9.4.3 [Link]
An approach to product development in which products are conceived, designed, manufactured, and marketed within a “systems
thinking” context. It is a way of framing environmental problems that recognizes the three parts of the sustainability paradigm, and
incorporates the concepts of sustainable manufacturing, marketing, utility-to-society, impacts of the use of the product, and end-of-
life disposition of the product.
Systems Thinking
In the context of sustainability, systems thinking is a way of conceiving human-created and natural systems as functional parts of a
larger, integrated system.
Triple Bottom Line
A reference to the value of a business going beyond dollar profitability to include social and environmental costs and benefits as
well.
9.4: Sustainability and Business is shared under a not declared license and was authored, remixed, and/or curated by LibreTexts.
9.4.4 [Link]
CHAPTER OVERVIEW
This page titled 10: Sustainability - Ethics, Culture, and History is shared under a CC BY license and was authored, remixed, and/or curated by
Heriberto Cabezas (GALILEO Open Learning Materials) .
1
10.1: The Human Dimensions of Sustainability- History, Culture, Ethics
10.1.1 [Link]
human community only a few times over the past five millennia. The goal of sustainability, in these terms, is clear and non-
controversial: to avoid a new and scaled-up Dark Age in which the aspirations of billions of people, both living and yet unborn,
face brutal constraints. The implications of sustainability, in this sense, extend well beyond what might ordinarily considered
“green” issues, such as preserving rainforests or saving whales. Sustainability is a human and social issue as much as it is
“environmental.” Sustainability is about people, the habitats we depend on for services vital to us, and our ability to maintain
culturally rich civic societies free from perennial crises in food, water, and energy supplies.
Glossary
Adaptation
Focuses on the need for strategies to deal with the climate change that is unavoidable because of increased carbon already in the
atmosphere
Mitigation
Refers to the importance of reducing carbon emissions so as to prevent further, catastrophic changes in the climate system.
This page titled 10.1: The Human Dimensions of Sustainability- History, Culture, Ethics is shared under a CC BY license and was authored,
remixed, and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
10.1.2 [Link]
10.2: It’s Not Easy Being Green- Anti-Environmental Discourse, Behavior, and
Ideology
Introduction
The consensus view among scientists and professional elites in the early twenty-first century, as it has been among environmental
activists for a much longer time, is that our globalized industrial world system is on an unsustainable path. Inherent in this view is a
stern judgment of the recent past: we have not adapted well, as a species, to the fruits of our own brilliant technological
accomplishments, in particular, to the harnessing of fossil fuels to power transport and industry.
Taking the long view of human evolution, it is not surprising to find that we are imperfectly adapted to our modern industrialized
world of cars, computers, and teeming cities, or that human societies organized for so many millennia around the problem of
scarcity should treat a sudden abundance of resources with the glee of a kid in a candy store. In evolutionary terms, we have simply
not had sufficient time to adapt to the windfall of change. Though rapid advances in the biophysical sciences in recent decades
mean that we mostly understandour maladaptation to industrialization and the great dangers it poses, our political decision-making
and consumption patterns have barely changed on the basis of this understanding. This sobering fact tells us that, at this moment in
human history, social behavior and political decision-making are not being driven by knowledge, but rather by entrenched attitudes
that perpetuate an unsustainable drawdown of earth’s resources. In short, human decision making and consumption of material
goods in our fossil-fuel age continues to largely take place outside of an awareness of the strained and finite nature of our planet’s
ecosystem services.
It is the character of modern consumer society to promote the idea that nothing is connected, that the jeans we wear, or the food we
eat, are matters of personal choice without any greater context beyond a concern for immediate pleasure and peer approval.
Sustainability, by contrast, teaches that everything is connected. That favorite pair of jeans, for instance, is dependent on cheap
labor in developing countries, on heavily fertilized cotton plantations, and enormous volumes of water expended throughout the
jeans’ lifecycle, from the irrigation to grow the cotton to the washing machine that cleans them. Or let’s take that “cheap” fast food
lunch from yesterday: it most likely contained processed soybeans from a recently cleared stretch of the Amazon rainforest, and
artificial sweeteners made from corn whose enormous production quotas are subsidized by government tax revenues. The corn-
based sweetener, in turn, turns out to be a principal cause of the national obesity epidemic, a key contributor to spiraling health care
costs. Thus the “value meal” turns out not to be so economical after all, once the systems-wide effects are factored in.
A twenty minute video, The Story of Stuff, tells the complicated story of how our "stuff" moves from extraction to sale to disposal.
an image of fast food
Figure 10.2.2 Fast Food Industry's Environmental Impact? Here’s food for thought. Though we are accustomed to measuring the
impact of a fast food diet on our physical health, there is much less readily available information on the global network of
agricultural providers that supports the fast food industry, and on its environmental impacts on land use, water resources, and
human communities. Source: Created by CrazyRob926
Connectivity
To think about sustainability in these terms may sound exhausting. But because we live in a world characterized by connectivity,
that is, bycomplex chains linking our everyday lives to distant strangers and ecosystems in far flung regions of the earth, we have
no choice. In the end, we must adapt our thinking to a complex, connected model of the world and our place in it. Persisting with
only simple, consumerist frames of understanding—“I look great!” “This tastes delicious!”—for a complex world of remote
impacts and finite resources renders us increasingly vulnerable to episodes of what ecologists call system collapse, that is, to the
sudden breakdown of ecosystem services we rely upon for our life’s staple provisions.
In the early twenty-first century, vulnerability to these system collapses varies greatly according to where one lives. A long-term
drought in India might bring the reality of aquifer depletion or climate change home to tens of thousands of people driven from
their land, while the life of a suburban American teenager is not obviously affected by any resource crisis. But this gap will narrow
in the coming years. Overwhelming scientific evidence points to rapidly increasing strains this century on our systems of food,
water, and energy provision as well as on the seasonable weather to which we have adapted our agricultural and urban regions. In
time, no one will enjoy the luxury of remaining oblivious to the challenges of sustainability. Drought, for example, is one of the
primary indices of global ecosystem stress, and arguably the most important to humans. According to the United Nations Food and
Agriculture Organization, without wholesale reformation of water management practices on a global scale, two-thirds of the
world’s population will face water shortages by 2025, including densely populated regions of the United States.
10.2.1 [Link]
So how did we arrive at this point? Without you or I ever consciously choosing to live unsustainably, how has it nevertheless come
about that we face environmental crises of global scale, circumstances that will so decisively shape our lives and those of our
children? Here’s one explanatory narrative, framed by the long view of human evolution.
Since the emergence of the first proto-human communities in Africa millions of years ago, we have spent over 99% of evolutionary
time as nomadic hunters and gatherers. A fraction of the balance of our time on earth spans the 10,000 years of human agriculture,
since the end of the last Ice Age. In turn, only a third of that fractional period has witnessed the emergence of the institutions and
technologies—writing, money, mathematics, etc.—that we associate with human “civilization.” And lastly, at the very tip of the
evolutionary timeline, no more than a blink of human species history, we find the development of the modern industrialized society
we inhabit. Look around you. Observe for a moment all that is familiar in your immediate surroundings: the streetscape and
buildings visible through the window, the plastic furnishings in the room, and the blinking gadgets within arm’s length of where
you sit. All of it is profoundly “new” to human beings; to all but a handful of the tens of thousands of generations of human beings
that have preceded us, this everyday scene would appear baffling and frightening, as if from another planet.
Normalization
In this sense, the real miracle of human evolution is that cars, computers, and cities appear so normal to us, even sometimes
“boring” and monotonous! Our perception of the extraordinary, rapid changes in human societies in the past two centuries—even
the past half-century—is deadened by virtue of what is our greatest evolutionary acquirement, namely normalization, an adaptive
survival strategy fundamental to human success over the millennia. The ability to accept, analyze, and adapt to often fluctuating
circumstances is our great strength as a species. But at this point in human history it is also a grave weakness, what, in the language
of Greek tragedy might be called a “fatal flaw.”
To offer an analogy, for many centuries slavery appeared normal to most people across the world—until the late eighteenth century,
when a handful of humanitarian activists in Britain began the long and difficult process of de-normalizing human bondage in the
eyes of their compatriots. The task of sustainability ethics is analogous, and no less difficult, in that it lays out the argument for
wholesale and disruptive attitude adjustment and behavior change in the general population. Given the long-term adaptation of the
human species to the imperatives of hunter-gathering, our decision-making priorities and consumption drives still tend toward the
simple necessities, based on the presumption of relative and seasonal scarcity, and with little emotional or social reward for
restraint in the face of plenty, for viewing our choices in global terms, or for measuring their impacts on future generations.
A working distinction between the historical evolution of human society and human culture is useful to understanding the social
and psychological obstacles to achieving sustainability. As both individuals and societies, we work hard to insulate ourselves from
unpleasant surprises, shocks, and disorder. We crave “security,” and our legal and economic institutions accordingly have evolved
over the millennia to form a buffer against what Shakespeare’s Hamlet called “the thousand natural shocks that flesh is heir to.” For
instance, the law protects us from violent physical harm (ideally), while insurance policies safeguard us from financial ruin in the
event of an unexpected calamity.
In one sense, this security priority has determined the basic evolution of human societies, particularly the decisive transition 10,000
years ago from the variable and risky life of nomadic hunter communities to sedentary farming based on an anticipated stability of
seasonal yields. Of course, the shift to agriculture only partially satisfied the human desire for security as farming communities
remained vulnerable to changing climatic conditions and territorial warfare. Global industrialization, however, while it has
rendered vast populations marginal and vulnerable, has offered its beneficiaries the most secure insulation yet enjoyed by humans
against “the slings and arrows of outrageous fortune.” This success has been a double-edged sword, however, not least because the
industrialized cocoon of our modern consumer lifestyles relentlessly promotes the notion that we have transcended our dependence
on the earth’s basic resources. As it stands, we look at our highly complex, industrialized world, and adapt our expectations and
desires to its rewards. It is never our first instinct to ask whether the system of rewards itself might be unsustainable and collapse at
some future time as a result of our eager participation.
10.2.2 [Link]
Luckily, our cultural institutions have evolved to offer a counterweight to the complacency and inertia encouraged by the other
simple, security-focused principles governing our lives. If society is founded upon the principle of security, and promotes our
complacent feeling of independence from the natural world, we might think of culture as the conscience of society. What culture
does, particularly in the arts and sciences, is remind us of our frailty as human beings, our vulnerability to shocks and sudden
changes, and our connectedness to the earth’s natural systems. In this sense, the arts and sciences, though we conventionally view
them as opposites, in fact perform the same social function: they remind us of what lies beyond the dominant security paradigm of
our societies—which tends to a simplified and binary view of human being and nature—by bringing us closer to a complex,
systemic understanding of how the natural world works and our embeddedness within it. Whether by means of an essay on plant
biology, or a stage play about family breakdown (like Hamlet), the arts and sciences model complex worlds and the systemic
interrelations that shape our lives. They expose complexities and connectivities in our world, and emphasize the material
consequences of our actions to which we might otherwise remain oblivious. The close relation between the arts and sciences in the
Western world is evidenced by the fact that their concerns have largely mirrored each other over time, from the ordered,
hierarchical worldview in the classical and early modern periods, to the post-modern focus on connectivity, chaos, and emergence.
Life in the pre-modern world, in the memorable words of the English philosopher Thomas Hobbes, was mostly “nasty, brutish, and
short.” By contrast, social and economic evolution has bestowed the inhabitants of the twenty-first century industrialized world
with a lifestyle uniquely (though of course not wholly) insulated from physical hardship, infectious disease, and chronic violence.
This insulation has come at a cost, however, namely our disconnection from the basic support systems of life: food, water and
energy. This is a very recent development. At the beginning of the 20th century, for example, almost half of Americans grew up on
farms. Now, fewer than two percent do. We experience the staples of life only at their service endpoints: the supermarket, the
faucet, the gas station. In this context, the real-world sources of food, water, and energy do not seem important, while supplies
appear limitless. We are not prepared for the inevitable shortages of the future.
On the positive side, it is possible to imagine that the citizens of the developed world might rapidly reconnect to a systems view of
natural resources. One product of our long species evolution as hunters and agricultural land managers is an adaptive trait the
ecologist E. O. Wilson has called “biophilia,” that is, a love for the natural world that provides for us. In the few centuries of our
fossil fuel modernity, this biophilia has become increasingly aestheticized and commodified—as landscape art, or nature tourism—
and consequently marginalized from core social and economic decision structures. In the emerging age of environmental decline
and resource scarcity, however, our inherited biophilia will play a key role in energizing the reform of industrialized societies
toward a sustainable, renewable resource and energy future.
Review Questions
1. How has the human capacity for normalization both helped and hindered social development, and what are its implications for
sustainable reform of our industries, infrastructure, and way of life?
2. Take an everyday consumer item—running shoes, or a cup of coffee—and briefly chart its course through the global consumer
economy from the production of its materials to its disposal. What are its environmental impacts, and how might they be
reduced?
Glossary
Connectivity
An important feature of complex systems. Connections exist between even apparently remote and disparate things. For example,
drought in Australia might impact the price of bread in Egypt, which in turn has repercussions for U.S. foreign policy.
Normalization
An acquired evolutionary trait characteristic of human beings, whereby even radical changes are quickly adapted to and represented
as normal.
This page titled 10.2: It’s Not Easy Being Green- Anti-Environmental Discourse, Behavior, and Ideology is shared under a CC BY license and
was authored, remixed, and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
10.2.3 [Link]
10.3: The Industrialization of Nature- A Modern History (1500 to the present)
Introduction
It is a measure of our powers of normalization that we in the developed world take the existence of cheap energy, clean water,
abundant food, and international travel so much for granted, when they are such recent endowments for humanity, and even now
are at the disposal of considerably less than half the global population. It is a constant surprise to us that a situation so “normal”
could be having such abnormal effects on the biosphere—degrading land, water, air, and the vital ecosystems hosting animals and
fish. How did we get here? How can we square such apparent plenty with warnings of collapse?
Population growth graph
Figure 10.3.1 Graph showing the rapid increase in human population since the beginning of the Industrial Age, with exponential
rise since the mid-twentieth century. Source: IGBP synthesis: Global Change and the Earth System, Steffen et al 2004
Raw figures at least sketch the proportions of global change over the last 500 years. In 1500, even after several centuries of rapid
population growth, the global population was no more than 500 million, or less than half the population of India today. It is now
fourteen times as large, almost 7 billion. Over the same period, global economic output has increased 120 times, most of that
growth occurring since 1820, and with the greatest acceleration since 1950. This combination of rampant population and economic
growth since 1500 has naturally had major impacts on the earth’s natural resources and ecosystem health. According to the United
Nations Millennium Ecosystem Assessment, by the beginning of the 21st century, 15 of the world’s 24 ecosystems, from rainforests
to aquifers to fisheries, were rated in serious decline.
Economic Development
Fundamental to significant changes in human history has been social reaction to resource scarcity. By 1500, Europeans, the first
engineers of global growth, had significantly cleared their forests, settled their most productive agricultural lands, and negotiated
their internal borders. And yet even with large-scale internal development, Europe struggled to feed itself, let alone to match the
wealth of the then dominant global empires, namely China and the Mughal States that stretched from the Spice Islands of Southeast
Asia to the busy ports of the Eastern Mediterranean. As a consequence of resource scarcity, European states began to sponsor
explorations abroad, in quest initially for gold, silver, and other precious metals to fill up their treasuries. Only over time did
Europeans begin to perceive in the New World the opportunities for remote agricultural production as a source of income. Full-
scale colonial settlement was an even later idea.
The new “frontiers” of European economic development in the immediate pre-industrial period 1500-1800 included tropical
regions for plantation crops, such as sugar, tobacco, cotton, rice, indigo, and opium, and temperate zones for the cultivation and
export of grains. The seagoing merchants of Portugal, France, Spain, Britain and the Netherlands trawled the islands of the East
Indies for pepper and timber; established ports in India for commerce in silk, cotton and indigo; exchanged silver for Chinese tea
and porcelain; traded sugar, tobacco, furs and rice in the Americas; and sailed to West Africa for slaves and gold. The slave trade
and plantation economies of the Americas helped shift the center of global commerce from Asia to the Atlantic, while the new
oceangoing infrastructure also allowed for the development of fisheries, particularly the lucrative whale industry. All these
commercial developments precipitated significant changes in their respective ecosystems across the globe—deforestation and soil
erosion in particular—albeit on a far smaller scale compared with what was to come with the harnessing of fossil fuel energy after
1800.
The 19th century witnessed the most rapid global economic growth seen before or mostly since, built on the twin tracks of
continued agricultural expansion and the new “vertical” frontiers of fossil fuel and mineral extraction that truly unleashed the
transformative power of industrialization on the global community and its diverse habitats. For the first time since the human
transition to agriculture more than 10,000 years before, a state’s wealth did not depend on agricultural yields from contiguous
lands, but flowed rather from a variety of global sources, and derived from the industrialization of primary products, such as cotton
textiles, minerals and timber. During this period, a binary, inequitable structure of international relations began to take shape, with a
core of industrializing nations in the northern hemisphere increasingly exploiting the natural resources of undeveloped periphery
nations for the purposes of wealth creation.
10.3.1 [Link]
Trade Map, Late 20th Century
Figure 10.3.2 Trade Map, Late 20th Century. This map shows the “core” industrialized nations of the northern hemisphere, and the
“periphery” nations of the tropics and south dependent on subsistence agriculture and natural resource extraction. This unequal
relationship is the product of hundreds of years of trade and economic globalization Source: Created by Naboc1, based on a list in
Christopher Chase-Dunn, Yukio Kawano and Benjamin Brewer, Trade Globalization since 1795, American Sociological Review,
2000 February, Vol. 65
Soil Degradation
Since the transition to agriculture 10,000 years ago, human communities have struggled against the reality that soil suffers nutrient
depletion through constant plowing and harvesting (mostly nitrogen loss). The specter of a significant die-off in human population
owing to stagnant crop yields was averted in the 1970s by the so-called “Green Revolution,” which, through the engineering of new
crop varieties, large-scale irrigation projects, and the massive application of petroleum-based fertilizers to supplement nitrogen,
increased staple crop production with such success that the numbers suffering malnutrition actually declined worldwide in the last
two decades of the 20th century, from 1.9 to 1.4 billion, even as the world’s population increased at 100 times background rates, to
6 billion. The prospects for expanding those gains in the new century are nevertheless threatened by the success of industrial
agriculture itself. Soil depletion, declining water resources, and the diminishing returns of fertilizer technology—all the products of
a half-century of industrial agriculture—have seen increases in crop yields level off. At the same time, growing populations in
developing countries have seen increasing clearance of fragile and marginal agricultural lands to house the rural poor.
It has been estimated that industrial fertilizers have increased the planet’s human carrying capacity by two billion people.
Unfortunately, most of the chemical fertilizer applied to soils does not nourish the crop as intended, but rather enters the
hydrological system, polluting aquifers, streams, and ultimately the oceans with an oversupply of nutrients, and ultimately draining
the oxygen necessary to support aquatic life. As for the impact of fertilizers on soil productivity, this diminishes over time,
requiring the application of ever greater quantities in order to maintain yields.
10.3.2 [Link]
Deforestation
Arguably the biggest losers from 20th century economic growth were the forests of the world’s tropical regions and their non-
human inhabitants. Across Africa, Asia, and the Americas, approximately one-third of forest cover has been lost. Because about
half of the world’s species inhabits tropical rainforests, these clearances have had a devastating impact on biodiversity, with
extinction rates now greater than they have been since the end of the dinosaur era, 65 million years ago. Much of the cleared land
was converted to agriculture, so that the amount of irrigated soils increased fivefold over the century, from 50 to 250m hectares.
Fully 40% of the terrestrial earth’s total organic output is currently committed to human use. But we are now reaching the ceiling of
productive land expansion, in terms of sheer area, while the continued productivity of arable land is threatened by salinity, acidity
and toxic metal levels that have now degraded soils across one third of the earth’s surface, some of them irreversibly.
Global Forest Map
Figure 10.3.3 Global Forest Map. Since the middle of the twentieth century, the global logging industry, and hence large-scale
deforestation, has shifted from the North Atlantic countries to the forests of tropical regions such as Indonesia and the Amazon
Basin in Latin America. This tropical “green belt” is now rapidly diminishing, with devastating consequences for local ecosystems,
water resources, and global climate. Source: NASA
Figure 10.3.4 Mississippi Watershed Map. The catchment area of the Mississippi River covers almost 40% of the U.S. continental
landmass, collecting freshwater from 32 states. Included in the runoff that feeds the river system are large quantities of agricultural
fertilizer and other chemicals that eventually drain into the Gulf of Mexico, creating an ever-growing “dead zone.” Source:
Environmental Protection Agency
Air Pollution
In many parts of the world, pollution of the air by industrial particles is now less a problem than it was a century ago, when
newspapers lamented the “black snow” over Chicago. This is due to concerted efforts by a clean air caucus of international scope
that arose in the 1940s and gained significant political influence with the emergence of the environmental movement in the 1970s.
The impact of the post-70s environmental movement on the quality of air and water, mostly in the West, but also developing
countries such as India, is the most hopeful precedent we have that the sustainability issues facing the world in the new century
might yet be overcome, given political will and organization equal to the task.
Climate Change
Air pollution is still a major problem in the megacities of the developing world, however, while a global change in air chemistry—
an increase of 40% in the carbon load of the atmosphere since industrialization—is ushering in an era of accelerated climate
change. This era will be characterized by increased droughts and floods, higher sea levels, and extreme weather events, unevenly
and unpredictably distributed across the globe, with the highest initial impact in regions that, in economic and infrastructural terms,
can least support climate disruption (for example, sub-Saharan Africa). The environmental historian J. R. McNeil estimates that
between 25 and 40 million people died from air pollution in the 20th century. The death toll arising from climate change in the 21st
century is difficult to predict, but given the scale of the disruption to weather systems on which especially marginal states depend, it
is likely to be on a much larger scale.
Summary
From the Portuguese sea merchants of the 16th century in quest of silver and spices from Asia, to the multinational oil companies of
today seeking to drill in ever more remote and fragile undersea regions, the dominant view driving global economic growth over
the last half millennium has been instrumentalist, that is, of the world’s ecosystems as alternately a source of raw materials (foods,
energy, minerals) and a dump for the wastes produced by the industrialization and consumption of those materials. The
10.3.3 [Link]
instrumentalist economic belief system of the modern era, and particularly the Industrial Age, is based on models of perennial
growth, and measures the value of ecosystems according to their production of resources maximized for efficiency and hence
profit. In this prevailing system, the cost of resource extraction to the ecosystem itself is traditionally not factored into the product
and shareholder values of the industry. These costs are, in economic terms, externalized.
A future economics of sustainability, by contrast, would prioritize the management of ecosystems for resilience rather than pure
capital efficiency, and would incorporate the cost of ecosystem management into the pricing of goods. In the view of many
sustainability theorists, dismantling the system of “unnatural” subsidization of consumer goods that has developed over the last
century in particular is the key to a sustainable future. Only a reformed economic system of natural pricing, whereby environmental
costs are reflected in the price of products in the global supermarket, will alter consumer behavior at the scale necessary to ensure
economic and environmental objectives are in stable alignment, rather than in constant conflict. As always in the sustainability
paradigm, there are tradeoffs. A future economy built on the principle of resilience would be very different from that prevalent in
the economic world system of the last 500 years in that its managers would accept reduced productivity and efficiency in exchange
for the long-term vitality of the resource systems on which it depends.
Review Questions
1. What are the major technological and economic developments since 1500 that have placed an increased strain on thge planet’s
ecosystem services? What is the role of carbon-based energy systems in that history?
2. What is the so-called Great Acceleration of the 20th century? What were its principal social features and environmental
impacts?
3. What is the Green Revolution? What were its successes, and what problems has it created?
Glossary
Biodiversity and Extinction
Thriving ecosystems are characterized by diverse plant and animal populations; there is, therefore, a strong correlation between
current ecosystem decline globally, and the rate of extinction of species, which is in the order of a thousand times that of
background rates. This has prompted scientists to label the current period the Sixth Mass Extinction in the long history of the
biosphere, and the first since the end of the dinosaurs.
Externalization
The process by which costs inherent to the production of goods—particularly environmental costs—are not included in the actual
price paid.
Instrumentalist
An attitude to environmental resources characteristic of the last 500-year period of global human economic development, whereby
ecosystem provisions—water, minerals, oil and gas, etc.—are perceived only in terms of their use value to human beings, rather
than as integral elements of a wider natural system.
Nonlinear
Changes in a system are nonlinear when they exhibit sudden changes in rate of increase or decline. The population of a particular
tropical frog species, for example, may suddenly crash as a result of warming temperatures, rather than show gradual decline.
This page titled 10.3: The Industrialization of Nature- A Modern History (1500 to the present) is shared under a CC BY license and was authored,
remixed, and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
10.3.4 [Link]
10.4: Sustainability Studies- A Systems Literacy Approach
Introduction
Transition to a sustainable resource economy is a dauntingly large and complex project, and will increasingly drive research and
policy agendas across academia, government, and industry through the twenty-first century. To theorize sustainability, in an
academic setting, is not to diminish or marginalize it. On the contrary, the stakes for sustainability education could not be higher.
The relative success or failure of sustainability education in the coming decades, and its influence on government and industry
practices worldwide, will be felt in the daily lives of billions of people both living and not yet born.
The core of sustainability studies, in the academic sense, is systems literacy—a simple definition, but with complex implications.
Multiple indicators tell us that the global resource boom is now reaching a breaking point. The simple ethos of economic growth
—“more is better”—is not sustainable in a world of complex food, water and energy systems suffering decline. The grand
challenge of sustainability is to integrate our decision-making and consumption patterns—along with the need for economic
viability— within a sustainable worldview. This will not happen by dumb luck. It will require, first and foremost, proper education.
In the nineteenth and twentieth centuries, universal literacy—reading and writing—was the catch-cry of education reformers. In the
twenty-first century, a new global literacy campaign is needed, this time systems literacy, to promote a basic understanding of the
complex interdependency of human and natural systems.
Here I will lay out the historical basis for this definition of sustainability in terms of systems literacy, and offer specific examples of
how to approach issues of sustainability from a systems-based viewpoint. Systems literacy, as a fundamental goal of higher
education, represents the natural evolution of interdisciplinarity, which encourages students to explore connections between
traditionally isolated disciplines and has been a reformist educational priority for several decades in the United States. Systems
literacy is an evolved form of cross-disciplinary practice, calling for intellectual competence (not necessarily command) in a variety
of fields in order to better address specific real-world environmental problems.
For instance, a student’s research into deforestation of the Amazon under a sustainability studies paradigm would require
investigation in a variety of fields not normally brought together under the traditional disciplinary regime. These fields might
include plant biology, hydrology, and climatology, alongside economics, sociology, and the history and literature of post-colonial
Brazil. Systems literacy, in a nutshell, combines the study of social history and cultural discourses with a technical understanding of
ecosystem processes. Only this combination offers a comprehensive view of real-world environmental challenges as they are
unfolding in the twenty-first century.
From the viewpoint of systems literacy sustainability studies works on two planes at once. Students of sustainability both
acknowledge the absolute interdependence of human and natural systems—indeed that human beings and all their works are
nothing if not natural—while at the same time recognizing that to solve our environmental problems we must often speak of the
natural world and human societies as if they were separate entities governed by different rules. For instance, it is very useful to
examine aspects of our human system as diachronic—as progressively evolving over historical time—while viewing natural
systems more according to synchronic patterns of repetition and equilibrium. The diachronic features of human social evolution
since 1500 would include the history of trade and finance, colonization and frontier development, and technology and urbanization,
while examples of nature’s synchronicity would be exemplified in the migratory patterns of birds, plant and animal reproduction, or
the microbial ecology of a lake or river. A diachronic view looks at the changes in a system over time, while the synchronic view
examines the interrelated parts of the system at any given moment, assuming a stable state.
While the distinction between diachronic and synchronic systems is in some sense artificial, it does highlight the structural
inevitability of dysfunction when the two interlocked systems operate on different timelines and principles. The early twentieth
century appetite for rubber to service the emerging automobile industry, for instance, marks an important chapter in the “heroic”
history of human technology, while signifying a very different transition in the history of forest ecosystems in Asia and Latin
America. Human history since the agricultural transition 10,000 years ago, and on a much more dramatic scale in the last two
hundred years, is full of such examples of new human technologies creating sudden, overwhelming demand for a natural resource
previously ignored, and reshaping entire ecosystems over large areas in order to extract, transport and industrialize the newly
commodified material.
10.4.1 [Link]
Biocomplexity
For students in the humanities and social sciences, sustainability studies requires adoption of a new conceptual vocabulary drawn
from the ecological sciences. Among the most important of these concepts is complexity. Biocomplexity—the chaotically variable
interaction of organic elements on multiple scales—is the defining characteristic of all ecosystems, inclusive of humans.
Biocomplexity science seeks to understand this nonlinear functioning of elements across multiple scales of time and space, from
the molecular to the intercontinental, from the microsecond to millennia and deep time. Such an approach hasn’t been possible until
very recently. For example, only since the development of (affordable) genomic sequencing in the last decade have biologists
begun to investigate how environments regulate gene functions, and how changes in biophysical conditions place pressure on
species selection and drive evolution.
Image of Biocomplexity Spiral
Figure 10.4.1 The Biocomplexity Spiral. The biocomplexity spiral illustrates the concept of biocomplexity, the chaotically variable
interaction of organic elements on multiple scales. Source: U.S. National Science Foundation
How is the concept of complexity important to sustainability studies? To offer one example, a biocomplexity paradigm offers the
opportunity to better understand and defend biodiversity, a core environmental concern. Even with the rapid increase in knowledge
in the biophysical sciences in recent decades, vast gaps exist in our understanding of natural processes and human impacts upon
them. Surprisingly little is known, for example, about the susceptibilities of species populations to environmental change or,
conversely, how preserving biodiversity might enhance the resilience of an ecosystem. In contrast to the largely reductionist
practices of twentieth-century science, which have obscured these interrelationships, the new biocomplexity science begins with
presumptions of ignorance, and from there goes on to map complexity, measure environmental impacts, quantify risk and
resilience, and offer quantitative arguments for the importance of biodiversity. Such arguments, as a scientific supplement to more
conventional, emotive appeals for the protection of wildlife, might then form the basis for progressive sustainability policy.
But such data-gathering projects are also breathtaking in the demands they place on analysis. The information accumulated is
constant and overwhelming in volume, and the methods by which to process and operationalize the data toward sustainable
practices have either not yet been devised or are imperfectly integrated within academic research structures and the policy-making
engines of government and industry. To elaborate those methods requires a humanistic as well as scientific vision, a need to
understand complex interactions from the molecular to the institutional and societal level.
A practical example of biocomplexity as the frame for studies in environmental sustainability are the subtle linkages between the
hypoxic “dead zone” in the Gulf of Mexico and farming practices in the Mississippi River watershed. To understand the impact of
hydro-engineered irrigation, nitrogen fertilizer, drainage, and deforestation in the Midwest on the fisheries of the Gulf is a classic
biocomplexity problem, requiring data merging between a host of scientific specialists, from hydrologists to chemists, botanists,
geologists, zoologists and engineers. Even at the conclusion of such a study, however, the human dimension remains to be
explored, specifically, how industry, policy, culture and the law have interacted, on decadal time-scales, to degrade the tightly
coupled riverine-ocean system of the Mississippi Gulf. A quantitative approach only goes so far. At a key moment in the process,
fact accumulation must give way to the work of narrative, to the humanistic description of desires, histories, and discourses as they
have governed, in this instance, land and water use in the Mississippi Gulf region.
To complexity should be added the terms resilience and vulnerability, as core concepts of sustainability studies. The resilience of a
system—let’s take for example, the wildlife of the Arctic Circle—refers to the self-renewing stability of that system, its ability to
rebound from shocks and threats within the range of natural variability. The vulnerability of Artic wildlife, conversely, refers to the
point at which resilience is eroded to breaking point. Warming temperatures in the Arctic, many times the global average, now
threaten the habitats of polar bear and walruses, and are altering the breeding and migratory habits of almost all northern wildlife
populations. The human communities of the Arctic are likewise experiencing the threshold of their resilience through rising sea
levels and coastal erosion. Entire villages face evacuation and the traumatic prospect of life as environmental refugees.
As mentioned earlier, we have grown accustomed to speaking of “nature” or “the environment” as if they were somehow separate
from us, something that might dictate our choice of holiday destination or wall calendar, but nothing else. A useful counter-
metaphor for sustainability studies, to offset this habitual view, is to think of human and natural systems in metabolic terms. Like
the human body, a modern city, for example, is an energy-dependent system involving inputs and outputs. Every day, millions of
tons of natural resources (raw materials, consumer goods, food, water, energy) are pumped into the world’s cities, which turn them
out in the form of waste (landfill, effluent, carbon emissions, etc.).
Unlike the human body, however, the metabolism of modern cities is not a closed and self-sustaining system. Cities are consuming
resources at a rate that would require a planet one and a half times the size of Earth to sustain, and are ejecting wastes into the land,
10.4.2 [Link]
water, and air that are further degrading the planet’s ability to renew its vital reserves. Here, another body metaphor—the
environmental “footprint”—has become a popular means for imagining sufficiency and excess in our consumption of resources.
The footprint metaphor is useful because it provides us an image measurement of both our own consumption volume and the
environmental impact of the goods and services we use. By making sure to consume less, and to utilize only those goods and
services with a responsibly low footprint, we in turn reduce our own footprint on the planet. In important ways, the problem of
unsustainability is a problem of waste. From a purely instrumentalist or consumerist viewpoint, waste is incidental or irrelevant to
the value of a product. A metabolic view of systems, by contrast, promotes sustainability concepts such as closed loops and carbon
neutrality for the things we manufacture and consume, whereby there are no toxic remainders through the entire lifecycle of a
product. In this sense, systems literacy is as much a habit or style of observing the everyday world as it is an academic principle for
the classroom. Because in the end, the fate of the world’s ecosystems will depend not on what we learn in the classroom but on the
extent to which we integrate that learning in our lives beyond it: in our professional practice and careers, and the lifestyle and
consumer choices we make over the coming years and decades. If systems literacy translates into a worldview and way of life, then
sustainability is possible.
Review Questions
1. What are synchronicand diachronic views of time, and how does the distinction help us to understand the relation between
human and natural systems, and to potentially rewrite history from an environmental point of view?
2. How is a bio-complex view of the relations between human and natural systems central to sustainability, in both theory and
practice?
Glossary
Biocomplexity
A defining characteristic of living things and their relationships to each other. The biocomplexity concept emphasizes the multiple
dependent connections within ecosystems, and between ecosystems and human societies.
Carbon Neutrality
To be carbon neutral, the carbon emissions of a consumable product or human activity must either not involve the consumption of
carbon-based energy (a difficult thing to achieve under our present regime), or offset that consumption through the drawdown of an
equivalent amount of atmospheric carbon during its lifecycle.
Closed Loops
The sustainable reform of industrial production and waste management emphasizes the recycling of materials back into the
environment or into the industrial cycle, that is, to eliminate the concept of waste entirely.
Diachronic/Synchronic
A diachronic view of a system examines it evolution over time, while a synchronic view is concerned with its characteristics at a
single point in time.
Interdisciplinarity
A trend in higher education research and teaching of the last thirty years that emphasizes the bridging of traditional disciplines, and
that is an essential framework for sustainability studies.
Lifecycle
In terms of sustainability, the entire lifecycle of a product must be measured for its environmental impact, not simply its point of
production, consumption, or disposal. A key aspect of general sustainability education is the understanding of where goods
originate, the industrial processes required for their manufacture and transport, and their fate after use.
Metabolism and Footprint
Two metaphors, related to the human body, for conceptualizing the relationship between consumption and waste at the social level.
Metabolism emphasizes a system of inputs and outputs dependent upon “energy” and measured according to the “health” of the
whole, while footprint is a popular metric for quantifying the environmental impacts of goods, services, and lifestyles.
Resilience and Vulnerability
10.4.3 [Link]
Important terms of measurement for the impact of environmental change, particularly on human communities. The goal of
sustainability analysis and policy, at all levels, is to enhance the resilience of communities to change, in other words, to mitigate
their vulnerability.
Systems Literacy
An educational philosophy that emphasizes a student’s competence in a wide variety of disciplines, so that he or she might better
understand the operations of those complex systems, both human and natural, that underpin sustainability.
This page titled 10.4: Sustainability Studies- A Systems Literacy Approach is shared under a CC BY license and was authored, remixed, and/or
curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
10.4.4 [Link]
10.5: The Vulnerability of Industrialized Resource Systems- Two Case Studies
Introduction
Sustainability is best viewed through specific examples, or case studies. One way of conceiving sustainability is to think of it as a
map that shows us connections between apparently unrelated domains or sequences of events. To cite an earlier example, what do
the cornfields of Illinois have to do with the decline of fisheries in the Gulf of Mexico? To the uneducated eye, there is no
relationship between two areas so remote from each other, but a sustainable systems analysis will show the ecological chain linking
the use of chemical fertilizers in the Midwest, with toxic runoff into the Mississippi Basin, with changes in the chemical
composition in the Gulf of Mexico (specifically oxygen depletion), to reduced fish populations, and finally to economic and social
stress on Gulf fishing communities. Here, I will look at two case studies in greater detail, as a model for the systems analysis
approach to sustainability studies in the humanities. The first concerns the alarming worldwide decline of bee populations since
2006, owing to a new affliction named Colony Collapse Disorder (CCD). The second case study examines the BP oil disaster in the
Gulf of Mexico in 2010, considered in the larger historical context of global oil dependency.
Figure 10.5.1 Faust and Mephistopheles. Mephistopheles, the devil figure in Goethe’s play Faust, tempts Faust with the
exhilaration of flight. From the air, it is easy for Faust to imagine himself lord of the earth, with no limits to his powers. Source:
Public Domain. Illustration by Alphonse de Neuville
In short, we are all Fausts now, not the insignificant, powerless creatures we sometime feel ourselves to be, but rather, the lords of
the planet. How this came to pass is an object lesson in complex diachronic evolution. Without any single person deciding, or any
law passed, or amendment made to the constitution, we have transformed ourselves over but a few centuries from one struggling
species among all the rest, to being planetary managers, now apparently exempted from the evolutionary struggle for survival with
other species, with the fate of animals, birds, fish, plants, the atmosphere, and entire ecosystems in our hands. This Faustian power
signals both our strength and vulnerability. We are dependent on the very ecosystems we dominate. That is, we have become
carbon-dependent by choice, but we are ecosystem-dependent by necessity. We may all be supermen and wonderwomen relative to
the poor powers of our forebears, but we still require food, clean water, and clean air. The billion or more people on earth currently
not plugged into the carbon energy grid, and hence living in dire poverty, need no reminding of this fact. Many of us in the
developed world do, however. Our civilization and lifestyles as human beings have changed beyond recognition, but our biological
10.5.1 [Link]
needs are no different from our species ancestors on the East African savannah a million years ago. In sum, the lesson of the Faust
story is hubris. We are not exempt from natural laws, as Faust recklessly hoped.
To understand the impact of our fossil fuel based, industrialized society on the planet we inhabit requires we think on dual time
scales. The first is easy enough, namely, the human scale of days and years. For example, consider the time it takes for liquid
petroleum to be extracted from the earth, refined, transported to a gas station, and purchased by you in order to drive to school or
the shopping Mall. Or the time it takes for that sweater you buy at the mall to be manufactured in China or Indonesia and
transported thousands of miles to the shelf you grab it from. This is an oil-dependent process from beginning to end: from the
petroleum-based fertilizers that maximized the productive efficiency of the cotton plantation, to powering the machinery in the
factory, to the massive goods ship transporting your sweater across the oceans, to the lights in the store that illuminate your sweater
at the precise angle for it to catch your eye.
Now consider the second time scale, to which we are usually oblivious—the thousands or millions of years it has taken for
terrestrial carbon to form those reserves of liquid petroleum that brought you your sweater. This is a process describable only on a
geological time scale, the costs of the disruption to which have been wholly omitted from the sticker price of the sweater. What are
the environmental, and ultimately human costs that have been externalized? In powering our modern societies through the
transference of the earth’s carbon reserves from long-term storage and depositing it in the atmosphere and oceans, we have
significantly altered and destabilized the earth’s carbon cycle. There is now 40% more carbon in the atmosphere and oceans than in
1800, at the outset of the industrial age. The earth’s climate system is reacting accordingly, to accommodate the increased
nonterrestrial carbon load. The result is altered weather patterns, increasing temperatures, glacial melt, and sharp increases in
droughts, floods, and wildfires. The cost to the global economy of these climate disruptions this century has been projected in the
trillions of dollars, even before we consider the human costs of climate change in mortality, homelessness, impoverishment, and
social instability.
Extracting carbon from the earth, and transforming it into energy, fertilizers, and products has enabled an almost magical
transformation of human lives on earth, as compared to those of our premodern ancestors. The house you live in, the clothes you
wear, the food you eat, the gadgets you use, and all the dreams you dream for your future, are carbon-based dreams. These amazing
fossil-fuel energy sources—oil, coal, gas—have created modernity itself: a crest of population growth, economic development,
prosperity, health and longevity, pulling millions out of poverty, and promoting, life, liberty, and happiness. This modernity is truly
a thing of wonder, involving the high-speed mass transport of people, goods, and information across the globe, day after day.
Regardless of the season, it brings us apples from New Zealand, avocadoes from Mexico, and tomatoes that have traveled an
average of 2000 miles to reach the “fresh produce” section of our supermarkets. Having bought our groceries for the week, we
jump in our car and drive home. Because our species ancestors were both nomads and settlers, we love our cars and homes with
equal passion. We value both mobility and rootedness. Done with roaming for the day, we cherish our indoor lives in
atmospherically controlled environments: cool when hot outside, toasty when cold, light when dark, with digital devices plugged in
and available 24/7. A miraculous lifestyle when one sits back to reflect, and all the result of ongoing carbon-intensive investments
in human comfort and convenience.
But it is also a 200-year chemistry experiment, with our planet as the laboratory. We are carbon beings in our own molecular
biology; we touch and smell it; we trade, transport, and spill it; we consume and dispose of it in the earth and air. Intensifying heat
and storms and acidifying oceans are carbon’s elemental answer to the questions we have posed to the earth system’s resilience.
Mother Nature is having her say, acting according to her nature, and prompting us now to act according to our own mostly
forgotten natures—as beings dependent on our ecosystem habitat of sun, rain, soil, plants, and animals, with no special allowance
beyond the sudden responsibility of reformed stewardship and management.
The 2010 BP oil spill in the Gulf of Mexico was a spectacular warning that the 200-year era of cheap fossil fuel energy is drawing
to a close. With viable oil reserves likely to be exhausted in the next decade or so, and the dangers to global climate associated with
continued reliance on coal and natural gas, the transition to a sustainable, low-carbon global economy—by means that do not
impoverish billions of people in the process—looms as nothing less than the Great Cause of the 21st century, and without doubt the
greatest challenge humanity has faced in its long residence on earth. The stakes could not be higher for this task, which is of
unprecedented scope and complexity. If enormous human and financial resources were expended in meeting the greatest challenges
faced by the international community in the 20th century—the defeat of fascism, and the hard-earned progress made against poverty
and infectious diseases—then the low-carbon sustainability revolution of our century will require the same scale of resources and
more. At present, however, only a tiny fraction of those resources have been committed.
10.5.2 [Link]
References
Jacobsen, R. (2008). Fruitless Fall: The Collapse of the Honey Bee and the Coming Agricultural Crisis. New York: Bloomsbury
Glossary
Anthropocene
A term bestowed by Noble Laureate Paul Crutzen to describe the last 200-year period of human industrialization. The prefix
“anthro” points to the decisive impact of human population growth and technological development on the planetary biosphere since
1800, as its principal agents of change superceding all other factors.
This page titled 10.5: The Vulnerability of Industrialized Resource Systems- Two Case Studies is shared under a CC BY license and was
authored, remixed, and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
10.5.3 [Link]
10.6: Case Study- Agriculture and the Global Bee Colony Collapse
Two thousand years ago, at the height of the Roman Empire, the poet Virgil wrote lovingly about the practice of beekeeping, of
cultivating the “aerial honey and ambrosial dews” he called “gifts of heaven” (Georgics IV: 1-2). Bees represent a gift to humanity
even greater that Virgil knew. In addition to satisfying the human appetite for honey, the Italian honeybee, Apis melliflora, is the
world’s most active pollinator, responsible for over 80 of the world’s most common nongrain crops, including apples, berries,
almonds, macadamias, pumpkins, melons, canola, avocadoes, and also coffee beans, broccoli and lettuce. Even the production
chain of the enormous meat and cotton industries relies at crucial points on the ministrations of the humble honeybee. We depend
on pollinated fruits, nuts and seeds for a third of our caloric intake, and for vital vitamins, minerals and antioxidants in our diet. In
total, around 80% of the foods we eat are to some degree the products of bee pollination, representing one third of total agricultural
output.
Given the $1 trillion value of pollinated produce, any threat to the health of honey bees represents a serious threat to the human
food chain—a classic sustainability issue. With the industrialization of the global agricultural system over the last 50 years—
including crop monoculture and mass fertilization—bees have indeed faced a series of threats to their ancient role, the most recent
of which, so-called Colony Collapse Disorder, is the most serious yet.
Busy Bee Hive
Figure 10.6.1 Busy Bee Hive. A forager honeybee comes in for landing at a healthy hive, her legs dusted with pollen. Colony
Collapse Disorder has devastated tens of thousands of such hives. Source: Ken Thomas
In his poetic primer on beekeeping, Virgil includes a moving description of a bee colony suffering mysterious decline:
“Observe the symptoms when they fall away” “ And languish with insensible decay.” “ They change their hue; with haggard eyes
they stare . . .” “ The sick, for air, before the portal gasp,” “ Their feeble legs within each other clasp,” “ Or idle in their empty
hives remain,” “ Benumbed with cold, and listless of their gain. (368-78”
Beekeepers worldwide faced an even worse predicament in late 2006: the mysterious disappearance of entire hives of bees. Over
the winter, honeybees enter a form of survival hibernation. Their populations suffer inevitable losses, but these are replenished by
the Queen’s renewed laying of eggs once winter thaws. In the spring of 2007, however, hundreds of thousands of colonies in the
United States did not survive the winter. A full 30% of all honeybee colonies died. Each spring since has witnessed even worse
declines. Similar losses afflicted Europe and Asia. Worldwide, millions of colonies and billions of bees have perished since 2006
on account of the new bee plague.
Because the global commercial value of bee pollination is so enormous, well-funded research into colony collapse began
immediately. A number of theories, some credible, some not, were quickly advanced. Several studies pointed to new or enhanced
viral strains, while others suggested the toxic effect of industrial fertilization. Still others claimed that mobile phone towers were
interfering with the bees’ navigations systems. Because the honeybee is a charismatic creature and features so prominently in our
cultural lore—we admire their industriousness, fear their stings, call our loved ones “honey,” and talk much of Queen Bees—the
story of colony collapse was quickly taken up by the media. A flurry of news stories announced CCD as an epic “disaster” and
profound “mystery,” which was true in simple terms, but which cast bee decline as a new and sudden calamity for which some
single culprit must be responsible.
The truth, as it is now unfolding, is more complex, and shows the importance of viewing the interactions between human and
natural ecologies in systemic terms. In strictly pathogenic terms, CCD is caused by the combination of a virus (called Iridoviridae
or IIV) and a microsporidian fungus called Nosema. The specific interaction between the pathogens, and why they cause bees in
their millions to vacate their hives, is not understood. What is becoming clear, however, is the increasing burden being placed on
bees by the human agricultural system, a burden that has rendered bees increasingly vulnerable to epidemic infection. Humans have
been keeping bees for eight thousand years, and European bees were at the vanguard of the successful crop colonization of the
Americas. But the numbers of bees in the United States had already declined by a third since 1950 before the arrival of CCD,
owing to various viral and mite infestations, and the large scale changes in bee habitat and lifestyle.
Before the industrialization of farming, bees came from neighboring wildlands to pollinate the diverse range of crops available to
them on small plots. But the conversion, for economic reasons, of arable land into enormous monocrop properties in the last sixty
years, and hence the diminishment of proximate wildflower habitats, has necessitated a different system, whereby bees are trucked
around the country to service one crop at a time, be it peppers in Florida, blueberries in Maine, or almonds in California. At the
height of the recent almond boom, the California crop required almost the entire bee population of the United States to be fully
pollinated. Wholesale suburbanization is also to blame for the destruction of the bees’ natural wildflower habitats. Be it a thousand
10.6.1 [Link]
acre cornfield or a suburban street of well-tended green lawns, to a bees’ eyes, our modern landscape, engineered to human needs,
is mostly a desert.
Studies that have not identified specific culprits for CCD have nevertheless shown the extent of the long-term decline in bee health
wrought by their conscription to industrial agriculture. For instance, researchers found no fewer than 170 different pesticides in
samples of American honeybees, while other studies found that even bees not suffering CCD habitually carry multiple viral strains
in their systems. The combined toxic and viral load for the average honeybee is enormous. In the words of Florida’s state apiarist,
“I’m surprised honey bees are alive at all.” (Jacobsen, 2008, p. 137) A further study showed a decline in the immune systems of
bees owing to lack of diverse nutrition. Pollinating only almonds for weeks on end, then travelling on a flatbed truck for hundreds
of miles in order to service another single crop, is not the lifestyle bees have adapted to over the near 80 million years of their
existence. As Virgil warned, “First, for thy bees a quiet station find.” The lives of modern bees have been anything but quiet, and
the enormous changes in their habitat and lifestyle have reduced their species’ resilience.
The most important lesson of recent research into CCD is not the identification of IIV and Nosema as the specific contributors, but
the larger picture it has provided of a system under multiple long-term stresses. Complex systems, such as bee pollination and
colony maintenance, are not characterized by linear development, but rather by sudden, nonlinear changes of state called tipping
points. CCD is an example of a potential tipping point in a natural system on which humans depend, in which sudden deterioration
overtakes a population beyond its ability to rebound. Everything seems fine, until it isn’t. One day we have almonds, berries,
melon, and coffee on our breakfast menu. The next day there’s a critical shortage, and we can’t afford them.
In sustainability terms, bee colony collapse is a classic “human dimensions” issue. CCD will not be “solved” simply by the
development of a new anti-viral drug or pesticide targeting the specific pathogens responsible. Part of what has caused CCD is the
immunosuppressive effects of generations of pesticides developed to counter previous threats to bee populations, be they microbes
or mites. Our chemical intervention in the lifecycle of bees has, in evolutionary terms, “selected” for a more vulnerable bee. That
is, bees’ current lack of resilience is a systemic problem in our historical relationship to bees, which dates back thousands of years,
but which has altered dramatically in the last fifty years in ways that now threaten collapse. And this is to say nothing of the impact
of bee colony collapse on other pollination-dependent animals and birds, which would indeed be catastrophic in biodiversity terms.
That we have adapted to bees, and they to us, is a deep cultural and historical truth, not simply a sudden “disaster” requiring the
scientific solution of a “mystery.” In the light of sustainability systems analysis, the bee crisis appears entirely predictable and the
problem clear cut. The difficulty arises in crafting strategies for how another complex system on a massive scale, namely global
agriculture, can be reformed in order to prevent its collapse as one flow-on effect of the global crisis of the vital honey bee. The
incentive for such reform could not be more powerful. The prospect of a future human diet without fruits, nuts and coffee is bleak
enough for citizens of the developed world and potentially fatal for millions of others in the long term.
References
1. Jacobsen, R. (2008). Fruitless Fall: The Collapse of the Honey Bee and the Coming Agricultural Crisis. New York: Bloomsbury
Review Questions
1. What is the long history of the human relationship to bees, and what radical changes in that relationship have occurred over the
last fifty years to bring it to the point of collapse?
2. What are the implications of bee colony collapse for the global food system?
Glossary
Tipping Point
The critical moment of nonlinear change whereby a system changes suddenly from one state to another.
This page titled 10.6: Case Study- Agriculture and the Global Bee Colony Collapse is shared under a CC BY license and was authored, remixed,
and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
10.6.2 [Link]
10.7: Case Study- Energy and the BP Oil Disaster
On the night of April 20, 2010, the Deepwater Horizon oil rig, one of hundreds operating in the Gulf of Mexico, exploded, killing
eleven men, and placing one of the most rich and diverse coastal regions on earth in imminent danger of petroleum poisoning. BP
had been drilling in waters a mile deep, and in the next two days, as the rig slowly sank, it tore a gash in the pipe leading to the oil
well on the ocean floor. Over the next three months, two hundred million gallons of crude oil poured into the Gulf, before the
technological means could be found to seal the undersea well. It was the worst environmental disaster in American history, and the
largest peacetime oil spill ever.
The Deepwater Horizon Oil Rig on Fire
Figure 10.7.1 The Deepwater Horizon Oil Rig on Fire The Deepwater Horizon oil rig on fire, April, 2010. It would later sink,
precipitating the worst environmental disaster in United States history. Source: Public Domain U.S. Coast Guard
The BP oil disaster caused untold short- and long-term damage to the region. The initial impact on the Gulf—the oil washing up on
beaches from Texas to Florida, and economic hardship caused by the closing down of Gulf fishing—was covered closely by the
news media. The longer term impacts of the oil spill on wetlands erosion, and fish and wildlife populations, however, will not
likely receive as much attention.
Much public debate over the spill has focused on the specific causes of the spill itself, and in apportioning responsibility. As with
the example of bee colony collapse, however, the search for simple, definitive causes can be frustrating, because the breakdown is
essentially systemic. Advanced industries such as crop pollination and oil extraction involve highly complex interactions among
technological, governmental, economic, and natural resource systems. With that complexity comes vulnerability. The more
complex a system, the more points at which its resiliency may be suddenly exposed. In the case of the Deepwater Horizon rig,
multiple technological “safeguards” simply did not work, while poor and sometimes corrupt government oversight of the rig’s
operation also amplified the vulnerability of the overall system—a case of governmental system failure making technological
failure in industry more likely, with an environmental disaster as the result.
In hindsight, looking at all the weaknesses in the Gulf oil drilling system, the BP spill appears inevitable. But predicting the
specific vulnerabilities within large, complex systems ahead of time can be next to impossible because of the quantity of variables
at work. Oil extraction takes place within a culture of profit maximization and the normalization of risk, but in the end, the lesson
of BP oil disaster is more than a cautionary tale of corporate recklessness and lax government oversight. The very fact that BP was
drilling under such risky conditions—a mile underwater, in quest of oil another three miles under the ocean floor—is an expression
of the global demand for oil, the world’s most valuable energy resource. To understand that demand, and the lengths to which the
global energy industry will go to meet it, regardless of environmental risk, requires the longer view of our modern history as a
fossil-fueled species.
Review Questions
1. In what ways is the BP Oil Disaster of 2010 an example of complex human systems failure, and what are its longer chains of
causation in the history of human industrialization?
10.7: Case Study- Energy and the BP Oil Disaster is shared under a not declared license and was authored, remixed, and/or curated by LibreTexts.
10.7.1 [Link]
10.8: Sustainability Ethics
Developing an Ethics of Sustainability
The 1987 United Nations Brundtland definition of sustainability embodies an intergenerational contract: to provide for our present
needs, while not compromising the ability of future generations to meet their needs. It’s a modest enough proposal on the face of it,
but it challenges our current expectations of the intergenerational contract: we expect each new generation to be better off than their
parents. Decades of technological advancement and economic growth have created a mindset not satisfied with “mere”
sustainability. We might call it turbo-materialism or a cornucopian worldview: namely that the earth’s bounty, adapted to our use by
human ingenuity, guarantees a perpetual growth in goods and services. At the root of the cornucopian worldview lies a brand of
technological triumphalism, an unshakeable confidence in technological innovation to solve all social and environmental problems,
be it world hunger, climate change, or declining oil reserves. In sustainability discourse, there is a wide spectrum of opinion from
the extremes of cornucopian optimism on one side and to the doom-and-gloom scenarios that suggest it is already too late to avert a
new Dark Age of resource scarcity and chronic conflict on the other.
California, the Cornucopia of the World
Figure 10.8.1 California, the Cornucopia of the World Cover of an 1885 promotional book prepared by the California Immigration
Commission. Source: The California State Library
For every generation entering a Dark Age, there were parents who enjoyed a better life, but who somehow failed to pass along their
prosperity. No one wants to fail their children in this way. To this extent, biology dictates multigenerational thinking and ethics.
Though it might not always be obvious, we are all already the beneficiaries of multi-generational planning. The world-leading
American higher education system, for example, depends upon an intergenerational structure and logic—a financial and human
investment in the future committed to by multiple generations of Americans going back to the 19th century. But conversely, in
terms of vulnerability, just as higher education in the United States is neither necessarily permanent nor universal, but a social
institution built on an unwritten contract between generations, so the lifestyle benefits of advanced society as we know it will not
simply perpetuate themselves without strenuous efforts to place them on a sustainable footing.
Our current problem lies in the fact that multigenerational thinking is so little rewarded. Our economic and political systems as they
have evolved in the Industrial Age reward a mono-generational mindset driven by short-term profits and election cycles. In the
West, for example, there is no significant political philosophy, regulatory system, or body of law that enshrines the idea that we act
under obligation to future generations, despite widely held views that we naturally must. One challenge of sustainability is to
channel our natural biological interest in the future into a new ethics and politics based on multigenerational principles. Many
indigenous communities in the world, marginalized or destroyed by colonialism and industrialization, have long recognized the
importance of sustainability in principles of governance, and provide inspiring models. The Great Law of the Iroquois Confederacy,
for example, states that all decisions made by its elders should be considered in light of their impact seven generations into the
future.
To embrace an ethics of sustainability is to accept that our rapid industrialization has placed us in the role of planetary managers,
responsible for the health, or ruinous decline, of many of the globe’s vital ecosystems. This ethics requires we activate, in the
popular sense, both sides of our brain. That is, we must toggle between a rational consideration of our environmental footprint and
practical issues surrounding the reinvention of our systems of resource management, and a more humble, intuitive sense of our
dependence and embeddness within the web of life. Both reason and emotion come into play. Without emotion, there can be no
motivation for change. Likewise, without an intellectual foundation for sustainability, our desire for change will be unfocused and
ineffective. We are capable of adapting to a complex world and reversing broad-based ecosystem decline. But to do so will require
technical knowledge wedded to an ethical imagination. We need to extend to the natural world the same moral sense we intuitively
apply to the social world and our relations with other people.
Sustainability ethics thus does not need to be invented from whole cloth. It represents, in some sense, a natural extension of the
ethical principles dominant in the progressive political movements of the 20th century, which emphasized the rights of historically
disenfranchised communities, such as women, African-Americans, and the global poor. Just as we have been pressed to speak for
dispossessed peoples who lack a political voice, so we must learn the language of the nonhuman animal and organic world, of
“nature,” and to speak for it. Not simply for charity’s sake, or out of selfless concern, but for our own sake as resource-dependent
beings.
10.8.1 [Link]
Remote Responsibilities
What distinguishes an ethics of sustainability from general ethical principles is its emphasis on remote responsibilities, that is, our
moral obligation to consider the impact of our actions on people and places far removed from us. This distance may be measured in
both space and time. First, in spatial terms, we, as consumers in the developed world, are embedded in a global web of commerce,
with an ethical responsibility toward those who extract and manufacture the goods we buy, whether it be a polo shirt from
Indonesia, or rare metals in our computer extracted from mines in Africa. The economic and media dimensions of our consumer
society do not emphasize these connections; in fact, it is in the interests of “consumer confidence” (a major economic index) to
downplay the disparities in living standards between the markets of the developed world and the manufacturing countries of the
global south (Africa, Asia, Latin America), which serve as the factories of the world.
Second, as for sustainability ethics considered in temporal terms, the moral imagination required to understand our remote
responsibilities poses an even greater challenge. As we have seen, the landmark United Nations Brundtland Report establishes an
ethical contract between the living and those yet to be born. For an industrial civilization founded on the no-limits extraction of
natural resources and on maximizing economic growth in the short term, this is actually a profoundly difficult challenge to meet.
More than that, the practical ethical dilemmas it poses to us in the present are complex. How, for instance, are we to balance the
objectives of economic development in poorer nations—the need to lift the world’s “bottom billion” out of poverty—with the
responsibility to conserve resources for future generations, while at the same time making the difficult transition from
industrialized fossil fuels to a low-carbon global economy?
The issue of fairness with regard to individual nations’ carbon emissions reduction mandates is a specific example of how ethical
issues can complicate, or even derail, negotiated treaties on environmental sustainability, even when the parties agree on the end
goal. In the view of the developing countries of the global south, many of them once subject to colonial regimes of the north, the
advanced industrialized countries, such as the United States and Europe, should bear a heavier burden in tackling climate change
through self-imposed restraints on carbon consumption. They after all have been, over the last 200 years, the principal beneficiaries
of carbon-driven modernization, and thus the source of the bulk of damaging emissions. For them now to require developing
nations to curb their own carbon-based modernization for the benefit of the global community reeks of neo-colonial hypocrisy.
Developing nations such as India thus speak of common but differentiated responsibilities as the ethical framework from which to
justly share the burden of transition to a low-carbon global economy.
From the point of view of the rich, industrialized nations, by contrast, whatever the appearance of historical injustice in a carbon
treaty, all nations will suffer significant, even ruinous contractions of growth if an aggressive mitigation agreement among all
parties is not reached. Some commentators in the West have further argued that the sheer scale and complexity of the climate
change problem means it cannot effectively be addressed through a conventional rights-based and environmental justice approach.
To this degree at least, the sustainability issue distinguishes itself as different in degree and kind from the landmark social
progressive movements of the 20th century, such as women’s emancipation, civil rights, and multiculturalism, to which it has often
been compared.
Disputes over the complex set of tradeoffs between environmental conservation and economic development have dominated
environmental policy and treaty discussions at the international level for the last half century, and continue to stymie progress on
issues such as climate change, deforestation, and biofuels. These problems demonstrate that at the core of sustainability ethics lies a
classic tragedy of the commons, namely, the intractable problem of persuading individuals, or individual nations, to take specific
responsibility for resources that have few or no national boundaries (the atmosphere, the oceans), or which the global economy
allows to be extracted from faraway countries, the environmental costs of which are thus “externalized” (food, fossil fuels, etc).
How the international community settles the problem of shared accountability for a rapidly depleting global commons, and
balances the competing objectives of economic development and environmental sustainability, will to a large extent determine the
degree of decline of the planet’s natural capital this century. One tragic prospect looms: If there is no international commitment,
however patchwork, to protect the global resource commons, then the gains in economic prosperity, poverty alleviation and public
health in the developing world so hard won by international agencies over the second half of the 20th century, will quickly be lost.
A comic demonstrating the principle of the tragedy of the commons
Figure 10.8.2 Tragedy of Commons The tragedy of the commons is evident in many areas of our lives, particularly in the
environment. The over-fishing of our oceans that causes some marine life to be in danger of extinction is a good example. Source:
Food and Agriculture Organization of the United Nations
10.8.2 [Link]
Precautionary Principle
The precautionary principle is likewise central to sustainability ethics. The margins of uncertainty are large across many fields of
the biophysical sciences. Simply put, there is a great deal we do not know about the specific impacts of human activities on the
natural resources of land, air, and water. In general, however, though we might not have known where the specific thresholds of
resilience lie in a given system—say in the sardine population of California’s coastal waters—the vulnerability of ecosystems to
human resource extraction is a constant lesson of environmental history. A prosperous and vital economic engine, the Californian
sardine fishery collapsed suddenly in the 1940s due to overfishing. The precautionary principle underlying sustainability dictates
that in the face of high risk or insufficient data, the priority should lie with ecosystem preservation rather than on industrial
development and market growth.
Great Fish Market by Jan Brueghel
Figure 10.8.3 Great Fish Market by Jan Brueghel Though we might not have known where the specific thresholds of resilience lie
in a given system—say in the sardine population of California’s coastal waters—the vulnerability of ecosystems to human resource
extraction is a constant lesson of environmental history. Source: Public Domain
Sustainability, in instances such as these, is not a sexy concept. It’s a hard sell. It is a philosophy of limits in a world governed by
dreams of infinite growth and possibility. Sustainability dictates that we are constrained by earth’s resources as to the society and
lifestyle we can have. On the other hand, sustainability is a wonderful, inspiring concept, a quintessentially human idea. The
experience of our own limits need not be negative. In fact, what more primitive and real encounter between ourselves and the world
than to feel our essential dependence on the biospheric elements that surround us, that embeddedness with the air, the light, the
warmth or chill on our skins, and the stuff of earth we eat or buy to propel ourselves over immense distances at speed unimaginable
to the vast armies of humanity who came before us.
Sustainability studies is driven by an ethics of the future. The word itself, sustainability, points to proofs that can only be projected
forward in time. To be sustainable is, by definition, to be attentive to what is to come. So sustainability requires imagination, but
sustainability studies is also a profoundly historical mode, committed toreconstruction of the long, nonlinear evolutions of our
dominant extractivist and instrumentalist views of the natural world, and of the “mind-forg’d manacles” of usage and ideology that
continue to limit our ecological understanding and inhibit mainstream acceptance of the sustainability imperative.
Sustainability studies thus assumes the complex character of its subject, multiscalar in time and space, and dynamically agile and
adaptive in its modes. Sustainability teaches that the environment is not a sideshow, or a scenic backdrop to our lives. A few more
or less species. A beautiful mountain range here or there. Our relation to our natural resources is the key to our survival. That’s why
it’s called “sustainability.” It’s the grounds of possibility for everything else. Unsustainability, conversely, means human
possibilities and quality of life increasingly taken away from us and the generations to come.
Review Questions
1. What does it mean to say that global environmental problems such as climate change and ocean acidification represent a
“tragedy of the commons?” How are global solutions to be tied to local transitions toward a sustainable society?
2. How does sustainability imply an “ethics of the future?” And in what ways does sustainability ethics both borrow and diverge
from the principles that drove the major progressive social movements of the 20th century?
Glossary
Common but Differentiated Responsibilities
An ethical framework, promoted particularly by developing nations, that recognizes mitigation of global warming as a shared
responsibility, but at the same time argues that the wealthy, industrialized countries of the West that have been the historical
beneficiaries of carbon-based development should accept a greater burden for both reducing global carbon emissions, and
providing developing nations with the technology and economic means to modernize in sustainable ways.
Cornucopian
The view that economic growth and technological innovation will continue to improve the conditions of humanity as they have
done for the past 500 years, and that no environmental constraints are important or permanent.
Precautionary Principle
The proposition that decision-making should be driven by a concern for the avoidance of bad outcomes. In environmental terms,
this means coordinating economic development and the profit motive with the need to maintain resilient ecosystems.
10.8.3 [Link]
Remote Responsibilities
An ethical extension of systems literacy and the principle of connectivity: we are linked to peoples and places remote from us
through the web of global industrial production and commerce, and thus have responsibility toward them.
10.8: Sustainability Ethics is shared under a not declared license and was authored, remixed, and/or curated by LibreTexts.
10.8.4 [Link]
CHAPTER OVERVIEW
This page titled 11: Sustainable Infrastructure is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto Cabezas
(GALILEO Open Learning Materials) .
1
11.1: Sustainable Infrastructure - Chapter Introduction
Introduction
At present 80% of the US population lives in urban regions, a percentage that has grown steadily over the past two hundred years.
Urban infrastructures have historically supported several needs of the population served: the supply of goods, materials and
services upon which we rely; collection, treatment and disposal of waste products; adequate transportation alternatives; access to
power and communication grids; a quality public education system; maintenance of a system of governance that is responsive,
efficient and fair; generation of sufficient financial and social capital to maintain and renew the region; and insurance of the basic
elements of safety and public health. Collectively, these needs have been perceived as the basic attributes needed to make an urban
region livable.
Urban infrastructures are designed and built in response to social needs and economies of scale that urbanization has brought about.
Although our urban infrastructures are in many ways remarkable achievements of engineering design that were conceived and built
during times of rapid urbanization, as they have aged and, inevitably, deteriorated; significant strains on their function and ability to
provide services have become evident. In its program to identify the “grand challenges” facing society in the near future, the
National Academy of Engineering has proposed several focus areas, among them the restoration and improvement of urban
infrastructures. Such a challenge involves the need for renewal, but also presents opportunities for re-envisioning the basis of
infrastructure design and function as we move forward. Urban infrastructures of the past were not generally conceived in concert
with evolutionary social and ecological processes. This has resulted in several characteristic attributes: conceptual models of
infrastructure that perceive local ecological systems either indifferently or as obstacles to be overcome rather than assets for
harmonious designs; a general reliance on centralized facilities; structures that often lack operational flexibility such that alternative
uses may be precluded during times of crisis; heavy use of impervious and heat absorbing materials; systems that have become
increasingly costly to maintain and that are often excessively consumptive of natural resources on a life cycle basis; and a built
environment the materials and components of which are often difficult to reuse or recycle.
The urban environment is an example of a complex human-natural system. The resiliency of such systems lies in their capacity to
maintain essential organization and function in response to disturbances (of both long and short duration). A complimentary view,
inspired by traditional ecological and economic thought focuses on the degree of damage a system can withstand without exhibiting
a “regime” shift, defined as a transition that changes the structure and functioning of the system from one state to another as a result
of one or more independent factors. Upon exceeding a given threshold, the system shifts to a new alternative state which may not
be readily reversed through manipulation of causative factors. In the context of human-natural systems, regime shifts can have
significant consequences, and not all shifts are preferred by the human component of the system. To the extent that change of some
order is a given property of essentially all dynamic systems, “preferred” resiliency might be viewed as the extent to which human
societies can adapt to such shifts with acceptable levels of impacts. Resilient infrastructures, then, are those which most readily
facilitate such adaptation. Much of the foregoing discussion also applies to sustainability, with the added constraints of the
sustainability paradigm: the equitable and responsible distribution of resources among humans, present and future, in ways that do
not harm, and ideally reinforce, the social and biological systems upon which human society is based. Although there are important
differences between those two concepts, there remains a close interrelationship that stems from the same need: to understand and
design urban infrastructural systems that enhance human interactions with the environment.
It is beyond the scope of this book to present an exhaustive treatment of the urban environment, indeed there are many books and
treatises on this topic. But in this chapter several important aspects of urban resiliency and sustainability are presented, beginning
with the concept of a sustainable city, and proceeding through various elements of urban systems: buildings, energy and climate
action planning, transportation, and stormwater management. The chapter concludes with a case study of a net zero energy home,
one in which perhaps you can envision yourself inhabiting one day.
This page titled 11.1: Sustainable Infrastructure - Chapter Introduction is shared under a CC BY license and was authored, remixed, and/or
curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
11.1.1 [Link]
11.2: The Sustainable City
Introduction
Figure 11.2.1 A diagram depicting the necessary conditions for an ideal living space
Sustainability, from science to philosophy to lifestyle, finds expression in the way we shape our cities. Cities are not just a
collection of structures, but rather groups of people living different lifestyles together. When we ask if a lifestyle is sustainable,
we’re asking if it can endure. Some archaeologists posit that environmental imbalance doomed many failed ancient civilizations.
What could the sustainable city look like, how would it function, and how can we avoid an imbalance that will lead to the collapse
of our material civilization? This module will make some educated guesses based upon the ideas and practices of some of today’s
bold innovators.
Throughout history settlement patterns have been set by technology and commerce. Civilizations have produced food, clothing and
shelter, and accessed foreign markets to purchase and sell goods. Workers traditionally had to live near their place of occupation,
although in modern industrial times advanced transportation systems have enabled us to live quite a distance from where we work.
In hindsight we can see how reliance on water and horse-drawn transportation shaped historical civilizations and how this equation
was radically altered with the rise of the automobile following World War II. While attempting to envision the “Sustainable City”
we must discern what factors will influence its shape and form in the future.
Energy
For the last century energy has been affordable and plentiful, limited mainly by our technological ability to use it. Contemporary
civilization consumes 474 exajoules (474×1018 J=132,000 TWh). This is equivalent to an average annual power consumption rate
of 15 terawatts (1.504×1013 W). The potential for renewable energy is: solar energy 1,600 EJ (444,000 TWh), wind power 600 EJ
(167,000 Wh), geothermal energy 500 EJ (139,000 TWh), biomass 250 EJ (70,000 TWh), hydropower 50 EJ (14,000 TWh) and
ocean energy 1 EJ (280 TWh). Even though it is possible to meet all of our present energy needs with renewables, we do not do so
because the way in which the market prices our fossil reserves. In the current framework, when a company exploits resources it
normally does not account for the loss of resource base or for environmental damage. Gasoline has been cheap in the United States
because its price does not reflect the cost of smog, acid rain, and their subsequent effects on health and the environment let alone
recognize that the oil reserves are being depleted. Scientists are working on fusion nuclear energy; if that puzzle is solved energy
will be affordable, plentiful and carbon neutral. See Environmental and Resource Economics and Sustainable Energy Systems for
more detail.
11.2.1 [Link]
that could eliminate the use of ball bearings in many products as well as industrial diamond dust in automobile air bags. The pearl
oyster uses carbon dioxide to construct its calcium carbonate shell, so a Canadian company developed a technology that reduces
large amounts of CO2 in cement production. 300,000 buildings in Europe use self-cleaning glass that mimics the way water balls
up on lotus leaves and simply rolls off. In the future, we should see more of a return to natural systems as well as the use of new
materials that mimic nature in our sustainable city.
Social Equity
Perhaps the most significant development that separates sustainability from its conservation antecedents is the element of social
equity. The environmental and conservation movements have been criticized in the past for being too "white collar" and promoting
the interests of the “haves”; these movements have traditionally not dealt with the needs of the underclass in the U.S. and,
especially, developing countries. In Agenda 21, the manifesto of the Earth Summit conference on the environment held in Rio de
Janeiro in 1992, sustainable development was viewed as the strategy that would be needed to increase the basic standard of living
of the world's expanding population without unnecessarily depleting our finite natural resources and further degrading the
environment upon which we all depend. The challenge, as viewed at the Earth Summit, was posed in terms of asking humanity to
collectively step back from the brink of environmental collapse and, at the same time, lift its poorest members up to the level of
basic human health and dignity.
The concept of ecological footprint asks each of us to limit resource use to our equitable share. Sitting on the apex of civilization
the Western world is being asked to share the earth’s bounty with the masses of Asia, South America and Africa. If technology
continues to advance we can do this without significant long-term degradation of our standard of living. Short-term economic
dislocations are inevitable, however, as increasing demand from China and India bring us to peak oil and rising transportation costs
will highlight the nexus between location efficiency and affordable housing. The Center for Neighborhood Technology, for
instance, has mapped 337 metropolitan areas covering 80% of the United States population showing how efficient (near mass
transit) locations reduce the cost of living (housing + utilities transportation) and vice versa. The reality of rising transportation
costs could have a significant impact on the shape of the city. Lookin target- g backward we also realize that racial politics were
one of the dynamics that fueled suburban expansion in the 50’s and 60’s decimating many of our urban centers. The Sustainable
City of the future, if it works, will have stably integrated mixed income neighborhoods.
Technology
Computers brought intelligence to the Machine Age and transformed it into the Information Age. Markets run on information and
better information will help the markets perform more efficiently. Whereas in the past surrogate measures were developed to guess
at impacts, information technology can track actual use. For example, do we charge everyone the same fee for water and sewers or
do we measure their use, charge proportionally, and thus encourage landowners to reduce their use? In Miami the (congestion
pricing) charge for driving in the special lanes goes up instantaneously with actual traffic conditions. As the old adage goes, “What
gets measured gets managed,” and as technology increases the precision to which environmental measures, consumption, and
behavior increases, our ability to manage, and therefore reduce negative impacts, will increase.
In the past humans have been one of the beasts of burden and workers have been needed to produce and move goods. Modern
factories reduce human labour needs and artificial intelligence will soon carry most of the load in our partnership with machines. In
the Information Age humans should no longer have to live and work near the factories and centers of commerce and jobs will move
from the production of goods to the provision of services. People may choose to live in exciting urban centers, but if one wants a
bucolic life style telework will offer an alternative.
Shrinking Cities
Many American cities have declined in population from highs immediately following World War II, even as the host metropolitan
area has continued to grow. While populations have declined poverty and other social problems have been concentrated. In the
United States, nearly 5,000,000 acres of vacant property (including brownfields) exist. This is equivalent to the combined land area
of the nation’s 60 largest cities.
Figure 11.2.2 A table depicting the amount of population lost over a 58 year span.
The traditional planning approach has been to focus on managing growth and new development through time-honored tools such as
comprehensive planning, zoning, subdivision regulations, and urban growth boundaries. This growth-oriented approach to
11.2.2 [Link]
addressing the shrinking city is now criticized. Some commentators postulate that green infrastructure offers a better model for
improving these cities health.
How could these factors manifest themselves in our sustainable city? They will influence its design and our settlement patterns will
influence our lifestyles. You are not only what you eat but also where you live. For instance:
Transportation
Most agree that walkability is a key component of any sustainable neighborhood. Walkability not only reduces energy use, but also
increases public health. How can we measure walkability? [Link] identifies and measures nearby amenities and provides a
rating for specific locations and neighborhoods. Try it for where you live.
Sustainable cities could consist of walkable neighborhoods that separate pedestrian, bike and vehicular traffic and are connected to
each other through multiple transportation modes, with biking and mass transit choices in addition to the automobile. Instead of
averaging 10 (auto) trips per unit per day most residents would not use an automobile daily and instead walk or bike or order
online. Trip generation would be reduced through telework and ecommerce. Goods would be brought to residents in bulk delivery
trucks using services like UPS, Fedex and Peapod. Under telework many would only visit an office to attend meetings once or
twice a week (or less). Streets and intersections would be made less daunting to pedestrians through traffic calming techniques.
Figure 11.2.3 Walkable Neighborhood Picture shows a bike path in Chapinero, Bogotá. Source: Tequendamia (Own work) [CC-
BY-SA-3.0] via Wikimedia Commons
Most trips to other parts of the city would be made via mass transit. When an individual car is need it would be provided through
car sharing, or taxis. Much to the dismay of science fiction fans, flying cars sound nice but it is difficult to see how they can be
sustainable until a non-polluting, renewable energy source for air travel is obtained. Traffic congestion would be relieved not
through artificial subsidies (overbuilding roads and providing free parking) but through congestion pricing, removal of free street
parking, and providing viable bicycle and mass transit alternatives.
Even rural centers could be planned as concentrated walkable neighborhoods with viable transportation options. Telework, e-
commerce and low impact cluster development would enable residents to enjoy country living without a guilty conscience. Think
in terms of a kibbutz (collective farm) where most members do not own their own cars and rarely have to use this type of
transportation.
Cities will continue to draw entrepreneurs and foster productivity. Most of the population will reside in mega-regions and linking
them through high-speed rail that connects to mass transit will be the key to long-term economic growth.
Water
Traditional approaches have sought to rapidly move stormwater away from what we’ve built via gutters, sewers and artificial
channels. While this approach on the micro scale is intended to prevent local flooding and undesired ponding, on the macro scale it
may actually cause area wide flooding. It also short-circuits the opportunity for water to naturally soak into the ground – to water
plants and recharge groundwater resources, and, with traditional planting of lawns and other exotics, necessitates bringing more
water in for irrigation.
Best water management practices for sustainable cities would include:
Green Roofs
Downspouts, Rain Barrels and Cisterns
Permeable Paving
Natural Landscaping
Filter Strips
Bioinfiltration using: Rain Gardens
Drainage Swales
Naturalized Detention Basins
Figure 11.2.4 Green Roof. The green roof of City Hall in Chicago, Illinois Source: TonyTheTiger [CC-BY-SA-3.0] via Wikimedia
Commons
These features are discussed in more detail in the Module Sustainable Stormwater Management. The sustainable city would
recharge its local aquifer and surface water would flow primarily from groundwater and not storm water discharge. Instead of
charging a flat tax for storm and sanitary sewer services, technology allows districts to charge usage fees based upon volume, thus
11.2.3 [Link]
providing a financial incentive for sustainable design. Governmental bodies could use these tools to encourage the reorientation and
designers could use the techniques outlined above as divert stormwater into the ground rather than directly into surface water.
Working together, local government and project planners could also retrofit older urban streets with attractive walkable
streetscapes, just as Lansing, Michigan has done as part of its combined sewer overflow project (below).
Some visionaries see even more dramatic transformations in the way we deal with water. Sarah Dunn and Martin Felsen of
Urbanlab envision Chicago’s evolution into a model city for “growing water” by creating a series of Eco-Boulevards that function
as a giant Living Machine – treating the city’s waste and storm water naturally, using micro-organisms, small invertebrates such as
snails, fish, and plants. Under their plan treated water would be returned to the Great Lakes Basin and create a closed water loop
within Chicago, instead of being exported to the Mississippi and Gulf Coast.
Food
The ancients would wonder at modern supermarkets with their food from all over the world and fresh fruits and vegetables all year
round. Yet most environmental activists advocate locally produced organic food. Peak oil will raise petrochemical costs and upset
the dynamics of modern agriculture, but it will be difficult to change the acquired tastes of the consuming public.
USEPA is encouraging urban agriculture as one of the solutions to the shrinking city. It sees urban agriculture as not only providing
a use for vacant land, (thus addressing blight and the deleterious affect of neglect on property values) but also as a potential cleanup
strategy for contamination. It addresses the problem of food deserts (lack of healthy, affordable, fresh produce) in blighted inner
city neighborhoods while educating children and adults about farming and local enterprise. Practitioners have found that urban
farming enhances social capital and community connections. Victory gardens produced about 40% of American vegetables
consumed during World War II and urban gardens could be a prime user of the compost that we could generate either through
individual compost bins or through collective efforts performed on a large scale by our waste haulers.
Figure 11.2.5 Urban Agriculture. Some new crops being started, protected by shade cloth barriers to the west. Note the new
construction in the background. This area used to be all public housing. The high rise "warehouses of the poor" were torn down and
are being replaced with mix of market-rate and low-income housing (also called mixed income housing.) The 1.5 acre parcel that
City Farm sits on is owned by the City of Chicago and provided, rent-free, to this non-profit initiative. The property is valued at $8
million, however, so it's anyone's guess as to when the city decides to terminate the agreement and City Farm must move again.
Source: Linda from Chicago, USA (New crops) [CC-BY-2.0] via Wikimedia Commons
Many of us might belong to food cooperatives where the members contract with organic farmers to purchase the food grown for
them. In Sustainable Cities farming will not just be an interim land use in blighted neighborhoods but, like Prairie Crossing in
Grayslake, Illinois, will be an integral part of the community plan. Some even forecast vertical farming in the great cities of the
world.
11.2.4 [Link]
LEED – Leadership in Energy and Environmental Design, a voluntary effort by the U.S. Green Building Council, includes more
than energy and also gives points for site, water, materials and resources, and indoor air quality. LEED in particular and sustainable
construction in general have found widespread acceptance as even the National Association of Homebuilders has rolled out its own
version of green construction. Since its inception in 1998, the U.S. Green Building Council has grown to encompass more than
10,000 projects in the United States and 117 countries covering 8 billion square feet of construction space and in April, 2011 LEED
certified its 10,000th home in its LEED for Homes program. The U.S government’s General Services Administration (GSA), the
part of the federal government that builds and manages federal space, currently requires LEED Gold certification for all of its new
buildings (up from Silver). In addition to using fewer resources sustainable buildings reduce absenteeism, improve employee
morale, and lead to improved educational performance.
In the Pacific Northwest the International Living Future Institute has set up a Living Building Challenge to go beyond LEED and
design and build triple net zero (storm water, energy, wastewater) structures. As of Fall 2010 there were 70 registered projects.
What about the neighborhoods where we live and raise our families? Many now recognize that our grandparents' mixed-use,
walkable neighborhoods were more sustainable than today’s reality. The Congress for New Urbanism promotes mixed use in
contrast to its predecessor, Congrès International d'Architecture Moderne (CIAM), which promoted separation of use. CIAM,
active in the first part of the 20th century, proposed that the social problems faced by cities could be resolved by strict functional
segregation, and the distribution of the population into tall apartment blocks at widely spaced intervals. This view found its
expression in Le Corbusier’s The Radiant City (1935). Separation of Use heavily influenced subdivision and building codes that, in
turn, shaped Post World War II suburban expansion. In our suburbs zoning dictates mutually exclusive uses in each district so that
Industrial use is exclusive of commercial, which is exclusive of residential. In the suburbs separation of use combined with the
platting of superblocks to replace the traditional grid network gives us a lifestyle that produces 10 auto trips per unit per day,
because you need one car per driver to get around where much of America lives.
CIAM’s view also formed the intellectual underpinning for large-scale high-rise public housing projects. Today we recognize that
safe, sound and sanitary housing is not just indoor plumbing and more bedrooms, and that affordable housing is not just rent but
includes utility and transportation costs and the right to live in a safe, mixed income, stably integrated neighborhood. Our
sustainable city should stand upon the leg of social equity and include ethnic and income diversity. Neighborhoods should be sited
at efficient locations. with broad transportation choices. What will they look like? Most new urbanists think they will be similar to
the diverse neighborhoods built at the turn of the last century. Other visionaries, such as Moshe Safdie, think it possible to integrate
the variety and diversity of scattered private homes with the economics and density of a modern apartment building. Modular,
interlocking concrete forms in Safdie’s Expo ’67 defined the space. The project was designed to create affordable housing with
close but private quarters, each equipped with a garden. In a different vein, in outlying Grayslake, Illinois, cluster development that
incorporates open space, wetlands, and a working organic farm enables residents to live (somewhat) sustainably in the country. Our
future must recognize that we don’t want everyone to live in Manhattan or Brooklyn, and we must provide for diverse tastes and
lifestyles.
Will everything look futuristic, like the sets of Blade Runner or Star Wars? Historic Buildings not only have a special charm but
they represent a great deal of embodied energy that is wasted if they are demolished. The government and marketplace will
probably continue to promote historic rehab, including adaptive reuse where new uses are found for old buildings through rehab
that installs modern utilities and fixtures while preserving the outer shell’s look and feel. In Chicago the Sears Powerhouse was
converted to a charter school and in Philadelphia Urban Outfitters took the old Navy Yard and transformed it into a new corporate
headquarters. In all five of the former Navy Yard buildings, employees work in light-filled interiors with open layouts. Most of the
furnishings are custom-made and contain recycled material (tabletops crafted from salvaged wood, for instance). Amenities such as
a gym, yoga studio, dog park, and farmers’ market further add to the lively and informal atmosphere.
All of these gestures to what the CEO calls “a quality of life thing” help Urban Outfitters boost employee satisfaction. Since
moving into the new headquarters, employee turnover has dropped to 11 percent, and fewer sick days are being used. “They feel
more linked to the community and culture of the company.” The campus has improved his company’s ability to attract new talent.
The informal atmosphere is alluring to Millennial-aged employees, who tend to value open, flexible work arrangements more than
previous generations of workers. “The campus has improved creative collaboration, which ultimately impacts our bottom line.”
11.2.5 [Link]
most of the work in manufacturing our goods and moving them from place to place. In the Information Age most tasks can be
performed anywhere within reach of the World Wide Web. In Understanding Media Marshall McLuhan showed how “improved
roads and transport have reversed the ancient pattern and made cities the centers of work and the country the place of leisure and of
recreation.” The new reality once again reverses roles by locating the factories on vast campuses away from the people and, the city
becomes a place to meet, be entertained and educated. Sustainable Cities of the future will probably still function as the center for
service industries such as health and beauty care, hospitality, tourism, travel, and government, and other service industries, such as
insurance, advertising and marketing, and financial services, are amenable to telework. Sure, we can eat our frozen dinners and get
TV or on line entertainment at home, but it’s still enjoyable to go out to eat and catch a live show, concert, sporting event or movie.
We get a better view of the players on television, but the excitement of thousands of fans under one roof is palpable. Many of us in
the postindustrial Information Age will not have to live near our factories, power plants or transportation centers, because we just
have to connect to the World Wide Web. In the Information Age many of us might never have to attend physical meetings, but
those of us who do might find ourselves going to the office only a few times a month. But it is important for those who tout cities
as more sustainable places to live (i.e. Cities pollute less per capita) to understand that rural areas can be just as benign as cities if
one has the will and controls resource use with the appropriate life style. Communal farms, for example, generally have small
ecological footprints, producing more resources than they consume.
Figure 11.2.7 Telecommuting Source: Gilangreffi (Own work) [CC0] via Wikimedia Commons
Power
Today huge plants generate electricity from coal (44.9%), natural gas (23.8%), atomic energy (19.6%), hydroelectric (6.2%), and
other renewable (e.g. wind, solar) (4%) sources. Power affects urban form in that urban centers must be connected to the grid, and
the ubiquitous power line tethers us to the power plant. Sustainable cities will probably not lose the grid, but should accommodate
those who want to produce their own power by running the meter backwards. Until and unless atomic fusion supplies cheap, safe,
reliable power, renewables will compete with fossil fuels. Even as we hit peak oil, coal will be plentiful for the foreseeable future.
Coal will continue to be cheap because its price will probably not reflect all the costs of smog, acid rain, and its subsequent effects
on health and the environment let alone recognize that the reserves are being depleted. As the technology evolves and as
government policy requires utilities to buy power from small decentralized sources we will all get used to wind mills and
photovoltaic arrays. Geothermal heat pumps will heat and cool space more efficiently. Zoning and building codes will have to be
revised to deal with solar access rights, noise from windmills and odors from biomass and biofuels.
Figure 11.2.8 Windmills Source: James McCauley from Enon, OH, United States of America (Flickr) [CC-BY-2.0], via Wikimedia
Commons
Figure 11.2.9 Local wind generator, Spain, 2010 Source: By Patrick Charpiat (Own work) [CC-BY-SA-3.0] via Wikimedia
Commons
Technology will also enable us to operate more efficiently. One of the problems with conventional power generation is that the
plants must be (over)sized to accommodate peak loads. In the future the “Smart Grid” should smooth peak loads by instructing
consumer appliances to perform tasks, such as laundry and dishwashing, in low demand periods (middle of the night), and will
offer lower rates as an incentive.
Figure 11.2.10 Microgrids A local microgrid in Sendai, Japan Source: See page for author [Public domain], via Wikimedia
Commons
Commerce
Our parents and grandparents have seen dramatic change in the area of commerce. Until 1950 we walked to the neighborhood store
and most of the goods we bought were produced locally. On special occasions you’d take the trolley downtown to the central
business district (CBD) to visit the large department stores (e.g. Marshall Fields in Chicago, Macy’s in New York). With the
suburbanization following World War II CBD’s were replaced by suburban malls. We drove to the malls. The CBD’s died out. For
the last thirty years big box chains have dominated retail, but most recently ecommerce entices consumers with better selection and
prices (and sometimes no sales tax). Most of us find that we shop online more efficiently and those that need the personal attention
that retail establishments currently offer might find that they will have to engage the services of a personal shopper (note the shift
from good to service). Most of the goods we purchase, even food, have been produced somewhere else. World trade, as measured
in US dollars at current prices, has grown astronomically, from $10.1 billion in 1900, to $62 billion in 1950, to $15.2 trillion in
2010. US imports and exports have risen from $1.4 billion (exports) in 1900 to $9.6/10.3 billion in 1950 and to $1.97/1.28 trillion
in 2010.
11.2.6 [Link]
In our Sustainable City of the Future ecommerce will probably rule, which will mean a reduction in actual physical commercial
floor area. When we feel we need to see something in person we will visit centralized showrooms, and pay for that service, but
most purchases will be made on line and the product delivered either by company delivery vehicles (e.g. Peapod) or through
common carriers (e.g. UPS, Fedex). New uses will be found for dead malls. Neighborhood stores will fare better in walkable
neighborhoods, especially those that offer services in place of or in addition to goods.
It is unreasonable for us to expect to obtain all of our goods locally, but regional specialization in the production of goods will
reflect climate, access to raw materials and markets, and locally developed expertise rather than cheap labor and the ability to avoid
environmental and workplace safety regulations. Modern robots do not have to follow repetitive assembly line logic of repeated
application of the same exact set of instructions. They can be programmed to intelligently consider each individual product and
thus can produce a wider variety of products, and many products will be special ordered to fit individual tastes. Once we even out
the playing field (of necessary government regulation to deal with the externalities of production) the cost of shipping should work
towards more local production. Goods will be made “just in time” thus reducing inventories and overproduction that is sent to
landfills. Efficiency will be a big part of the Sustainable City.
Planned obsolescence and its impact upon material culture is more problematic. Fashion defies logic but speaks to the most basic
instincts of human behavior. How do we avoid the need for more closet space, let alone offsite storage? New materials will enable
us to change the look and feel of clothing. Advances in crystal technology, for example, will allow us to change its color and/or
pattern. Interlocking parts using material similar to Velcro could let us change lapels, sleeves and other components of garments,
much as we now add or subtract liners for warmth. More complicated goods will be designed for disassembly and recycling, or
replacement of key parts while keeping most of the old components. Retrofit can add life to old buildings and machines, if we learn
to view old and retro as cool and in.
A good example of how we can move towards sustainability is provided by Interface, which repositioned their carpeting business
from the sale of goods to the leasing of floor covering services. Under the old paradigm the consumer purchased a new carpet a few
years after the old one began to show wear. The selection was based upon the perception that the carpet would last (e.g. because it
felt thicker), but the reality was that the carpet manufacturer made out better when new carpets had to be repurchased more
frequently. Under the new leasing paradigm the manufacturer owns the carpet and therefore it is in their best interest to have it last
longer. Interface’s Solenium lasts four times longer and uses 40% less material than ordinary carpets, an 86% reduction in materials
intensity. When marketed as floor covering services under its Evergreen Lease, modular floor tiles are replaced as soon as they
show any wear and, since 80% of wear takes place on 20% of the area, this reduces material intensity by another 80%. In other
words, 3% of the materials are used under the new paradigm (yes, a 97% reduction) than the old one. And the worn out panels are
recycled, and no chlorine or other toxic materials are used in its manufacture.
Education
Our children will find their place in tomorrow’s workplace based upon their brains, not brawn. Education is the most important
component in preparing for tomorrow’s workplace. Classrooms today link via television and the internet to amazing resources.
More importantly, artificial intelligence has the capacity to treat each student as an individual and to tailor instruction to meet his or
her individual abilities and needs, in contrast to the classroom that moves, at best, at the speed of the average student.
Figure 11.2.11 E-Learning A ubiquitous-learning (u-learning) classroom where students use electronic textbooks in the form of
tablet PCs. Source: By B.C. (Own work) [CC-BY-3.0] via Wikimedia Commons
The sustainable school should still contain classrooms, but it will probably be supplemented by individualized computer learning
labs. Each student would have their own personal computer (see Cyborgean Man below) that would link them to the internet.
Classrooms would have multi-media capabilities that would link to other classrooms around the world. Webinars would make
expert instruction available to all. As noted earlier, the biophilia of green classrooms would improve learning and test scores.
11.2.7 [Link]
Miniaturization in the Computer Age now promises to let us reinternalize some of the external abilities we’ve created. Over ten
years ago Thad Starner, a research assistant at MITs media lab, garnered a lot of publicity by calling himself a cyborg because he
incorporated his computer monitor into his glasses, and let his keyboard and computer hang as appendages by his side. Today half
the civilized world has smartphones that link us via the Facebook, the World Wide Web, texts and tweets. In our sustainable city we
could all be wired, possibly though implants directly into out brainstem. Artificial intelligence could provide a virtual butler that
would be available to schedule and keep track of our appointments, order our groceries and other items, and help us with things we
can’t even imagine today. Advertising would be directed at us individually, as in Philip Dick’s Minority Report, as advertisers use
software like doubleclick and links to national databases (see Big Brother below) to track who we are and our buying preferences.
This should be good for sustainability, as there will be no need to hoard things for possible future use, and to discard them when we
no longer need them. Government would find it easier to poll citizen’s preferences and opinions.
Figure 11.2.12 Cyborgean Man Source: PIX-JOCKEY (Roberto Rizzato) [Link]
In George Orwell’s 1984 everyone is constantly reminded, “Big Brother is watching you.” In our sustainable city everyone will be
watching everyone else. Video surveillance webcams might be everywhere, and everyone would have access to them. Everyone
could have a reality show in which those wanting to follow you would merely tune into the appropriate URL and be able to choose
the cameras and microphones that are in range. Some people would not even turn off the feeds, ever. Artificial intelligence (or
actual humans for the more popular content) could provide edited condensed feeds. Crime would go down in those areas with eyes
and ears, but crime will evolve and persist. It will be easy to research everyone you meet and to stay connected. Since this will be
by and large electronic it will promote a sustainable lifestyle by not consuming additional resources.
Footnotes
1. Montgomery, David, Dirt, The Erosion of Civilizations, University of California Press, 2007
2. Energy - Consumption'!A1 "Consumption by fuel, 1965 - 2008" (XLS). Statistical Review of World Energy 2009, BP. July 31,
2006. Retrieved 2009-10-24.
3. State of the world 2009, Worldwatch institute, 2009
4. Hawken, Paul; The Ecology of Commerce, a Declaration of Sustainability; Harper Buisness, 1993, p.76
5. [Link]/cityparks/br...[Link], Accessed 4/29/11
6. [Link]/sites/defaul...t/[Link], accessed 4/30/11
7. [Link] accessed 4/30/11
8. [Link]/wp-dyn...[Link], accessed 4/29/11
9. Sitarz, Daniel, editor; 1993; AGENDA 21: The Earth Summit Strategy to Save Our Planet; Earthpress; Boulder, Co; 1993; p.4
10. 2005-2007 American Community Survey three-year estimates
11. Joe Dufficy, USEPA, [Link]/portals/30/B...ks%[Link] accessed 4/25/11
12. Dawkins, Casey J., and Arthur C. Nelson. (2003). “State Growth Management Programs and
13. Central City Revitalization.” Journal of the American Planning Association, 69(4), 381-396
14. Oswalt, Philipp. 2006. Shrinking Cities Volume 2: Interventions. Ostfildern, Germany: Hatje, Cantz Verlag
15. Joseph Schilling, Blueprint Buffalo—Using Green Infrastructure to Reclaim America’s Shrinking Cities,
[Link]/pub...g_PA_final.pdf, accessed 4/25/11
16. based upon Institute of Traffic Engineers Trip Generation Manual, 1997 from [Link]/mdot/planning-p...rams/[Link]
accessed 4/25/11
17. Lincoln institute of Land Policy, April 26, 2100, [Link]/news/[Link], accessed 4/27/11 quoting
Edward Glaeser, author of "The Triumph of Cities,"
18. Lincoln institute of Land Policy, April 26, 2100, [Link]/news/[Link], accessed 4/27/11 quoting Petra
Todorovich, America 2050,
19. Conservation Design Forum from A Guide to Stormwater Best Management Practices,
[Link]/xm_client/clie...water_BMPs.pdf, accessed 4/25/11
20. Newport, Robert, USEPA lecture to UPP 594 class 9/7/10
21. Newport, Robert, USEPA lecture to UPP 594 class 9/7/10
22. [Link]/brownfields/urbanag/[Link] accessed 4/25/11
23. [Link]/2011/03/10...ctory-gardens/ accessed 4/25/11
24. [Link] accessed 4/25/11
25. Margot Adler, NPR, Behind the Ever-Expanding American Dream House, [Link]/templates/story/s...toryId=5525283,
accessed 4/25/11
11.2.8 [Link]
26. Lincoln institute of Land Policy, April 26, 2100, [Link]/news/[Link], accessed 4/27/11 quoting Arthur
C. Chris Nelson, University of Utah
27. Zhu Xiao Di, Nancy McArdle and George S. Masnick, Second Homes: What, How Many, Where and Who, February 2001,
[Link] accessed 4/25/11
28. [Link] accessed 4/25/11
29. [Link]/your_hom.../mytopic=10360, accessed 4/25/11
30. [Link]/[Link]?CategoryID=19, accessed 4/28/11
31. [Link] accessed 4/28/11
32. see [Link]/publication_deta...icationID=3887
33. [Link]/Docs/News/LEED%...ril%[Link], accessed 4/28/11
34. [Link] accessed 4/28/11
35. [Link]/Libraries/...AL_1.[Link], accessed 4/20/11
36. [Link] accessed 4/28/11
37. [Link]
38. [Link]/wiki/Congr%C...ecture_Moderne
39. [Link]/news/category/location-efficiency/, accessed 4/25/11
40. [Link]/case-studi...porate-campus/, Accessed 4/29/11
41. McLuhan, Marshall, Understanding Media, McGraw Hill, NY, 1964, p. 38
42. Hoornweg D., Sugar L., Lorena Trejos Gomez C., “Cities and greenhouse gas emissions: moving forward,” 2011,
[Link] accessed 4/25/11
43. US Energy Information Administration, Net Generation by Energy Source: Total (All Sectors), 1996 through December 2010,
[Link]/cneaf/electricity.../table1_1.html, accessed 4/25/11
44. Hawken; op cit p.76
45. See Section 2.7.2 & also CNT [Link]/news/media/brockw...[Link], accessed 4/25/11
46. United Nations, INTERNATIONAL TRADE STATISTICS 1900 – 1960, [Link]
accessed 4/26/11
47. extracted from world Trade Organization database 4/26/11, [Link]/StatisticalProgr...spx?Language=E
48. [Link] accessed 4/26/11
49. [Link]/wiki/Just-in-time_(business), accessed 4/27/11
50. [Link] accessed 4/29/11
51. Sustainable Design - Not Just for Architecture Any More, [Link]/~roy/classno...rks_design.pdf, accessed 4/26/11
52. Lovins, Lovins, and Hawken, A Road Map to Natural Capitalism, 1999, Harvard Business REview
53. McLuhan, Marshall, Understanding Media; The Extensions of Man, McGraw Hill, NY, 1964
54. Paradise lost – the Hebrew word for forest is Pardais - paradise
55. CNN at [Link]/1998-07-23/t...eld?_s=PM:TECH, accessed 4/27/11
56. As in Philip Dick’s Minority Report, published in Fantastic Universe, January 1956,
57. [Link] accessed 4/27/11
58. Orwell, George (1949). Nineteen Eighty-Four. A novel. London: Secker & Warburg
Review Questions
1. Select an aspect of your day-to-day existence that has environmental consequences. Describe the environmental consequences,
and briefly discuss more sustainable alternatives.
2. What does a complete street look like? How does it differ from the street outside of your home
3. Describe a sustainable neighborhood that you’re familiar with and explain what makes it sustainable.
4. If sustainability is so beneficial why isn’t everything sustainable? Name one market barrier to sustainability and explain what
can be done to overcome it.
Glossary
Biomimicry
Biomimicry or biomimetics is the examination of nature, its models, systems, processes, and elements to emulate or take
inspiration from in order to solve human problems. The terms biomimicry and biomimetics come from the Greek words bios,
11.2.9 [Link]
meaning life, and mimesis, meaning to imitate. Examples include adhesive glue from mussels, solar cells made like leaves, fabric
that emulates shark skin, harvesting water from fog like a beetle, etc.
E-commerce
Electronic commerce, commonly known as e-commerce, eCommerce or e-comm, refers to the entire online process of developing,
marketing, selling, delivering, servicing and paying for products and services.
Ecological Footprint
Ecological footprint is a measure of human demand on the Earth's ecosystems. It is a standardized measure of demand for natural
capital that may be contrasted with the planet’s ecological capacity to regenerate. It represents the amount of biologically
productive land and sea area necessary to supply the resources a human population consumes, and to mitigate associated waste.
Using this assessment, it is possible to estimate how much of the Earth (or how many planet Earths) it would take to support
humanity if everybody followed a given lifestyle.
Low Impact Cluster Development
Low impact cluster development is the grouping of buildings on a portion of the site and devoting the undeveloped land to open
space, recreation or agriculture. Though cluster development lowers development cost through savings on roads and infrastructure
(sewers, electric and water lines, etc.), it has issues such as conflicts with many older zoning ordinances, perceptions of personal
space (lower individual lot size) and maintenance of common areas.
McMansion
A slang term that describes a large, opulent house that may be generic in style and represents a good value for a homebuyer in
terms of its size. This type of home is built to provide middle and/or upper middle class homeowners with the luxurious housing
experience that was previously only available to high-net-worth individuals.
Postindustrial Information age
Is a way of capturing the nature of western economies, in which most people are no longer engaged in the production of goods
(which is highly automated) but rather deal with the publication, consumption, and manipulation of information, especially by
computers and computer networks. A post-industrial society has five primary characteristics: the domination of service, rather than
manufacturing, the pre-eminence of the professional and technical classes, the central place of theoretical knowledge as a source of
innovations, the dominating influence of technology, and levels of urbanization higher than anywhere else in the world.
This page titled 11.2: The Sustainable City is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto Cabezas
(GALILEO Open Learning Materials) .
11.2.10 [Link]
11.3: Sustainability and Buildings
Introduction
Buildings present a challenge and an opportunity for sustainable development. According to the most recent available Annual
Energy Outlook from the U.S. Environmental Information Administration, buildings account for about 39% of the carbon dioxide
emissions, 40% of primary energy use, and 72% of the electricity consumption in the U.S. Additional information from the U.S.
Geological Survey indicates that 14% of the potable water consumption occurs in buildings.
Globally, buildings are the largest contributors to carbon dioxide emissions, above transportation and then industry. The
construction of buildings requires many materials that are mined, grown, or produced and then transported to the building site.
Buildings require infrastructure including roads, utility lines, water and sewer systems. People need to be able to get to and from
buildings to work, live, or take advantage of the services provided within them. They need to provide a safe and comfortable
environment for the people that inhabit them.
Table 11.3.1 Impacts of the Built Environment Source: U.S. Environmental Protection Agency [Link]/greenbuilding/pubs/[Link]
Aspects of Built Environment Consumption Environmental Effects Ultimate Effects
Noise
It is possible to design and construct fully functional buildings that have far fewer negative environmental impacts than current
norms allow. Beyond benefiting the environment, green buildings provide economic benefits including reduced operating costs,
expanded markets for green products and services, improved building occupant productivity, and optimized life-cycle performance.
Green buildings also offer social benefits that range from protecting occupant comfort and health, to better aesthetic qualities, less
strain on local infrastructure, and overall improvement in quality of life.
In 1994, a group of experts was brought together by the National Renewable Energy Laboratory (NREL) to develop a pathway and
specific principles for sustainable development. According to these principles, building should be:
Ecologically Responsive: The design of human habitat shall recognize that all resources are limited, and will respond to the
patterns of natural ecology. Land plans and building designs will include only those with the least disruptive impact upon the
natural ecology of the earth. Density must be most intense near neighborhood centers where facilities are most accessible.
Healthy, Sensible Buildings: The design of human habitat must create a living environment that will be healthy for all its
occupants. Buildings should be of appropriate human scale in a non-sterile, aesthetically pleasing environment. Building design
must respond to toxicity of materials, care with EMF, lighting efficiency and quality, comfort requirements and resource
efficiency. Buildings should be organic, integrate art, natural materials, sunlight, green plants, energy efficiency, low noise
levels and water. They should not cost more than current conventional buildings.
Socially Just: Habitats shall be equally accessible across economic classes.
Culturally Creative: Habitats will allow ethnic groups to maintain individual cultural identities and neighborhoods while
integrating into the larger community. All population groups shall have access to art, theater and music.
Beautiful: Beauty in a habitat environment is necessary for the soul development of human beings. It is yeast for the ferment of
individual creativity. Intimacy with the beauty and numinous mystery of nature must be available to enliven our sense of the
sacred.
Physically and Economically Accessible: All sites within the habitat shall be accessible and rich in resources to those living
within walkable (or wheelchair-able) distance.
11.3.1 [Link]
Evolutionary: Habitats' design shall include continuous re-evaluation of premises and values, shall be demographically
responsive and flexible to change over time to support future user needs. Initial designs should reflect our society's
heterogeneity and have a feedback system.
What is meant by a sustainable or green building? The U.S. EPA defines green building as “the practice of creating structures and
using processes that are environmentally responsible and resource-efficient throughout a building's life-cycle from siting to design,
construction, operation, maintenance, renovation and deconstruction. This practice expands and complements the classical building
design concerns of economy, utility, durability, and comfort.” (U.S. Environmental Protection Agency, 2010)
The benefits of sustainable buildings have already been documented. These buildings can reduce energy use by 24-50%, carbon
dioxide emissions by 33-39%, water use by 40%, and solid waste by 70% (Turner & Frankel, 2008; Kats, Alevantis, Berman,
Mills, & Perlman, 2003; Fowler & Rauch, 2008). Green building occupants are healthier and more productive than their
counterparts in other buildings, and this is important because in the U.S, people spend an average of 90% or more of their time
indoors (U.S. Environmental Protection Agency, 1987). Green buildings tend to have improved indoor air quality and lighting.
There are also numerous perceived business benefits to green buildings, including decreased operating costs and increased building
value, return on investment, occupancy ratio, and rent ratio.
11.3.2 [Link]
Energy-saving Building Features
Energy efficient measures have been around a long time and are known to reduce the use of energy in residential and commercial
properties. Improvements have been made in all of these areas and are great opportunities for further innovation. Green buildings
incorporate these features to reduce the demand for heating and cooling.
Insulation
The building should be well insulated and sealed so that the conditioned air doesn’t escape to the outside. Insulation can be
installed in floors, walls, attics and/or roofs. It helps to have more even temperature distribution and increased comfort as well.
High-performance Windows
Several factors are important to the performance of a window (see Figure 11.3.1):
Thermal windows are at least double-paned and vacuum-filled with inert gas. This gas provides insulation
Improved framing materials, weather stripping and warm edge spacers reduce heat gain and loss
Low-E coating block solar heat gain in the summer and reflect radiant heat indoors during the winter
(north in the southern hemisphere) exposed to full sun. Heat is trapped, absorbed and stored by materials with high thermal mass
(usually bricks or concrete) inside the house. It is released at night when needed to warm up the building as it loses heat to the
cooler outdoors. Shading provided by trees or shades keeps the sun out in the hot months.
Figure 11.3.2 Passive Solar Design Source: [Link]/technical/[Link]#what
Lighting
Well-designed lighting can minimize the use of energy. This includes enhancing day lighting (natural light), through windows,
skylights, etc. Using energy efficient lighting such as compact fluorescent light bulbs and LEDs (light-emitting diodes) can save
11.3.3 [Link]
energy as well. Using occupancy sensors also means that lights will only be on when someone is in a room. See Module 13.4 for
more energy-saving technologies that can be incorporated into buildings.
Water
Water usage can be minimized by using low-flow fixtures in restrooms, bathrooms, and kitchens. Dual-flush toilets allow for the
user to have the option of select less water (e.g. for liquid waste) and more water (e.g. for solid waste) when flushing (See Figure
11.3.3). These have long been in use in Europe, the Middle East and other places where water conservation is paramount. Fresh
water consumption can be reduced further through the use of greywater systems. These systems recycle water generated from
activities such as hand washing, laundry, bathing, and dishwashing for irrigation of grounds and even for flushing toilets.
Figure 11.3.3 Dual-flush Toilet This toilet has two flush controls on the water tank. Pushing only the circular button releases half
as much (0.8 gallons, 3 liters) water as pushing the outer button. Source: By Eugenio Hansen, OFS (Own work) [CC-BY-SA-
3.0],via Wikimedia Commons
Integrated Design
Integrated design is a design process for a building that looks at the whole building, rather than its individual parts, for
opportunities to reduce environmental impact. Incremental measures would include those approaches described above. To
accomplish integrated design of a building, all parties involved in the design--architects, engineers, the client and other
stakeholders--must work together. This collaborative approach results in a more harmonious coordination of the different
components of a building such as the site, structure, systems, and ultimate use.
Standards of Certification
Most countries establish certain standards to assure consistency, quality and safety in the design and construction of buildings.
Green building standards provide guidelines to architects, engineers, building operators and owners that enhance building
sustainability. Various green building standards have originated in different countries around the world, with differing goals, review
processes and rating. In this section we will discuss a few examples.
A good certification system should be developed with expert feedback. In addition, it should be transparent, measurable, relevant
and comparable.
Expert-based: Was input acquired from experts and professionals in the fields of design, construction, building operation and
sustainability?
Transparent: Is information readily available to the public about how buildings are rated?
Measurable: Does the rating system use measurable characteristics to demonstrate the extent of sustainable design incorporated
into the building? Does the system use life-cycle analysis to evaluate?
Relevance: Does the rating system provide a “whole building evaluation” rather than an evaluation of an individual design
feature?
Comparable: Is the rating system able to compare building types, location, years, or different sustainable design features?
Comparison of Certification Systems Source: Klein-Banai, C.
Year Country of Measurable/
System Trans- parent Expert-based Relevance Comparable
established origin Uses LCA
BREEAM 1990 UK √* - √ √ √
ENERGY
1999 US √ √# √ Only energy √
STAR
*Only assessment prediction check lists available publicly
#Benchmarking tool developed by US EPA
11.3.4 [Link]
Year Country of Measurable/
System Transparent Expert Based Relevance
Established Origin Uses LCA
BREEAM 1990 UK √* - √ √
ENERGY
1999 US √ √** √ Only energy
STAR
* Only assessment prediction check lists available publicly
** Benchmarking tool developed by US EPA
Conclusion
The built environment is the largest manifestation of human life on the planet. Buildings have been essential for the survival of the
human race, protecting us from the elements and forces of nature. However, they also consume a lot of material, energy and water,
and they occupy land that might otherwise be undeveloped or used for agriculture. There are many ways to reduce that impact by
building to a higher standard of conservation and reuse. There are a number of systems that can help architects, engineers, and
planners to achieve those standards, and they should be selected with a full awareness of their limitations.
References
1. Fowler, K.M. & Rauch, E.M. (2008). Assessing green building performance. A post occupancy evaluation of 12 GSA
buildings. (U.S. General Services Administration). PNNL-17393 Pacific Northwest National Laboratory Richland, Washington.
Retrieved from [Link]/graphics/pbs/GSA_...ull_Report.pdf
2. Horvath, A. (2004). Construction materials and the environment. Annual Review of Energy and the Environment, 29 , 181-204.
3. Humphreys, K. & Mahasenan, M. (2002). Toward a Sustainable Cement Industry. Substudy 8, Climate Change. World Business
Council for Sustainable Development. Retrieved from [Link]/web/publication...[Link]
4. Kats, G., Alevantis, L., Berman, A., Mills, E. & Perlman, J. (2003). The costs and financial benefits of green building: A report
to California’s sustainable building task force. Retrieved from [Link]/Docs/News/[Link]
5. Turner, C. & Frankel, M. (2008). Energy performance of LEED for New Construction Buildings, Final Report. Retrieved from
[Link]
6. U.S. Department of Energy. (2011). Energy savers: Furnaces and boilers. Retrieved from
[Link]/your_hom.../mytopic=12530
7. U.S. Environmental Protection Agency. (1987). The total exposure assessment methodology (TEAM) study (EPA 600/S6-
87/002). Retrieved from [Link]/pub/repor..._book_1987.pdf
8. U.S. Environmental Protection Agency. (1998). Characterization of building-related construction and demolition debris in the
United States. (Report No. EPA530-R-98-010). Retrieved from [Link]/wastes/hazard/gen...sqg/[Link]
9. U.S. Environmental Protection Agency. (2010). Green Building Basic Information. Retrieved from
[Link]/greenbuilding/pubs/[Link].
Review Questions
1. What are the positive and negative impacts that buildings have on the environment and society?
2. How can those impacts be reduced?
3. What would be the advantages and disadvantages of demolishing an old building and replacing it with a new, highly
“sustainable” building vs. renovating an old building to new standards?
Glossary
Deconstruction
The selective dismantling or removal of materials from buildings prior to or instead of conventional demolition.
Envelope
11.3.5 [Link]
The physical barrier between the interior and exterior of a building including the walls, roof, foundation, and windows.
Greywater
The water generated from activities such as handwashing, laundry, bathing, and dishwashing that can be recycled on-site to be used
for irrigation of grounds and even for flushing toilets.
Low-emissions
Materials that have little to no volatile organic compounds and other toxic chemicals that are released into the environment after
installation.
Thermal Mass
The ability of a material to absorb heat energy. High density materials like concrete, bricks and tiles need a lot of heat to change
their temperature and thus have a high thermal mass. Lightweight materials such as wood have a low thermal mass.
This page titled 11.3: Sustainability and Buildings is shared under a CC BY license and was authored, remixed, and/or curated by Heriberto
Cabezas (GALILEO Open Learning Materials) .
11.3.6 [Link]
11.4: Sustainable Energy Practices - Climate Action Planning
Introduction
Traditionally, the United States has relied on fossil fuels with minimal use of alternatives to provide power. The resources appeared
to be unlimited and they were found within our borders. As our population has grown and our reliance on power increased, our
resources are decreasing. As discussed in Module 10.2, this is particularly true of petroleum oil, which primarily powers
transportation. Our electrical grid and transportation infrastructure of roads and highways support these fossil fuel dependent
technologies. Fossil fuels store energy well, are available upon demand (not weather dependent), and are inexpensive. However, as
we saw in Module 10.2 there are many environmental, social, and even economic impacts of using these nonrenewable fuel sources
that are not accounted for in the traditional methods of cost accounting. Further, the oil industry has been provided with many
subsidies or tax incentives not available to other energy industries.
How do we move to a more sustainable energy economy? We need to pay more attention to the environment, humans, biodiversity,
and respecting our ecosystems. It means finding ways to share our resources equitably both now and in the future so all people can
have an equal opportunity to derive benefits from electricity, motorized transportation systems, industry, and conditioned indoor
environments. At the same time, we must preserve human health and protect the natural world.
Energy use is one big piece of the sustainability puzzle, but it is not the only one. Changing the way we use energy is not easy
because of infrastructure, the vision of the American Dream (own a house with a big yard, a big car, independence), changing
government policy, lack of economic incentives, etc. Goals need to be set, plans made, and policy set to change the way we use
energy. This chapter will discuss some of the commonly held views of where we can start and how we can change.
11.4.1 [Link]
reduction to prevent greenhouse gases from reaching levels that will have severe effects. A climate action plan is made of a number
of strategies to achieve that goal. To examine the impact of each strategy the wedge approach is used. Developed by two professors
at Princeton, Socolow and Pacala, the approach proposes that in order to reach those levels, emissions must be decreased globally
by seven gigatons of carbon (not carbon dioxide) compared to "business as usual" (BAU) scenarios which would increase
emissions over time due to growth and increased demand for energy (Figure 11.4.1) These professors identified 15 proposed
actions that could each reduce emissions by 1 gigaton, and if we could enact seven of them we would achieve the goal (Figure
11.4.2). Each of those technologies is represented by a "wedge" of the triangle, hence the designation of the "wedge approach."
Figure 11.4.1 The Wedge Approach. Represents the current path of increasing carbon emissions and the lower figure
a chart of the effects of many different strategies used to reduce the
emissions (a wedge of the triangle)
Figure 11.4.2 The Wedge Approach. Represents the effects of many different strategies used to reduce the emissions (a wedge of
the triangle). Source: The Carbon Mitigation Initiative, Princeton University
Sustainable Solutions
All of the proposed solutions in Sokolov and Pacala’s proposal are existing technologies. However, for a solution to be sustainable
it must be economically viable. Another aspect of developing a plan is the cost of the solutions. Figure 11.4.3 shows the amount of
greenhouse gas emissions that can be abated beyond "business as usual" in 2030, along with the costs of different abatement
strategies. Those technologies that fall below the 0 line will actually have a negative cost or positive economic benefit. Those
technologies that rise above the 0 line will have positive cost associated with them which could be offset by the technologies that
fall below the line.
Global GHG Abatement Cost Curve Beyond Business-As-Usual – 2030
Figure 11.4.3 Global GHG Abatement Cost Curve Beyond Business-As-Usual – 2030 Source: McKinsey & Company, Pathways to
a Low-Carbon Economy. Version 2 of the Global Greenhouse Gas Abatement Cost Curve, 2009
The types of technologies that fall below the line are primarily energy conservation and efficiency technologies. Energy
conservation is the act of reducing energy use to avoid waste, save money, and reduce the environmental impact. In the framework
of sustainable energy planning it allows the more expensive alternatives, such as renewables, to become more advanced and cost-
effective, while we conserve as much as possible. Conservation has a behavioral aspect to it, such as turning off lights when not
needed, lowering thermostats in the winter, or keeping the proper air pressure in a vehicle’s tires. There is a very low cost to
conservation but it entails behavioral change. There are technologies such as motion detectors that can control lights or
programmable thermostats that adjust temperature that can help overcome the behavioral barrier. Energy efficiency can be seen as a
subset of conservation as it is really about using technological advancements to make more efficient energy-consuming equipment.
In the United States we use twice as much energy per dollar of GDP as most other industrialized nations (see Figure Energy
Demand and GDP Per Capita (1980-2004)). There are many reasons for this. One reason is that we use less efficient vehicles and
use proportionally more energy to heat and cool buildings, behaviors that could be modified to be more efficient.
Energy Demand and GDP Per Capita (1980-2004)
Figure 11.4.4 Energy Demand and GDP Per Capita (1980-2004) Each line represents a different country and the points are for the
years 1980-2004, which the exception of Russian which is 1992-2004. Source: U.S. Department of Energy, Sustainability and
Maintaining US Competitiveness (June 2010), p. 4
U.S. Energy Consumption by Source
Figure shows United States energy consumption by source, with breakdown for buildings. Source: U.S. Department of Energy,
Berkeley Lab
Another reason that the United States uses so much more energy than other industrialized countries has to do with heating, cooling,
and illuminating buildings. Buildings account for about 40 percent of total energy consumption in the United States (costing $350
billion per year) and greenhouse gas emissions (see Figure 11.4.5). Energy use in buildings is primarily for heating, cooling, and
illumination, with significant differences between commercial and residential buildings. The rest of the energy use is for equipment
such as office equipment, electronics, refrigeration, cooking, and washing. There are many ways to save energy in existing
buildings and most of them have a good financial benefit of saving money on the energy costs, i.e. they have a short term financial
payback, or return on investment (ROI).
11.4.2 [Link]
with compact fluorescent light bulbs (CFLs) that can reduce energy use by 75 percent. The light bulbs also last 10 times as long,
reducing waste and maintenance costs. In commercial buildings more efficient fluorescent light bulbs (T-8s) and ballasts are
replacing the older T-12s. In 2002, the U.S. Department of Energy required that T-12 ballasts no longer be manufactured, ending a
five year phase out of this technology.
Compact Fluorescent Light Bulb
Figure 11.4.6 Compact Fluorescent Light Bulb Over its lifetime, each standard (13 watt) CFL will reduce electricity bills by about
$30 and emissions from 200 lbs of coal over its lifetime. Source: Kevin Rector
Already newer, more efficient technologies are hitting the market – light-emitting diodes (LEDs) – which use less than 25 percent
of the energy of an incandescent light and last at least 15 times longer, if it has the ENERGY STAR rating. ENERGY STAR is the
government-backed symbol for energy efficiency recognition. LEDs are small light sources that become illuminated by the
movement of electrons through a semiconductor material (Figure 11.4.7). The technology is still evolving and not all LED lights
are created equally. LEDs are more expensive and will have a longer pay back time, but they also last longer.
Solid State Lighting (SSL)
Figure 11.4.7 Solid State Lighting (SSL) Solid state lighting (SSL) is comprised of many small LEDs. Since they release very little
energy as heat, they are cool to the touch and highly efficient. Source: Ocrho
If CFLs were used in all homes, the most advanced linear fluorescent lights in office buildings, commercial outlets and factories,
and LEDs in traffic lights would reduce the percentage of electricity used for lighting in the world from 19 percent to sever percent.
That’s equivalent to 705 coal-fired power plants.
Figure 11.4.8 Average Efficiency of New Refrigerators in the United States (1972-1997) Graph shows the efficiency of an average
new refrigerator in the United States from 1972 to 1997. Source: U.S. Energy Information Administration
11.4.3 [Link]
pane glass windows can also provide similar insulation to double-pane without the need for the larger financial investment and the
creation of waste that replacement entails.
Insulation in the attic or roof of a building and at the "seam" of the building between the basement and first floor, as well as the
walls can be installed or increased to retain the heated or cooled air for building or home. Related to insulation is sealing of opening
to prevent air from leaking out (see Figure 11.4.9).
Diagram of a Leaky Home
Figure 11.4.9 Diagram of a Leaky Home Diagram shows the various points in a home where energy may leak. Source: ENERGY
STAR
Figure 11.4.10 Comparison of Energy Efficiency of Standard Power Plant and Combined Heat and Power Plant Diagram compares
the energy efficiency of a standard power plant with a combined heat and power plant.
Figure 11.4.11 Lincoln Hall, LEED Gold Certified, University of Illinois at Chicago Lincoln Hall, LEED Gold certified building
on the University of Illinois at Chicago campus. Features include geothermal heating and cooling, solar photovoltaic rooftop
system, low-emittance, high-U windows, daylighting, native planting, bioswales for stormwater management, and use of recycled
materials. Source: UIC Office of Sustainability
11.4.4 [Link]
Implementing Renewable Energy Technologies
When buildings have been retrofitted to be more energy efficient and combined heat and power systems are used more broadly, we
will have reduced energy demand significantly and cost effectively, while creating more jobs domestically. We can then look at the
mass deployment of renewable energy technologies. Over time these technologies will mature and become more affordable. This
process can be enhanced through policy implementation that incentivizes renewable energy development.
The electric grid will need to be expanded. This will allow for more interstate transmission of renewable electricity from the areas
where the resources are good such as the southwest, for solar, and the central and plains states, for wind, to the areas where the
population centers are such as the east and west coasts. If these grids are smart and allow for real-time energy pricing then demand
will be leveled out. This unified national smart grid would include more efficient, higher-voltage long-distance transmission lines;
"smart" distribution networks that use the internet to connect smart meters in homes and along the grid; energy storage units (i.e.
batteries) throughout the network; and two-way communication between the equipment that consumes electricity and the
equipment that produces it.
We can envision a future where most cars on the road are electric. At night, consumption across the grid is lower because lights are
off, buildings are closed, and less manufacturing occurs. Owners of electric cars will plug their cars into the grid at night and
recharge them after being driven during the day. Since, demand for electricity is lower, prices for this utility will be lower. With
smart meters, residents will be charged for the actual cost of electricity at time of use rather than an average price. They will be
incentivized to set washing machines and dishwashers to run at night when electricity demand is lowest. All of this evens out the
demand on the grid, which means that power plants do not need to operate at peak capacity and reduces the need for new plants.
Figure 11.4.12 Carbon Dioxide Emissions and Fuel Economy by Model Year Two graphs show carbon dioxide emissions and fuel
economy by model year from 1975-2010. Source: U.S. EPA, Light-Duty Automotive Technology, Carbon Dioxide Emissions, and
Fuel Economy Trends: 1975 through 2010 (Nov. 2010), p. iv
Other ways to increase efficiency can be found through innovative alternative vehicle technologies, improved internal combustion
engines, exhaust gas recycling, variable valve timing, vehicle downsizing, lightweighting, and behavior. Government policies need
to make the cost of driving evident through full amortization, fuel/road tax, and insurance costs.
Another tactic to reduce fuel consumption is increasing the use of transportation alternatives. The use of active transportation will
cause a change from environmentally harmful, passive travel to clean, active travel by bicycle, foot, and public transit. Convenient
and safe public transit is not available in all communities, as it requires a certain population density to be viable. Moreover, since
Americans often associate the car they drive with their material success and our communities are spread out, many people do not
view public transportation favorably. Most metropolitan areas have some kind of transit system to provide transportation to those
who cannot afford cars or cannot drive and/or to relieve traffic congestion. Historically, the United States has not invested equally
in road and public transportation infrastructure meaning that often it is slower and more complicated to travel by transit. However,
transit use is generally more economical than owning and driving a car. The American Public Transportation Association has
calculated the savings based on a two person-two car household going to one-car. They found that riding public transportation
saves individuals, on average $9,656 annually, and up to $805 per month based on the January 5, 2011 average national gas price
($3.08 per gallon-reported by AAA) and the national unreserved monthly parking rate. Savings for specific cities are shown here.
Bicycling and walking are two forms of alternate transit that have no environmental impact on energy demand. Many local
governments are devoting resources to adding bike routes and parking facilities to encourage bicycling as a mode of transportation.
Sidewalks and safe cross-walks are prerequisites for safe walking.
11.4.5 [Link]
There are some options for those who must drive to reduce their energy use. Carpooling and car sharing are also options that lower
the number of cars on the road, while providing opportunities to travel by car when needed. Improved social network-based car-
pooling programs can help to match riders with drivers in a dynamic way. Car sharing is a decentralized, hourly car rental system
that allows people who do not own cars, but occasionally need one, to access a vehicle in proximity to their workplace or home.
Summary
There is no one silver bullet when it comes to solving the "energy problem" or planning for climate action. There are many viable
solutions and the problem is so large that multiple pathways must be forged. The primary challenge is to use energy more
efficiently so little goes to waste. From small actions like changing a light bulb, to large projects like CHP, the potential is great and
the financial payback rewarding. Increased vehicle efficiency and active transportation are also strategies for reducing energy use.
Both within the building sector and the transportation sector we have the greatest challenges to and potential for changing how we
use energy today. We have already started to make that transition from more stringent CAFÉ standards to more green buildings.
The challenge is to upscale all the strategies to make a significant impact.
References
1. Brown, L.R. (2008). Plan B 3.0: Mobilizing to save civilization. New York: Earth Policy Institute.
2. City of Chicago. (2011). Chicago Climate Action Plan. Retrieved September 12, 2011 from
[Link]/...[Link]
3. Eagan, D.J., Calhoun, T., Schott, J. & Dayananda, P. (2008). Guide to climate action planning: Pathways to a low-carbon
campus. National Wildlife Federation: Campus Ecology. Retrieved September 12, 2011 from [Link]
Warming/Ca...-[Link]
4. Gore, A. (2009). Our choice: A plan to solve the climate crisis. New York: Melcher Media.
5. Pacala, S. & Socolow, R. (2004). Stabilization wedges: Solving the climate problem for the next 50 years with current
technologies. Science, 305, 968-972.
6. University of Illinois at Chicago. (2009). UIC Climate Action Plan. Retrieved September 12, 2011 from
[Link]
Review Questions
1. What does the chart in Figure 11.4.4 tell us about developing countries such as China, India and Brazil’s energy use? a) In
comparison to developed countries. b) Over time.
2. Briefly describe a path to reducing our dependency on fossil fuels for transportation energy consumption.
3. Why is energy efficiency considered a sustainable energy choice?
Glossary
Active Transportation
Means of transportation that involve more physical activity, typically considered walking, biking, and use of public transit (bus and
rail).
Building Automation System (BAS)
Controls and monitors a buildings mechanical and lighting systems through a computerized, intelligent network
[[Link]/wiki/Computer_networking] of electronic devices.
Carpooling
When two or more people travel to and from proximal departure and arrival destinations in the same vehicle.
Car Sharing
A program that allows for more than one person to have use of a car. Generally, it works like a short-term (hourly) car rental
service. Cars are located near residences and work places to facilitate the access to the vehicles and to reduce the need for
individual car ownership.
Lightweighting
Making a product out of materials that weigh less than were previously used
Low-Emittance Coatings
11.4.6 [Link]
Microscopically thin, virtually invisible, metal or metallic oxide layers deposited on a window or skylight glazing surface primarily
to reduce the U-factor by suppressing radioactive heat flow.
Phantom Load or Vampire Power
Refers to the electrical load of appliances and chargers when they are not in use but plugged in, as they still draw power but
provide no service.
U-factor
The rate of heat loss is indicated in terms of the of a window assembly. The lower the U-factor, the greater a window's resistance to
heat flow and the better its insulating properties.
Wedge Approach
A way of expressing the concept that there is no one solution to the challenge of reducing greenhouse gas emissions. Each
technology, action or change is represented by a triangular wedge in a chart of time vs. emissions.
This page titled 11.4: Sustainable Energy Practices - Climate Action Planning is shared under a CC BY license and was authored, remixed, and/or
curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
11.4.7 [Link]
11.5: Sustainable Transportation- Accessibility, Mobility, and Derived Demand
What is Sustainable Transportation?
Transportation is a tricky thing to analyze in the context of sustainability. It consists in part of the built environment: the physical
infrastructure of roads, runways, airports, bridges, and rail lines that makes it possible for us to get around. It also consists in part of
individual choices: what mode we use to get around (car, bus, bike, plane, etc.), what time of day we travel, how many people we
travel with, etc. Finally, it also is made up of institutions: federal and state agencies, oil companies, automobile manufacturers, and
transit authorities, all of whom have their own goals and their own ways of shaping the choices we make.
Most importantly, transportation is complicated because it's what is called a derived demand. With the exception of joyriding or
taking a walk or bicycle ride for exercise, very rarely are we traveling just for the sake of moving. We're almost always going from
Point A to Point B. What those points are—home, work, school, shopping—and where they're located—downtown, in a shopping
mall, near a freeway exit—influence how fast we need to travel, how much we can spend, what mode we're likely to take, etc. The
demand for transportation is derived from other, non-transportation activities. So in order to understand transportation
sustainability, we have to understand the spatial relationship between where we are, where we want to go, and the infrastructure
and vehicles that can help get us there.
Is our current transportation system in the U.S. sustainable? In other words, can we keep doing what we're doing indefinitely? The
answer is clearly no, according to professional planners and academics alike. There are three main limitations: energy input,
emissions, and social impacts (Black, 2010).
Energy Inputs
The first reason that our current transportation system is unsustainable is that the natural resources that power it are finite. The
theory of peak oil developed by geologist M. King Hubbert suggests that because the amount of oil in the ground is limited, at
some point in time there will be a maximum amount of oil being produced (Deffeyes, 2002). After we reach that peak, there will
still be oil to drill, but the cost will gradually rise as it becomes a more and more valuable commodity. The most reliable estimates
of the date of peak oil range from 2005 to 2015, meaning that we've probably already passed the point of no return. New
technologies do make it possible to increase the amount of oil we can extract, and new reserves, such as the oil shale of
Pennsylvania and the Rocky Mountains, can supply us for some years to come (leaving aside the potential for environmental and
social damage from fully developing these sites). However, this does not mean we can indefinitely continue to drive gasoline-
powered vehicles as much as we currently do.
Scientists are working on the development of alternative fuels such as biofuels or hydrogen, but these have their own limitations.
For example, a significant amount of land area is required to produce crops for biofuels; if we converted every single acre of corn
grown in the U.S. to ethanol, it would provide 10% of our transportation energy needs. Furthermore, growing crops for fuel rather
than food has already sparked price increases and protests in less-developed countries around the world (IMF, 2010). Is it fair to ask
someone living on less then two dollars a day to pay half again as much for their food so we can drive wherever and whenever we
want?
Emissions or Outputs
The engine of the typical automobile or truck emits all sorts of noxious outputs. Some of them, including sulfur dioxides, carbon
monoxide, and particulate matter, are directly harmful to humans; they irritate our lungs and make it hard for us to breathe. (Plants
are damaged in much the same way). These emissions come from either impure fuel or incomplete burning of fuel within an
engine. Other noxious outputs cause harm indirectly. Nitrous oxides (the stuff that makes smog look brown) from exhaust, for
example, interact with oxygen in the presence of sunlight (which is why smog is worse in Los Angeles and Houston), and ozone
also damages our lungs.
Carbon dioxide, another emission that causes harm indirectly, is the most prevalent greenhouse gas (GHG), and transportation
accounts for 23% of the CO2 generated in the U.S. This is more than residential, commercial, or industrial users, behind only
electrical power generation (DOE, 2009). Of course, as was explained above, transportation is a derived demand, so to say that
transportation itself is generating carbon emissions is somewhat misleading. The distance between activities, the modes we choose
to get between them, and the amount of stuff we consume and where it is manufactured, all contribute to that derived demand and
must be addressed in order to reduce GHG emissions from transportation.
11.5.1 [Link]
Social Impacts
If the definition of sustainability includes meeting the needs of the present population as well as the future, our current
transportation system is a failure. Within most of the U.S., lack of access to a personal automobile means greatly reduced travel or
none at all. For people who are too young, too old, or physically unable to drive, this means asking others for rides, relying heavily
on under-funded public transit systems, or simply not traveling. Consider, for example, how children in the U.S. travel to and from
school. In 1970, about 50% of school-aged children walked or biked to school, but by 2001, that number had dropped to 15%
(Appleyard, 2005). At the same time that childhood obesity and diabetes are rising, children are getting less and less exercise, even
something as simple as walking to school. Furthermore, parents dropping off their children at school can increase traffic levels by
20 to 25%, not just at the school itself, but also throughout the town in question (Appleyard, 2005). At the other end of the age
spectrum, elderly people may be functionally trapped in their homes if they are unable to drive and lack another means of getting to
shopping, health care, social activities, etc. Finally, Hurricane Katrina made it clear that access to a car can actually be a matter of
life or death: the evacuation of New Orleans worked very well for people with cars, but hundreds died because they didn't have the
ability to drive away.
Another serious social impact of our transportation system is traffic accidents. Road accidents and fatalities are accepted as a part
of life, even though 42,000 people die every year on the road in the U.S. This means that cars are responsible for more deaths than
either guns, drugs, or alcohol (Xu et al., 2010). On the bright side, there has been a steady reduction in road fatalities over the last
few decades, thanks to a combination of more safety features in vehicles and stricter enforcement and penalties for drunk or
distracted drivers. Nevertheless, in many other countries around the world, traffic accidents are in the top ten or even top five
causes of death, leading the World Health Organization to consider traffic accidents a public health problem.
An additional problem with our current unsustainable transportation system is that much of the rest of the world is trying to
emulate it. The U.S. market for cars is saturated, meaning that basically everyone who can afford or is likely to own a car already
has one. This is why automobile manufacturers vie so fiercely with their advertising, because they know they are competing with
each other for pieces of a pie that's not getting any bigger. In other countries such as China and India, though, there are literally
billions of people who do not own cars. Now that smaller, cheaper vehicles like the Tata are entering these markets, rates of car
ownership are rising dramatically. While the same problems with resources, emissions, and social impacts are starting to occur in
the developing world, there are also unique problems. These include a lack of infrastructure, which leads to monumental traffic
jams; a need for sharing the road with pedestrians and animals; and insufficient regulation to keep lead and other harmful additives
out of gasoline and thus the air.
11.5.2 [Link]
If transportation is a derived demand, another way to meet our current transportation needs is by changing the demand. There are
two related aspects to this. First, there is a clear causal link between having more transportation infrastructure and more miles
traveled on that infrastructure, and greater economic growth. This is true between regions of the world, between individual
countries, and between people and regions within countries. This causal connection has been used as a reason to finance
transportation projects in hundreds of different contexts, perhaps most recently in the American Reinvestment and Recovery Act
that distributed federal funds to states and localities to build infrastructure in the hopes that it would create jobs. Policymakers,
businesspeople, and citizens therefore all assume that we need more transportation to increase economic growth.
However, it is also true that more transportation does not automatically mean more economic growth: witness the state of West
Virginia, with decades' worth of high-quality road infrastructure bestowed upon it by its former Senator Robert Byrd, but still at the
bottom of economic rankings of states. Furthermore, at some point a country or region gains no significant improvements from
additional infrastructure; they have to focus on making better use of what they already have instead. We therefore need to decouple
economic growth from transportation growth (Banister and Berechman, 2001). We can substitute telecommunication for travel,
work at home, or shop online instead of traveling to a store (although the goods still have to travel to our homes, this is more
efficient than each of us getting in our own cars). We can produce the goods we use locally instead of shipping them halfway
around the world, creating jobs at home as well as reducing resource use and emissions. All of these options for decoupling are
ways to reduce the demand for transportation without also reducing the benefits from the activities that create that demand.
The other way to think about changing the derived demand of transportation is via the concepts of accessibility and mobility.
Mobility is simply the ability to move or to get around. We can think of certain places as having high accessibility: at a major
intersection or freeway exit, a train station, etc. Company headquarters, shopping malls, smaller businesses alike decide where to
locate based on this principle, from the gas stations next to a freeway exit to the coffee shop next to a commuter rail station. At
points of high accessibility, land tends to cost more because it's easier for people to get there and therefore more businesses or
offices want to be there. This also means land uses are usually denser: buildings have more stories, people park in multi-level
garages instead of surface lots, etc.
We can also define accessibility as our own ability to get to the places we want: where we shop, work, worship, visit friends or
family, see a movie, or take classes. In either case, accessibility is partially based on what the landscape looks like—width of the
roads, availability of parking, height of buildings, etc.—and partially on the mode of transportation that people have access to. If a
person lives on a busy four-lane road without sidewalks and owns a car, most places are accessible to him. Another person who
lives on that same road and doesn't have a car or can't drive might be literally trapped at home. If her office is downtown and she
lives near a commuter rail line, she can access her workplace by train. If her office is at a major freeway intersection with no or
little transit service, she has to drive or be driven.
Figure 11.5.2 A modern subdivision near Markham, Ontario. The suburb is residential only, and cars are the only visible means of
transport; accessibility for those without personal vehicles is low. Photo by IDuke, November 2005. Source: IDuke (English
Wikipedia) [CC-BY-SA-2.5], via Wikimedia Commons
Unfortunately, in the U.S. we have conflated accessibility with mobility. To get from work to the doctor's office to shopping to
home, we might have to make trips of several miles between each location. If those trips are by bus, we might be waiting for
several minutes at each stop or making many transfers to get where we want to go, assuming all locations are accessible by transit.
If those trips are by car, we are using the vehicle for multiple short trips, which contributes more to air pollution than a single trip
of the same length. Because of our land use regulations, which often segregate residential, retail, office, and healthcare uses to
completely different parts of a city, we have no choice but to be highly mobile if we want to access these destinations. John Urry
has termed this automobility, the social and economic system that has made living without a car almost impossible in countries like
the US and the UK (2004).
So how could we increase accessibility without increasing mobility? We could make it possible for mixed uses to exist on the same
street or in the same building, rather than clustering all similar land uses in one place. For example, before a new grocery store
opened in the student neighborhood adjacent to the University of Illinois campus in Champaign, people living there had to either
take the bus, drive, or get a friend to drive them to a more distant grocery store. Residents of Campustown had their accessibility to
fresh produce and other products increase when the new grocery store opened, although their mobility may have actually gone
down. In a larger-scale example, the Los Angeles Metropolitan Transit Authority (MTA) was sued in the 1990s for discriminating
against minorities by pouring far more resources into commuter rail than into buses. Commuter rail was used mainly by white
suburbanites who already had high levels of accessibility, while the bus system was the only means of mobility for many African-
American and Hispanic city residents, who had correspondingly less accessibility to jobs, shopping, and personal trips. The courts
ruled that the transit authority was guilty of racial discrimination because they were providing more accessibility for people who
11.5.3 [Link]
already had it at the expense of those who lacked it. The MTA was ordered to provide more, cleaner buses, increase service to
major job centers, and improve safety and security. More sustainable transportation means ensuring equitable accessibility — not
mobility — for everyone now and in the future.
New Technology
This is the hardest category to rely on for a solution, because we simply can't predict what might be invented in the next five to
fifty years that could transform how we travel. The jet engine totally changed air travel, making larger planes possible and
increasing the distance those planes could reach without refueling, leading to the replacement of train and ship travel over long
distances. However, the jet engine has not really changed since the 1960s. Is there some new technology that could provide more
propulsion with fewer inputs and emissions? It's possible. But at the same time, it would be unreasonable to count on future
inventions magically removing our sustainability problems rather than working with what we already have.
Technology is more than just machines and computers, of course; it also depends on how people use it. When the automobile was
first invented, it was seen as a vehicle for leisure trips into the country, not a way to get around every day. As people reshaped the
landscape to accommodate cars with wider, paved roads and large parking lots, more people made use of the car to go to work or
shopping, and it became integrated into daily life. The unintended consequences of technology are therefore another reason to be
wary about relying on new technology to sustain our current system.
11.5.4 [Link]
transportation is connected to other activities, and that focusing on how the demand for transportation is derived is the key to
making and keeping it sustainable.
References
1. Appleyard, B. S. 2005. Livable Streets for School Children: How Safe Routes to School programs can improve street and
community livability for children. National Centre for Bicycling and Walking Forum, available online:
[Link]
2. Banister, D. and Berechman, Y. 2001. Transport investment and the promotion of economic growth. Journal of Transport
Geography 9:3, 209-218.
3. Black, W. 2010. Sustainable Transportation: Problems and Solutions. New York: Guilford Press.
4. Deffeyes, K. 2002. Hubbert's Peak: The Impending World Oil Shortage. Princeton, NJ: Princeton University Press.
5. DOE (Department of Energy). 2009. Emissions of greenhouse gases report. DOE/EIA-0573, available online:
[Link]/oiaf/1605/ggrpt/[Link]
6. Downs, A. 1992. Stuck in Traffic: Coping With Peak-Hour Traffic Congestion. Washington, DC: Brookings Institution Press.
7. IMF (International Monetary Fund). 2010. Impact of high food and fuel prices on developing countries. Available online:
[Link]
8. Maring, G. 2007. Surface transportation funding issues and options. Presentation to the National Surface Transportation
Infrastructure Financing Commission. Available online: [Link]
9. Marsh, B. 2011. Kilowatts vs. Gallons. New York Times, May 28. Available online:
[Link]
10. UPI (United Press International). 2011. Global biofuel land area estimated. Available online:
[Link]
11. Urry, J,. 2004. The 'System' of Automobility. Theory Culture and Society 21:4-5, 25-39.
12. Usón, A.A., Capilla, A.V., Bribián, I.Z., Scarpellini, S. and Sastresa, E.L. 2011. Energy efficiency in transport and mobility for
an eco-efficiency viewpoint. Energy 36:4, 1916-23.
13. Xu, J., Kochanek, K., Murphy, S., and Tejada-Vera, B. 2010. Deaths: Final Data for 2007. National Vital Statistics Reports,
58:19, available online: [Link]
Review Questions
1. Explain the concept of a derived demand and how it accounts for the connections between transportation and land use planning.
2. What is the concept of embodied energy? Why does it suggest that switching to electric cars is not a surefire way to make
transportation more sustainable?
3. Give an example in your daily life that could be used to explain the difference between accessibility and mobility.
Glossary
Accessibility
In transportation, a measure of the ease with which people are able to get places they want or need to go.
Derived Demand
Demand for a good or service that comes not from a desire for the good or service itself, but from other activities that it enables or
desires it fulfills.
Embodied Energy
The sum of all energy used to produce a good, including all of the materials, processes, and transportation involved.
Externality
Cost of an activity not paid by the person doing the activity.
Mobility
The ability to move or to get around.
This page titled 11.5: Sustainable Transportation- Accessibility, Mobility, and Derived Demand is shared under a CC BY license and was
authored, remixed, and/or curated by Heriberto Cabezas (GALILEO Open Learning Materials) .
11.5.5 [Link]
11.6: Sustainable Stormwater Management
Learning Objectives
After reading this module, students should be able to
describe how stormwater runoff affects water quality in urban watersheds
explain how stormwater is currently managed in the United States
analyze some of the conventional and innovative techniques that have been developed to address the water
pollution and flood risks associated with urban stormwater runoff
This module reviews some of the complex issues of urban stormwater management. It first examines the hydrological issues
affecting the discharge of stormwater runoff to our urban rivers and streams, and then provides an overview of how urban
stormwater is managed under the Clean Water Act. After describing the conventional approaches to urban stormwater management,
the final section provides an overview of various "sustainable" strategies, especially the use of "green infrastructure," that can be
considered to reduce the water pollution and flooding risks generated by urban stormwater runoff.
11.6.1 [Link]
Degrees of Imperviousness and its Effects on Stormwater Runoff
Figure 11.6.1 Degrees of Imperviousness and its Effects on Stormwater Runoff. These four images show increasing amount of
stormwater runoff as the area becomes developed with more impervious surfaces. Source: In Stream Corridor Restoration:
Principles, Processes, and Practices (10/98) By the Federal Interagency Stream Restoration Working Group (FISRWG) (15 Federal
agencies of the U.S.)
When flowing downhill within a watershed, stormwater runoff can pick up pollutants from various anthropogenic sources and
activities. It can also collect pollutants from the atmospheric deposition of particulates and air pollutants carried to the earth's
surface by precipitation, by windblown dust, or by simply settling out of the atmosphere. Urban runoff can also dissolve or
transport chemicals that may be found naturally in soil or nutrients which may have been deliberately added to lawns. Common
urban pollutants can include such things as pesticides and fertilizers applied to residential lawns, parks and golf courses, enteric
microbes from animal waste, industrial chemicals that may have been accidentally spilled on the ground or improperly stored, or
oils and greases leaking from cars parked in lots or on driveways.
As stormwater runoff flows towards lower-lying areas of the watershed, it carries these contaminants with it and therefore
contributes to the pollution of the stream, river or lake into which it is discharging. Once it reaches a river or stream, the
concentrations of pollutants in the receiving waters are naturally reduced as the contaminants are carried downstream from their
sources, largely through dilution but also by settlement, by uptake by posure to sunlight and oxygen, and by interactions with
various chemical and physical proplants and animals (including bacteria and other microorganisms), through degradation by
excesses occurring within the waterway and its streambed.
11.6.2 [Link]
river that did not meet the water quality standards for their intended uses. Models were used to calculate the "total maximum daily
load" (TMDL) of pollutants entering the waterway through both point and non-point sources that would enable the stream
segments to achieve their highest proposed use. The Clean Water Act's new TMDL program provides a more sophisticated
framework for evaluating the impacts of non-point pollution on water quality. However, given the limitations of trying to put more
and better BMPs into place, environmental protection agencies have begun to refocus some of their attention from reducing the
total amount of pollutants being released within a watershed to also reducing the amount of stormwater runoff.
Environmental protection agencies have developed strategies for urban stormwater management that involve modifying a
development site so that more precipitation would be retained on-site rather than flowing off of it into nearby waterways or
waterbodies. These stormwater retention strategies initially stressed traditional engineering solutions, such as installing a
stormwater collection system that temporarily stores the stormwater on-site in order to reduce the rate and amount of stormwater
being released to a waterway. The strategies were later expanded to include various site modifications, such as constructing
vegetated buffer strips or swales (ditches),in order to encourage more stormwater to infiltrate into the ground.
Reducing the volume of urban stormwater leaving a site as runoff also offers an additional hydrologic benefit in urban watersheds –
reducing flood risks (NRC 2008). Besides having the potential to carry pollutants, stormwater runoff discharge increases the
amount of water entering into a lake, stream or river, increasing both the water volume and flow velocity of the waterway. A
relatively large amount of stormwater runoff entering a waterway over a relatively short time can quickly raise a stream's water
levels beyond its banks, causing flooding that could threaten adjacent development. Stormwater contribution to a river or stream
can also increase the velocity of the stream's flow, causing increased channel and bank erosion, undercutting or damaging dikes,
levees and other water control structures, and scouring the stream or river bed. Stream edge or streambed erosion can impair water
quality by increasing the cloudiness (or turbidity) of the waterway, which can also damage aquatic and riparian habitats.
Stormwater-induced flood risks are managed by the
National Flood Insurance Act, where hydrologic models (adjusted by historical flood events) are used to forecast the potential
flooding caused by a 100-year storm (a storm that has a one percent chance of occurring in any given year). The Act forces
financial institutions to require homeowners within the designated 100-year floodplains to purchase flood insurance in order to get
a mortgage, with the federal government subsidizing the insurance premiums if the community adopts a flood management
program restricting development from extremely hazardous areas and instituting building code changes to lessen flood damage.
In assessing flood risks, it is important to realize that managing the volume and rate of urban stormwater being discharged from
developed areas does not affect the total amount of stormwater that is being discharged to a river or stream within a watershed –
they only affect the timing of when a storm's precipitation will be discharged to the waterway (NRC, 2008). Both the conventional
and the newer, more sustainable, ways of managing stormwater discussed below seek to delay the time it takes for stormwater
runoff to reach a waterway in order to reduce the water levels and flow velocities of the receiving streams after a storm. Slowing
the rate by which stormwater is being contributed to a stream spreads out the peak of the resultant flood levels over a longer time
period, allowing many flood risks to be substantially reduced.
11.6.3 [Link]
might infiltrate into the soils in the bottom of the detention basin, and some pollutants might be taken up by grass lining the basin,
but many pollutants still end up being carried into the waterway along with the released stormwater.
Since the 1990s, environmental protection agencies have begun to consider the water pollution impacts of releases from stormwater
detention facilities, after the Clean Water Act was amended to require states to treat stormwater discharges from detention basins as
a type of direct source and to require that NPDES permits be phased in for discharges from Municipal Separate Stormwater Sewer
Systems ("MS4") in cities and urban areas above certain population thresholds (NRC, 2008). The NPDES permits issued under the
U.S. Environmental Protection Agency's (U.S. EPA) MS4 program now require the water pollution loads from stormwater
detention basin discharges to be assessed through the creation and adoption of local stormwater management plans and that the
contaminants carried by the stormwater runoff to the basins for later re-release to a waterway be better managed and reduced
through the adoption of local BMPs. MS4 permit regulations issued by state environmental protection agencies usually involve the
issuance of a "general permit" by the agency, applying to all applicable Municipal Separate Stormwater Sewer Systems located
within the state’s designated urban areas.
Stormwater Sewer Systems Located within the State's Designated Urban Areas
A different set of stormwater management issues arise in older urban areas that are already developed. Most of the United States'
older cities and suburbs, especially those established in the late-19th and early 20th centuries, do not have Municipal Separate
Stormwater Sewer Systems. Instead, they have what are known as combined sewer systems – sewers that carry both the stormwater
runoff from paved streets and the wastewater (sewage) from homes, stores and factories. These combined sewers transport the
mixed wastewater and stormwater to municipal sewage treatment plants where the diluted sewage is treated and then discharged to
a waterway under an NPDES permit (NRC, 2008).
Water quality problems arise when rainstorms deposit more precipitation in the city than can be handled by the sewage treatment
plant. As the diluted wastewater begins to fill up the combined sewer system at a faster rate than it can be treated, the sewage
treatment plant operators are faced with a difficult choice – they can either allow the diluted sewage to continue to back up in the
sewers, eventually flooding residents' basements (a politically unpopular as well as unhealthy option), or they can allow the diluted
wastewater to bypass the sewage treatment plant and be discharged directly into the waterway, with the untreated wastewater's
pollutant levels usually exceeding the limits set forth in the plant's NPDES permit. Most treatment plant operators choose the more
politically acceptable option of releasing the wastewater in violation of their NPDES permit, creating water pollution incidents
called combined sewer overflows (CSOs).
11.6.4 [Link]
Managing Urban Stormwater More Sustainably
There is a fourth approach to dealing with CSO problems, which involves intercepting and delaying the discharge of precipitation
from a parcel of land before it flows off-site to a separate or combined sewer system, or to an adjacent waterway. Encouraging on-
site storage or infiltration reduces the stormwater contribution to a combined sewer's flow in developed areas, thereby reducing the
amount of diluted wastewater being generated and enabling combined sewer systems to better handle their wastewater loads during
rainstorms. These decentralized on-site approaches to managing stormwater could also be used to reduce the amount of
conventional stormwater infrastructure needed in new developments using separate stormwater sewer systems. Because these on-
site approaches are less resource-intensive and more cost-effective than conventional stormwater management approaches, they are
also more sustainable investments.
On-site stormwater management techniques are also often known as "green infrastructure" (Jaffe et al., 2010). Development
projects using "green infrastructure" for urban stormwater management are commonly known as "Low Impact Developments."
Low Impact Development projects using green infrastructure usually allow stormwater to be managed at lower costs than by using
conventional detention practices (US EPA, 2007).
There are essentially three strategies for on-site stormwater management: (1) techniques that encourage the infiltration of
stormwater into soils to reduce its volume before it reaches a sewer system, or which employ more selective grading and the
planting of vegetation to reduce its rate of flow from the site; (2) techniques that encourage the temporary storage of stormwater
on-site, instead of transporting it off-site for centralized detention within a development project or a municipality; and (3)
techniques, such as the construction of artificial wetlands, which also allow some degree of longer-term retention and treatment of
the stormwater by natural processes before it is discharged. Infiltration techniques might also provide some water treatment
capabilities due to the longer retention times of groundwater before discharge, but the degree of such treatment would largely
depend on soil characteristics, the amount of overlying vegetation and the depth of the soil's unsaturated zone.
Figure 11.6.2 Permeable Paving & Vegetated Swales. Permeable paving drains into a vegetated swale as part of Elmhurst College's
(in Illinois) parking lot's "green" stormwater management system. Source: Jaffe, M., et al. (2010), Figure 14, p. 117.
"Rain gardens" can also be used to encourage stormwater to infiltrate into the soils, where it can be taken up by plants and
transpired to the atmosphere, evaporated from the soils, or allowed to infiltrate deeper into the soils to become groundwater. Rain
gardens are created in areas of low-lying terrain that are expressly designed for, or engineered with, well-drained soils and are
usually planted with deep-rooted native vegetation that often can survive the drier soil conditions between rains. Rain gardens can
be quite effective in intercepting and infiltrating stormwater being discharged from roofs, with roof downspouts directing the
discharge of stormwater into a rain garden instead of allowing it to flow across the lot and into the street sewer system. Some native
vegetation, however, may have special maintenance requirements, such as the periodic burning needed to manage some prairie
plants.
Vegetated ditches or swales can also be used to transport stormwater runoff to a conventional stormwater management system, with
the vegetation planted in the ditch slowing the rate of stormwater flow while also allowing a portion of the runoff to be infiltrated
into the soils or taken up by plants. In many cases, vegetated swales and rain gardens can provide less-expensive alternatives to the
installation of separate stormwater sewer system, since it reduces the need for the construction of street gutters, grates, street
catchment basins and sewer pipes (US EPA, 2007). Interception of the stormwater by infiltration and plant uptake in a rain garden
or vegetated swale may also reduce the amount, capacity and size of the sewers that would have to be built to manage a predicted
volume of stormwater, if these green infrastructure techniques are used to supplement a conventional stormwater collection system.
11.6.5 [Link]
Increasing Interim On-site Storage
Sustainable management techniques that can temporarily store stormwater on-site until it can be released off-site to a sewer system
or to conventional stormwater detention facilities include the use of "green roofs" and rain barrels connected to roof downspouts.
Rain barrels allow precipitation to be collected and stored, and then used for non-potable purposes (lawn irrigation, for instance)
allowing the captured stormwater to substitute for more expensive, treated water (see Figure 11.6.3).
A Rain Barrel Collection System Used at Ryerson Woods Welcome Center, Lake County (Illinois)
Forest Preserve District
Figure 11.6.3 A Rain Barrel Collection System This "green" building (Ryerson Woods Welcome Center, Lake County (Illinois)
Forest Preserve District )uses both a rain barrel to collect stormwater draining from the roof, and a rain garden to help infiltrate
precipitation. Source: Jaffe, M., et al. (2010), Figure 12, p. 116.
A green roof is a flat roof surface that uses amended soil materials installed above a layer of waterproof roofing materials to allow
shallow-rooted plants to be planted. While still being an impermeable feature of a development site (because of its waterproof
layer), a green roof can temporarily store rainwater before it is discharged to the ground by the roof gutters and downspouts (see
Figure 11.6.4). Just as a rain barrel can store (and re-use) a portion of the stormwater precipitation being discharged from
impervious roofs, the soils of a green roof can capture and temporarily store stormwater precipitation as the pores between the soil
particles fill up with rainwater. Green roofs can even partially reduce the runoff's pollution load through plant uptake and by other
biological and physical processes within the roofs' soil materials while they are saturated. Because of the need to both water-proof
the roof while installing a biological system on top of it, green roofs tend to cost more than conventional roofs, even ignoring the
additional structural engineering that might be necessary to accommodate the weight of the green roof's soil and plantings.
Green Roof Installed on Village of Villa Park, Illinois, Police Station
Figure 11.6.4 A Green Roof The green roof on this police station in Village of Villa Park, Illinois has shallow-rooted plants placed
in a thin layer of growing medium installed on top of a waterproof roof membrane. Source: Jaffe, M. et al. (2010), Figure 13, p.
116.
The stormwater management benefits of rain barrels and green roofs depend on their storage capacity relative to the amount of
impervious surface area of the roof with which they are associated. Rain barrels might be able to capture only a fraction of an inch
of the stormwater falling on a roof and being discharged from a downspout, while several inches of amended soils on a rooftop
might be able to store substantially more precipitation before it evaporates, is taken up by the roof's plants, or is discharged from
the green roof via its gutters and downspouts. In both cases, however, the interception and temporary retention of stormwater by
these green technologies may allow conventional stormwater management systems to function more efficiently by reducing the
amount of stormwater being discharged into the systems. They would also certainly reduce some of the "peakiness" of stream
flooding by being able to temporarily store and then release stormwater from impermeable roof surfaces later after a storm event.
11.6.6 [Link]
climate change impacts. Moreover, trees provide a valuable soil amendment as their fallen leaves decay into mulch, improving the
infiltration rate and biological activity of surrounding soils, while larger broken branches falling into urban streams can slow
stream velocities and provide improved riparian and aquatic habitat. The shading of streams by riparian trees is particularly
important in ensuring that a stream's ecological functions remain resilient in the face of rising temperatures caused by global
climate change.
Conclusions
All of the green infrastructure and Low Impact Development techniques that provide interim on-site stormwater storage to reduce
flood risks can also provide some pollution removal capabilities, as well. The American Society of Civil Engineers and U.S. EPA
maintain an International Stormwater BMP Database of development projects using green infrastructure. This on-line resource
reviews the effectiveness of various stormwater management practices and makes these sustainable techniques more accessible to
local officials and municipal public works departments charged with managing stormwater runoff in their communities.
There is increasing public interest in using sustainable stormwater management techniques to replace or supplement conventional
stormwater facilities. The U.S. federal government, for example, is now requiring that green infrastructure be used in all federal
projects above a certain size to manage urban stormwater runoff. Local officials are also showing a greater interest in these
sustainable approaches, since they are often less expensive to install and maintain over their life-spans than conventional
stormwater sewer systems and detention facilities. Finally, state governments are beginning to set aside money in their revolving
loan funds for public infrastructure that is earmarked for green infrastructure projects. It is likely that this interest in sustainable
urban stormwater management will continue to grow.
Review Questions
1. Which of the sustainable urban stormwater management practices can best be used in existing neighborhoods, and which are
best suited for new development?
2. The performance of many of the green infrastructure practices often depends on how well they are maintained over their life-
spans. What are some effective strategies that local officials can consider in order to ensure that the green infrastructure being
used to manage urban stormwater in their communities is adequately maintained and continues to perform as designed?
Resources
For more information about the:
Clean Water Act, visit [Link]/agriculture/[Link].
International Stormwater BMP Database, visit [Link]
Metropolitan Water Reclamation District of Greater Chicago, visit [Link]/irj/portal/anonymous/tarp.
Milwaukee Metropolitan Sewerage District, visit [Link]/[Link].
National Flood Insurance Act, visit [Link]/library/[Link]?id=2216.
References
1. Gulliver, G.S. & Anderson, J.L. (eds.). (2008). Assessment of Stormwater Best Management Practices. Stormwater
Management Practice Assessment Study. Minneapolis: University of Minnesota.
2. Jaffe, M., Zellner, M., Minor, E., Gonzalez-Meler, M., Cotner, L., Massey, D., Ahmed, H., Elbert M., Wise, S., Sprague, H., &
Miller, B. (2010). Using Green Infrastructure to Manage Urban Stormwater Quality: A Review of Selected Practices and State
Programs. Springfield, IL: Illinois Environmental Protection Agency. Retrieved June 23, 2011 from
[Link]
3. National Research Council. (2008). Urban Stormwater Management in the United States. Washington, DC: National Academies
Press. Retrieved June 23, 2011 from [Link]
4. U.S. Environmental Protection Agency. (2000, October). Street Storage for Combined Sewer Surcharge Control: Skokie and
Wilmette, Illinois (Factsheet). (EPA Publication No. EPA-841-B-00-005C). Washington, D.C. Retrieved May 17, 2011 from
[Link]/..._Factsheet.pdf
5. U.S. Environmental Protection Agency. (2007, December). Reducing Stormwater Costs through Low Impact Development
(LID) Strategies and Practices. (EPA Publication No. EPA 841-F-07-006). Washington. D.C. Retrieved June 23, 2011 from
[Link]/owow/NPS/lid/costs07/documents/[Link]
6. U.S. Global Climate Change Research Program (USGCCRP). 2009. Global Climate Change Impacts in the United States.
Cambridge: Cambridge University Press. Retrieved May 18, 2011 from
11.6.7 [Link]
[Link]
Glossary
Ambient Water Quality
The concentration of pollutants found within waterbodies and waterways.
Combined Sewer Overflows (CSOs)
The overflow and discharge of excess wastewater to surface waters during storms, when diluted wastewater flows exceed the
capacity of a combined sewer systems or sewage treatment plant.
Combined Sewer Systems
Sewer systems that are designed to collect stormwater runoff and domestic and industrial wastewater within the same sewer pipes.
"First Flush" Phenomenon
The higher pollutant concentrations found at the beginning of a storm or spring snowmelt.
Green Roof
Vegetation and planting media installed on a rooftop in order to store and delay stormwater runoff from the roof's surface.
Hydrology
The scientific examination of the occurrence, distribution, movement and properties of water within the natural environment.
Low Impact Development
An approach to land development (or re-development) that uses natural drainage and environmental processes to manage
stormwater as close to its source as possible.
Native Vegetation
"Wild" plants that have naturally evolved and successfully adapted to a region's environmental conditions.
Non-point Source
The term "nonpoint source" is defined to mean any source of water pollution that does not meet the legal definition of "point
source" in section 502(14) of the Clean Water Act (see "Point Source" definition below)
"Peaky" Waterways
The "peakiness" of a waterway describes the more rapid increase and decline in stream flow and the higher stream levels after a
storm in urbanized watersheds compared to the more gradual rise and decline in stream volumes and lower water levels in less-
developed drainage basins after the same storm event, largely because of the greater amounts of impervious surfaces and runoff
generated within urban areas.
Point Source
Defined by Section 502(14) of the Clean Water Act as any single identifiable and discrete source of pollution from which pollutants
are discharged, such as from a pipe, ditch, channel, culvert, confined animal feeding operation, or discharged from a floating
vessel.
Pollution Prevention
Reducing or eliminating waste at the source by modifying production processes, promoting the use of non-toxic or less-toxic
substances, implementing conservation techniques, and re-using materials rather than putting them into the waste stream.
Rain Barrel
A cistern, barrel or storage system that collects and stores the rainwater or snowmelt from roofs that would otherwise be diverted to
storm drains and streams as stormwater runoff.
Stormwater Runoff
The overland flow of precipitation generated by that portion of rain and snowmelt that does not infiltrate into the ground, is not
taken up by plants, and is not evaporated into the atmosphere.
11.6.8 [Link]
Swales
Graded and engineered landscape features designed as vegetated, shallow, open channels or ditches that are usually planted with
flood tolerant and erosion resistant plants.
Watershed
A geographic area that naturally drains to a specific waterway or waterbody.
11.6: Sustainable Stormwater Management is shared under a not declared license and was authored, remixed, and/or curated by LibreTexts.
11.6.9 [Link]
11.7: Case Study- A Net-Zero Energy Home in Urbana, Illinois
How much fossil fuel does it take to operate a comfortable home for a couple of retired American baby-boomers?
None.
That’s according to Ty and Deb Newell of Urbana, Illinois. Moreover, they hope the example of their home, the Equinox House,
will awaken others to the opportunity of constructing a net-zero energy house in the Midwest using technology available today.
Equinox House
Figure 11.7.1 The Equinox House. The picture of the house on a sunny November day shows the passive solar entering the
clerestory windows. Source: © 2012 Equinox Built Environment Engineering; a division of Newell Instruments Inc.
The Newells celebrated the first anniversary of life in the Equinox House in late 2011, so they now possess more than a year’s
worth of data about how much electricity they used on day-to-day basis, as well as how much electricity their solar panels
produced.
According to Ty Newell, who is professor emeritus of mechanical engineering at the University of Illinois at Urbana-Champaign,
the Equinox House required about 12,000 kilowatt-hours of electricity to operate from December 2010 through November 2011.
That total includes electricity for heating and air conditioning, hot water heat, clothes washing and drying, and all other appliances.
No natural gas is used in the house.
Newell noted that energy use in the Equinox House for the first year was approximately 20 percent greater than it will be in 2012
and subsequent years. That’s because he was using the least efficient of three different heating systems that will be tested in the
home.
During the first year, the solar panels that power the Equinox House produced approximately 11,000 kilowatt-hours of electricity.
This would have made the Newells purchasers of 1,000 kilowatt-hours, in net terms, had it not been for the fact that the solar panels
were on line for some time before they moved into the house.
Thanks to the more efficient heating system now in place, the Equinox House will produce surplus electricity in 2012 and in the
future. That’s by design. The surplus will be used to power their all-electric Ford Focus for the 8,000 miles of in-town driving they
do annually.
In conjunction with its solar panels, the Equinox House achieves net-zero energy use because it requires far less energy than even a
well-built conventional home—about one-fifth as much. It does so through the use of design and technology that did not add a
significant burden to the cost of construction.
The walls and roof of the Equinox House are constructed with twelve-inch thick structural insulated panels, which are four to five
times more effective at preventing thermal transfer than the walls of a typical house. Great care has also been taken to minimize
any leakage of air through envelope of the house.
The Equinox House uses high performance, triple-pane windows, which also help to prevent thermal transfer. Beyond that, the
windows are oriented to allow direct sunlight into living space for the heat it provides during the cooler half of the year—beginning
on the Fall equinox—and to exclude direct sunlight during the warmer half of the year—beginning on the Spring equinox—when it
would increase the load on the cooling system.
Ultimately, the demands of the Equinox House for heating, cooling, ventilation, and humidity control will all be met by a single,
heat-pump based system, developed by Ty Newell and his son Ben through their company, Newell Instruments. Aside from the fact
that it maintains a comfortable temperature and level of humidity in the house, this system also delivers a constant flow of fresh air
from the outside, and it does that without the loss of conditioned air that occurs in a drafty house.
Of course the Equinox House will be outfitted in other ways that emphasize conservation, including LED lighting, low-flow
plumbing fixtures, etc. It even features a system for collecting rainwater that is designed to meet 80 percent of the annual water
needs for a family of four.
When he talks about the Equinox House, Ty Newell emphasizes how well it works from an economic perspective, since the
couple’s average daily cost for energy is a mere $3.00. That’s based on a twenty-year life for the solar array, which cost a net of
$20,000 installed.
In addition, Newell enjoys the fact that a significant part of their up-front expenditure supported job creation, the labor that went
into the manufacture and installation of their solar panels. That’s in contrast to money they might have otherwise spent on fossil
11.7.1 [Link]
fuel.
You might think that the Newells must be sacrificing comfort for the sake of energy savings, but that’s not the case. Their house
boasts 2,100 square feet of living space and all of the amenities you would expect in a contemporary suburban residence.
On top of that, they enjoy much better indoor air quality than people who live in conventional homes, thanks to a constant flow of
conditioned fresh air from the outside.
You can find photos and further information about the Equinox House and net-zero living at [Link]/equinox
11.7: Case Study- A Net-Zero Energy Home in Urbana, Illinois is shared under a not declared license and was authored, remixed, and/or curated
by LibreTexts.
11.7.2 [Link]
Index
A I W
Aral Sea IPAT equation water pollution
5.3: Case Study- The Aral Sea - Going, Going, Gone 1.2: The IPAT Equation 5.4: Water Pollution
1 [Link]
Glossary
Sample Word 1 | Sample Definition 1
1 [Link]
Detailed Licensing
Overview
Title: Book: Sustainability - A Comprehensive Foundation (Cabezas)
Webpages: 84
All licenses found:
CC BY 4.0: 77.4% (65 pages)
Undeclared: 22.6% (19 pages)
By Page
Book: Sustainability - A Comprehensive Foundation 4.3: Biodiversity, Species Loss, and Ecosystem
(Cabezas) - CC BY 4.0 Function - CC BY 4.0
Front Matter - Undeclared 4.4: Soil and Sustainability - CC BY 4.0
TitlePage - Undeclared 5: Physical Resources- Water, Pollution, and Minerals -
InfoPage - Undeclared CC BY 4.0
Table of Contents - Undeclared 5.1: Physical Resources- Water, Pollution, and
Foreword - CC BY 4.0 Minerals - Chapter Introduction - CC BY 4.0
Licensing - Undeclared 5.2: Water Cycle and Fresh Water Supply - CC BY 4.0
Preface - CC BY 4.0 5.3: Case Study- The Aral Sea - Going, Going, Gone
1: Introduction to Sustainability Humanity and the - CC BY 4.0
Environment - CC BY 4.0 5.4: Water Pollution - CC BY 4.0
1.1: What is Sustainability? - Undeclared 5.5: Case Study- The Love Canal Disaster - CC BY
1.2: The IPAT Equation - CC BY 4.0 4.0
1.3: Human Consumption Patterns and the 5.6: Mineral Resources- Formation, Mining,
“Rebound” Effect - CC BY 4.0 Environmental Impact - CC BY 4.0
1.4: Challenges for Sustainability - CC BY 4.0 5.7: Case Study- Gold- Worth its Weight? -
Undeclared
2: The Evolution of Environmental Policy in the United
States - CC BY 4.0 6: Environmental and Resource Economics - CC BY 4.0
6.1: Environmental and Resource Economics -
2.1: The Evolution of Environmental Policy in the
Chapter Introduction - CC BY 4.0
United States - Chapter Introduction - CC BY 4.0
6.2: Tragedy of the Commons - CC BY 4.0
2.2: The American Conservation Movement - CC BY
6.3: Case Study- Marine Fisheries - CC BY 4.0
4.0
6.4: Environmental Valuation - CC BY 4.0
2.3: Environmental Risk Management - CC BY 4.0
6.5: Evaluating Projects and Policies - CC BY 4.0
2.4: Sustainability and Public Policy - CC BY 4.0
6.6: Solutions- Property Rights, Regulations, and
2.5: Public Health and Sustainability - CC BY 4.0
Incentive Policies - CC BY 4.0
3: Climate and Global Change - CC BY 4.0
7: Modern Environmental Management - CC BY 4.0
3.1: Climate and Global Change – Chapter
Introduction - CC BY 4.0 7.1: Modern Environmental Management – Chapter
3.2: Climate Processes; External and Internal Introduction - CC BY 4.0
Controls - CC BY 4.0 7.2: Systems of Waste Management - CC BY 4.0
3.3: Milankovitch Cycles and the Climate of the 7.3: Case Study- Electronic Waste and Extended
Quaternary - CC BY 4.0 Producer Responsibility - CC BY 4.0
3.4: Modern Climate Change - CC BY 4.0 7.4: Government and Laws on the Environment - CC
3.5: Climate Projections - CC BY 4.0 BY 4.0
7.5: Risk Assessment Methodology for Conventional
4: Biosphere - CC BY 4.0
and Alternative Sustainability Options - CC BY 4.0
4.1: Prelude to the Biosphere - CC BY 4.0
8: Sustainable Energy Systems - CC BY 4.0
4.2: Biogeochemical Cycles and the Flow of Energy
in the Earth System - CC BY 4.0
1 [Link]
8.1: Sustainable Energy Systems - Chapter 10.5: The Vulnerability of Industrialized Resource
Introduction - CC BY 4.0 Systems- Two Case Studies - CC BY 4.0
8.2: Environmental Challenges in Energy, Carbon 10.6: Case Study- Agriculture and the Global Bee
Dioxide, Air, Water and Land Use - CC BY 4.0 Colony Collapse - CC BY 4.0
8.3: Case Study- Greenhouse Gases and Climate 10.7: Case Study- Energy and the BP Oil Disaster -
Change - CC BY 4.0 Undeclared
8.4: Energy Sources and Carriers - Undeclared 10.8: Sustainability Ethics - Undeclared
8.5: Energy Uses - Undeclared 11: Sustainable Infrastructure - CC BY 4.0
8.6: Applications of Phase Change Materials for 11.1: Sustainable Infrastructure - Chapter
Sustainable Energy - Undeclared Introduction - CC BY 4.0
9: Problem-Solving, Metrics, and Tools for Sustainability 11.2: The Sustainable City - CC BY 4.0
- CC BY 4.0 11.3: Sustainability and Buildings - CC BY 4.0
9.1: Problem-Solving, Metrics, and Tools for 11.4: Sustainable Energy Practices - Climate Action
Sustainability - Chapter Introduction - CC BY 4.0 Planning - CC BY 4.0
9.2: Life Cycle Assessment - CC BY 4.0 11.5: Sustainable Transportation- Accessibility,
9.3: Derivative Life Cycle Concepts - CC BY 4.0 Mobility, and Derived Demand - CC BY 4.0
9.4: Sustainability and Business - Undeclared 11.6: Sustainable Stormwater Management -
10: Sustainability - Ethics, Culture, and History - CC BY Undeclared
4.0 11.7: Case Study- A Net-Zero Energy Home in
Urbana, Illinois - Undeclared
10.1: The Human Dimensions of Sustainability-
History, Culture, Ethics - CC BY 4.0 Back Matter - Undeclared
10.2: It’s Not Easy Being Green- Anti-Environmental Index - Undeclared
Discourse, Behavior, and Ideology - CC BY 4.0 Glossary - Undeclared
10.3: The Industrialization of Nature- A Modern Detailed Licensing - Undeclared
History (1500 to the present) - CC BY 4.0
10.4: Sustainability Studies- A Systems Literacy
Approach - CC BY 4.0
2 [Link]