0 ratings0% found this document useful (0 votes) 400 views723 pagesIntroduction To The Mechanics of A Continuous Medium
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content,
claim it here.
Available Formats
Download as PDF or read online on Scribd
INTRODUCTION
TO THE MECHANICS
OF A CONTINUOUS MEDIUM
Lawrence E. Malvern
Professor of Mechanics
College of Engineering
Michigan State University
Prentice-Hall, Inc.
Englewood Cliffs, New JerseyEnglewood Chis, N J
All rights reserved No part of this book
may be reproduced in any form or by any means
without permission in writing from the publisher.
Current printing (last digit):
0987
13-487603-2
Library of Congress Catalog Card Number 69-13712
Printed in the United States of AmericaPreface
This book offers a unified presentation of the concepts and general princi-
ples common to all branches of solid and fluid mechanics, designed to appeal
to the intuition and understanding of advanced undergraduate or first-year
postgraduate students in engineering or engineering science.
The book arose from the need to provide a general preparation in contin-
uum mechanics for students who will pursue further work in specialized fields
such as viscous fluids, elasticity, viscoelasticity, and plasticity. Originally the
book was introduced for reasons of pedagogical economy—to present the com-
mon foundations of these specialized subjects in a unified manner and also to
Provide some introduction to each subject for students who will not take
courses in all of these areas. This approach develops the foundations more
carefully than the traditional separate courses where there is a tendency to
hurry on to the applications, and moreover provides a background for later
advanced study in modern nonlinear continuum mechanics.
The first five chapters devoted to general concepts and principles applicable
to all continuous media are followed by a chapter on constitutive equations, the
equations defining particular media. The chapter on constitutive theory begins
with sections on the specific constitutive equations of linear viscosity, linearized
elasticity, linear viscoelasticity, and plasticity, and concludes with two sections
on modern constitutive theory. There are also a chapter on fluid mechanics and
one on linearized elasticity to serve as examples of how the general principles
of the first five chapters are combined with a constitutive equation to formu-
late a complete theory. Two appendices on curvilinear tensor components
follow, which may be omitted altogether or postponed until after the main
exposition is completed.
Although the book grew out of lecture notes for a one-quarter course for
first-year graduate students taught by the author and several colleagues during
the past 12 years, it contains enough material for a two-semester course and is
written at a level suitable for advanced undergraduate students, The only
vvi Preface
prerequisites are the basic mathematics and mechanics equivalent to that usu-
ally taught in the first two or three years of an undergraduate engineering
program. Chapter 2 reviews vectors and matrices and introduces what tensor
methods are needed. Part of this material may be postponed until needed, but
it is collected in Chap. 2 for reference.
The last 15 to 20 years have seen a great expansion of research and publi-
cation in modern continuum mechanics. The most notable developments have
been jn the theory of constitutive equations, especially in the formulation of
very general principles restricting the possible forms that constitutive equations
can take. These new theoretical developments are especially addressed to the
formulation of nonlinear constitutive equations, which are only briefly touched
upon in this book. But the new developments have also pointed up the limita-
tions of some of the widely used linear theories. This does not mean that any
of the older linear theories must be discarded, but the new developments pro-
vide some guidance to the conditions under which the older theories can be
used and the conditions where they are subject to significant error. The last
two sections of Chap. 6 survey modern constitutive theory and provide refer-
ences to original papers and to more extended treatments of the modern theory
than that gi in this introductory text.
The booK is a carefully graduated approach to the subject in both content
and style. The earlier part of the book is written with a great deal of illustrative
detail in the development of the basic concepts of stress and deformation and
the mathematical formulation used to represent the concepts. Symbolic forms
of the equations, ‘using dyadic notation, are supplemented by expanded Carte-
sian component forms, matrix forms, and indicial forms of the same equations
to give the student abundant opportunity to master the notations. There are
also many simple exercises involving interpretation of the general ideas in con-
crete examples. In Chaps. 4 and 5 there is a gradual transition to more reliance
on compact notations and a gradual increase in the demands on the reader's
ability to comprehend general statements.
Until the end of Sec. 4.2, each topic considered is treated fairly completely
and (except for the brief section on stress resultants in plate theory) only
concepts that will be used repeatedly in the following sections are introduced.
Then there begin to appear concepts and formulations whose full implementa-
tion is beyond the scope of the book. These include, for example, the relative
description of motion, mentioned in Sec. 4.3 and also in some later sections,
and the finite rotation and stretch tensors of Sec. 4.6, which are important in
some of the modern developments referred to in the last two sections of Chap.
6. The aim in presenting this material is to heighten the reader's awareness
that the subject of continuum mechanics is in a state of rapid development,
and to encourage his reading of the current literature. The chapters on fluids
and on elasticity also refer to published methods and results in addition to
those actually presented.Preface ‘vii
The sections on the constitutive equations of viscoelasticity and plasticity
are introduced by accounts of the observed responses of real materials in order
to motivate and also to point up the limitations of the idealized representations
that follow. The second section on plasticity includes work-hardening. a part
of the theory not in a satisfactory state, but so important in engineering appli-
cations that it was believed essential to mention and point out some of the
shortcomings of the available formulations.
A one-quarter course might well include most of the first five chapters, only
part of Chap. 6, and either Chap. 7 on fluids or Chap. 8 on elasticity. Section
3.6 on stress resultants in plates and those parts of Secs. 5.3 and 5.4 treating
couple stress can be omitted without destroying the continuity, as also can Secs.
6.5 and 6.6 on plasticity. Section 4.6 can be given only minor emphasis, or
omitted altogether if the last two sections of Chap. 6 are not to be covered,
The second appendix, presenting only physical components in orthogonal cur-
vilinear coordinates might be included if time permits; although not needed in
the text, it is useful for applications.
A two-term course could include the first appendix on general curvilinear
tensor components, useful as a preparation for reading some of the modern lit-
erature. There is sufficient textual material in the book for a full year course,
but it should probably be supplemented with some challenging applications
problems. Most of the exercises in the text are teaching devices to illuminate
the theory, rather than applications.
The book is a textbook, designed for classroom teaching or self-study, not
@ treatise reporting new scientific results. Obviously the author is indebted to
hundreds of investigators over a period of more than two centuries as well as
to earlier books in the field or in its specialized branches. Some of these inves-
tigators and authors are named in the text, but the bibliography at the end of
the book includes only the twentieth-century writings cited. Extensive bibli-
ographies may be found in the two Encyclopedia of Physics treatises: “The
Classical Field Theories,” by C. Truesdell and R. A. Toupin, Vol. I1],’l, pp.
226-793 (1960), and “The Non-Linear Field Theories of Mechanics.” by C.
Truesdell and W. Noll, Vol. 111/3 (1965), published by Springer-Verlag,
Berlin. These two valuable comprehensive treatises are among the references
for collateral reading cited at the end of the introduction. Many of the
historical allusions in the text are based on these two sources.
The author is indebted to several colleagues at Michigan State University
who have used preliminary versions of the book in their classes. These include
Dr. C. A. Tatro (now at the Lawrence Radiation Laboratory. Livermore,
California)and Professors M. A. Medick, R. W. Little. and K. N. Subramanian.
Professors John Foss and Merle Potter read the first version of the material
‘on fivid mechanics. Encouragement and helpful criticism have been provided
by these colleagues and also by the dozens of students who have taken the
course.vill Preface
The author is also indebted to Michigan State University for sabbatical
leave during 1966-67 to work on the book and to Prentice-Hall, Inc., for their
cooperation and assistance in preparing the final text and illustrations.
Finally, thanks are due to the author's wife for inspiration, encouragement
and forbearance.
Lawrence E, MALVERN
East Lansing, Michigan3.
4.
Contents
Introduction 1
1.1 The Continuous Medium = 1
Vectors and Tensors 7
21 Introduction 7
2.2 Vectors; Vector Addition; Vector and Scalar Components;
Indicial Notation; Finite Rotations not Vectors 10
2.3 Scalar Product and Vector Product 17
2.4 Change of Orthonormal Basis (Rotation of Axes); Tensors
as Linear Vector Functions; Rectangular Cartesian Tensor
Components; Dyadics; Tensor Properties; Review of Ele-
mentary Matrix Concepts 25
2.5 Vector and Tensor Calculus; Differentiation; Gradient, Di-
vergence and Cus 48
Stress 64
3.1 Body Forces and Surface Forces 64
3.2 Traction or Stress Vector; Stress Components 69
3.3 Principal Axes of Stress and Principal Stresses; Invariants;
Spherical and Deviatoric Stress Tensors 85
3.4 Mohr's Circles 95
3.5 Plane Stre fohr’s Circle 102
3.6 Stress Resultants in the Simplified Theory of Bending of Thin
Plates 112
Strain and Deformation 120
4.1 Small Strain and Rotation in Two Dimensions 120
4.2 Small Strain and Rotation in Three Dimensions 129
ixContents
43 Kinematics of a Continuous Medium; Material Derivatives
138
44 Rate-of-Deformation Tensor (Stretching); Spin Tensor (Vor-
ticity); Natural Strain Increment 145
4.5 Finite Strain and Deformation; Eulerian and Lagrangian
Formulations; Geometric Measures of Strain; Relative De-
formation Gradient 154
46 Rotation and Stretch Tensors 172
4.7 Compatibility Conditions; Determination of Displacements
When Strains are Known = 183
General Principles
5.1 Introduction; Integral Transformations; Flux 197
5.2 Conservation of Mass; The Continuity Equation 205
5.3 Momentum Principles; Equations of Motion and Equilib-
tium; Couple Stresses 213
5.4 Energy Balance; First Law of Thermodynamics; Energy
Equation 226
5.5 Principle of Virtual Displacements 237
5.6 Entropy and the Second Law of Thermodynamics; the
Clausius-Dubem Inequality 248
5.7 The Caloric Equation of State; Gibbs Relation; Thermody-
namic Tensions; Thermodynamic Potentials; Dissipation
Function 260
Constitutive Equations
6.1 Introduction; Ideal Materials 273
62 Classical Elasticity; Generalized Hooke's Law; Isotropy;
Hyperelasticity; The Strain Energy Function or Elastic Po-
tential Function; Elastic Symmetry; Thermal Stresses 278
6.3 Fluids; Ideal Frictionless Fluid; Linearly Viscous (New-
tonian) Fluid; Stokes Condition of Vanishing Bulk Vis-
cosity; Laminar and Turbulent Flow 295
64 Linear Viscoelastic Response 306
6.5 Plasticity I. Plastic Behavior of Metals; Examples of Theo-
ties Neglecting Work-Hardening: Levy-Mises Perfectly
Plastic; Prandtl-Reuss Elastic, Perfectly Plastic; and Visco-
plastic Materials 327
66 Plasticity Il. More Advanced Theories; Yield Conditions;
Plastic-Potential Theory; Hardening Assumptions; Older
Total-Strain Theory (Deformation Theory) 346
6.1 Theories of Constitutive Equations 1: Principle of Equi-
presence; Fundamental Postulates of a Purely Mechanical
Theory; Principle of Material Frame-Indifference 378
197
273Contents xi
68 Theories of Constitutive Equations Il: Material Symmetry
Restrictions on Constitutive Equations of Simple Materials;
Isotropy 406
7. Fluid Mechanics 423
7.1 Field Equations of Newtonian Fluid: Navier-Stokes Equa-
tions; Example: Parallel Plane Flow of Incompressible Fluid
Between Flat Plates 423
7.2 Perfect Fluid: Euler Equation; Kelvin's Theorem; Bernoulli
Equation; Irrotational Flow; Velocity Potential; Acoustic
Waves; Gas Dynamics 434
7.3 Potential Flow of Incompressible Perfect Fluid 448
7.4 Similarity of Flow Fields in Experimental Model Analysis;
Characteristic Numbers; Dimensional Analysis 462
7.5 Limiting Cases: Creeping-Flow Equation and Boundary-
Layer Equations for Plane Flow of Incompressible Viscous
Fluid 475
8. Linearized Theory of Elasticity 497
8.1 Field Equations 497
8.2 Plane Elasticity in Rectangular Coordinates 505
83 Cylindrical Coordinate Components; Plane Elasticity in
Polar Coordinates 525
84 Three-Dimensional Elasticity; Solution for Displacements;
Vector and Scalar Potentials; Wave Equations; Galerkin
Vector; Papkovich-Neuber Potentials; Examples, Including
Boussinesq Problem $48
Appendix 1. Tensors 569
1.1 Introduction; Vector-Space Axioms; Linear Independence;
Basis; Contravariant Components of a Vector; Euclidean
Vector Space; Dual Base Vectors; Covariant Components
ofa Vector 569
1.2 Change of Basis; Unit-Tensor Components 576
1.3 Dyads and Dyadics; Dyadics as Second-Order Tensors;
Determinant Expansions; Vector (cross) Products 588
1.4 Curvilinear Coordinates; Contravariant and Covariant
‘Components Relative to the Natural Basis; The Metric
Tensor 596
1.5 Physical Components of Vectors and Tensors 606
1.6 Tensor Calculus; Covariant Derivative and Absolute De-xii Contents
; Christoffel Symbols; Gradient,
tivative of a Tensor Fie!
Divergence, and Curl; Laplacian 614
1.7 Deformation; Two-Point Tensors; Base Vectors; Metric
Tensors; Shifters; Total Covariant Derivative 629
1.8 Summary of General-Tensor Curvilinear-Component
Forms of Selected Field Equations of Continuum Me-
chanics 634
Appendix Il. Orthogonal Curvilinear Coordinates,
Physical Components of Tensors
IL 1 Coordinate Definitions; Scale Factors; Physical Compo-
nents; Derivatives of Unit Base Vectors and of Dyadics
64l
IL. 2 Gradient, Divergence, and Curl in Orthogonal Curvilinear
Coordinates 650
IL3 Examples of Field Equations of Continuum Mechanics,
‘Using Physical Components in Orthogonal Curvilinear
Coordinates 659
U4 Summary of Differential Formulas in Cylindrical and
Spherical Coordinates 667
Bibliography. Twentieth-Century Authors Cited
in the Text
Author Index
Subject Index
641
673
691INTRODUCTION
TO THE MECHANICS
OF A CONTINUOUS MEDIUMCHAPTER 1
Introduction
1.1 The continuous medium
The mechanics of a continuous medium is that branch of mechanics concerned
with the stresses in solids, liquids, and gases and the deformation or flow of
these materials. The adjective continuous refers to the simplifying concept
underlying the analysis: we disregard the molecular structure of matter and
picture it as being without gaps or empty spaces. We further suppose that
all the mathematical functions entering the theory are continuous functions,
except possibly at a finite number of interior surfaces separating regions of
continuity. This statement implies that the derivatives of the functions are
continuous too, if they enter the theory, since all functions entering the theory
are assumed continuous. This hypothetical continuous material we call a con-
tinuous medium or continuum,
The concept of a continuous medium permits us to define stress at a point,
geometric point in space conceived as occupying no volume, by a mathema-
tical limit like the definition of the derivative in differential calculus, This ap-
proach immediately makes the powerful methods of calculus available for the
study of nonuniform distributions of stress, and at the same time provides an
easily visualized physical model which agrees with the testimony of everyday
observation of matter in the large. Thus this approach allows the mathematical
analysis to be guided by intuition. The theories of elasticity, plasticity, and
fluid mechanics based on the concept of continuous material lead moreover to
quantitative predictions which agree closely with experience over a wide range
of conditions. These theories and the further simplified engineering theories of
beams and plates and shells, which are also based on the conunuum concept
of matter, are adequate for analysis of stress and deformation in most engi-
neering problems.
There are exceptions of course. lt would be too much to expect that a
theory which disregards the molecular nature of matter could account for all
12 Introduction Chap. 1
the observed properties of matter, even in the large. There is, for example,
nothing in the theories that accounts for the formation of a fatigue crack in a
body after many cycles of reversed loading. Problems in super aerodynamics
at altitudes where the air is extremely rarefied may require molecular theories.
Even in marginal situations like these, where continuum mechanics alone is
inadequate, it is sometimes possible to use the continuum theory in combination
with empirical information or with information derived from a physical theory
based on the molecular nature of the material.
The use of the continuum concept to construct a working theory unifying a
large body of observational knowledge and permitting the deduction of useful
conclusions obviously has nothing to do with any assumption as to the real
nature of matter. The existence of areas in which the theory is not applicable
does not destroy its usefulness in other areas.
In most of the analyses in continuum mechanics, two further assumptions
about the nature of the material are made, namely, that it is homogeneous and
isotropic. It should be clearly understood that the three assumptions are com-
pletely independent.
Continuity. A material is continuous if it completely fills the space that it
occupies, leaving no pores or empty spaces, and if furthermore its properties
are describable by continuous functions.
Homogeneity. A homogeneous material has identical properties at all
points,
Isotropy. A material is isotropic with respect to certain properties if these
Properties are the same in all directions.
It is quite possible, for example, to conceive of the atmosphere as a con-
tinuous fluid but still suppose that its density decreases with altitude. It is also
possible to conceive of a rolled steel plate that is both continuous and homo-
geneous but has different tensile strengths or different elastic moduli in tension
specimens cut from the plate, one with its axis in the direction of rolling and
one with its axis at right angles to the direction of rolling. This kind of an-
isotropy can be introduced into continuum mechanics, and has led to useful
results, The developments of the concepts of stress and strain in Chaps, 3 and
4 and the general principles of Chap. 5 use only the basic assumption of con-
tinuity. The other two simplifying assumptions are not needed until constitu-
tive equations (stress-strain relations) or strength properties of a material are
considered.
It is sometimes said that solutions to engineering problems obtained through
the simplifying assumptions of the engineering theory of beams are only ap-See. 1d Introduction 3
proximations to the “exact” solutions of the theory of elasticity. The fore-
going paragraphs should have made it clear that the adjective “exact” is not
justified. There may be exact mathematical solutions to the equations formulated
in elasticity or in other branches of continuum mechanics, but the equations
themselves are not exact descriptions of nature. In this respect the difference
between the elementary theory and the advanced theory is one of degree rather
than of kind. When the elementary theory is formulated consistently and
logically, it is just as respectable as the advanced theory from a mathematical
or logical point of view. And from a practical point of view it is just as good
in those areas where its predictions agree closely enough with experience. The
bounds of applicability of these elementary theories are determined by experi-
ence, either from experimental verification, or from comparisons with predic-
tions of the more advanced continuum theories.
The last fifteen years have seen a spectacular growth in general continuum.
theory as well as in its specialized branches. The theory divides logically into
three parts: (1) general principles, assumptions and consequences applicable to
all continuous media, (2) constitutive equations defining the particular idealized
material, and (3) specialized theories of each individual idealized material
built on the foundations of the general principles and the constitutive equations
of that material to the point where boundary-value problems are formulated
and solved for application to specific problems. The third part would require
a book at least as long as this one for each specialized theory. Moreover,
despite some superficial resemblance between the application of Laplace's equa-
tion in fluid mechanics and its application in some parts of linearized elasticity,
there is very little common ground in the way the complete specialized theories
are developed and applied. Hence specialized treatises must be consulted for
the individual materials. Two small samples are included in Chap. 7 on fluid
mechanics and Chap. 8 on linearized elasticity, to give some introduction to
the way these two theories are put together and applied. References to spe-
cialized treatises on fluids and elasticity are given in those chapters. Some works
‘on viscoelasticity and plasticity are also cited in the sections dealing with the
constitutive equations of viscoelasticity and plasticity in Chap. 6. Until recently
continuum mechanics was presented and studied almost exclusively in its
separate specialized branches, but during the last ten years a definite trend has
developed toward a unified treatment of the first two parts of the theory, the
general principles and the constitutive theory.
The general principles are best studied within the framework of a general
continuum theory. This approach emphasizes their general applicability, which
is easily lost from sight when the general principles are interspersed with
results applicable only to a specific theory. Some pedagogical economy is also
achieved by presenting them only once instead of in two or three courses that
might be taken by the same student. But an even more important advantage is4 Introduction Chap. 1
that the general principles can be presented more carefully, avoiding the tend-
ency to hurry through them to get to the applications.
Most of the material in the first five chapters of this book, covering the
general principles, is not new. Important concepts date back to Euler (1770)
and much of the formal development was carried out during the early part of
the nineteenth century by such pioneers as Cauchy, Navier, and Green, to
mention only a few. The foundations of the theory have been subjected 10 ex-
tensive recent critical review. C. Truesdell (1952) published an important
study, “The Mechanical Foundations of Elasticity and Fluid Mechanics,”
reprinted in book form in Continuum Mechanics, Vol. I (1966). Two monu-
mental treatises, Classical Field Theories by Truesdell and R. Toupin (1960),
and Nonlinear Field Theories of Mechanics by Truesdell and W. Noll (1965),
have included comprehensive accounts of the foundations of the theory as well
as of the newer developments in constitutive theory, with the most complete
historical and bibliographical information on the subject up to that time.
Because of the availability of these reference works no attempt will be made in
this book to give a complete historical account, especially of the early work.
For those early scientists mentioned in this text in connection with develop-
ments attributed, to them, complete bibliographical information can be found
in the three works cited above, These invaluable reference works and a few
recent books at various levels on continuum mechanics are included in the list
of references at the end of this chapter.
The most noteworthy recent developments in continuum mechanics have
been in the theories of constitutive equations. The classical constitutive equa-
tions presented in the first six sections of Chap. 6 were developed indepen-
dently before the advent of the modern theories of constitutive equations,
although a few modifications to them have been made as a result of the modern
theories. The modern theories are briefly referred to in Secs, 6.7 and 6.8, but
their full development and application, especially to nonlinear constitutive
equations, is beyond the scope of this introductory book. The modern theories
are in essence theories of theories, setting some limitations on the forms the
individual theories can take, once the proposed independent and dependent
variables have been chosen. This reduces the amount of experimentation and
intuition required to furnish the information needed for completing the def-
inition of the individual idealized material. Historically the linear theories and
the theory of plasticity developed by a different route, beginning with assump-
tions based on intuition and experience and only later being subjected to critical
scrutiny to determine whether the form of the theory satisfies the requirements
of a theory of theories. Some new constitutive equations will no doubt continue
to spring up in this fashion, growing out of the needs of engineers and scientists.
But the modern constitutive theories promise valuable aid in formulating new
theories, especially of nonlinear material behavior.Sec. 1.4 Introduction 5
Some of the pioneering contributions to the modern theories, especially the
work of R. S. Rivlin and his associates and of W. Noll, dating from about 1956,
are cited in Secs. 6.7 and 6.8. Many of the early papers and subsequent papers
continuing the development of the theory have been reprinted in Continuum
Mechanics, Vols. 1, II, 111, 1V (1965-66), a four-volume series reprinting 49
papers published by 18 authors between 1945 and 1961, edited by C. Truesdell,
referenced at the end of this chapter. The most complete account of the de-
velopment growing out of the work of Noll is in the Nonlinear Field Theories
of Mechanics already cited. For more introductory recent accounts see
Jaunzemis (1967), Truesdell (1965), Leigh (1968), and Eringen (1967).
The general principles of Chap. 5 applicable to all continuous media, and
constitutive equations defining the particular material, for example one of those
given in Chap. 6, form the field equations of a continuum theory. The me-
chanical variables appearing in these equations are the stress and kinematic
variables such as strain and rate of deformation. In elementary mechanics of
materials, tensile strain is defined as elongation per unit length of the test
specimen and is a dimensionless measure of the deformation. Stress is the in-
ternal force brought into play by the deformation, per unit cross-sectional area:
it is equal to the externally applied force, per unit cross-sectional area. The
stress tensor of a three-dimensional combined-stress field is a generalization of the
elementary concept of stress, which will be developed in Chap. 3. The small-
strain tensor of Secs. 4.1 and 4.2, a generalization of the elementary concept
of strain, is the only kinematic variable appearing in the constitutive equation
of linearized elasticity. But additional kinds of kinematic variables are needed
for more general continua. Some of these additional kinematic concepts, in-
cluding rate of deformation and finite strain and rotation tensors, will be
developed in Chap. 4. The stress tensor will be treated first, since it is simpler
and casier to visualize for most engineers, and it affords an opportunity to
develop some skill in the use of tensors before the more complicated kinematic
tensor development is encountered.
As a preparation for the stress tensor development of Chap. 3 and the
kinematic tensor development of Chap. 4, a summary of vectors and tensors is
given in Chap. 2, which can serve either as a quick review for the reader
already acquainted with tensors or as a self-contained introduction to tensors.
Additional material on curvilinear components of tensors is given in Appendices
Tand I}, but these are not needed for the main body of the text. Some addi-
tional vector and tensor concepts are also presented where they are first needed
in Chaps. 3,5, and 7, and the reader to whom the tensor methods are new
will find that the repeated use of them required throughout the book and the
Many cross references to Chap. 2 will greatly strengthen his understanding
beyond that acquired in a first reading of Chap. 2.6 Introduction Chap. 1
Selected references for collateral reading
COLLECTIONS OF ORIGINAL PAPERS AND TREATISES
Truesdell, C., ed., Continuum Mechanics (reprints). New York: Gordon and Breach,
vy. 1 (1966), v. II, IIT, LV (1965).
Truesdell, C. and R. A. Toupin, “The Classical Field Theories,” : Encyclopedia of
Physics, v. W1/1, ed. S. Fliigge. Berlin: Springer-Verlag, 1960.
Truesdell, C. and W. Noll, “The Non-Linear Field Theories of Mechanics,” in Ency-
clopedia of Physics, v. 11/3, ed. S. Fliigge. Berlin: Springer-Verlag, 1965.
IntRopucTory TexTsooxs Not INCLUDING MODERN CONSTITUTIVE THEORIES
Aris, R., Vectors, Tensors, and the Basic Equations of Fluid Mechanics. Englewood
Cliffs, N. J.: Prentice-Hall, Inc., 1962.
Frederick, D. and T.S. Chang, Continuum Mechanics. Boston: Allyn and Bacon,
Inc., 1965.
Fung, ¥.C., Foundations of Solid Mechanics, Englewood Cliffs, N.J.: Prentice-
Hall, Inc., 1965."
Long, R.R., Mechanics of Solids and Fluids. Englewood Cliffs, N.J.: Prentice-
Hall, Inc., 1961.
Prager, W., Introduction to Mechanics of Continua. New York: Ginn and Com-
pany, 1961.
Scipio, L. A., Principles of Continua with Applications. New York: John Wiley &
Sons, Inc., 1967.
TEXTBOOKS INCLUDING MODERN CONSTITUTIVE THEORIES
Eringen, A. C., Mechanics of Continua. New York: John Wiley & Sons, Inc., 1967.
Eringen, A. C., Nonlinear Theory of Continuous Media. New York: McGraw-Hill
Book Company, 1962.
Jaunzemis, W., Continuum Mechanics. New York: The Macmillan Company, 1967.
Leigh, D. C., Nonlinear Continuum Mechanics. New York: McGraw-Hill Book
Company, 1968.
Truesdell, C., The Elements of Continuum Mechanics. New York: Springer-Verlag,
1965.CHAPTER 2
Vectors and Tensors
2.1 Introduction
Physical laws, if they really describe the physical world, should be independent
of the position and orientation of the observer. That is, if two scientists using
different coordinate systems observe the same physical event, it should be pos-
sible to state a physical law governing the event in such a way that if the law is
true for one observer, it is also true for the other. For this reason, the equa-
tions of physical laws are vector equations or tensor equations, since vectors
and tensors transform from one coordinate system to another in such a way
that if a vector or tensor equation holds in one coordinate system, it holds in
any other coordinate system not moving relative to the first one, i.e., in any
other coordinate system in the same reference frame. Invariance of the form
of a pirysical law referred to two frames of reference in accelerated motion rela-
tive to each other is more difficult and requires the apparatus of general rela~
tivity theory, tensors in four-dimensional space-time. For simplicity, we limit
ourselves in this book to tensors in three-dimensional Euclidean space and for
most of the discussion use only rectangular Cartesian coordinates for the com-
ponents of tensors and vectors. The way in which the physical laws defining
material response transform from one frame of reference to another frame
moving relative to the first will be explained in Sec. 6.7 by means of the pos-
tulated principle of material frame indifference.
Most readers will have used vectors before; however, we review in Secs. 2.2
to 2.5 and Sec. 5.1 some of the ideas of vector algebra and calculus. The con-
cept of a tensor asa linear vector operator will be presented in this introduction
and further developed in Sec. 2.4, where some properties of tensors are sum-
marized and Cartesian tensor components are defined. The Cartesian com-
ponent formulation is sufficient for the development of the theory of continuum
mechanics and will be used throughout the main body of the text. For the
solution of specific problems, orthogonal curvilinear coordinates suitable to the
ans8 Vectors and Tensors Chap. 2
geometry of the problem may lead to simplification of the analysis. Appendix
TJ presents the physical components of vectors and tensors in orthogonal cur-
vilineat coordinates and summarizes the equations of continuum mechanics
in this form. It is recommended that on a first reading the material of this
appendix be omitted or postponed unti] the continuum theory has been com-
pleted. Appendix I presents the more complicated formulation of general
tensor components referred to an arbitrary curvilinear coordinate system; this
general formulation is not actually necessary for the development of continuum
mechanics, and is included only because many modern writers on the subject
have chosen to use this language. The advanced student can profit from some
acquaintance with the language of general tensors in order to read the literature,
but it is recommended that App. I be omitted in an introductory course.
Tensors as linear vector operators. In three-dimensional space. a vector
may be visualized as an arrow that has length equal to the magnitude of the
vector and that is pointing in the direction of the vector; it thus is seen to have
an existence quite independent of any coordinate system in which it might be
observed. Although a tensor cannot be visualized simply as an arrow, it also
has an existence independent of the coordinate system. Tensors frequently arise
as physical entities whose components are the coefficients of a linear relationship
between vectors. For example, in Chap. 3 we will find that for every direc-
tion at a point, specified by a vector f, there may be associated a second vector
t, the traction vector (force per unit area) acting at the point across the surface
whose unit normal is A. At a given point, t is thus a vector function of fi;
symbolically t = t(f). The functional dependence in this case is an example of
a linear vector function. (A vector function F is said to be a linear vector
function if F(an + by) = 2F(u) + bF(v) for every choice of the vectors u, ¥,
and real numbers a, b.) Since the argument fi of the function and the func-
tional value t associated with it are both vectors, and therefore have an exis-
tence independent of any coordinate system, the functional relationship must
then be independent of any coordinate system. Actually, in practice, we
observe the two vectors in a particular coordinate system and express the func
tional relationship by means of coefficients referred to that coordinate system.
These coefficients will be the components of the tensor that relates one vector
to the other. And we will derive tensor transformation formulas giving the
tensor components in any other coordinate system, such that when these com-
ponents are used as coefficients of the vector function, the function yields the
same vector as before, but the vector is represented by its components in the
new coordinate system.
In any one rectangular coordinate system, the linear vector functional rela-
tionship may be represented as a matrix equation involving the square matrixSec. 2.1 Vectors and Tensors 9
of tensor components either premultiplied by the argument vector expressed as
a row matrix or postmultiplied by the argument vector expressed as a column
matrix. (See Part 4 of Sec. 2.4, for a review of elementary matrix concepts.)
This is a very important kind of relationship in all branches of physics. Some
examples are:
1. Tis stress tensor (square matrix).
nis unit vector normal of surface (row matrix).
tis traction vector on surface (row matrix):
t=nT,
2. Tis inertia tensor of a rigid body (square matrix).
is angular velocity vector of body (column matrix).
A is angular momentum vector of body (column matrix):
h=Io.
3. E_ is small strain tensor (square matrix).
dX is infinitesimal initial relative position vector of adjacent material
particles (column matrix).
du is relative displacement vector produced by straining (column
matrix):
du = Edx.
4. K_ is thermal conductivity of anisotropic medium (square matrix).
q_ is heat flux vector (column matrix).
‘V8 is gradient of the temperature (column matrix with components
26 20 26).
ax’ By 32”
q=~—KV6.
5. 3 is electrical conductivity of anisotropic medium (square matrix).
e iselectric field intensity vector (column matrix).
i is current density vector (column matrix):
i=De.
These are just a few of the many physical laws in which a tensor is used as
the function associating one vector with another. Remember that the relation-
ship between vectors must be independent of the coordinate system, although
we usually observe it in some one system; and remember that it is precisely this
independence of the coordinate system that motivates us to study tensors at all.
Before beginning our study of tensors we first review (in Secs. 2.2 and 2.3)
some of the notions of elementary vector algebra.10 Vectors and Tensors Chap. 2
2.2 Vectors;+ vector addition; vector and scalar components; indicial
notation; summation convention; finite rotations are not vectors
In a three-dimensional Euclidean space, vector quantities (or briefly, rectors)
are entities possessing both magnitude and direction and obeying certain laws;
a vector is conventionally represented by an arrow that is pointing in the
direction associated with the vector and that has a length proportional to the
magnitude of the vector. The simplest example is a directed line joining two
points. Other familiar examples are force, velocity, and acceleration. A scalar
quantity, on the other hand, does not have any direction associated with it,
although some scalar quantities may have positive and negative values, e.g.,
Fahrenheit temperature. In printed text, a vector is usually denoted by a
boldface letter. The magnitude of the vector is then usually denoted by the
same letter in lightface italic type. In typewritten or handwritten text, the wavy
underscore used by copyreaders to designate boldface type is often used for
both vectors and tensors, or the vector may be identified by an arrow above
the fetter, or by underlining. The magnitude of vector a is denoted by a (the
same letter in italic instead of boldface and not underlined), or by [al, and is
always a positive number or zero. In script @ = Jal.
Two sectors are equal if they have the same direction and the same mag-
nitude. Thus a vector is unchanged if it is moved parallel to itself. In some
applications, however, the point or the line of action of a vector quantity is
important. For example, a force is a vector quantity with a definite point of
action; it is a fixed or hound vector. In rigid-body dynamics, only the line of
action of the force is important, not the point; force is then a sliding vector.
When we say that two forces are equal, however, we mean only that they have
the same magnitude and direction; this does not imply that they are mechani-
cally equivalent.
Vector addition. Any two vector quantities of the same kind (e.g., two
forces or two velocities) may be represented as
two vectors a and h so placed that the initial
point of b coincides with the terminal point of
a, as in Fig. 2.1. The swum of a and b is then
defined as the vector ¢ extending from the ini-
tial point of a to the terminal point of b:
+b. (2.2.1)
Fig. 2.1 Addition of vectors.
Only those definitions and formulas used in the text are summarized here; they are
so important in mechanics that they should be mastered by every serious student of the
subject. For additional details, or for an axiomatic development of the theory, see books
‘on vector analysis and vector spaces.See. 2.2 Vectors and Tensors W
The actual addition of two vectors is most conveniently performed by using
their rectangular components in some coordinate system as in Eq. (8). but the
definition and properties of vector addition do not depend on the introduction
of a coordinate system.
From the definition, it follows that
b+a=a-+t
as is indicated by the dashed arrows in Fig. 2.1. Thus vector addition obeys
the commutative law. It also follows from the definition that the associative law
(a+b) +ce=a+ (b+ c)
is satisfied. The common value of the two expressions is written as the sum
2 +b + of the three vectors
Ifa vector is reversed in direction with no change in magnitude, the re-
sulting vector is called the negative of the original vector. For example,
BA = — AB in Fig. 2.1. To subtract b from a, add the negative of b to a:
a—b=a+(—b). (2.2.2)
A vector a may be multiplied by a scalar c to yield a new vector, ca or ac,
of magnitude |ca|. If c is positive, ca has the same direction as a; if ¢ is nega-
tive, ca has the direction of —a. If c is zero, the product-is the zero vector
with zero magnitude and undefined direction. Thus (0)a = 0. Sometimes 0 is
written as 0, not in boldface and not underlined in script.
A vector of unit length is called a unit vector. Any vector may be written
as the product of the scalar magnitude of the given vector by a dimensionless
unit vector in the given direction. Thus a = ai, where i is a dimensionless
unit vector in the direction of a. A unit vector may be signified by the caret
(*) placed over its letter symbol. The caret is often omitted, when the context
makes it clear that the vector is a unit vector.
Vector components and scalar components. Two or more vectors whose
sum is a given vector are called rector components of the given vector. It is
often convenient to resolve a given vector into three components parallel to
three noncoplanar base vectors; the resolution is unique for a given vector and
given base vectors. The set of base vectors is called a basis. If the basis con-
sists of mutually orthogonal unit vectors, it is called an orthonormal basis. In
this book, the base vectors are usually specified as unit vectors in the positive
directions of a set of rectangular Cartesian coordinate axes. The components
are then called the rectangular vector components of the given vector in that
coordinate system. The name rectangular components is also used for the
numerical coefficients of the base vectors. that is, for a,,.a, and a. in Eq. (4)
below.12 Vectors and Tensors Chap. 2
Fig. 2.2. Rectangular components and unit vectors.
In Fig. 2.2, for example, PA, AD, and DE are the rectangular vector com-
ponents of the vector PE, in the directions of the unit base vectors i, j,k. We
denote the vector by a, its magnitude by a, and the angles it makes with the
positive coordinate directions by @,,@,,@,. Then the three rectangular com-
ponents of the vector, that is, the numerical coefficients of the base vectors, are
@, = acosa,, a, = acosa,, a, = @Cos a, (2.2.3)
and the rectangular vector components are a,i, a,j, @,k so that
a=a,i+aj+a.k. (2.2.4)
Note that a, is a number (positive, negative, or zero), while a.i is a vector,
4 directed physical quantity parallel to i. The rectangular components are
the orthogonal projections of the given vector onto the three coordinate axes.
The word component is also often used as a synonym for “projection on an
arbitrary direction,” For example, a unit vector fi may be given (not necessarily
parallel to a coordinate axis); “the component a, of » in the direction of a”
then simply means the orthogonal projection of a onto a line having the direc:
tion of f, namely
a, = @ cos 6, (2.2.5)
where @ is the angle between the directions of a and i. This equation applies
whether or not a and f are drawn from the same point, but the angle @ is more
easily visualized if one of the vectors is moved parallel to itself to the same
initial point as the other.Sec. 22 Vectors and Tensors 13
The three cosines, cos az, cos @y, cos @,, are the direction cosines of the
vector. By repeated application of the theorem of Pythagoras we see that the
magnitude of the vector is given in terms of the scalar components by
at = ab +a} + al. (2.2.6)
Substitution of Eqs. (3) into this relationship yields
cos? a, + cos? a, + costa, = 1, (2.2.7)
which expresses the fact that the sum of the squares of the direction cosines of
any direction is unity.
The addition of two or more vectors is most simply performed in terms of
their rectangular components. For example, if
c=a+b and d=a—b
then the components of c and d are given by
es=a,tb, d,=a,—b,
ty=ay+b, dy=a,—b, (2.2.8)
¢:=a,+b, d, = a, — b,.
Equations (8) are special cases of the relationship
Cy = an + by, 2.2.9)
which is valid for the components in an arbitrary direction fi (projections), as
is evident from Fig. 2.3, where c = a + b and where a, = OA,b, = AB and
c= OB.
Fig. 2.3 Projection OB of sum ¢ equals sum of
projections.14 Vectors and Tensors Chap. 2
Indicial notation. Instead of denoting the rectangular Cartesian coordi-
nates of a point by (x, y, 2), we may call them (x;, X2, x) as is indicated by the
labeling of the coordinate axes in Fig. 2.4.
eee
K
Fig. 2.4 Indicial notation: x,-, xr, xyaxes. Vector a.
The components of a vector a are then denoted by a,,/ = 1,2,3. The
three components are (a, ds, 23). The symbol a, is used for a typical one of
the components or for the set of three components. It is also sometimes used
instead of a to represent the vector itself; the context prevents serious mis-
understanding of this usege. Then if @ is a unit vector in the direction of the
vector a, its components are
Ny = COS Ay,
where the @, are the angles between a and the x,-axes. Hence the components
of the vector a are given by
a = an, (2.2.10)
Equation (10) represents a set of three equations, obtained by giving i, succes-
sively, the values 1, 2,3. In fact Eq. (10) is a compact form of Eq. (3). Simi-
larly the three equations (8) for the components of ¢ can be written compactly
as
ey ay + by
Summation convention. The indicial notation permits us to write many of
the formulas of continuum mechanics in a much shorter form than would
otherwise be possible. This brevity makes the formulas easier 10 remember and
also easier to understand, once the notation is learned. Many of our formulasSec. 22 Vectors and Tensors 15
require a summation over all values of the index. For example, the magnitude
a of a vector a, is given by Eq. (6) as
(2.2.11)
a
Lael
We adopt the summation convention which states that in Cartesian coordinates
whenever the same letter subscript occurs twice in a term, that subscript is to
be given all possible values and the results added together. The symbol 5 may
then be omitted because the repeated subscript indicates the summation. For
exampie, in three dimensions,
ay = ai + a + ai,
De = Oy + Ose 7 Assy
Gybsy = aids + Geabss + Gabry
When general tensor components in curvilinear coordinates are used, summa-
tion is only implied by repeated indices if one of the repeated indices appears
as a superscript and one as a subscript; ¢.2.,
by + ath, + abby.
by =
Equation (11) can now be written as
a= aa, (2.2.12)
This is a single equation, while Eq. (10) is a set of three equations, because
Eq. (10) has a free subscript, ic., a letter subscript not repeated twice in the
same term. Note that c, = a; + ; is also a set of three equations because the
‘same subscript does not occur twice in one term. The notation ¢,; = a1; + bry
Tepresents nine equations, one for each of the nine possible combinations of nu-
merical values 1, 2,3 that the free subscripts i and j can have.
Finite rotations are not vectors. Not ail entities representable as directed
lines are vectors; to qualify as vectors they must satisfy the laws of vectors. In
particular, they must combine according to the laws of vector addition. For
example, it is one of the basic postulates of Newtonian mechanics that two
forces applied at the same point may be combined according to these laws, i.c.,
the vector sum applied at the point produces the same effect as the two forces
would produce. And this postulate is crucial in establishing that forces are
vector quantities. Finite rigid-body rotations are not vector quantities, since
they do not combine vectorially, as the following counterexample shows, de-
spite the fact that any one rigid-body rotation about a fixed axis may be repre-
sented as an arrow laid off to scale along the axis of rotation.16 Vectors and Tensors Chap. 2
=
zk
Fig. 2.5 Example showing that finite rotations are
‘not vectors.
In Fig. 2.5, the rigid body is taken to be the rectangle initially lying in the
xp-plane with diagonal OA». It is given the following sequence of 90-degree
rotations, each about a fixed axis:
1. About x-axis to position OA,, represented as an arrow (z/2)i in the x-
direction,
2. About y-axis to position O.4,, represented by (x/2)j,
3. About z-axis to position OAs, represented by (2/2)k
Tf finite rotations were vector quantities, then the single rotation about a fixed
axis, defined by the vector sum (2/2)(i + j + k) of the three individual rota-
tions, would take the rectangle from the initial position OA, to the final posi-
tion OAs. But this is not the case, since evidently the single rotation which
takes OA, to OA; is a 90-degree rotation about the y-axis, represented as
(x/2)j, which is not the vector sum of the three individual rotations. This
shows that finite rotations are not vectors.
If a vector quantity is conceived of as a physical entity (e.g., force or ve-
locity), it is clear that it has an existence independent of the coordinate system.
If another rectangular Cartesian system is used, the components of the vector
in the second system are different from those in the first, but it is still the same,
vector. lt must therefore be possible to express the vector’s components in the
new coordinate system in terms of those in the old one; the formulas for such
a change of axes will be given in Sec. 2.4. Thus, if the components of a vector
are given in one coordinate system, the vector is completely specified. Before
considering change of axes in Sec. 2.4, we will review the notions of scalar
product and vector product: two different ways of multiplying two vectors.Sec. 23 Vectors and Tensors 17
2.3 Scalar product and vector product
In Sec. 2.2, multiplication of a vector by a scalar was defined to yield a vector.
We now define two different kinds of multiplication of one vector by another.
In one case, the product is not a vector but a scalar; it is called the scaler
product or dot product. The second case yields a vector; it is called the vector
product of cross product, A third kind of multiplication yielding a tensor is
defined in Eq. (2.4.21).
The scalar product is defined as the product of the two magnimdes times
the cosine of the angle between the vectors:
arb = abcosé. (2.3.1)
It follows immediately from the definition that if m and 7 are scalars (numbers),
then
(ma)+(nb) = mn(a+b). (2.3.2)
It is also evident from the definition that the scalar product is commutative,
acb = bea. (2.3.3)
‘We may interpret ab cos @ either as b times the projection of a onto the direc-
tion of b or as a times the projection of b onto a.
‘The scalar product is also distributive, i.c.,
a+(b + c) = (ab) + (ac), (23.4)
as the following argument shows. Since the projection of b + ¢ onto a equals
the sum of the projections of b and ¢ onto a, according to Eq. (2.2.9); and
since a+(b + c) may, by the scalar product definition, be interpreted as the
product of a by the projection of b ++ ¢ onto a, Eq. (4) takes the form
abe + Ca) = abs + ala,
which is known to be true by the distributive property of the arithmetic prod-
uct of numbers. Here b, and c, are the components (projections) of b and ¢
in the direction of a. The distributive property of the scalar product permits
us to derive a formula for the scalar product in terms of the rectangular com-
Ponents by using the unit base vectors i, j, k.
From the defining equation (1), it follows that the scalar product of any
two different unit base vectors is zero, since cos 90° = 0, while the scalar
Product of any unit vector by itself equals unity.18 Vectors and Tensors Chap. 2
i= jjakk=I
bjs pk =ki=0. (235)
Then the scalar product
acb = (a,1-+ a,j + ask)-(bei + byj + bk)
yields
a,b, + ayby + a,b
or (2.3.6)
in
idicial notation.
Remember this very important formula. It furnishes the method by which we
actually evaluate scalar products. But do not forget Eq. (1), which gives physi-
cal significance to what we accomplish by using Eq. (6).
One of the most common applications of the scalar product is the projection
of a given vector onto a line having a given direction. If a is the given vector
and fis a unit vector in the given direction, then the projection a, is given by
ay = ah = ayy ayny ban, OF ay = am 3.7)
The three components (11,, My, 12) of the unit vector fi are the direction cosines
of the given direction.
Another important geometrical application is the determination of the angle
between two vectors given in terms of their components. From Eq. (1), we
obtain
(2.3.8)
‘The numerator of this formula is given by Fq. (6), and the two magnitudes in
the denominator may be calculated in terms of the components by Eq. (2.2.6).
It follows from Eq. (8), that a necessary and sufficient condition for two nonzero
vectors to be perpendicular is
0. (2.3.9)
Another important application of the scalar product is the calculation of
the work done by a force. If force f acts on a point of application which moves
through the infinitesimal displacement dr, then the work dIV done is
AW =: fede ~ fos 0 ds, (2.3.10)Sec. 2.3 Vectors and Tensors 19
where @ is the angle between f and dr, and where fand ds are the magnitudes
of f and dr, respectively. Since, in general, the force may vary both in mag-
nitude and direction during any finite motion, it will be necessary to express
the force asa function of position and evaluate the total work as a line integral
along the path of the motion, When f is constant and the motion is along a
straight line, this results in a total work IV given by
W=f-Ar (2.3.11)
where Ar is the finite displacement. We note again that the work IV isa scalar
quantity, obtained by multiplying two vectors, g the scalar product. We
now consider another way of multiplying two vectors; this time the product
itself be a vector.
The second kind of product of one vector multiplied by another is the
vector product ot cross product:
c=axb, (2.3.12)
defined as a vector ¢ perpendicular to both a and b in the sense that makes a, b,
ca right-handed system. Magnitude c is given by
= absin 6. (2.3.13)
(This always yields a nonnegative value for c because the angle is between
zero and 180°.) Because of the sense prescription,
vector multiplication is not commutative. We
have instead
bx a= —(a x b), (2.3.14)
as illustrated in Fig. 2.6. The magnitudes are the
same; in either case the magnitude c is numeri-
cally equal to the area absin 6 of the parallelo-
gram formed with the two given vectors as sides;
but the senses are opposite. :
For example, the unit vectors i, j, k of a right- Yoc= bxa
handed system form the following products: Fig. 2.6 Vector products.
ixi xj=kxk=0 (the zero vector),
ixjak, Jxk= kxia} (2.3.15)
jxt=-k, kx] . ix k=].
The algebraic signs for the products (15) can be remembered by the scheme
shown in the figure on page 20. If the two factors occur in the same sequence20 Vectors and Tensors Chap. 2
as i, j,k on the circle, the product is positive. If the
order is reversed, the product is negative. (All the signs
7 # are reversed for a left-handed system.)
The vector product is also distributive, but we omit
the proof.t Thus
ax(b+e)=(ax b+ (axe). (2.3.16)
7 The distributive character of the product permits us to
evaluate it in terms of the components of the given vectors. This is done by
writing each vector in the form given in Eq. (2.2.4), multiplying the two vectors
together, and simplifying by means of Eqs. (15). The result is, for a right-
handed system,
a X b= (a,b, — a,by)i + (a,bz — azb,)j + (a2by — aybz)k, (2.3.17)
which may be remembered conveniently as the formal expansion of the deter-
minant
ij ok
axb=|a, a, a, (2.3.18)
jb, byob
in terms of the minors of the elements in the first row. For a left-handed
coordinate system a minus sign is prefixed to the formulas. The cross product
would be an example of an axial vector if we did not employ this con-
vention of prefixing a minus sign to Eqs. (17) and (18) in a left-handed system.
With this convention, the cross product behaves as a polar rector or true vector
under transformation of coordinates (see Sec. 2.4). When the vectors have
several zero components, it is often more convenient to apply Eq. (15) directly
instead of using Eqs. (17) or (18). For example,
2X (3} — 2k) = 61 x J— AE x k= 6k + 4}.
Here and in the derivation of Eq. (17), we have tacitly assumed in addition to
the distributive property that
(ma) x (nb) = rin(a x b) (2.3.19)
for arbitrary scalars m and n, a consequence immediately evident from the
definition of the vector product. However, the vector product is not associative
aX (bX c)# (a xb) xc.
See, for example, Nathaniel Coburn, Vector Analysis (New York: The Macmillan
Company, 1955), p. 14.Sec. 2.3 Vectors and Tensors 2
Hence parentheses must be used to indicate the order of multiplication in the
vector triple product.
With the scalar triple product a x bec, parentheses are not needed, since
this product has meaning only if the cross product is evaluated first. However,
the dot and cross may be interchanged without changing the (scalar) value of
the product. You can easily verify this by writing each vector in the standard
component form of Eq. (2.2.4) and carrying out the multiplications for a right-
handed coordinate system:
a, ay a,
axbe=abxec=|b, by by}. (2.3.20)
Cr Cy Ce
In fact if a, b, ¢ are oriented so that they form a right-handed system (so that
this product is positive), the scalar triple product of Eq. (20) gives the volume
of the parallelepiped formed with a, b and ¢ as three edges.
Applications of the vector product include the calculation of the moment
of a force, the determination of the displacements produced by a small rotation,
or the particle velocities induced by rotation. One important geometric appli-
cation is the determination of a direction perpendicular to two given directions.
A unit vector ft perpendicular to both a and b is obtained as
A= 2xb. (2.3.21)
The rectangular components thus obtained wilt be the direction cosines of the
common perpendicular, This is especially useful in relating new coordinate axes
to the old ones, An example will be given in Chap. 3; when we are finding the
principal axes of stress, we will obtain the third axis as the common perpendic-
ular to the first two.
In order to write the vector product in indicial notation, we introduce the
permutation symbol enn, defined to have the value 0, --1, or —1 as follows:
0 when any two indices are equal;
+1 when m,n,r are I, 2,3 or an even
nar permutation of 1, 2, 3; (2.3.22)
when m,n, 7 are an odd permutation
of 1, 2,3.
Thus, for example,
€1n9 = Cant = Can =
in = Cn = Orn = —E
eu = ems = ers = 0, ele,22 Vectors and Tensors Chap. 2
The sign convention can be remembered by marking the numbers I, 2,3 on a
circle, Then any even permutation will be in cyclic order, ic., it will go around
the circle in the same sense as 1, 2, 3, while an odd permutation will go in the
reverse direction or acyclic order, Then the three components c, are given in
a right-handed coordinate system by
C=AXD Cy = Cyardyby = Cryer, (2.3.23)
as you can verify by writing out the sums represented in Eq. (23) and compar-
ing the results with Eq. (17). Note that because there is one free subscript p,
there are three Eqs. (23), and each equation contains nine terms: one for each
possible choice of the two summation indices q and r. But when these nine
terms are written out, all but two will have a repeated index on the permuta-
tion symbol, which means that there ure only two nonzero terms in each
nine-term sum—one requiring a plus sign and one a minus sign, The vector
equation (17) may be written for a right-handed system as
aX b= Cyerdeb,t, (vector sum on p). (2.3.24)
For actual numerical evaluations, the determinant scheme of Eq. (18) or direct
use of the unit vector products, as in the example after Eq. (18), is more con-
venient than the indicial Eqs. (23) or (24), but the indicial expressions are
useful in theoretical derivations because they can be substituted conveniently
into other equations.
The Kronecker delta 8,q is another symbol useful to simplify and shorten
equations in index notation. It is defined by
oa ee (2.3.25)
la By 3.
ronecker dela Ba [oer
where pand g can each have any positive integer value appearing in the dis-
cussion, that is 1, 2, or 3 for the usual case of three dimensions, Asan example
of the use of the Kronecker delta, we note that the statement that 4,, 1,, 1, form
an orthonormal set, ie. a set of three mutually orthogonal vectors each of unit
magnitude, can be expressed by the equation
iri, = By (2.3.26)
ing out the summations
Eq. (25) we can ea
implied by the repeated indices, and using
y establish the following useful identities.
By wri
the definition ofSec. 2.3 Vectors and Tensors 23
@) Ban = 3 (0) 8nn Bnn = 3
(©) Um Ban = Hn) Tran Ban = Tax oe
The e-B identity of Eq, (28) below can also be verified by writing out the
expression, although this requires a more lengthy enumeration of the indepen-
dent choices of the free indices.
entity cs en
By (2.3.28)
This identity relating the three-index symbols ¢,,,, of Eq. (22) to the Kronecker
delta of Eq. (25) is useful for proving many vector identities. For example, the
following identity for triple vector product
a X (bX ¢) = (arc)b — (ab)e (2.3.29)
may be proved as follows. Let vy; denote the rectangular Cartesian components
of the triple product on the left, Then
tH = €114(Cirsb-e,) by Eq. (23)
Fer Cirayb,C, = (By8ir — By5..)ajb¢. by Eq. (28).
ty = (By yC)8ird, — (Byes beBises
= (a,¢5)be — (ayby)ce by Eqs. (27c) and (27d).
But this is the component form of the right side of Eq. (29), which establishes
the required identity. A similar derivation shows that
(a Xb) X € = (aeyb — (bee)a, (2.3.30)
which is a completely different vector from a x (bX ¢). The vector product
therefore is not associative, so that the parentheses must not be omitted on
the left sides of Eqs. (29) and (30).
We have emphasized the physical or geometric character of a vector, which
exists independently of any coordinate system. However, most of our actual
computations are made with rectangular components, which require the intro-
duction of a coordinate system. We now turn our attention to the relationship
between the components of a vector or a tensor in one rectangular coordinate
system and the components in any other rectangular coordinate system.24 Vectors and Tensors Chap. 2
EXERCISES
1, Express in terms of I, j, k a unit vector parallel to a= 41 — 2) + 4k.
2, A vector of magnitude 100 is directed along the line from point A(10, ~5, 0) to
point B(9, 0, 24). Express the vector in terms of i, J, k.
3, Find: wey, u X v, and v Xu for
(a) u= 61 — 4)— 6k, v= 4i—2)— 8k
(b)u=3+ J—2k, v= —i+2)— 3k.
Find the projection of ¢ = 181 — 27) + 81k onto d, if
@) d= 1-24 2k, (b) d= 1-4 2h — 2k.
. Show that the vector Al + Bj + Ck is normat to the plane whose equation is
Ax + By + Cz =D. Hint: Show that it is perpendicular to the vector joining
any two points (x1, ¥1 21) and (xs, ¥2, 22) lying in the plane,
Find the component of the force f = 6001 -- 1200) — 900k pounds in the ditec-
tion of the normal to the plane 2x — 2y + 2 = 3, choosing the positive normal
to be on the opposite side of the plane from the origin,
. Find the equation of the locus of all points (x.,', 2) such that a vector from point
(2, =1, 4) to (x, », 2) Is perpendicular to the vector from (2, 1, 4) to (3, 3, 2).
»
a
2
Find the angle between the two vectors from the origin to the two points
A(A, 3, 2) and B(—2,
9. Find the area of the parallelogram formed on the two vectors of Ex. 8.
10. Find the unit vector perpendicular to both vectors of Ex. 8.
11, Find a unit vector parallel to the line x — z = y = 2z. Hint: The line is the in-
tersection of the two planes x — y — z = O and y — 2z=0.
12, New &, ‘axes are chosen so that the new X-axis makes equal angles with the
positive x- and y-axes and the negative z-axis, while the new j-axis lies in the x2~
plane and bisects the angle between the old x- and z-axes. Show that these new
5- and j-axes ate perpendicular, and find the direction cosines of the new 3-
axis for a right-handed system.
|. Write out the following expressions:
@u=Tan @ Ty =264,
(emen FE + pir = pa,
(©) W= Tye
=
14, Verify the identities of Eqs. (2.3.27).
15, Verify Eq. (2.3.20).
16. Verify that Eq. (2.3.23) gives the same results as Eq. (2.3.18), when Eq. (2.3.18)
is written with fy, fy, fy for J, J, k and ay for ae, etc.Sec. 2.4 Vectors and Tensors 25
17. Prove the identity of Eq. (2.3.30).
18. Prove the e-8 identity e1y €1rs = 85,51, — 84554, of Eq. (2.3.28). Hint: Two of
the four free indices jkrs must be equal. Show first that if » then
both sides vanish identically. [For example, eres) = 0, and Bay Buy. — SayeSuyr
= 0, where the parentheses around k are used to indicate that there is no sum
on k.] Then show that even with j/k and r % 5, both sides vanish when j =r
unless also & = 5, and then prove that
exer etirney = Soren Bexn — SinwySieny fore # k.
What cases remain to consider?
19. If vis any vector and fi is any unit vector, show that
v= (vA) fA x (¥ xO).
This gives a resolution of v into vector components parallel and perpendicular to
A.
20. Show that three concurrent vectors a, b, ¢ are coplanar if eyardybecr
2.4 Change of orthonormal basis (rotation of axes); tensors as linear
vector functions; rectany 7 conlponents; dyadics;
tensor properties; ary matrix concepts
Part 1. Change of Orthonormal Basis
A change of rectangular coordinate systems can be made by a translation of
the axes without rotation followed by a rigid-body rotation of the axes keeping
the origin fixed (and possibly followed by a reflection in a coordinate plane if
we allow transformations between right-handed and left-handed systems). The
Cartesian components of a vector are not changed by translation of axes, since
the projections of a vector onto parallel axes are the same in both magnitude
and sense. Hence we need to consider only transformations with fixed origin.
Rotation of axes. If new right-handed rectangular Cartesian ¥,, 1, ¥s-axes
are chosen with the same origin as the original right-handed x,, xs, Xs-aXxes, as
in Fig. 2.7, the direction of each new axis can be specified by giving its direc-
tion cosines referred to the old axes, denotedt aj = cos (%,x-). The subscript
refers to a new (barred) axis ¥,, while the superscript refers to the original
(unbarred) axis x. The direction cosines defining the change of axes are con-
veniently tabulated as in Table 2-1. The components of the new unit vector
Some authors use ays, and some use air, for the direction cosine. We denote it by
af to emphasize that it is not a Cartesian tensor component.26 Vectors and Tensors Chap. 2
Fig. 2.7 Rotation of axes.
1, for example, referred to the old axes, are found in the column below the
symbol {,.
‘TABLE 2-1
Direction Costnes FOR ROTATION OF AXES
foe
if ol al
“ha at
by ay ay ay
Each new unit vector J, is thus given as a linear combination of the old unit
vectors i, with coefficients listed in the column. Inversely, each old unit vec-
tor is a linear combination of the new unit vectors with coefficients given in a
row. Thust
{,= ati, and i, = atl, (2.4.1)
(vector sum on s in the first case and on r in the second). Until you learn that
the subscript refers to the barred system, you can write the cosines as a}.
Because of the orthogonality of the unit vectors we must have
il, =0 for pq, while {,-1, = 1.
This is conveniently expressed by introducing the Kronecker delta 8,9, defined
tin the curvilinearcomponent notation of App. I, the second Eq, (2.4.1) would be
written as {f= a#I’,so that s would appear as a superscript on both sides of the equa-
tion. But in Cartesian components the free indices need not appear on the same level,
and we repiace {# by I.Sec. 2.4 Vectors and Tensors 27
to be zero when p and q differ and to be unity when p = g. See Eq. (2.3.25).
7
2.4.2)
if pq. oe
wt
Then the fact that each basis (set of base vectors) is orthonormal, i.e., com-
posed of mutually orthogonal unit vectors can be expressed by the orthogonality
conditions
and in Ban (2.4.3a)
or
aha = 8s, and aPa? = Sun, (2.4.36)
where Eqs. (3b) follow from Eqs. (3a) by use of Eqs. (1), [Note that the left
side of each of the Eqs. (3b) is a sum of three terms because of the repeated
index s or r.] An additional identity involving the direction cosines follows
from the fact that the determinant det |a7| with elements a¥ in the mth row
and nth column is equal to the scalar triple product i, X iz+4s of the old unit
vectors in terms of their components referred to the new axes, according to
Eq. (2.3.20). But this triple product equals unity, the volume of the unit cube
whose edges are the three unit vectors 4, I, 15, provided the old axes are right-
handed. The triple product equals —1 if the old axes are left-handed. Hence
for rotation: det fan] = 1; (2.4.48)
while
for rotation and refiection: det ja] = —1. (2.4.4b)
A linear change of basis of the form of Eqs. (1) is called an orthogonal
transformation if the coefficients a{ satisfy Eqs. (3b). Every orthogonal trans-
formation satisfies one of the two Eqs. (4). If it satisfies Eq. (4a) [positive de-
terminant], it is called a proper orthogonal transformation; if the determinant
is negative, the transformation is improper. Hence rotation of axes is repre-
sented by a proper orthogonal transformation, while reflection is represented
by an improper orthogonal transformation. For most of our purposes we need
only consider proper orthogonal transformations, but occasionally an improper
transformation is useful. [See, for example, Eqs. (6.2.22).]
Polar vectors and axial vectors, transformation of Cartesian components.
When we consider how the rectangular Cartesian components of a vector
transform under the change of basis given above, and admit the possibility of28 Vectors and Tensors Chap. 2
improper orthogonal transformations, it is necessary to distinguish two kinds
of vectors called polar vectors and axial vectors. These vectors are distinguished
by the way their components change under an improper orthogonal trans-
formation. The axial vector transformation formula differs in having an extra
minus sign prefixed if the transformation is improper. No distinction is needed
if only proper orthogonal transformations are used. An example of an axial
vector would be furnished by the cross product, if we used the definition of
Eqs. (2.3.17) and (2.3.23) without prefixing a minus sign in a left-handed
system. But with the convention we adopted in Sec. 2.3, the components of
the cross product will transform as a polar vector.
The polar vector is the typical vector or true vector. Its component trans-
formation formula follows from the requirement that the representations of the
vector in the two systems have exactly the same form whether or not the
transformation is proper. Thus the vector equality
i, (2.4.5)
ol, =
must hold between the two representations, Substituting in Eq. (5) the second
Eg. (1) and collecting terms, we obtain
G, — vat)h, = 0.
Since all the components of the zero vector 0 must be zero, the expression in
parentheses must be zero for each r, whence we obtain the three equations
Polar Vector v
6 = atv, or inversely v, = até. (2.4.6a)
The inverse equations follow in a similar fashion from substituting the first
Eg. (1) on the left side of Eq. (5). We sce that the transformation equations
for the vector components (6a) have exactly the same form as the transforma-
tion equations for the base vectors (1). Despite the similarity in form, the two
transformation equations [(1) and (6a)] are quite different, since Eqs. (6a) give
the components of the same vector y in two different systems, while the vector
equations (1) pertain to two different sets of vectors, one set for each system.
Equations (6a) may be written in the following matrix form (see Part 4 of this
section for a review of matrix notation):
t=ATr v= Ab, (2.4.6b)
where A is the matrix shown in the array of Table 2-1, with element a} in row
sand column r, A” is the transpose of A, and § and v are column matricesSec. 2.4 Vectors and Tensors 29
with elements @, and v,. Note that in the matrix formulation the matrix A is
the matrix used for the inverse transformation from 6 to v. For the ortho-
gonal transformation, the inverse A~' of the matrix A is equal to its transpose
‘A’, In the indicial form, the second Eq. (6a) represents multiplication by the
matrix A, since the summation is on the column index r, while in the first Eq.
(6a) the summation is on the row index s.
An axial vector differs from a polar vector in that for an improper orthog-
onal transformation of an axial vector a minus sign must be introduced in the
defining Eq. (5), which becomes é,1, = —v,i,. Since the minus sign is not
needed for a proper orthogonal transformation, both cases may be included by
prefixing (det |a™|) in place of the minus sign. Thus
axial vector y: Gi, = (det |az|)o,
whence as in the derivation of Eqs. (6) we obtain the component transformations
Axial Vector ¥
6 = (det |az|are, —-v, = (det lar |)ate,; (2.4.7a)
or in matrix form
6 = (det ja"|)A7v v = (det |at|)A6. (2.4.76)
We have already noted that one example of an axial vector would be the
vector product of two polar vectors if we did not prefix a minus sign to Eqs.
(2.3.17) and (2.3.23) in a left-handed system. Another example would be the
curl of a vector field, to be discussed in Sec. 2.5, if we did not employ the same
convention about the minus sign as in the cross-product formulas. With our
conventions, all the vectors we encounter will transform as polar vectors. But
the reader should be warned that some writers do not use the minus sign in the
cross-product formulas; then the cross product is an axial vector.
Coordinate transformation formulas for the change of axes under rotation
and reflection are merely special cases of Eqs. (6) in which the vector v is the
position vector r with components , or x, in the two systems. If a translation
is added, the most general transformation of rectangular Cartesian coordinates
has the form
%, = ay(x, — 6,) or inversely x, = b, + ati, (2.4.8)
where the a! satisfy Eqs. (3b) and either (4a) or (4b).
The transformation property of rectangular components under rotation of
axes is sometimes made a part of the formal definition of a vector quantity as
follows. A physical entity, defined by three components in every system of
rectangular Cartesian coordinates, is a vector quantity if30 Vectors and Tensors Chap. 2
1, the three components in any one system are related to the three com-
ponents in any other (meaning every other) system by Eqs. (6), and if
2. two such physical entities of the same type, say a and b with compo-
nents a, and 5, in one coordinate system, together produce the same
effect as would be produced by the single one, c with components c, =
a, + 6, in the same system.
In short, to be a vector quantity, the physical entity must transform like a
vector and compound like a vector. We remark that it is possible to invent
physical entities with three components not satisfying these two conditions.
We might take the triplet (p, v, 1) representing the pressure, specific volume,
and temperature at a point in a gas. These values do not transform as the
components of a vector under rotation of axes; each one maintains its value,
ie., each is a scalar point function under rotation of axes. We could always
construct a vector artificially by specifying its components in one system of
axes and prescribing that its components in any other system should be given
by the vector transformation formulas, When we are dealing with physical
entities, however, we do not have this much freedom of definition, as is shown
by the absurd example (p, v, 1).
Part 2. Second-Order Tensors as Linear Vector Functions
(Transformations)
A vector function is a correspondence which associates with each argument
vector v in a certain set of vectors (the domain of definition of the function)
another vector, which is the value of the function. The function is also called
a transformation, since it transforms one set of vectors (the domain) into
another set of vectors. A vector function F is a linear vector function if it
satisfies the two requirements that for any vectors u and y in the domain and
any real number c,
F(u + v) = Fu) + F(v) and F(cu) = cF(u), (2.4.9)
For the purposes of this book we define a second-order tensor (also called a
second-rank tensor) to be a linear vector function. A vector is a first-order
tensor, while a scalar is a tensor of order zero. Higher-order tensors will also
be defined later, but the second-order tensors are the ones of primary interest
in this book. For brevity, we will often use the word tensor to mean second-
order tensor. Since the two sets of vectors have an existence quite independent
of any coordinate system in which they might happen to be observed, the ten-
sor correspondence or pairing of each antecedent element in the domain with
its image element in the second set must also have an existence independent of
any coordinate system. In order to make explicit the way in which any partic-See. 24 Vectors and Tensors 3M
ular tensor operates on the argument vector, however, we will introduce a
coordinate system. For simplicity, we limit ourselves for the time being to
rectangular Cartesian coordinate systems, but the tensor itself is not limited to
any coordinate system. We introduce a notation for the second-order tensor
itself (as distinct from any possible component representation). A boldface
capital letter symbol will be used in printed text for a second-order tensor,
while a boldface lower-case letter will identify a vector. Some exceptions to
this distinction between capitals and lower-case symbols may be made with
commonly used tensors or vectors whose physical nature will help us to re-
member whether they are vectors or second-order tensors. In typed or hand-
written text, the second-order tensor will be identified by wavy underscores,
eg, TorE. In the following, the word “tensor” will mean “second-order
tensor” unless otherwise stated,
Linear transformations of rectangular Cartesian vector components, tensor
components. The second Eg. (9) requires that the zero vector be transformed
into the zero vector (put ¢ = 0 to see this). In any one rectangular Cartesian
system, the most general linear transformation of the components v, of the
argument vector v into the components u, of the image vector u, which always
transforms the zero vector into the zero vector, has the form
% =Tyv, with matrixt notation u = Tv (2.4.10a)
or the form, for S = TT and T = S*,
4 = Sy, with matrix notation uv = oS. (2.4.10b)
If Eqs. (10b) always give the same output vector components as are given by
(10a) for an arbitrary input vector v, then the two forms are not fundamentally
different, since they differ only in how the same nine coefficients are labeled:
‘Sy, in (b) being equal to 7,, in (a), ie., in the matrix notation the matrix Sin
the second equation is the transpose of the one which should be used in the
first equation. Each of the two tensors T and S with rectangular Cartesian
component matrices T and S = T” is said to be the tensor transpose of the
other, denoted
S=T and T=S". (2.4.10c)
When we do not plan to write both forms for the same tensor, we let T denote
the matrix of components in the form which we actually use. The nine coef-
ficients are the rectangular Cartesian components of the tensor T. They are
denoted by the same letter printed in ordinary type instead of boldface (without
tA brief review of some elementary matrix properties is given in Part 4 of this
section.32 Vectors and Tensors Chap. 2
the wavy underscore in handwritten text) and with two subscripts attached,
eg., Ti Tr, etc. Some exceptions will be made to the rule that components
are denoted with the same letter as the tensor. For example, the strain tensor
is often denoted by E with components e,;. The linear transformation of Eq.
(10) is the representation of the tensor T (linear vector function) in one system
of Cartesian components. This transformation of the components of one vec-
tor into the components of another vector referred to the same system should
not be confused with the change of components of one vector under rotation
of axes, given by Eq. (6a).
Symmetric tensors and skew or antisymmetric tensors. A tensor is said to
be symmetric, if its matrix of rectangular Cartesian components is symmetric,
ie, if Ty, = Ty). If the matrix of a tensor is symmetric in one Cartesian coor-
dinate system, it is symmetric in all such systems; thus symmetry is really a
tensor property. A skew or antisymmetric tensor has T,, = —T,, in any
Cartesian system, which implies that the diagonal elements, 7, Tr, T'ss are all
zero. We will now derive the rule for change of components of a tensor under
rotation of axes.
Rotation of axes, change of tensor components. The rule for changing
second-order tensor components under rotation of axes follows from the rule
for changing vector components when the transformation (10) is required to
apply in each of the two systems. For example, Eq. (62) yields
a = alu,
aiT,% by Eq. (108)
=alT,,a$6, by Eq. (6a).
Also ii, = T,,8, by Eq. (10a) in the barred system. By equating these two
different expressions for i, and by collecting terms on one side of the equation,
we conclude that
(Tip — ala$T,)8, = 0
for every possible choice of the argument vector vy. Hence the quantity in
parentheses on the left must vanish, whence
Tip = alagT,, and inversely T,, = alatT,,, (2.4.11)
or in matrix notation
To=ATA T= ATA, (2.4.12)
where for the proper orthogonal transformation A~! = A’, The inverse relationSec. 24 Vectors and Tensors 33
can be derived in a way similar to the derivation of the first Eq. (11). Com-
paring these formulas for the change of components of a second-order tensor
with Eqs. (6a) for a polar vector (first-order tensor), we see that they follow a
similar pattern. There is on the right side one free index on each direction
cosine, corresponding to the free indices on the left, and each direction cosine
also has a summation index corresponding to one on a tensor component on
the right. Each of the nine first equations (11) has a sum of nine terms on the
right, corresponding to the nine possible values of the pair of indices jg. For
example,
Fay = alalTy, + abalTy + alah Ty,
+ dlaiTy + aah + atai Ts (2.4.13)
+ ataiTy, + alaiTs: + atasT as.
‘We remark that exactly the same change-of-component formulas apply
whether the components are taken in the form of Eq. (10a) or of Eq. (10b),
provided the same choice is made in both coordinate systems.
These change-of-component formulas are sometimes made the basis for the
definition of a “Cartesian tensor.” Briefly, a second-order Cartesian tensor
would be characterized as an entity with nine components in any Cartesian
coordinate system, such that the components in any one system are related to
the components in any other system by Eqs. (11) and (12). Higher-order ten-
sors can be similarly defined, for example, with three subscripts T,,., and with
change-of-axis formulas similar to Eq. (11) but with three direction cosines in
each term. We will be concerned primarily with vectors and second-order
tensors. For our purposes it is more convenient to define the second-order
tensor as @ linear vector function and to deduce that Eqs. (11) and (12) relate
its components in any two rectangular Cartesian systems. We postpone until
the appendix any consideration of how the tensor would be represented by
components in a curvilinear coordinate system, but our basic definition will
still apply, since it is not dependent on any coordinate system.
Dyadic (single-dot) product notation for a second-order tensor. Many of
the general equations containing tensors have a simpler appearance when writ-
ten in terms of the symbolic representation T instead of in terms of the com-
ponents, We use the Gibbs dyadic notation (see Part 3 of this section) to give
a symbolic representation independent of any choice of coordinate system for
the linear vector operation represented in rectangular Cartesians by Eqs. (10).
Dyadic Notation for Tensors as Linear Vector Operators
u=Tey
usvS — whereS = 7
(2.4.14a)
(2.4.14b)34 Vectors and Tensors Chap. 2
The symbolic representation T+v is not merely another notation for the matrix
product Tv of Eq. (10a). T+v symbolizes an operation pairing one physical or
geometric vector with another one, while Ty is the mathematical product of
appropriate matrices of components of T and of v in one coordinate system. In
general curvilinear tensor components some care is necessary in evaluating T+v
as a matrix product [see App. I, Eq. (1.2.14)]. For example, the matrix prod-
uct of two matrices of covariant components is not in general equal to the
matrix of covariant components of u. The covariant component matrix of o
can be obtained as the matrix product of the covariant components of T by
the contravariant components of y instead of by the covariant components of
vy. In rectangular coordinates there is no difference between the different kinds
of components, and the matrix forms of Eqs. (10) can always be used to rep-
resent in any one coordinate system the operation symbolized by Eqs. (14).
Some authorst use a symbolic operational product notation without the
dot. For example the relationship we have expressed by Eq. (14a) would be
written
=Ty, (2.4.15)
an operation completely independent of any component representation.
Some properties of second-order tensors. The sum S = T + U of two
tensors is the single tensor which operates on any vector v to yield the sum of
the two vectors obtained by operating with T and U separately; in rectangular
Cartesian components, S;; = T,, + U,,. The zero tensor 0 assigns the zero
vector as image to every argument vector y; its components are all zero in
every coordinate system. Multiplication by a scalar c is defined, for example,
by the requirement that (cT)-v = c(T+y) for every vector vy. The rectangular
Cartesian components of cT = Te are cT,,=T jc. Tensor addition and
multiplication by a scalar obey the following four addition axioms and four
scalar-multiple axioms, characteristic of a generalized vector space; hence the
set of all tensors is a vector space (see Sec. I.1).
ADDITION AXIOMS
@) T+U=U4+T commutative
(ob) T+(U+V)=(T+U)+V associative
© T+0=T
(4) For each T, there exists another, called —T, such that
T+(-T)=0.
(2.4.16)
See, for example, Truesdell and Noll (1965) and Jaunzemis (1967).Sec.24 Vectors and Tensors 35
SCALAR-MULTIPLE AXIOMS
(Note aT = Ta)
(a) a(bT) = (ab)T
@) IT=T
(©) (a+ dT =aT +06
@) a&T + U)=aT+aU
(2.4.17)
The scalar product of two tensors is a scalar, denoted T : U, which can be
calculated in terms of components of the two tensors in any one rectangular
Cartesian system, by
T:U=TyUy (2.4.18a)
ie. by the sum of the nine products of components. This formula applies only
to rectangular Cartesian components, but the result, a scalar, is independent of
any coordinate system. Some authors define the scalar product as T,,U,, in-
stead of T,,U,, and some use the notation T++U instead of T:U. In this
book, T: U will always mean 7,,U,, in rectangular Cartesians, while
THU = TyUy (2.4.18b)
will represent another possible scalar product, different from T : U unless one
of the two tensors T or U is symmetric. The scalar products satisfy the fol-
lowing four axioms, which with (16) and (17) characterize an inner-product
space. See also Eq. (25b) below, where the scalar products are expressed as
traces of the product of two tensors.
Inner-PRODUCT AXIOMS
@) T:U=U:T
(b) T3U+V)=T:U4¢T:V
© &T:U)=(@T):U =T: (aU)
@) T:T>0 unless T=0
(2.4.19)
The identity tensor, also called the unit tensor and denoted by 1, is de-
fined by the requirement that
levovl=y (2.4.20)
for all vectors y. The rectangular Cartesian components of 1 are 8,,; its matrix
of rectangular Cartesian components is the identity matrix with zero in each
off-diagonal position and one in each diagonal position.36 Vectors and Tensors Chap. 2
The tensor product or open product of two vectors, denoted ab, is a tensor
called a dyad, defined by the requirement that
(ab)-7 = a(b-v) (24.218)
for all vectors vy. That is, if T = ab, then T-v = a(b-y) for all vectors v; in
rectangular Cartesian components
Ty = aby. (2.4.21b)
Some writers use the notation 2 © b instead of ab.
The product of two second-order tensors, denoted T+ U in this book, means
the composition of the two operations T and U, with U performed first, de-
fined by the requirement that
(T-U)-y = T-(U-y) (2.4.22a)
for all vectors v. If
TU (2.4,22b)
then Pi,
Ur or
in matrix notation, where P, 7, and U are the matrices of components in any
one rectangular Cartesian system. Some authors omit the dot and write the
tensor product as -
P=TU. (2.4.22c)
The product satisfies the linearity rules and is therefore also a second-order
tensor. It also satisfies the following five axioms, which together with (16) and
(17) characterize the set of all second-order tensors as an algebra. Similar def-
initions and axioms apply to ve(U-T) = (v-U)-T.
TENSOR ALGEBRA AXIOMS
(a) (T+U)*R = T-(U-R)
(b) T(R+U)=TeR4T-U
© (R+U)-TH=R-T+U-T (2.4.23)
@) ATU) = (aT)-U = T-(aU)
© PT=Te1=T
The product is not commutative; in general U-T # T-U. Instead, Eq. (49)
below, when applied to the matrices of rectangular Cartesian components,
shows thatSec. 24 Vectors and Tensors 37
BeA = [AT-BT}r (2.4.24)
where the superscript T denotes the transpose as in Eqs. (10). In exceptional
cases, when T-U = U-T, we say that T and U commute. For example, the
unit tensor 1 commutes with any other tensor.
The trace, denoted trT, of a second-order tensor T is a scalar invariant
function of the tensor, having the same numerical value in all coordinates
systems. In rectangular Cartesians (see Sec. 3.3)
uT=Tn. (2.4.25)
Thus tr T is equal to the trace of the matrix of rectangular Cartesian com-
ponents, the sum of the elements on the main diagonal.
The scalar products of Eqs. (18) can be represented as traces of tensor
products. It is left es an exercise for the student to show that
As+B = tr(A-B)
A:B = tr(A-B’) = tr(A7-B)
tr(A-B) = tr(B-A)
tr (A+B-C) = tr (B+C+A) = tr (C+A-B)
(2.4.25b)
The last line demonstrates the cyclic property of the trace of the continued
tensor product of any number of factors; the trace is unaltered in value by any
cyclic rearrangement of the factors.
A test for tensor character. An entity T defined by nine components in
any coordinate system is a tensor, if and only if T:(av) is a scalar invariant
for arbitrary vectors u and y, where the symbol “:” denotes the scalar product
of Eq. (18a), while uv denotes the tensor product of Eq. (21). This means that
in rectangular Cartesian components
Tesi, = Toate (2.4.26)
in every pair of Cartesian systems, for every choice of the vectors a and v. If
the components being tested satisfy the conditions for symmetry of a tensor in
all coordinate systems, it is sufficient to show that
T: (vv) is a scalar invariant for arbitrary v and symmetric T, (2.4.27)
which means that
F,0,6, =Tyevpr, and Ty =Ty Ty = Te (2.4.28)
in every pair of Cartesian systems for arbitrary choice of y.38 Vectors and Tensors Chap. 2
Part 3. Algebra of Dyads and Polyads; Dyadies
In Eq. (21), an operational definition was given for a dyad or open product of
two vectors. Higher-order polyads can be formed by open products of more
than two vectors, .g., triads abe or tetrads abed. Without assigning any phys-
ical significance to these polyads for the moment, we postulate that open
multiplication obeys all the rules of ordinary algebra except that open multi-
plication is not commutative.
ab ¥ ba. (2.4.29a)
We define the conjugate dyad (ab), to ab by
(ab). = ba. (2.4.29)
A linear combination of dyads with real coefficients is called a dyadic, e.g.,
mab + ned.
Contraction: single-dot product, double-dot (scalar) product of dyads. If
in a polyad we replace one of the open products by a scalar product of the
two vectors, we obtain a polyad of order two less than that of the original
polyad. This process is called contraction. For example, with the tetrad abed,
contraction can be done in three ways with adjacent vectors to yield three dif-
ferent dyadics:
a-bed = (a+b)cd
ab-cd = (b-c)ad (2.4.30)
abe+d = (c-d)ab,
where the dot product in parentheses is a scalar factor multiplying the dyad.
The second of these contractions defines the single-dot product of the dyad
ab by the dyad ed. Note that
ab-ed % cd-ab (2.4.31)
s0 that (unlike the dot product of two vectors) the single-dot product of two
dyads is not commutative.
The scalar product of two dyads (or donble-dot product) denoted ab: cd
is defined as the scalar obtained by multiplying together the two scalar products
a-cand bed, Note that the first vector of the first dyad multiplies the first
vector of the second dyad, and the second vector of the first dyad multiplies
the second vector of the second dyad. The scalar product is comnmtative.
ab: ed = (are)(bed) = (c+a)(d-b) = ed : ab. (2.4.32)Sec, 2.4 Vectors and Tensors 39
Some authors denote the scalar product of Eq. (32) by ab ++ cd instead of
ab: cd. Also, some authors define the scalar product differently, multiplying
the two outside vectors together and the two inside vectors together. In this
book, the double-dot notation with the two dots on the same line will denote
the product obtained by multiplying the two outside vectors together and the
two inside vectors together.
ab ++ cd = (a-d)(bee), (2.4.33)
Dyadics as second-order tensors. Since the single-dot product of a dyad
ab by @ vector y yields another vector, as in Eq. (21), and since this operation
is a linear vector function, a dyad operates as a second-order tensor. Is the
converse true? Can every second-order tensor be represented as a dyad? The
answer is no, but every second-order tensor can be represented as a dyadic, a
linear combination of the dyads formed with the base vectors. Thus, in rec-
tangular Cartesians,
T= Tabb + Tabs + Tits
+ Trbh + Tabl + Tabls
+ Tybt) + Tatsis + Trisis or
T=T,,1,4,.
(2.4.34)
‘The algebra of dyads and polyads, as outlined in Eqs. (29) through (33) can
then be shown to imply the results for tensors given earlier, when the tensors
are written in the form of Eq. (34), Moreover, if the change of basis Eqs. (1)
are substituted into Eq. (34), and the tensor is required to have the same form
in the new coordinate system, i.e.,
Trl, =T = Tall, (2.4.35)
it can be shown that this implies that the tensor component transformation
formulas of Eq. (I!) must hold between the coefficients 7,, and 7,,. Thus
the coefficients of the dyadic representation of Eq. (34) ate simply the rectan-
gular Cartesian components of the tensor.
Section 1.3 considers dyadic representation with an arbitrary basis b, instead
of the orthonormat basis },. The dyadic representation is especiatly convenient
for introducing derivatives of tensor fields, In Cartesian coordinates, since i,
is independent of position, 2T/2x, = (3T;,/2x,)i,1,, but in curvilinear coor-
dinates the unit base vector orientations vary with position. Thus in curvilinear
coordinates with unit base vectors @, it is necessary to differentiate each term
of the dyadic representation
T=7"8,8, (2.4.36)40 Vectors and Tensors Chap. 2
as a product of three functions. For orthogonal curvilinear coordinates this is
discussed in App. II, which avoids the general tensor complications of App. I.
(It is not necessary to read either appendix before proceeding with the text.)
Part 4. Review of Elementary Matrix Concepts
Definition of a matrix. A matrix is a rectangular array of elements. We
will consider here only matrices whose elements are real numbers.
Ay Aw Aw
A= (2.4.37,
Fe Ay a] ( ;
Ay, is the element in row ¢ and column j. (We will sometimes denote the element
Aj.) Two matrices are equal only if their elements are equal term by term. Un-
like a determinant, a matrix does not have a numerical value. Also it may not
have the same number of rows and columns. An m x n matrix has m rows and
n columns. The example above is a 2 x 3 matrix. Other common notations
are || lor ( ) instead of [ ].
Addition of matrices, Matrices of the same order add term by term.
Aut Bu Aw+ Be A+ By
A+B= 2.4.38)
7 [i + By Ant Ba An + | : ’
Multiplication by a numerical factor. Multiply every term.
kAy kKAy kA
kA= 2.4.39)
‘An kAn | ; y
Multiplication of two matrices. Matrices must be conformable or else they
cannot be multiplied. Matrices A and B are conformable and can be multiplied
in the order AB if the first’ matrix has the same number of columns as the
second matrix has rows. For example,
iccaae :
: ah and ap
are conformable in the order AB, but not in the order BA. The product of
the 2 x 3 matrix A by the 3 x 1 matrix Bisa 2 x 1 matrix C.Sec. 2.4 Vectors and Teasors 41
2-1 3°
|
an i ae [o + (14) + ree _ [il =c
(4)(6) + (0\(4) + 208)! ~ Lad
The element in the first row and first column of C is obtained as the inner
product (like a vector scalar product) of the first row of A by the first col-
umn of B, as indicated above. The element in the second row and first column
of the product matrix C is similarly obtained from the inner product of the
second row by the first column,
Formulas for the elements of the product matrix. Consider the product
AB, schematically shown as
AB = [m x nj[n x s}=[m x sJ=C.
The two matrices are conformable, since the first matrix has the same
number of columns as the second matrix has rows, namely n. The element C,,
in the ith row and jth column of the product is
Cy = 4uBy (summed). (2.4.40)
If s = m, it is possible to multiply in the opposite order, and
BA = [Bu Ags) (summed). (2.4.41)
In general BA + AB.
Square matrices and linear transformations. A square matrix has the
same number of rows as columns. The system of equations
Mux + Mix + Misxs = i
Max, + Maxi + Maxi =r (2.4.42)
Myx, + Max, 4 Masxa = Ys
may be abbreviated as
Mx=y, (2.4.43)
where M is the square matrix of coefficients and x and y represent the two
column matrices
xelml y= (2.4.44)42 Vectors and Tensors Chap. 2
Equations (42) and (43) are said to transform the column matrix x into the
column matrix y, This isan example of a linear transformation. Solving (43)
consists of finding the column matrix x transformed into a given column matrix
y by the given square matrix M, The example illustrated is for a 3 x 3 system,
but the ideas apply for any n x n system.
A diagonal matrix is a square matrix with nonzero elements only on the
main diagonal Mj,, Mz, Mss,.... If Mis diagonal, the system (42) may be
solved immediately, one equation at a time. When M is not diagonal, we may
try to diagonalize it by performing on it a succession of operations equivalent
to the kind of operations performed in solving (42) by elimination.
An identity matrix, denoted by I, is a diagonal matrix with each diagonal
element equal to unity, The elements of the identity matrix are 8,,. An iden-
tity matrix transforms a column vector into itself (the identity transformation).
Ix = x, (2.4.45)
ilarly for any row matrix x, i.c., any matrix with only one row,
xt = x. (2.4.46)
The inverse of a square matrix M (if it has an inverse) is denoted by M=*
and defined by
MM"'= M'M =I. (2.4.47)
Multiplying (43) by M7? (from the left), we get
M-'Mx = M-ty
or
Ik = M"y
whence
x=Mty, (2.4.48)
Thus, finding the inverse of M is equivalent to solving the system of equations,
which can be done provided the determinant is nonzero. A square matrix
with a nonzero determinant is called nonsingular. Thus, every nonsingular
square matrix has an inverse. A formula for the inverse is given in Eq. (51).
The transpose M? of a square matrix M is the matrix obtained by inter-
changing rows and columns. For example,
My My Ms
M'=|My Mi Mss}.
Mia Mi MasSee. 2.4 Vectors and Tensors 43
A symmetric matrix is a square matrix with equal elements in the positions
symmetrically placed with respect to the main diagonal, i.e, with M,,= My
so that M” = M. A skew or antisymmetric matrix has My = —M,, so that
M? = —M, which implies that the diagonal elements are all zero. For square
matrices A and B
A = (AT BY, (2.4.49)
The proof of Eq. (49) is left as an exercise.
Formula for the inyerse. For a nonsingular matrix M, if
D=deM
is the determinant of the matrix, the solution of Eq. (42) by Cramer's rule
gives, for example,
yi Ma Mis ie Wie Ma M,
=4ln Ma Mal = (M" "| - aie Hag y [Ma ‘|
Mee Mi Mss Mix Mul Ma Mas}
or
- ee + Dy, = Diy
where D™® is the cofactor (signed minor subdeterminant) of the element Mun
in the matrix M. Thus, the solution of Eqs. (42) is
= a Wn (2.4.50a)
or, in matrix form,
(2.4,50b)
Hence the inverse matrix M~! of Eq, (48) is given by
a to T : pe
Mt = [ D with element D (2.4.51)
in row n and column m, where D™* is the cofactor of Mna. This method of
finding the inverse is prohibitively complicated for a large matrix, For numer-
ical methods of inverting a large matrix, see a text on numerical analysis.44 Vectors and Tensors Chap. 2
Similar matrices. If M and N are two n x 1 matrices, then N is said to
be similar to M if there exists a nonsingular matrix A such that
N= A"'MA (2.4.52)
Then M is also similar to N, since
M = ANA" =(A")'N(A').
Orthogonal matrix. A matrix @ is said to be orthogonal if
QFQ= QO" =I whence Q" =". (2.4.53)
Equations (3b) show that the matrix A of direction cosines for change of rec-
tangular Cartesian axes is orthogonal. Hence the matrices 7 and T of rec-
tangular Cartesian components of a tensor T are similar matrices under the
class of orthogonal transformation matrices, according to Eqs. (12) and (52).
Eigenvectors and elgenvalues of a matrix. The transformation y = Mx of
Eqs. (42) and (43) may be interpreted as associating to each point P(x1, xs, Xs)
another point Q()1, Ys, ys). Since the transformation is homogeneous, it as-
sociates to any other point (rx,,rx2,rxs) on the line OP another point
(ry ryn rys) on the line OQ, and we may consider the transformation to be @
transformation of the line OP into the line OQ. We may ask whether there
are any lines not changed by this transformation, i.e., any lines transformed
into themselves so that (x,,x;,%s) transforms into (Ax, xs, A%3) for some
scalar X. Any line transformed into itself is called an eigenvector of the matrix.
(Other names for this are proper vector, characteristic vector, or latent vector.)
If there is such an eigenvector x, then y = Xx and
Mx = dx = Alx
or
(M—A)x =0, (2.4.54)
and the eigenvector is given by the solution of the set of homogeneous alge-
braic equations (54). A nontrivial solution will exist if and only if the deter-
minant vanishes:
IM — a] =0. (2.4.55)
When the 3 x 3 determinant is expanded, Eq. (55) is a polynomial equation,
a cubic in three dimensions, whose roots 4,4, are called the eigenvalues
(proper numbers, characteristic values, or latent roots) of the matrix. When »Sec. 2.4 Vectors and Tensors 45
is put equal to one of these cigenvalues, Eq. (54) can be solved for the x,
which are determined only up to a scalar multiple. They are determined up
toa factor of +1, if we require in addition that the x, be the components of
a unit vector. The whole discussion can be generalized to n dimensions.
A real symmetric matrix has only real eigenvalues. For, if there were 2
complex root & of the polynomial equation (54) (with real coefficients), then
Mx = dx, whence MX = XX (where overbars denote complex conjugates).
Multiplying the first of these equations by 37 and the second by x”, we have
S°Mx=d8"x and xTMF = aks. (2.4.56)
But 37x = x78 = x18) + %y-+ +++ + x48; and, since M is symmetric
x" MZ = 3° Mx, Hence, subtracting the two Eqs. (56) we find
0.
(A — )3
Since x is nontrivial #7x + 0, and it follows that
N=d (2.4.57)
so that X must be real. It is also known that when no two of the eigenvalues
are equal, the eigenvectors are uniquely determined up to a scalar multiple of
each vector, so that the three lines in space (one associated with each cigen-
value) are uniquely determined. Furthermore, if any two eigenvalues are
distinct, say 4, % A», then the two associated eigenvectors must be orthogonal.
This will be demonstrated in Chap. 3 for the particular example where the
eigenvectors are the principal axes of the stress tensor. If all three eigenvalues
are equal, then every direction is an eigenvector; in this case the eigenvectors
are not unique, but it is still possible to choose a set of three mutually orthog-
onal eigenvectors. If two are equal, say Ax = Ay # Ay, then the eigenvector
associated A, is unique (up to a scalar multiple) and any direction orthog-
onal to it is an eigenvector, so that the other two are not uniquely determined.
In any case, it is possible to choose a set of mutually orthogonal eigenvectors.
If the three orthogonal eigenvectors are normalized to be unit vectors, then
the square matrix L whose columns are the column matrices of the orthonormal
eigenvectors will be an orthogonal matrix, such that
D=L'ML (2.4.58)
is a diagonal matrix whose elements are the eigenvalues. Thus every real sym-
metric matrix M is similar to a diagonal matrix whose elements are the eigen-
values, and the transformation matrix L required for the similarity relation
is an orthogonal matrix whose columns are a set of mutually orthonormal
eigenvectors of M.46 Vectors and Tensors Chap. 2
This completes our introduction to tensor algebra. Some additional tensor
properties having to do with principal axes of tensors and the algebraic
invariants of a tensor will be presented in Sec. 3.3 in connection with the
development of the stress tensor. Although developed specifically for the stress
tensor, these properties are also relevant to other tensors. Also presented in
Sec, 3.3 will be the separation of a tensor into the sum of a spherical and a
deviatoric tensor. And in Sec, 4.6 the concept of the inverse of a tensor and
of positive, negative, and fractional powers of a tensor will be presented. The
differential calculus of vectors and tensors, which will be needed in our study
of strain and deformation, is considered in Sec. 2.5.
EXERCISES
1, New right-handed coordinate axes are chosen with new origin at (4,1, —2)
and with Fy = (2h - le + i:)/3 and Ty = (51 — ts)/W2-
(a) Express [, in terms of the ip.
(b) If t = 101, + 10h, — 20K, express { in terms of the new basis,
{c) Express the old coordinates x1, x2, xs in terms of ¥1 51, 25.
. Complete the following table of direction cosines for rotation of axes. If
S2 = Sy =f, = 1000 1b, find the components of f in the barred system.
g 3
x 0 i
i: 1 0
. Verify that Eq. (2.4.66) gives the same results as Eq. (6a).
Carry out the details of deriving the inverse transformation of the second Eq.
(2.4.6a).
Derive the first Eq, (2.4.72).
If (ab)-v means a(b+¥), show that its rectangular Cartesian components are the
same as those of T+¥, where T is the tensor with components aib,.
. Write out the nine-term sum of the first Eq. (2.4.11) for Ts1, for Tas.
AY
ae
23
. Rotated Zm-axes are chosen making angles with the x,-axes as shown in the
following table.
g 5 z
x 90° 45° 135°Sec. 2.4 Vectors and Tensors 47
(a) Verify that the new axes are a right-handed orthogonal system, and display
the matrix A of direction cosines.
(b) If tensor T has the components shown in the matrix below, referred to the
raxes, compute the matrix of fm-components of T.
2-2 0
2 6
ccc:
). Derive the second Eq. (2.4.11).
10, Show that the first Eq. (2.4.12) is equivalent to the first Eq. (11).
11, Show that the two scalar products T:U and T ++ U, defined in Eqs. (2.4.18),
are equal if U is eymmetric,
12, Show that T: W = 0, if T is symmetric and W is skew.
13, Show by use of Eqs. (2.4.6a) that the test of Eq. (26) implies the first transforma-
tion of Eq. (11).
14, Show by use of Eqs. (2.4.6a) that the test of Eq. (28) implies the first Eq. (11).
15, Show that if Eq. (2.4.35) holds for arbitrary rotations of axes, then the Tp, are
elated to the Trs by Eq. (11).
16. Find the coefficients af for the rotation of axes, if the new %,-axes are obtained
by a counterclockwise rotation through angle @ about the old xs-axis.
17. Find the rotation matrix A for the following coordinate changes: (a) rotate 180°
about x-axis, (b) rotate 45° about x,-axis, (c) rotate 90° about £1-axis and then.
rotate 90° about the position assumed by the x2-axis after the first rotation.
18. Determine the product matrices AB and BA, if A and B are the matrices shown.
' 000
A=|0 00 B=|0 00
10 0,
19. Multiply 8, if
12
fo 1
45[ | B=|3 0
24-1
2-1
Can you multiply as BA?
20, Multiply AB and BC,
ma -12 6 4
A=([} —-b 4) B=/-12. 90 c=|-3|.
6 0 6. +48 Vectors and Tensors Chap. 2
21, Verify that, if
321 ot
A=|4 65 then At=| ys —$ 48:
oo 4 oie te
22, Show that if u,v, w are orthonormal eigenvectors of matrix M, corresponding to
eigenvalues 21, 21, As, respectively, and L is the square matrix whose columns
are u,v, w, then L'(ML) is a diagonal matrix, as asserted in Eq. (2.4.58).
23, and 24, Determine the eigenvalues and eigenvectors of the matrices shown
below, and verify that the orthogonal matrix L of Eq. (2. 4. 58) diagonalizes the
given matrix in each case.
uw4o0 70 -2
23. 430 24, 05 0
oe ee
25. Verify Eq. (2.4.49).
26. Verify the first two Eqs. (2.4.25b) for A ++ Band A:B.
27. Verify the last two lines of Eqs. (2. 4. 25b).
2.5 Vector and tensor calculus; differentiation; gradient, divergence,
and curlt
The preceding sections have dealt with the algebra of vectors at a point. The
physical entities defined as vectors, in general, vary from point to point and
with time. The vector quantity is then a vector field or vector function, Study
of the rate of change of such a vector function leads naturally to the study of
derivatives with respect to space coordinates. A brief summary of vector dif-
ferential calculus is given in this section; for more details and proofs see a
textbook on vector analysis. Some integral theorems will be given in Sec. 5.1,
since these are useful in deriving the general principles of continuum mechanics
introduced there, We introduce the total time derivative of a vector function
of ¢ by an example, the differentiation of the position vector of a moying
particle.
If p(e) is the position vector at time # of a particle (material point), then
its velocity y is given by dp/dt, which is defined in the same way as the de-
tivative of a scalar function, namely
This section may be postponed until after Chap. 3, since it will not be needed until
the development of the concepts of strain and deformation in Chap. 4.See. 2.5 Vectors and Tensors 49
dp — tim AP (2.5.1)
It ato
If the particle moves on a smooth path, as in Fig. 2.8, then
Wal as (2.8.2)
Fig, 2.8 Differentiation of position vector p.
and in the limit, assuming that |Ap|/As approaches unity and that the direc-
tion of the secant PQ approaches that of the tangent to the path at P
dp ds 4 (2.5.3)
where @, is the unit vector tangent to the path, For the evaluation of the
derivative, if we express the Cartesian coordinates of the particle in a “fixed”
system of reference as functions of t,
p= x()i + (Oj + 20k, (2.5.4)
we obtain
dp
dt
dx, dy) dzy_ y
Git alta = Simba (2.5.5a)
since i, j,k do not change with time. Inversely the time integral is50 Vectors and Tensors Chap. 2
‘ ‘ 4
p—pe=fivde =if} tadt + If) oat + kf dt. (2.5.5b)
The time derivative of any other vector a is evaluated in the same way,
dia _ dag 4 day y 4 dos y
a dey 5 Mie 4 Mek = daly (25.6)
and a similar interpretation to that of Fig. 2.8 can be made if we plot succes-
sive values of a in a space with coordinates (a,, ay, 4,).
It is easy to show that the derivatives of a product of a scalar by a vector,
of the dot product of two vectors, of the cross product of two vectors, and of
the tensor products all obey the usual calculus rule for the derivative of a
product. Proofs will not be given here; the essential ingredients for the proofs
are the distributive character of the products considered and certain limit
theorems—such as that the limit of a product is the product of the limits when
both limits exist. Under the assumption that the vector functions considered
have continuous derivatives, the whole algebra of derivatives of scalar functions
carries over to the derivatives of vector functions. In applying the product
tule to the cross product of two vectors, however, it it necessary to keep the
two factors always in the same order, since the cross product is not commuta-
tive, Thus
dx yaux ts Bx y, 25.7)
and it would be incorrect to write the last term as v x(du/dt) in analogy with
the way the scalar formula is often written.
If A and B in Fig, 2.9 are two points on a rigid body with instantaneous
angular velocity w, the relative velocity va), of B relative to A is given entirely
by the rotation, Thus, the magnitude
Waal = 42 cB] = 4 legs sin 6, (2.5.8)
and the direction of vg. is tangent to the circle of radius CB in a plane
perpendicular to the axis of instantaneous relative rotation drawn through A.
Hence
Vaya =O X Faye. (2.59)
If any other vector of constant magnitude, say a, is fixed to a rigid body, then
it follows also that
a =axa, (2.5.10)Sec. 2.5 Vectors and Tensors SL
Fig. 2.9 Relative velocity vaya for two
points on a rigid body.
@ result which can be used to express vector derivatives in curvilinear coordi-
nates, or with respect to moving axes.
In cylindrical coordinates (r, 8, 2) (Fig. 2.10), where
x=reos6, y=rsin8,
the position vector p is given by
pare, + 28, (2.5.11)
where @, and @, are unit vectors in the direction of increasing r and z, respec-
tively. Note that @, = k is constant, but that in a general motion @, changes
with time as does the third unit vector, @, in the transverse direction of in-
creasing 6.
If we differentiate Eq. (11), we obtain
v= 2= ta Bey He, (2.5.12)
The triad of unit vectors @,, &,, @, may be considered as a rigid body moving
with the particle; since @, always intersects the z-axis at a right angle, the52 Vectors and Tensors Chap. 2
Fig. 2.10 Vector p in cylindrical coordinates r, 6, 2.
angular velocity @ of this little rigid body is (d@/dt)@,, and the rates of change
of the unit vectors are given by Eq. (10) as
a, 4 oe
FAUX eHGuxba
dt a
TUX eH Gaxea= (2.5.13)
since
=@,x@, while @ x & = —@,.
Hence
_dr a a
= Fe tr Pet Fe, (2.5.14)
This last equation may be differentiated again to yield the acceleration, using
once more Eqs. (13) for the derivatives of the unit vectors. For further ex-
amples of this kind see a textbook on vector dynamics. We turn our attention
now to differentiation with respect to space coordinates.Sec. 2.5 Vectors and Tensors 53
Scalar point function. A scalar point function has a single component in
any coordinate system, which has the same numerical value in all coordinate
systems. Thus
FG, 5,2) = Fy), (2.5.15)
where, by definition,
FR, I, DS FUR, Fs Ds WEI, Ds 28 FD). (2.5.16)
It is common practice to write Eq. (15) as
FUR, Fs 2) = Fs ys 2),
but it must be remembered that F docs not represent the same function of the
barred coordinates as it docs of the unbarred coordinates. For example, if
F(x, y, 2) = 25(x* — y*) + 2°, and if the transformation is x = (4% — 39)/5,
y = (38 + 49/5, and z= Z, then F(x, J, 2) = 7(8% — f) — 4B 8p 4+ 27, an
apparently quite different function from F(x, y, z) although it has the same
value at every point.
The gradient of a scalar F is defined to be the vector, denoted VF or
grad F, such that for any unit vector 0, the directional derivative dF/ds in the
direction of @ is given by the scalar product
dF _ op,
F = vr, (2.5.17)
where
=
a=.
Note that this definition makes no reference to any coordinate system. The
gradient is thus a rector invariant independent of the coordinate system. To
find its components in any rectangular Cartesian coordinate system we note
that
(2.5.18)
while, by the chain rule of calculus,
ae Olax (2.5.19)
ds ~ Ox, ds *54 Vectors and Tensors Chap. 2
If (VF), denotes the x,-component of VF, then Eq. (17) yields
F = (WF),
Comparing this with Eq. (19), we see that
[ery - x] tr =o
for arbitrary choice of the dx,/ds (arbitrary direction 0). Hence
(WF) = x or
5.
grad F = VF = zh a
In indiciat notation partial derivatives are frequently denoted by a subscript
preceded by a comma: 8F/@x, = Fy. Another notation places the operator 2,
ahead of the function. Thus
OFaF.= x= and VF =hOF =F, (25.206)
A geometric interpretation of the gradient follows from Eq. (17) if we take
0 = tas a tangent vector to the level surface F(x, y, 2) = constant.
Then dF/ds = 0, and Eq. (17) implies that
(VF)-t = 0. (2.5.21)
Since according to Eq. (21), the gradient vector is perpendicular to every
tangent vector at the point, the gradient vector must be normal to the level
surface through the point. Moreover the rate of change of the function F with
respect to distance in the direction fi of increasing F normal to the surface
F = constant, namely, (VF)-+A, is equal to |VF], since VF is parallel to the
normal and since (VF)-f is positive because f was taken as the direction of
increasing F. The unit normal vector fi to the level surface, taken in the direc-
tion of increasing F, is therefore given by
dF
f= ea (2.5.22)
Other important applications of the gradient in mechanics relate the force
vector to the gradient of the potential of a conservative force field and the
velocity vector to the gradient of a velocity potential in irrotational fluid flow.Sec.2.5 Vectors and Tensors 55
The symbolic vector operator del, denoted by V,
tA a i
v= a tla tka =in = bm (2.5.23)
may be used formally to obtain Eq. (20)
fee (pe Oyo, OF aF
end FHV = (IP +p +kg)F= i ties eeea(25 24)
The operator operates on the quantity written to the right of it. When the
quantity to the right is a scalar function F, the result is grad F. When the
quantity to the right is a vector function, it is necessary to specify how
the vector operator operates on it. Two common applications are V+y and
vx y, the first yielding a scalar, the second a vector. These will be intro-
duced here briefly and formally; applications and physical interpretations are
given later. Vv without dot or cross, is a tensor; see Eq. (31).
Divergence V-v. When the vector operator V operates in a manner anal-
ogous to scalar multiplication, the result is a scalar point function, divy,
called the divergence of the vector field y. In rectangular Cartesians,
divy = Vor = (2 es ig a 2)-Gv. + Ju, + ke.) or
(2.5.25a)
divy = Ses Bea y Be = Mey ey.
A physical interpretation of div v is given in Sec, 5.2, where it is shown that
if v is the velocity field in a flow of a continuous medium, then divy =
—(1/p)(dp/dt), where p is density, Thus the divergence of the velocity vector
measures the rate of flow of material away from the particle and is equal to
the unit rate of decrease of density p in the neighborhood of the particle.
The Laplacian Operator V* = V+V gives a scalar point function when it
operates on a twice-differentiable scalar field
(oes Or ror
VFS VIVE = Fa = oat ort oF (2.5.25b)
in rectangular Cartesian coordinates. For orthogonal curvilinear coordinates,
see Eq. (1.2.14). For a general-tensor component form, see Eq. (1.6.28). For
the Laplacian V*y of a vector v, see Eq. (2.5.35).
Curl V x vy. When the vector operator V operates in a manner analopous
to vector multiplication, the result is a vector, curl v, called the curl «' the
vector field v (sometimes called the rotation and denoted Rot v).