Skip to content

Commit

Permalink
Initial revision
Browse files Browse the repository at this point in the history
  • Loading branch information
Brent Baccala committed Jan 26, 2007
1 parent 4b755c0 commit 4328ad1
Show file tree
Hide file tree
Showing 10 changed files with 946 additions and 0 deletions.
87 changes: 87 additions & 0 deletions 00-PREFACE.tex
@@ -0,0 +1,87 @@

\chapter*{Preface}

This book grew out of an abortive class in Risch Integration that I
taught at University of Maryland at College Park in the spring of
2006,\footnote{I am not a professor at UMCP, and am not affiliated
with the University of Maryland in any way other than having studied
physics there as an undergraduate and being a member of the University
Alumni Association.} which I canceled after three weeks when I
had no students left. Aside from the lack of student interest (it was
a non-credit class), another deficiency in the class became apparent
to me --- the lack of a good textbook. So I am writing this book to
fill this perceived gap, the need for a senior level undergraduate
text on differential algebra, developing the subject so far as the
solution of the problem of integration in finite terms (the
integration problem), the theory's most famous application to date.

Why, first of all, should math students study this subject, and why
near the end of an undergraduate mathematics program?

First and foremost, for pedagogical reasons. Almost all modern
college math curricula include higher algebra, yet this subject seems
to be taught in a very abstract context. The integration problem puts
this abstraction into concrete form. We have a specific goal in
mind --- the development of an algorithm that, given an integral
constructed from elementary functions, either solves that integral by
expressing it using elementary functions, or else proves that no such
expression is possible. One of the best ways to learn a subject, or
at least to convince yourself that you understand it, is to apply it
in a specific and concrete way. The greatest difficulties I have
encountered in math is when faced with abstract concepts lacking
concrete examples. Such, in my mind, is the primary goal of studying
differential algebra near the end of an undergraduate program. The
student has no doubt been exposed to higher algebra, now we want to
make sure we understand it by taking all those rings, fields, ideals,
extensions and what not and applying them to this specific goal.

Secondly, there is a sense of both historical and educational
completion to be obtained here. Not only has the integration problem
challenged mathematicians since the development of the calculus, but
there is a real danger of getting through an entire calculus sequence
and be left thinking that if you really want to solve an integral, the
best way is to use a computer! Due to the intricacy of the
calculations involved, the best way probably is to use a computer, but
the student is left at a vague but quite definite disadvantage without
the understanding that the integration problem has been solved and
without some familiarity with the techniques used to solve it.

Third, an introduction to differential algebra may be quite
appropriate at a point where students are starting to think about
research interests. Though this field has profitably engaged the
attentions of a number of late twentieth century mathematicians, it is
still a young field that may turn out to be a major breakthrough in
the solution of differential equations. It may also turn out to be a
dead end (``interesting but not compelling'' in the words of one
commentator), which I why I hesitate to list this reason first on my
list. The big question, in my mind, is whether this theory can be
suitably extended to handle partial differential equations, as both
integrals and ordinary differential equations can now be adequately
handled using numerical techniques. This question remains unanswered
at this time.

Finally, I have a strong personal motivation in writing this book.
I am not an expert in this field, really a student myself at this
point. Another very good way to learn a subject, or at least to
convince two people that you understand it, is to explain it to
somebody else.

Since the available material on this subject is too sparsely spread
around among a variety of texts and research papers, I decided for all
of these reasons to compile, more so than write, a book targeted at an
undergraduate audience with some exposure to higher algebra. However,
in keeping with my primarily pedagogical aims, I re-introduce all the
key concepts of algebra as they are needed. This serves both to
refresh and reinforce concepts already learned and also to act a
convenient reference without having to flip constantly back and forth
between books. This book should not be taken as a substitute for a
broader theory text, as I introduce only the concepts needed for my
particular application, and only at a level of detail that seems
appropriate.

Since the book is still a work in progress, I can't hope to
properly conclude this preface at this time. I would,
however, like to specifically thank Dr. Denny Gulick, Undergradate
Chair of the UMCP Mathematics Department, for giving me
the opportunity to teach the class which lead directly
to this book.
156 changes: 156 additions & 0 deletions 01-INTRO.tex
@@ -0,0 +1,156 @@

\chapter{Introduction}

In high school, we study what the Arabs called ``al-jabr'', or what
the Encyclopaedia Britanncia calls ``a generalization and extension of
arithmetic''. ``Elementary algebra," the encyclopedia goes on, ``is
concerned with properties of arbitrary numbers,'' and cites the
commutative law of addition $(a+b=b+a)$ as an example of such a
property. We use only a few others: the commutative law of
multiplication; associative laws of both addition and multiplication;
the distributive law. The key point is that all of these laws are
valid for any numbers whatsoever, so we are justified in applying them
to unknown numbers.

In addition to these basic laws, there is a language to be learned, as
well as the more general Principle of Equality: given two identical
quantities, the same operation applied to both must given identical
results. This hold true no matter what the operation is, so long as
it is deterministic (i.e, has no randomness). Thus, combining the
Principle of Equality with the commutative law of addition, I can
conclude that $\sin(a+b)=\sin(b+a)$, without any additional knowledge of
what ``$\sin$'' might be.

For example, consider the following sequence:

\begin{tabular}{r c l l @{\vbox to20pt{}}}
$(ax+{b\over2})^2$ &=& $(ax+{b\over2})(ax+{b\over2})$ & definition of square \cr
&=& $ax(ax+{b\over2}) + {b\over2}(ax+{b\over2})$ & distributive law \cr
&=& $axax+ax{b\over2} + {b\over2}(ax+{b\over2})$ & distributive law \cr
&=& $axax+ax{b\over2} + {b\over2}ax+{b\over2}{b\over2}$ & distributive law \cr
&=& $aaxx+{1\over2}abx + {1\over2}abx+{b\over2}{b\over2}$ & commutative law of multiplication (3 times)\cr
&=& $a^2x^2 + {1\over2}abx+ {1\over2}abx + {b^2\over4}$ & definition of square\cr
&=& $a^2x^2 + ({1\over2}+{1\over2})abx + {b^2\over4}$ & distributive law\cr
&=& $a^2x^2 + abx + {b^2\over4}$ & basic arithmetic\cr
$(ax+{b\over2})^2 - {b^2\over4} + ac$ &=& $a^2x^2 + abx + {b^2\over4}- {b^2\over4}+ ac$ & principle of equality\cr
$(ax+{b\over2})^2 - {b^2\over4} + ac$ &=& $a^2x^2 + abx + ac$ & definition of subtraction\cr
\end{tabular}
\vfill\eject

So, if $ax^2+bx+c=0$, then

\begin{tabular}{r c l l @{\vbox to20pt{}}}
$ax^2+bx+c$ &=& $0$ & \cr
$a(ax^2+bx+c)$ &=& $0a$ & principle of equality \cr
$a(ax^2+bx+c)$ &=& $0$ & zero theorem\footnote{$0a=0a+0a-0a=(0+0)a-0a=0a-0a=0$, showing that zero's unique behavior under multiplication is a direct result of the distributive law and zero's role as the identity element under addition}\cr
$a^2x^2+abx+ac$ &=& $0$ & distributive law\cr
$(ax+{b\over2})^2 - {b^2\over4} + ac$ &=& $0$ & principle of equality\footnote{using the last equality from the previous page}\cr
$(ax+{b\over2})^2 - {b^2\over4} + ac + {b^2\over4} - ac$ &=& ${b^2\over4} - ac$ & principle of equality\cr
$(ax+{b\over2})^2 $ &=& ${b^2\over4} - ac$ & definition of subtraction\cr
$4(ax+{b\over2})^2 $ &=& $4{b^2\over4} - 4ac$ & principle of equality\cr
$4(ax+{b\over2})^2 $ &=& $b^2 - 4ac$ & definition of division\cr
$2^2(ax+{b\over2})^2 $ &=& $b^2 - 4ac$ & definition of square\cr
$(2(ax+{b\over2}))^2 $ &=& $b^2 - 4ac$ & commutative law of multiplication\footnote{In the form $a^2b^2=aabb=abab=(ab)^2$}\cr
$(2ax+2{b\over2})^2 $ &=& $b^2 - 4ac$ & distributive law \cr
$(2ax+b)^2 $ &=& $b^2 - 4ac$ & definition of division \cr
$\sqrt{(2ax+b)^2} $ &=& $\sqrt{b^2 - 4ac}$ & principle of equality \cr
$(2ax+b) $ &=& $\sqrt{b^2 - 4ac}$ & !?!?!??! \cr
$(2ax+b)-b $ &=& $\sqrt{b^2 - 4ac} - b$ & principle of equality \cr
$2ax $ &=& $\sqrt{b^2 - 4ac} - b$ & definition of subtraction \cr
${1\over2a}2ax $ &=& ${1\over2a}(\sqrt{b^2 - 4ac} - b)$ & principle of equality \cr
$x $ &=& ${1\over2a}(\sqrt{b^2 - 4ac} - b)$ & definition of division \cr

\end{tabular}

At each step in the sequence (except one), we're just applying one of
the basic rules above. The problem with the ``mystery step'' isn't so
much that we're taking the square root, since the principle of
equality tells us that we can perform the same operation on both sides
of the equal sign, but rather that it cancels out the square in some
undefined way. So, assuming that we can perform the mystery step, and
noting that the division in the next to last step is only defined if
$a\ne0$, we can legitimately conclude that the final result is true
for any $a$, $b$, and $c$ whatsoever.

The mystery step leads us to introduce complex numbers,
typically when we want to use this equation to solve polynomials such
as $x^2+1=0$. At this point, the alert student, having been lured in
to a false sense of security by the encyclopedia's ``numbers'', and
now finding himself facing a whole new type of number entirely, can
rightly ask, ``What is a number?''

To which we wave our hands and reply, ``It's, you know, a number!''
I am reminded of the time that I was asked to sub in a
seventh grade pre-algebra class, and was promptly asked by one of the
students to explain the difference between ``3'' and ``2.9999999\ldots''
I think I mumbled something lame like ``I don't know, what do you
think?'' I certainly hadn't come to class prepared to discuss Cauchy
sequences!

In college we are no longer satisfied with this answer, and here is
really the launching point for ``higher'' algebra. Our ``numbers''
become objects in a set, and our simple concepts of addition and
multiplication morph into operations which map pairs of objects into
other objects. When asked, ``What is a number?'', we now confidently
reply, ``Anything whose operations obey the axioms!'', which really
isn't all that surprising an answer (anymore) because our entire
theory had been built around those axioms to begin with.

The program of higher algebra (in fact much of modern mathematics)
goes thus. We postulate the existance of one or more sets of objects
and one or more operations, which are simply mappings defined on the
objects of those sets. We write out a list of axioms that we assume
those sets and operations obey. Which axioms are those? Whichever we
find useful (or at least interesting). Then we develop as little or
much of a theory as we can, reasoning always from the base axioms.
Finally, we take some specific set of objects (like the integers),
demonstrate that they obey our set of axioms, and conclude that the
entire theory developed for those axioms must apply, therefore, to the
integers. Sometimes we reverse the process by finding axioms obeyed
by some specific set of objects we wish to study, then developing a
theory around them.\footnote{How do we demonstrate that a certain set
obeys certain axioms? By using more axioms, of course! Mathematics
is probably the most self-contained of all major academic fields of study.
Many other fields use its results, but math itself references nothing.
It's impossible to get started without assuming {\it something}, so
the entire process becomes a bit of a chicken-and-egg operation, which
leads you to wonder$...$ which {\it did} come first?}

The most important (i.e, repeatedly used) sets of axioms are given
names, or more precisely the sets and operators which obey them are
given names. Thus, a ``group'' is any set and operator which obey three
or four certain axioms. A ``ring'' is any set and pair of operators
which obey about six axioms. Add another axiom or two and it
becomes a ``field''. If a different axiom is obeyed, it is a
``Noetherian ring''.

It's easy to get bogged down with terminology, especially in a
classroom environment where you can't raise your hand during a test
and ask, ``Excuse me, what's a semigroup again?'' Far more important,
I think, is to grasp the central idea that any of these terms refers
simultaneously to three things: a set of axioms, a theory logically
developed from those axioms, and any particular object(s) that obeys
those axioms, and therefore the theory. The ultimate goal is to
develop far more sophisticated theories than are possible using the
``numbers'' of elementary algebra.

Our goal in this book is the development of an algebraic system that
allows us to represent as a single object any expression written using
elementary functions, putting $\sqrt{1 + \sin x}$ on par with
$3\over2$, introducing the concept of a derivative so that we can
write differential equations using these objects (it now becomes {\it
differential} algebra), and equipping this system with a theory
powerful enough to either integrate anything so expressed, or prove
that it can't be done, at least not using elementary functions. This
is how computer programs like {\it Mathematica} or {\it AXIOM} solve
``impossible'' integrals. Along the way, we will have cause to
at least survey some of the deepest waters of modern
mathematics. Differential algebra is very much a 20$^{\rm th}$
century theory --- the integration problem was not solved until
roughly 1970; a really workable algorithm for the toughest cases
wasn't available until 1990; a key sub-problem (testing the
equivalence of constants) remains unsolved still. Yet one thing is
for sure. Three hundred years after the development of calculus, one
of its most basic and elusive problems has finally yielded not to
limits, sums, and series, but to rings, fields and polynomials. Quite a
triumph for ``al-jabr''.

0 comments on commit 4328ad1

Please sign in to comment.