Tuesday, September 2, 2014

Will the Real Geometry of Nature Please Stand Up?

Is fractal geometry “the geometry of nature”? I was asked this question recently in an email from someone who had watched the PBS video Hunting the Hidden Dimension that I worked on, and appeared in, a few years ago.

It would have been easy to simply reply “Yes,” and for many audiences I would (and have) done just that—for this was by no means the first time I had been asked that question, or others very much like it. But the context in which this recent questioner raised the issue merited a less superficial response. So I wrote back to say that there is no such thing as the geometry of nature, or more generally, the mathematics of W, where W is some real world domain.

The strongest claim that can be made is something along the lines of “Mathematical theory T is the best mathematical description (or model) we currently have of the real world domain (or phenomenon) W.” But even then, this statement is less definitive than it might first appear: In particular, what do we mean by “best”?

Best in terms of understanding? (If so, then understanding by whom?)

Best in terms of building something in W? (If so, then building out of what, using what tools, and for what use?)

Best in terms of teaching someone about W? (If so, then teaching what kind of person in terms of age, background, education, motivation, etc.?)

Slightly edited and extended, the next few paragraphs are what I wrote back to my correspondent:

Nature is just what it is. Mathematics provides various ways to model our perception and experience of reality. Different parts of mathematics provide different models, some better than others. Fractal geometry provides one model that seems to accord with our observations, measurements, and experiences. But so too do the cellular automata models on which Steve Wolfram bases his “New Kind of Science.”

Many of us think fractal geometry does a better job than cellular automata in helping us understand the natural world by virtue of its nature, but that reflects an assumed patterns/relationship conception of what constitutes science.

I would prefer to call Wolfram’s framework a computational theory (of the world), rather than science. But the distinction is, I think, purely one of the meaning we attach to the relevant words (particularly “science”).

Both approaches can be said to begin by looking at how nature works, but the moment you start to create a model, you leave nature and are into the realm of human theorizing. From then on, the only available metrics are (1) degree of fit to observations and measurements, (2) degree of utility, and (3) degree to which we find the model’s assumptions reasonable.

There is lots of slack here.

In (1), what are we observing and measuring? (They are often entities created by those very mathematical theories, e.g. mass, length, volume, velocity, momentum, temperature, etc.)

In (2), how do we define utility? Doing stuff, building stuff, understanding stuff, teaching stuff, or something else? (Each with the various audience/use/purpose caveats I raised earlier.)

Then there is (3). Unless we make some initial assumptions, we cannot get a theory off the ground. And make no mistake about it, we do begin with assumptions. Not arbitrary ones, to be sure—not even close to being arbitrary. For the resulting theory to be fully accepted (as a plausible explanation or model), it has to accord to any and all the available facts, and it has to be falsifiable—it should make claims or imply conclusions that we can attempt to prove wrong.

For instance, a mathematical theory that implied 3 = 4 (as an identity of integers) would be immediately rejected.

What about a theory that implies 0.999… = 1.0, where those three dots indicate that the decimal series continues for ever? According to the widely accepted, standard definitions that mathematicians use to provide meaning to the concept of an infinite sequence of decimal digits, this identity is correct. Indeed, it can be proved to be correct, starting from the reasonable, plausible, and accepted basic principles (axioms) for the real number system.

Most university math students learn about the framework within which 0.999… is indeed equal to 1.0. (Though many of the popular “proofs” you come across are not rigorous.) As a result, many mathematically educated people will state, as if it were an absolute fact of the world, that 0.999 = 1.0. But that is not true. The identity holds because we have made some assumptions about how to handle infinity. It’s easy to overlook that fact. So let me provide a further example where it may be less easy to miss an underlying assumption.

Graduate students of mathematics are introduced to further assumptions (about handling the infinite, and various other issues), equally reasonable and useful, and in accord both with our everyday intuitions (insofar as they are relevant) and with the rest of mainstream mathematics. And on the basis of those assumptions, you can prove that

1 + 2 + 3 + … = –1/12.

That’s right, the sum of all the natural numbers equals –1/12.

This result is so much in-your-face, that people whose mathematics education stopped at the undergraduate level (if they got that far) typically say it is wrong. It’s not. Just as with the 0.999… example, where we had to construct a proper meaning for an infinite decimal expansion before we could determine what its value is, so to we have to define what that infinite sum means.

It turns out that there is an entire branch of mathematics, called analytic continuation theory, that provides us with a “natural” meaning for (in particular) that sum. And when we calculate the value using that meaning, we arrive at the answer -1/12. See this Wikipedia article for a brief account.

Incidentally, just as with the 0.999… example, you will find purported “intuitive proofs” floating around, among them this video that went viral earlier this year, but those arguments too are not rigorous.

Both frameworks, the one that yields a value for 0.999… and the one that produces a value for 1 + 2 + 3 + … , satisfy all the requirements of being reasonable, plausible, consistent with the rest of mainstream mathematics, and useful (in studies of real world phenomena, including physics). If you accept one, you really cannot reasonably deny the other. Rather, you have to accept the implications they yield, even if they at first seem counter to your expectations.

True, neither identity accords with our experiences in the physical world, since those experiences do not involve any infinite quantities or processes. (So there is nothing to accord with!)

One of the things surprising examples involving infinity remind us of is that mathematics is not “the true theory of the real world” (whatever that might mean). Rather, mathematical theories are mental frameworks we construct to help us make sense of the world. They survive or wither according to the degree to which they continue to accord with our real world experiences and to prove useful to us in conducting our individual and collective lives.

To return to geometry. For most people, throughout human history the geometry of the world experienced was planar Euclidean geometry, which accords extremely well with our everyday experiences.

But for the global air traveler (such as long distance airplane pilots), and for the astronauts in the International Space Station, spherical geometry is “the geometry.” In still other circumstances (for the most part, physics and cosmology), hyperbolic and elliptic geometries are the best frameworks.

For the artist trying to represent three dimensions on a two-dimensional canvas (or the movie or video-game animator trying to represent three dimensions on a screen), projective geometry is the best framework.

Picking up on my opening example, when you adopt a geometric perspective to try to understand growth in the natural world, you find that fractal geometry is the most appropriate one to hand.

And, finally, when you adopt a geometric perspective to try to make sense of social life in today’s multi-cultural societies, you may find that higher dimensional Euclidean geometries seem to work best, as I explain in this video (30 minutes) taken from a talk I gave at a conference in New Mexico earlier this year. (The relevant segment starts at 3:20 and ends at 11:00.)

The fact is, there is not just one geometry, and there is no such thing as “the geometry of W,” where W is a real world phenomenon or domain.

Likewise for other branches of mathematics we develop and use to understand our world and to do things in our world.

This means that, whereas, within mathematics there are “right answers,” when you apply mathematics to the world, that certainty and accuracy is only as good as the fit between the mathematics (as a conceptual framework) and the world.

And now we are back, more or less, at the topic of my previous Devlin’s Angle post. It merits a second look. Given the nature of the modern world, with mathematical models playing such a major role, with major consequences (in banking, information storage, communication, transportation, national security, etc.), we should not lose track of the fact that mathematics is not the truth.

Rather, it provides us with useful models of the world. As a result, it is a powerful and useful way of making sense of the world, and doing things in the world.

This distinction was not particularly significant for anyone growing up in the 20th century and earlier. Back then, there was usually no danger in viewing mathematics as if it were the truth. But it is an absolutely critical distinction to keep in mind for those coming of age today.

That New Mexico talk video I referred to a moment ago was in fact from a conference on middle school mathematics education, and was an attempt to raise awareness among middle school math teachers of the need to make their students aware of the way mathematics is used in the world they will live in and help shape, emphasizing not only mathematics’ strengths but also its limitations.

When you think about what is at stake here, much of the current debate (largely uninformed on the opposition side) about the Common Core State Standards resembles nothing more than two elderly bald men arguing over ownership of a comb.

In the case of the UK’s Falkland’s War of 1983, where this analogy originated, both sides appeared equally stupid. The sad aspect to the CCSS debate is that the level of ignorance (or malicious intent) on the “Stop” side forces many well-informed teachers and mathematics learning experts to devote time to the debate, lest ignorance prevail and our kids find themselves unable to survive in the world they inherit. (What the debate should focus on is how to properly implement the Standards. There be dragons, and someone needs to slay them.)

WORTH LISTENING TO: American RadioWorks has just aired an excellent radio documentary about the Common Core, in which we hear from real teachers who have been using it, both in states where it has been implemented according to plan and others where the implementation has been modified.

Friday, August 1, 2014

Most Math Problems Do Not Have a Unique Right Answer

One of the most widely held misconceptions about mathematics is that a math problem has a unique correct answer.

(Some of those who hold that view also think that there is just one correct way to get that answer. A far smaller group, to be sure, but still a worryingly large number. Still, my focus here is on the first false belief.)

Having earned my living as a mathematician for over 40 years, I can assure you that the belief is false. In addition to my university research, I have done mathematical work for the U. S. Intelligence Community, the U.S. Army, private defense contractors, and a number of for-profit companies. In not one of those projects was I paid to find "the right answer." No one thought for one moment that there could be such a thing.

So what is the origin of those false beliefs? It's hardly a mystery. People form that misconception because of their experience at school. In school mathematics, students are only exposed to problems that (a) are well defined, (b) have a unique correct answer, and (c) whose answer can be obtained with a few lines of calculation.

But the only career in which a high school graduate can expect to continue to work on such problems is academic research in pure mathematics—and even then (and again speaking from many years of personal experience), cleanly specified problems that have (obtainable) "right answers" are not as common as you might think.

Since the vast majority of students who go through school math classes do not end up as university research mathematicians, whereas many do find themselves in careers that require some mathematical ability, it's reasonable to ask why their entire school mathematics education focuses exclusively on one tiny fraction of all possible mathematics problems.

The answer can be found by looking at the history of mathematics. Starting with the invention of numbers around 10,000 years ago, people developed mathematical methods to solve problems they faced in the world: arithmetic and algebra to use in trade and engineering, geometry and trigonometry for building and navigation, calculus for scientific research, and so forth.

While some of that mathematics was required only by specialists (e.g. calculus), arithmetic and parts of algebra in particular were essential for everyday living. As a consequence, mathematicians wrote books from which ordinary people could learn how to calculate. From the very earliest textbooks (Babylonian tablets, Indian manuscripts, etc.), two kinds of problems were presented: algorithm ("recipes") problems that showed the steps to be carried out to do a particular kind of computation, presented without any context, and word problems, designed to help people learn how to apply a particular algorithm to solve a real world problem. Ancient and medieval textbooks had many hundreds of such problems, so that a trader (say) could find a problem almost identical in form to the one he (and back then use of mathematics was primarily a male activity) actually wanted to solve in his business. If he were lucky, all he would have to do is substitute his own numbers for those in the book's worked word problem. In other cases, the book might not provide an exact match, but by working through five or six problems that were close in form, the individual could learn how to solve his real problem.

For the majority of people, that was enough. Life simply did not require anything more. The problems they faced in their everyday activities for which mathematics was needed were simple and routine. The mathematical word problems that today seem so unrealistic were by and large remarkably similar to the problems ordinary citizens faced every day.

"When do I need to leave home in order to catch that train?" There wasn't an app to tell you the answer; you had to calculate it yourself. That word problem about trains leaving stations in your math class showed you how.

Arithmetic, in particular, was an essential, basic life skill that remained so until the development of devices that automated the process in the 1960s. I am a member of the last generation for whom the question "What do I need arithmetic for?" simply did not arise. (We asked it about other parts of mathematics.)

But that computer technology that eliminated the need for people to be good calculators led to a world in which there is a huge demand for higher order mathematical skills, starting with algebra. I wrote about this change in this column back in 1998, in a piece titled "Forget 'Back to Basics.' It's Time for 'Forward to (the New) Basics.'" Looking back at what I wrote then, I am amazed at just how much things have changed in the intervening 16 years. In September of that year, Google was founded, and the Web became a dominant force in our lives and our work.

Today, we have instant access to vast amounts of information and to unlimited computing power. Both are now utilities, much like water and electricity. And that has led to a revolution in the mathematics ordinary citizens need in order to lead a fulfilling, productive life. In a world where procedural (i.e., algorithmic) mathematics is available at the push of a button, the need has shifted to what I and others have been calling mathematical thinking.

I wrote about this in my September 2012 Devlin's Angle. Broadly speaking, mathematical thinking is a way of approaching problems that is based on classical mathematics, but takes account of the fact that computation (both numeric and symbolic) can be readily done by machines.

In practical terms, what this means is that people can now focus all their attention on real-world problems in the form they are encountered. Knowing how to solve an equation is no longer a valuable human ability; what matters now is formulating the equation to solve that problem in the first place, and then taking the result of the machine solution to the equation and making use of it.

In the 1960s, we got used to the fact that the arithmetic part of solving a mathematical problem could be done by machines. Now we are in a world where almost all the procedural mathematics can be done by machines.

Of course, this does not mean we should stop teaching procedural mathematics to the next generation, any more than the introduction of pocket calculators meant we should stop teaching arithmetic. But in both cases, the reason for teaching changes, and with it the way we should teach it. The purpose shifts from mastering procedures—something that was necessary only when there were no machines to do that part—to understanding the concepts sufficiently well to make good use of those machines.

Though this change in emphasis has been underway for some years now, it did not garner much attention in the United States until the rollout of the Common Core State Standards, which are very much geared towards the mathematical thinking needs of the 21st century. The degree to which many parents were shortsighted by the shift was made clear when some of them took to social media to complain about the kinds of homework questions their children were being asked to do. While some of those questions were truly, truly awful, others garnering a lot of critical SM comments were actually extremely good.

What was particularly ironic was that many parents, faced with being unable to assist their child with elementary grade arithmetic homework, did not draw the obvious conclusion: "Gee, if I cannot understand something as basic as integer arithmetic—however it is done—there must have been something really lacking in my own education." Instead, they jumped to the totally off-the-wall conclusion that the current educational system must be wrong.

That's like waking up in the morning to find your car won't start and saying, "Oh dear, the laws of physics don't work." The smart person says, "I need to replace the battery."

I'll tell you something. I was taught math the "old-fashioned way" too, and some of those student arithmetic worksheets were new to me when I first saw them. But regardless of any views I might have as to how it is best taught in today's world, it didn't take a lot of effort to figure out what those kids were doing on those worksheets posted on Facebook. It was just whole number arithmetic for heavens sake! Anyone who understands the basic ideas of whole number arithmetic can figure it out.

It was not my training as a professional mathematician that helped me here. It was the simple fact that I understand whole number arithmetic, something that goes back to my early childhood, when I did not even know there was such a thing as a professional mathematician, let alone aspire to be one. Unfortunately, many Americans were never taught to understand arithmetic, they were just trained to execute procedures. It's not their kids who are being short-changed. They—the parents—were!

Breezing into this fray is University of Wisconsin mathematics professor Jordan Ellenberg, with his new book How Not To Be Wrong. I knew I would find a kindred spirit when I read the book's subtitle: “The Power of Mathematical Thinking.” With a Stanford MOOC and an associated textbook both called Introduction to Mathematical Thinking, how could I not?

Ellenberg's title is superb. In one fell swoop, it casts aside that old misconception that mathematics provides "right answers," replacing it with the far more accurate description that it is a great way to stop you being wrong. For, like me, he focuses not on the internal activities of pure mathematics, rather on how mathematics is used in today's real world.

To be sure, also like me, Ellenberg has devoted a lot of his career to working in pure mathematics, so he loves searching for those "right answers," and he enjoys the subject in its own terms. We both know that there are eternal truths within mathematics (a better term would be "tautologies") and have experienced the thrill of going after them. But we both realize that what we do as pure mathematicians is a very specialist pursuit. The society that supports us when we do that does so largely because of the payoff in terms of the benefits that emerge when mathematical thinking is applied to real world problems.

Ellenberg's book is chock full of examples of those benefits, from many walks of life, presented with a delightfully light touch. He grabs the reader's attention with his very first example, taken from the Second World War. The U. S. military chiefs wanted to reduce the number of warplanes that were being shot down. The obvious solution was to add more armor to protect them. But armor adds weight, which limits the distances that can be flown and the duration of the mission, as well as increasing the production cost. So the question was, where is the most effective place to put that extra protection?

To answer this question, the chiefs brought in a team of mathematicians to analyze the evidence and determine what parts of the aircraft were most likely to be hit. They examined the fuselages of all the damaged planes that had flown back after being hit to see where the most damage was. It turned out that the engines had an average of 1.11 bullet holes per square foot, the fuel system had 1.55, the fuselages 1.73, and the rest of the plane 1.8.

So where was the optimal place to add extra armor? According to the data, the fuselages took a lot of hits, while engines suffered the least damage. So an obvious suggestion was to add armor to the fuselages. But that was not what the mathematicians suggested. Their solution was to add the armor to the engines, the part that had fewer hits when the planes got back.

And they were right. I'll leave you to figure out why that is the best solution. It's a great example of mathematical thinking. After you have convinced yourself why adding armor to the engines was the best strategy, you should buy a copy of Ellenberg's book and gain some understanding of just what mathematical thinking is, and why it is a crucial ability in today's world.

(My own book on mathematical thinking is more of a "how to" guide, as is my MOOC. Another, excellent book on mathematical thinking, that is somewhere between Ellenberg's and mine, is Burger and Starbird's The 5 Elements of Effective Thinking.)

Finally, and to some extent switching gears (and definitely switching media), I want to draw your attention to a new video game, DragonBox Elements, by the Norwegian-based educational technology company WeWantToKnow. The company made a splash with its first game, DragonBox (Algebra) a couple of years ago.

Unlike my own work in educational videogames, through my company BrainQuake, which is very strongly focused on real-world mathematical thinking, the DragonBox folks are seeking to enhance and strengthen school mathematics.

When I first played the new Elements game, I was initially confused, since I approached it with a Geometer's Sketchpad expectation. But Elements is not a geometry construction/exploration tool. The focus is on the importance of providing justification for steps in a proof. Knowing why something is true. And that is not only a key feature of GOFM (“Good Old Fashioned Math”), as was taught for two thousand years, it's one of the aspects of mathematics that is characteristic of mathematical thinking (as used in the real world). Euclid, the author of the first Elements (the book), would surely have approved.

The modern world has not made GOFM redundant. What has changed, and drastically, is the way GOFM fits in with the rest of human activities. Unless you are going to make a career for yourself in pure mathematics research, GOFM today is simply an amazingly powerful tool for acquiring one of the most important cognitive capacities in the 21st century: mathematical thinking.

In today's world, most of the important problems are complex and multi-faceted. There are few right answers. As Ellenberg demonstrates, mathematical thinking can help you choose better answers—and avoid being wrong.

Tuesday, July 1, 2014

The Power of Dots


On June 29, the New York Times ran a story about the Common Core Mathematics Standards. If ever you wanted proof of the dismal mathematics education most Americans have been provided, you will find it in the story’s “human interest lede,” which described one mother’s response to seeing her daughter’s homework. By taking the daughter out of school to teach her herself “the old fashioned way” she herself had been subjected to, this well-meaning parent was ensuring that, as had clearly been the case for the mother, the daughter too would not be exposed to real mathematical thinking—the kind that in today’s world is a key to the most attractive jobs. Instead she would be subjected to the same, dreary, rote-skills-drills inflicted on previous generations—a process designed to train people for routine work in the pre-computer era, but so hopelessly inadequate for the 21st century that parents are un-equipped to figure out for themselves the simple (albeit unfamiliar) math homework their children are assigned.

Surely, if mathematics education should achieve one thing, it is develop the ability to figure things out for yourself. We’re not talking the Riemann Hypothesis here; the focus is basic school arithmetic, for heaven’s sake.

To continue with the Times article, arrays of dots seemed to figure large in this parent’s dislike of the Common Core. She felt it was pointless to spend time drawing and staring at arrays of dots.

True, it would be possible—and I am sure it happens—to generate tedious, and largely pointless, “busywork” exercises involving drawing arrays of dots. But the image of a Common Core math worksheet the Times chose to illustrate its story showed a very sensible, and deep use of dot diagrams, to understand structure in arithmetic. Much like the (extremely deep) dot array at the top of this article, which I’ll come to in a moment.

To the girl’s parent, mathematics is about numbers, but that’s just a surface feature. It’s really about structure. And throughout the ages, mathematicians have used the most simple symbols possible to bring out and understand that structure: namely, dots and lines.

The Times’ parent, so dismissive of time spent drawing and reflecting on dot diagrams, would, I am sure, think it a waste of time to devote any effort trying to make sense of the dot diagram I used to open this post. She would, I have no doubt, find it incomprehensible that an individual with a freshly-minted Ph.D. in mathematics would spend many months—at taxpayers’ expense—staring day-after-day at either that one diagram, or seemingly minor variations he would start each day by sketching out on a sheet of paper in front of him.

Well, I am that mathematician. That diagram helped me understand the framework that would be required to specify an infinite mathematical object of the third order of infinitude (aleph-2) by means of a family of infinite mathematical objects of the first order of infinitude (aleph-0). The top line of dots represents an increasing tower of objects that come together to form the desired aleph-2 object, and each of the lower lines of dots represent shorter towers of aleph-0 objects. In the 1970s, a number of us used those dot diagrams to solve mathematical problems that just a few years earlier had seemed impossible.

That particular kind of dot diagram was invented by a close senior colleague (and mentor) of mine, Professor Ronald Jensen, who called it a “morass.” He chose the name wisely, since the structure represented by those dots was extremely complex and intricate.

In contrast, the simple, rectangular array implicitly referred to in the New York Times article is used to help learners understand the much simpler (but still deep, and far more important to society) structure of numbers and the basic operations of arithmetic, as was well explained in a subsequent blog post by mathematics education specialist Christopher Danielson. The fact is, dot diagrams are powerful, for learners and world experts alike.

The problem facing parents (and many teachers) today, is that the present student generation is the one that, for the first time in history, is having to learn the mathematics the professionals use—what I and many other pros have started to call “mathematical thinking” in order to distinguish it from the procedural skills so important in past times.

The reason for that is that in the world today’s students will graduate into, computation is as plentiful as water or electricity. The smartphone we carry around with us is much faster, and more accurate, in carrying out mathematical procedures than any human.

In a single generation, society’s need for mathematical mastery has gone from procedural computation, to being able to make effective and reliable use of an effectively unlimited amount of automated computation. To put it bluntly, mastery of computational skills is no longer a marketable asset. The ability to make good use of computational power is where it’s at in math today.

For almost all the three thousand years of mathematical development, the focus in mathematics was calculation (numerical, symbolic, or geometric). Learning mathematics meant learning how to perform those calculations, which boiled down to achieving mastery of various procedures. Mastery of any one procedure could be achieved by rote learning—doing many examples, all essentially the same—leaving the only truly creative mental task that of recognition of which procedure to apply to solve which problem.

Numerical and symbolic calculation (arithmetic and algebra) are so simple and routine that we can program computers to do it for us. That is possible because calculation is essentially trivial. Perceiving and understanding structure, on the other hand, is something that (at least at the present time) requires human insight. It is not trivial and it is difficult. Dot diagrams can help us come to terms with that difficulty.

When movie director Gus Van Sant was faced with introducing the lead character, Will Hunting (played by Matt Damon) in the hit 1997 film Good Will Hunting, establishing in one shot that the hero was an uneducated (actually, self-educated) mathematical genius, the first encounter we had with Will showed him drawing a dot diagram on a blackboard in an MIT corridor.


You can be sure that when an experienced movie director like Gus Van Sant selects an establishing shot for the lead character, he does so with considerable care, on the advice of an expert. By showing Will writing a network of dots on a blackboard, Van Sant was right on the button in terms of portraying the kind of thing that professional mathematicians do all the time.

The one bit of license Van Sant took was that the diagram we saw Matt Damon writing was not the solution to a problem that had taken an MIT math professor two years to solve. (Unless MIT math professors are a lot less smart than we are led to believe!) It was a real solution to a real math problem, all right. I am pretty sure it was chosen because it fitted nicely on one blackboard and looked good on the screen. It absolutely conveyed the kind of (dotty) activity that mathematicians do all the time—the kind of (dotty) thing I did in my early post-Ph.D. years when I was working with Prof Jensen’s morasses.

But it’s actually a problem that anyone who has learned how to think mathematically should be able to solve in at most a few hours. Numberphile has an excellent video explaining the problem.

So, New York Times story parent, I hope you reconsider your decision to take your daughter out of school to teach her the way you were taught. The kind of mathematics you were taught was indeed required in times past. But not any more. The world has changed dramatically as far as mathematics is concerned. As with many other aspects of our lives, we have built machines to handle the more routine, procedural stuff, thereby putting a premium on the one thing where humans vastly outperform computers: creative thinking.

Those dot diagrams are all about creative thinking. A computer can understand numbers, and process millions of them faster than a human can write just one. But it cannot make sense of those dot diagrams. Because it does not know what any particular array of dots means! And it has no way to figure it out. (Unless a human tells it.)

Next month I’ll look further into the distinction between old-style procedural mathematics and the 21st-century need for mathematical thinking. In particular, I’ll look at an excellent recent book, Jordan Ellenberg’s How Not to be Wrong.

The book’s title is significant, since it recognizes that the vast majority of real-world mathematical problems do not have a unique right answer, and that the real power of mathematical thinking is making sure you are not wrong. (The book’s subtitle is “The power of mathematical thinking.”)

I’ll also look at a new mathematics video game that also focuses on mathematical thinking, this time, school-room Euclidean geometry. It’s called DragonBox Elements.

You might want to check out both.

Sunday, June 1, 2014

Déjà vu all over again: Fibonacci and Steve Jobs — Part 2

This month’s column is the second of a two-part video presentation of a public address I gave recently at Princeton, where I have been spending this semester as a Visiting Professor.

The talk was based on my 2011 e-book Leonardo and Steve, which itself was a supplement to my print book The Man of Numbers, published the same year.

Both the e-book and my presentation show how Jobs’s introduction of the Macintosh computer in 1984 was an almost exact replay of Leonardo of Pisa’s (Fibonacci) 13th Century introduction to Europe of Hindu-Arabic arithmetic.

Part 1 appeared last month.

Thursday, May 1, 2014

Déjà vu all over again: Fibonacci and Steve Jobs

This month’s column is the first of a two-part video presentation of a public address I gave recently at Princeton, where I have been spending this semester as a Visiting Professor.

The talk was based on my 2011 e-book Leonardo and Steve, which itself was a supplement to my print book The Man of Numbers, published the same year.

Both the e-book and my presentation show how Jobs’s introduction of the Macintosh computer in 1984 was an almost exact replay of Leonardo of Pisa’s (Fibonacci's) 13th Century introduction to Europe of Hindu-Arabic arithmetic.


Tuesday, April 1, 2014

What good is math and why do we teach it?

This month’s column comes in lecture format. It’s a narrated videostream of the presentation file that accompanied the featured address I made recently at the MidSchoolMath National Conference, held in Santa Fe, NM, on March 27-29. It lasts just under 30 minutes, including two embedded videos.

In the talk, I step back from the (now largely metaphorical) blackboard and take a broader look at why we and our students are there is the first place.


Download the video here.

Saturday, March 1, 2014

How Mountain Biking Can Provide the Key to the Eureka Moment

Because this blog post covers both mountain biking and proving theorems, it is being simultaneously published in Devlin’s more wide ranging blog profkeithdevlin.org


In my post last month, I described my efforts to ride a particularly difficult stretch of a local mountain bike trail in the hills just west of Palo Alto. As promised, I will now draw a number of conclusions for solving difficult mathematical problems.

Most of them will be familiar to anyone who has read George Polya’s classic book How to Solve It. But my main conclusion may come as a surprise unless you have watched movies such as Top Gun or Field of Dreams, or if you follow professional sports at the Olympic level.

Here goes, step-by-step, or rather pedal-stroke-by-pedal-stroke. (I am assuming you have recently read my last post.)

BIKE: Though bikers with extremely strong leg muscles can make the Alpine Road ByPass Trail ascent by brute force, I can't. So my first step, spread over several rides, was to break the main problem—get up an insanely steep, root strewn, loose-dirt climb—into smaller, simpler problems, and solve those one at a time.

MATH: Breaking a large problem into a series of smaller ones is a technique all mathematicians learn early in their careers. Those subproblems may still be hard and require considerable effort and several attempts, but in many cases you find you can make progress on at least some of them. The trick is to make each subproblem sufficiently small that it requires just one idea or one technique to solve it.

In particular, when you break the overall problem down sufficiently, you usually find that each smaller subproblem resembles another problem you, or someone else, has already solved.

When you have managed to solve the subproblems, you are left with the task of assembling all those subproblem solutions into a single whole. This is frequently not easy, and in many cases turns out to be a much harder challenge in its own right than any of the subproblem solutions, perhaps requiring modification to the subproblems or to the method you used to solve them.

BIKE: Sometimes there are several different lines you can follow to overcome a particular obstacle, starting and ending at the same positions but requiring different combinations of skills, strengths, and agility. (See my description last month of how I managed to negotiate the steepest section and avoid being thrown off course—or off the bike—by that troublesome tree-root nipple.)

MATH: Each subproblem takes you from a particular starting point to a particular end-point, but there may be several different approaches to accomplish that subtask. In many cases, other mathematicians have solved similar problems and you can copy their approach.

BIKE: Sometimes, the approach you adopt to get you past one obstacle leaves you unable to negotiate the next, and you have to find a different way to handle the first one.

MATH: Ditto.

BIKE: Eventually, perhaps after many attempts, you figure out how to negotiate each individual segment of the climb. Getting to this stage is, I think, a bit harder in mountain biking than in math. With a math problem, you usually can work on each subproblem one at a time, in any order. In mountain biking, because of the need to maintain forward (i.e., upward) momentum, you have to build your overall solution up in a cumulative fashion—vertically!

But the distinction is not as great as might first appear. In both cases, the step from having solved each individual subproblem in isolation to finding a solution for the overall problem, is a mysterious one that perhaps cannot be appreciated by someone who has not experienced it. This is where things get interesting.

Having had the experience of solving difficult (for me) problems in both mathematics and mountain biking, I see considerable similarities between the two. In both cases, the subconscious mind plays a major role—which is, I presume, why they seem mysterious. This is where this two-part blog post is heading.

BIKE: I ended my previous post by promising to

"look at the role of the subconscious in being able to put together a series of mastered steps in order to solve a big problem. For a very curious thing happened after I took the photos to illustrate this post. I walked back down to collect my bike from…where I'd left it, and rode up to continue my ride.
It took me four attempts to complete that initial climb!
And therein lies one of the biggest secrets of being able to solve a difficult math problem."

BOTH: How does the human mind make a breakthrough? How are we able to do something that we have not only never done before, but failed many times in attempts to do so? And why does the breakthrough always seem to occur when we are not consciously trying to solve the problem?

The first thing to note is that we never experience the process of making that breakthrough. Rather, what we experience, i.e., what we are conscious of, is having just made the breakthrough!

The sensation we have is a combined one of both elation and surprise. Followed almost immediately by a feeling that it wasn’t so difficult after all!

What are we to make of this strange process?

Clearly, I cannot provide a definitive, concrete answer to that question. No one can. It’s a mystery. But it is possible to make a number of relevant observations, together with some reasonable, informed speculations. (What follows is a continuation of sorts of the thread I developed in my 2000 book The Math Gene.)

The first observation is that the human brain is a result of millions of years of survival-driven, natural selection. That made it supremely efficient at (rapidly) solving problems that threaten survival. Most of that survival activity is handled by a small, walnut-shaped area of the brain called the amygdala, working in close conjunction with the body’s nervous system and motor control system.

In contrast to the speed at which our amydala operates, the much more recently developed neo-cortex that supports our conscious thought, our speech, and our “rational reasoning,” functions at what is comparatively glacial speed, following well developed channels of mental activity—channels that can be built up by repetitive training.

Because we have conscious access to our neo-cortical thought processes, we tend to regard them as “logical,” often dismissing the actions of the amygdala as providing (“mere,” “animal-like”) “instinctive reactions.” But that misses the point that, because that “instinctive reaction organ” has evolved to ensure its owner’s survival in a highly complex and ever changing environment, it does in fact operate in an extremely logical fashion, honed by generations of natural selection pressure to be in sync with its owner’s environment.

Which leads me to this.

Do you want to identify that part of the brain that makes major scientific (and mountain biking) breakthroughs?

I nominate the amygdala—the “reptilian brain” as it is sometimes called to reflect its evolutionary origin.

I should acknowledge that I am not the first person to make this suggestion. Well, for mathematical breakthroughs, maybe I am. But in sports and the creative arts, it has long been recognized that the key to truly great performance is to essentially shut down the neo-cortex and let the subconscious activities of the amygdala take over.

Taking this as a working hypothesis for mathematical (or mountain biking) problem solving, we can readily see why those moments of great breakthrough come only after a long period of preparation, where we keep working away—in conscious fashion—at trying to solve the problem or perform the action, seemingly without making any progress.

We can see too why, when the breakthrough (or the great performance) comes, it does so instantly and surprisingly, when we are not actively trying to achieve the goal, leaving our conscious selves as mere after-the-fact observers of the outcome.

For what that long period of struggle does is build a cognitive environment in which our reptilian brain—living inside and being connected to all of that deliberate, conscious activity the whole time—can make the key connections required to put everything together. In other words, investing all of that time and effort in that initial struggle raises the internal, cognitive stakes to a level where the amygdala can do its stuff.

Okay, I’ve been playing fast and loose with the metaphors and the anthropomorphization here. We’re really talking about biological systems, simply operating the way natural selection equipped them. But my goal is not to put together a scientific analysis, rather to try to figure out how to improve our ability to solve novel problems. My primary aim is not to be “right” (though knowledge and insight are always nice to have), but to be able to improve performance.

Let’s return to that tricky stretch of the ByPass section on the Alpine Road trail. What am I consciously focusing on when I make a successful ascent? 

BIKE: If you have read my earlier account, you will know that the difficult section comes in three parts. What I do is this. As I approach each segment, I consciously think about, and fix my eyes on, the end-point of that segment—where I will be after I have negotiated the difficulties on the way. And I keep my eyes and attention focused on that goal-point until I reach it. For the whole of the maneuver, I have no conscious awareness of the actual ground I am cycling over, or of my bike. It’s total focus on where I want to end up, and nothing else.

So who—or what—is controlling the bike? The mathematical control problem involved in getting a person-on-a-bike up a steep, irregular, dirt trail is far greater than that required to auto-fly a jet fighter. The calculations and the speed with which they would have to be performed are orders of magnitude beyond the capability of the relatively slow neuronal firings in the neocortex. There is only one organ we know of that could perform this task. And that’s the amygdala, working in conjunction with the nervous system and the body’s motor control mechanism in a super-fast constant feedback loop. All the neo-cortex and its conscious thought has to do is avoid getting in the way!

These days, in the case of Alpine Road, now I have “solved” the problem, the only things my conscious neo-cortex has to do on each occasion are switching my focus from the goal of one segment to the goal of the next. If anything interferes with my attention at one of those key transition moments, my climb is over—and I stop or fall.

What used to be the hard parts are now “done for me” by unconscious circuits in my brain.

MATH: In my case at least, what I just wrote about mountain biking accords perfectly with my experiences in making (personal) mathematical problem-solving breakthroughs.

It is by stepping back from trying to solve the problem by putting together everything I know and have learned in my attempts, and instead simply focusing on the problem itself—what it is I am trying to show—that I suddenly find that I have the solution.

It’s not that I arrive at the solution when I am not thinking about the problem. Some mathematicians have expressed their breakthrough moments that way, but I strongly suspect that is not totally true. When a mathematician has been trying to solve a problem for some months or years, that problem is always with them. It becomes part of their existence. There is not a single waking moment when that problem is not “on their mind.”

What they mean, I believe, and what I am sure is the case for me, is that the breakthrough comes when the problem is not the focus of our thoughts. We really are thinking about something else, often some mundane detail of life, or enjoying a marvelous view. (Google “Stephen Smale beaches of Rio” for a famous example.)

This thesis does, of course, explain why the process of walking up the ByPass Trail and taking photographs of all the tricky points made it impossible for me to complete the climb. True, I did succeed at the fourth attempt. But I am sure that was not because the first three were “practice.” Heavens, I’d long ago mastered the maneuvers required. It was because it took three failed attempts before I managed to erase the effects of focusing on the details to capture those images.

The same is true, I suggest, for solving a difficult mathematical problem. All of those techniques Polya describes in his book, some of which I list above, are essential to prepare the way for solving the problem. But the solution will come only when you forget about all those details, and just focus on the prize.

This may seem a wild suggestion, but in some respects it may not be entirely new. There is much in common between what I described above and the highly successful teaching method of R. L. Moore. For sure you have to do a fair amount of translation from his language to mine, but Moore used to demand that his students not clutter their minds by learning stuff, rather took each problem as it came and then try to solve it by pure reasoning, not giving up until they found the solution.

In terms of training future mathematicians, what these considerations imply, of course, is that there is mileage to be had from adopting some of the techniques used by coaches and instructors to produce great performances in sports, in the arts, in the military, and in chess.

Sweating the small stuff will make you good. But if you want to be great, you have to go beyond that—you have to forget the small stuff and keep your eye on the prize.

And if you are successful, be sure to give full credit for that Fields Medal or that AMS Prize where it is rightly due: dedicate it to your amygdala. It will deserve it.