tag:blogger.com,1999:blog-37486255103489613422019-11-16T02:36:30.472-08:00solidangl.esBill Shillitohttp://www.blogger.com/profile/17774101901445053590noreply@blogger.comBlogger14125tag:blogger.com,1999:blog-3748625510348961342.post-20609431438906592652019-10-15T13:40:00.000-07:002019-10-15T15:25:43.744-07:00Orbit-Stabilizer TheoremThis is going to be a different sort of post from the ranting-about-math-education kinds that usually I make.<br /><br />As many of you know, when I'm not doing math (yes, I do take a break sometimes), I write electronic music, some of which makes its way into rhythm games. Most recently, I wrote a track called <b>Orbit Stabilizer </b>for the Korean dance game Pump It Up XX.<br /><br /><div class="separator" style="clear: both; text-align: center;"><iframe allowfullscreen="" class="YOUTUBE-iframe-video" data-thumbnail-src="https://i.ytimg.com/vi/uYRJwzH043U/0.jpg" frameborder="0" height="266" src="https://www.youtube.com/embed/uYRJwzH043U?feature=player_embedded" width="320"></iframe></div><br />If you didn't figure it out from watching the background animation, there's a mathematical connection to the song. This isn't the first time I've made references to mathematics in my music:<br /><ul><li>One of my songs on Dance Dance Revolution, called <b><a href="https://www.youtube.com/watch?v=x2PQuUe2KG0">ΔMAX</a></b>, changes by one BPM every beat. It goes up to 573 BPM, which is significant because the number 573 can be <a href="https://en.wikipedia.org/wiki/Japanese_wordplay">read out loud as</a> "Konami" , who makes DDR.</li><li>My track on J-Rave Nation is called <b><a href="https://www.youtube.com/watch?v=WK7lVan2duw">π・ρ・maniac</a></b>, which is really just a play on the word "pyromaniac" but uses the well-known constant \(\pi\). (Also, there's a <a href="https://www.youtube.com/watch?v=c5v0as3KWT8">mission</a> in Pump It Up Infinity that makes you do math problems in the middle of the song.)</li><li>Another song from Pump It Up Prime, called <b><a href="https://www.youtube.com/watch?v=Wy5Sjd0ATLA">Annihilator Method</a></b>, is a reference to <a href="https://en.wikipedia.org/wiki/Annihilator_method">a technique for solving differential equations</a>. (You have to admit that's a pretty hardcore-sounding name for something from a math class!)</li></ul><div>This time, the reference is to something from a branch of mathematics called <u style="font-weight: bold;"><a href="https://en.wikipedia.org/wiki/Group_theory">group theory</a></u>, which is part of the deep abstract foundations for algebra. It was pioneered simultaneously by French mathematician <a href="https://en.wikipedia.org/wiki/%C3%89variste_Galois">Évariste Galois</a> (who shows up in the background of the Orbit Stabilizer video!) and Norwegian mathematician <a href="https://en.wikipedia.org/wiki/Niels_Henrik_Abel">Niels Henrik Abel</a>. Both are worth reading up on ― heartbreaking stories of geniuses who died tragically young.</div><div><br /></div><div>Group theory is, in essence, the study of <i>symmetry</i>. It can be used to describe everything from the shuffling of cards and the positions of a Rubik's cube to the fundamental laws of physics that govern our universe. I'm not going to go into an in-depth explanation of group theory ― there are <a href="https://www.youtube.com/playlist?list=PLL0ATV5XYF8AQZuEYPnVwpiFy0jEipqN-">plenty of YouTube videos</a> that do that (<a href="https://www.youtube.com/watch?v=WwndchnEDS4">including my own</a>). Rather I'm going to go just far enough to explain what the <b><u>orbit-stabilizer theorem</u></b> is.</div><div><br /></div><div>Imagine you have a square sheet of paper in front of you.<br /><br /><div class="separator" style="clear: both; text-align: center;"><a href="https://1.bp.blogspot.com/-mSVajUr0alU/XaX73fovSGI/AAAAAAAAaps/DCT99yLIETcjUSeQm0hZDgdwWxJ6fm0RgCLcBGAsYHQ/s1600/smallgraysquare.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" data-original-height="182" data-original-width="182" src="https://1.bp.blogspot.com/-mSVajUr0alU/XaX73fovSGI/AAAAAAAAaps/DCT99yLIETcjUSeQm0hZDgdwWxJ6fm0RgCLcBGAsYHQ/s1600/smallgraysquare.png" /></a></div><div class="separator" style="clear: both; text-align: center;"></div><div class="separator" style="clear: both; text-align: center;"></div><br /></div><div>You look away, and then I sneak in and potentially move the paper in some way. When you look back, the square looks the same as it did before. What could I have done to the paper?<br /><br />The set of possible transformations I could have done are called <u style="font-weight: bold;">symmetries</u> of the square. In this case, there are three overall types of such symmetries:<br /><ul><li>I could have left the paper alone entirely, doing nothing to it. This is called the <u style="font-weight: bold;">identity transformation</u>.</li><li>I could have rotated the paper by \(90^\circ\), \(180^\circ\), or \(270^\circ\).</li><li>I could have flipped the paper over across one of its four axes of symmetry ― the horizontal axis, the vertical axis, or one of the two diagonal axes.</li></ul><div>All in all, I have <i>eight</i> possible transformations that would take the square back to itself. Notice that translations (shifts) aren't included ― if I'd moved the paper a bit to the left, you would have noticed (we assume)! We're only considering transformations that would leave you completely unable to tell what I did. To visualize the effects of these transformations better, we can label the square and watch what happens:</div><div><br /></div><div class="separator" style="clear: both; text-align: center;"><a href="https://1.bp.blogspot.com/-2V0PdayTch4/XaYBJxE_6pI/AAAAAAAAap4/dN07hZ45WZodnBYDAyixavTl7gD88fVBwCLcBGAsYHQ/s1600/squares.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" data-original-height="723" data-original-width="917" src="https://1.bp.blogspot.com/-2V0PdayTch4/XaYBJxE_6pI/AAAAAAAAap4/dN07hZ45WZodnBYDAyixavTl7gD88fVBwCLcBGAsYHQ/s1600/squares.png" width="100%" /></a></div><div><br /></div><div><br /></div><div>These transformations form what mathematicians call a <u style="font-weight: bold;">group</u>, which means that they obey four fundamental laws (axioms):</div><div><ol><li>Combining any two of these transformations ― such as, say, rotating by \(90^\circ\) counterclockwise and then flipping across the horizontal axis ― also takes the square back to itself, and is equivalent to one of the original eight transformations ― in this case, flipping across the "backslash" diagonal axis. We call this the <u style="font-weight: bold;">closure</u> law.</li><li>Combining transformations is <b><u>associative</u></b>, which means that if I have any transformations \(a\), \(b\), and \(c\), then \((ab)c\) ― that is, doing \(c\) first and then doing whatever \((ab)\) is equivalent to ― gives the same result as \(a(bc)\) ― that is, doing whatever \((bc)\) is equivalent to first and then doing \(a\) afterward. This is just like how you can move parentheses around when doing addition and multiplication.</li><li>The set of transformations includes an <u style="font-weight: bold;">identity</u> transformation, which essentially does nothing. (I say "essentially" because you could also think of a \(360^\circ\) rotation as an identity transformation since it has no overall effect.)</li><li>Every transformation has an <u style="font-weight: bold;">inverse</u> transformation which "undoes" it. For example, a rotation of \(90^\circ\) can be undone by a rotation of \(270^\circ\), or a reflection can be undone by doing that same reflection again.</li></ol></div><div>We usually use the letter \(G\) to talk about a group in general. We'll also use the notation \(|G|\) to talk about the "size" of a group ― so in this case, we have \(|G|=8\).<br /><br />So we know that these transformations take the square back to itself, but what if we ask what happens to just a single point inside of the square?<br /><br />Suppose we let \(x\) mark the midpoint of segment \(AB\). Then we can see where \(x\) lands after each of the transformations:<br /><div class="separator" style="clear: both; text-align: center;"><a href="https://1.bp.blogspot.com/-oMziX-M3ag0/XaYTZGXIivI/AAAAAAAAaqE/Hxk8A64egloiyBBHnCCcPelRQqL0ckWUgCLcBGAsYHQ/s1600/midpoint.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" data-original-height="723" data-original-width="917" src="https://1.bp.blogspot.com/-oMziX-M3ag0/XaYTZGXIivI/AAAAAAAAaqE/Hxk8A64egloiyBBHnCCcPelRQqL0ckWUgCLcBGAsYHQ/s1600/midpoint.png" width="100%" /></a></div><br />The set of possible places where \(x\) can land using our transformations is called the <u style="font-weight: bold;">orbit</u> of \(x\) under our group, and we denote it as \(\text{orb}(x)\). Since there are four different places that \(x\) can land ― it can end up on any one of the four midpoints of one of the sides ― we say that \(|\text{orb}(x)|=4\).<br /><br />Looking a little closer, we can find that there are two particular transformations that do something special. Focus on the identity transformation and the vertical axis flip. What happens to \(x\) under these two transformations? Well, it stays where it is! The other six transformations, on the other hand, all move \(x\) to a different point of the square. These two special transformations form what's called the <u style="font-weight: bold;">stabilizer</u> of \(x\) under our group, and we denote it as \(\text{stab}(x)\). The stabilizer forms what's called a <u style="font-weight: bold;">subgroup</u> of our original group, because if you limit yourself to only doing those things, you still satisfy the same basic group laws. Again, we can measure the "size" of the stabilizer, which in this case is \(|\text{stab}(x)|=2\).<br /><br />Now, look at the three "sizes" we've computed.<br />\[<br />\begin{align*}<br />|G|&=8\\<br />|\text{orb}(x)|&=4\\<br />|\text{stab}(x)|&=2<br />\end{align*}<br />\]Those three numbers (\(8\), \(4\), and \(2\)) are just begging to be related to each other!<br /><br />\[|\text{orb}(x)|\cdot|\text{stab}(x)|=|G|\]<br /><br /></div><div>Is this a coincidence? Well, you can find out for yourself by picking different places for \(x\) (like, say, a corner point, or the very center, or just some random point in the square), then calculating the sizes of its orbit and its stabilizer. Or you can even try it for a different shape, like a triangle or a rhombus or a pentagon, which will have a different size group. You'll notice that you always get that same relationship!<br /><br />That relationship is called the <u style="font-weight: bold;">Orbit-Stabilizer Theorem</u>.<br /><br />...well, almost.<br /><br />We actually run into a bit of a problem if our group is <i>infinite</i>. If that's the case, then multiplying and dividing using infinite numbers can get a bit hairy and make it a bit difficult to show that the relationship is meaningful. To deal with that case, we usually write the Orbit-Stabilizer Theorem in a slightly different way:<br /><br />\[|\text{orb}(x)|=[G : \text{stab}(x)]\]</div><div><br /></div><div>The quantity \([G : \text{stab}(x)]\) is called the <u style="font-weight: bold;">index</u> of the stabilizer as a subgroup of \(G\). Essentially this stands for how many copies of \(\text{stab}(x)\) can fit inside \(G\). (Technically it's how many "cosets" the stabilizer has, if you want to look that up.) This is the version of the equation that shows up in the video.</div><div><br /></div><div>Okay, so that explains what the Orbit-Stabilizer Theorem <i>is</i>. But, you might be wondering, is it also <i>useful</i>?<br /><br />Well, one famous result in combinatorics (the study of counting) that makes great use of the Orbit-Stabilizer Theorem is <u style="font-weight: bold;"><a href="https://en.wikipedia.org/wiki/Burnside%27s_lemma">Burnside's Lemma</a></u>. Burnside's Lemma can be used to solve problems like counting the number of possible ways that the sides of a cube can be painted with three different colors. (Painting the top side blue and all the others white is considered the same as, say, painting the front side blue and all the others white, since you can just rotate one to get the other.) Looking for something more "real-world"? Burnside's Lemma has applications in <a href="https://www.springer.com/us/book/9781461291053">chemistry</a> ― like if you need to find out how many different ways certain groups can be placed around a central carbon atom ― as well as other areas such as electronic circuits and even (to bring things full circle) music theory!</div><div><br /></div><div>Hopefully this gives you an idea of why the Orbit-Stabilizer Theorem is interesting, and why I would name a song after it (besides just being a cool-sounding name). Feel free to let me know if you have any questions. And of course, if you play Pump It Up, give the track a shot!<br /><br /><i>P.S. If you want to know more about group theory, I would highly recommend <a href="https://www.youtube.com/playlist?list=PLL0ATV5XYF8AQZuEYPnVwpiFy0jEipqN-">this YouTube series</a> by Dr. Matt Salomone at Bridgewater State University. He has a knack for making these sorts of abstract algebraic topics very accessible with excellent examples and intuitive explanations!</i></div></div>Bill Shillitohttp://www.blogger.com/profile/17774101901445053590noreply@blogger.com0tag:blogger.com,1999:blog-3748625510348961342.post-6415839226314388482019-08-01T10:27:00.000-07:002019-08-02T10:38:33.010-07:00When Does Nothing Mean Something?<br />I thought I was done writing about this topic, but it just keeps coming back. The internet just cannot seem to leave this sort of problem alone:<br /><blockquote class="twitter-tweet"><div dir="ltr" lang="en">oomfies solve this <a href="https://t.co/0RO5zTJjKk">pic.twitter.com/0RO5zTJjKk</a></div>— em ♥︎ (@pjmdolI) <a href="https://twitter.com/pjmdolI/status/1155598050959745026?ref_src=twsrc%5Etfw">July 28, 2019</a></blockquote><script async="" charset="utf-8" src="https://platform.twitter.com/widgets.js"></script> I don't know what it is about expressions of the form \(a\div b(c+d)\) that fascinates us as a species, but fascinate it does. I've <a href="http://www.solidangl.es/2014/07/the-implications-of-being-implicit.html">written about this before</a> (as well as <a href="http://www.solidangl.es/2014/07/poorly-executed-mnemonics-definitely.html">why "PEMDAS" is terrible</a>), but the more I've thought about it, the more sympathy I've found with those in the minority of the debate, and as a result my position has evolved somewhat.<br /><br />So I'm going to go out on a limb, and claim that the answer <i style="font-weight: bold;">should</i> be \(1\).<br /><br />Before you walk away shaking your head and saying "he's lost it, he doesn't know what he's talking about", let me assure you that I'm obviouly not denying the left-to-right convention for how to do explicit multiplication and division. Nobody's arguing that.* Rather, there's something much more subtle going on here.<br /><br />What we may be seeing here is evidence of a mathematical "language shift".<br /><br />It's easy to forget that mathematics did not always look as it does today, but has arrived at its current form through very human processes of invention and revision. There's an <a href="http://jeff560.tripod.com/mathsym.html">excellent page</a> by Jeff Miller that catalogues the earliest recorded uses of symbols like the operations and the equals sign -- symbols that seem timeless, symbols we take for granted every day.<br /><br />People also often don't realize that this process of invention and revision still happens to this day. The <a href="http://www.solidangl.es/2014/09/throw-outdated-notation-to-floor.html">modern notation for the floor function</a> is a great example that was only developed within the last century. Even today on the internet, you occasionally see discussions in which people debate on how mathematical notation can be improved. (I'm still holding out hope that my <a href="http://www.solidangl.es/2015/04/a-radical-new-look-for-logarithms.html">alternative notation for logarithms</a> will one day catch on.)<br /><br />Of particular note is the evolution of <a href="http://jeff560.tripod.com/grouping.html">grouping symbols</a>. We usually think only of parentheses (as well as their variations like square brackets and curly braces) as denoting grouping, but an even earlier symbol used to group expressions was the <u style="font-weight: bold;">vinculum</u> -- a horizontal bar found over or under an expression. Consider the following expression: \[3-(1+2)\] If we wrote the same expression with a vinculum, it would look like this: \[3-\overline{1+2}\] Vincula can even be stacked: \[13-\overline{\overline{1+2}\cdot 3}=4\] This may seem like a quaint way of grouping, but it does in fact survive in our notation for fractions and radicals! You can even see both uses in the quadratic formula: \[x=\dfrac{-b\pm\sqrt{b^2-4ac}}{2a}\]<br /><br />Getting back to the original problem, what I think we're seeing is evidence that <u style="font-weight: bold;">concatenation</u> -- placing symbols next to each other with no sort of explicit symbol -- has become another way to represent grouping.<br /><br />"But wait", you might say, "concatenation is used to represent <i>multiplication</i>, not <i>grouping</i>!" That's certainly true in many cases, for example in how we write polynomials. However, there are a few places in mathematics that provide evidence that there's more to it than that.<br /><br />First of all, as <a href="https://twitter.com/EoN_Tweets/status/1156757840880656384">a beautifully-written Twitter thread by EnchantressOfNumbers (@EoN_tweets)</a> points out, we use concatenation to show a special importance of grouping when we write out certain trigonometric expressions without putting their arguments in parentheses. Consider the following identity:<br />\[\sin 4u=2\sin 2u\cos 2u\] When we write such an equation, we're saying that not only do \(4u\) and \(2u\) represent multiplications, but that this grouping is so tight that they constitute the entire arguments of the sine and cosine functions. In fact, the space between \(\sin 2x\) and \(\cos 2x\) can also be seen as a somewhat looser form of concatention. Then again, so does the space between \(\sin\) and \(x\), which represents a different thing -- the connection of a function to its argument. Perhaps this is why the popular (and amazing) online graphing calculator <a href="https://www.desmos.com/">Desmos</a> is only so permissive when it comes to parsing concatenation:<br /><br /><div class="separator" style="clear: both; text-align: center;"><a href="https://1.bp.blogspot.com/-Ne57iJO2yGY/XUMUP2-EkWI/AAAAAAAAaic/unI8YtWxEr8C6GhrpOxuL5d-GO0VFabeQCLcBGAs/s1600/concatsin.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" data-original-height="192" data-original-width="457" src="https://1.bp.blogspot.com/-Ne57iJO2yGY/XUMUP2-EkWI/AAAAAAAAaic/unI8YtWxEr8C6GhrpOxuL5d-GO0VFabeQCLcBGAs/s1600/concatsin.png" /></a></div>In contrast, where we do draw the line is with an expression like the following:\[\sin x+y\] We always interpret this as \((\sin(x))+y\), never \(\sin(x+y)\). To drive home just how much stronger implicit multiplication feels to us than explicit multiplication, just take a look at the following expression: \[\sin x\cdot y\] Does this mean \((\sin(x))\cdot y\) or \(\sin(x\cdot y)\)? If that expression makes you writhe uncomfortably, while if it had been written as \(\sin xy\) it would be fine, then you might see what I'm getting at.<br /><br />An even more curious case is <i>mixed numbers</i>. When writing mixed numbers, concatenation actually stands for addition, not multiplication. \[3\tfrac{1}{2}=3+\tfrac{1}{2}\] In fact, concatenation actually makes addition come <i>before</i> multiplication when we multiply mixed numbers! \[3\tfrac{1}{2}\cdot 5\tfrac{5}{6}=(3+\tfrac{1}{2})\cdot(5+\tfrac{5}{6})=20\tfrac{5}{12}\]<br /><br />Now, you may feel that this example shows how mixed numbers are an inelegance in mathematical notation (and I would agree with you). Even so, I argue that this is evidence that we <i>fundamentally</i> view concatenation as a way to represent <i>grouping</i>. It just so happens that, since multiplication takes precedence over addition anyway in the absence of other grouping symbols, we use concatenation when we write it. This all stems from a sort of "laziness" in how we write things -- - laying out precedence rules allows us to avoid writing parentheses, and once we've established those precedence rules, we don't even need to write out the multiplication at all.<br /><br />So how does the internet's favorite math problem fit into all this?<br /><br />The most striking feature of the expression \(8\div 2(2+2)\) is that <i>it's written all in one line</i>.<br /><br />Mathematical typesetting is difficult. <a href="https://en.wikibooks.org/wiki/LaTeX/Mathematics">LaTeX</a> is powerful, but has a steep learning curve, though various other editors have made it a bit easier, such as <a href="https://www.youtube.com/watch?v=oOtZW95ZcGg">Microsoft Word's Equation Editor</a> (which has much improved since when I first used it!). Calculators have also recognized this difficulty, which is why TI calculators now have <a href="https://www.youtube.com/watch?v=Qqc3zPNfJGw">MathPrint</a> templates (though its entry is quite clunky compared to Desmos's "as-you-type" formatting via <a href="http://mathquill.com/">MathQuill</a>).<br /><br />Even so, all of these input methods exist in very specific applications. What about when you're writing an email? Or sending a text? Or a Facebook message? (If you're wondering "who the heck writes about math in a Facebook message", the answer at least includes "students who are trying to study for a test".) The evolution of these sorts of media has led to the importance of one-line representations of mathematics with easily-accessible symbols. When you don't have the ability (or the time) to neatly typeset a fraction, you're going to find a way to use the tools you've got. And that's even more important as we realize that <i>everybody</i> can (and should!) engage with mathematics, not just mathematicians or educators.<br /><br />So that might explain why a physics student might type "hbar = h / 2pi", and others would know that this clearly means \(\hbar=\dfrac{h}{2\pi}\) rather than \(\hbar=\dfrac{h}{2}\pi\). Remember, mathematics is not about just answer-getting. It's about communication of those ideas. And when the medium of communication limits how those ideas can be represented, the method of communication often changes to accomodate it.<br /><br />What the infamous problem points out is that while almost nobody has laid out any explicit rules for how to deal with concatenation, we seem to have developed some implicit ones, which we use without thinking about them. We just never had to deal with them until recently, as more "everyday" people communicate mathematics on more "everyday" media.<br /><br />Perhaps it's time that we address this convention explicitly and admit that <i style="font-weight: bold;">concatenation really has become a way to represent grouping</i>, just like parentheses or the vinculum. This is akin to taking a more descriptivist, rather than prescriptivist, approach to language: all we would be doing is recognizing that this is <i>already </i>how we do things everywhere else.<br /><br />Of course, this would throw a wrench in PEMDAS, but that just means we'd need to <a href="http://www.solidangl.es/2014/07/poorly-executed-mnemonics-definitely.html">actually talk about the mathematics behind it</a> rather than memorizing a silly mnemonic. After all, as inane as these internet math problems can be, they've shown that (whether they admit it or not) people really <i>do</i> want to get to the bottom of mathematics, to truly understand it.<br /><br />I'd say that's a good thing.<br /><br /><br /><i>* If your argument for why the answer is \(16\) starts with "Well, \(2(2+2)\) means \(2\cdot(2+2)\), so...", then you have missed the point entirely.</i>Bill Shillitohttp://www.blogger.com/profile/17774101901445053590noreply@blogger.com1tag:blogger.com,1999:blog-3748625510348961342.post-18947808448249202942019-01-12T19:19:00.002-08:002019-02-12T14:54:25.566-08:00Who Says You Can't Do That? --- Trig Identities<br />Ahh, trig identities... a rite of passage for any precalculus student.<br /><br />This is a huge stumbling block for many students, because up until this point, many have been perfectly successful (or at least have gotten by) in their classes by learning canned formulas and procedures and then doing a bunch of exercises that just change a \(2\) to a \(3\) here and a plus to a minus there. Now, all of a sudden, there's no set way of going about things. No "step 1 do this, step 2 do that". Now they have to rely on their intuition and "play" with an identity until they prove that it's correct.<br /><br />And to make matters worse, many textbooks --- and, as a result, many teachers --- make this subject arbitrarily and artificially harder for the students.<br /><br />They insist that students are not allowed to work on both sides of the equation, but instead must specifically start at one end and work their way to the other. I myself once subscribed to this "rule", because it's how I'd always been taught, and I always fed students the old line of "you can't assume the thing you're trying to prove because that's a logical fallacy".<br /><br />Then one of my Honors Precalculus students called me on it.<br /><br />He asked me to come up with an example of a trig non-identity where adding the same thing on both sides would lead to a false proof that the identity was correct. After some thought, I realized that not only couldn't I think of one, but that mathematically, there's no reason that one should exist.<br /><br />To begin with, one valid way to prove an identity is to work with each side of the equation <i>separately</i> and show that they are both equal to the same thing. For example, suppose you want to verify the following identity:<br /><br />\[\dfrac{\cot^2{\theta}}{1+\csc{\theta}}=\dfrac{1-\sin{\theta}}{\sin{\theta}}\]<br />Trying to work from one side to the other would be a nightmare, but it's much simpler to show that each side is equal to \(\csc{\theta}-1\). This in fact demonstrates one of the oldest axioms in mathematics, as written by Euclid: "things which are equal to the same thing are equal to each other."<br /><br />But what about doing the same thing to <i>both</i> sides of an equation?<br /><br />There are two important points to realize about what's going on behind the scenes here.<br /><br />The first is that if your "thing you do to both sides" is a <i>reversible step</i> --- that is, if you're applying a <i>one-to-one function</i> to both sides of an equation --- then it's perfectly valid to use that as part of your proof because it establishes an <i>if-and-only-if</i> relationship. If that function is not one-to-one, all bets are off. You can't prove that \(2=-2\) by squaring both sides to get \(4=4\), because the function \(x\mapsto x^2\) maps multiple inputs to the same output.<br /><br />It baffles me that most Precalculus textbooks mention one-to-one functions in the first chapter or two, yet completely fail to understand how this applies to solving equations.* A notable exception is UCSMP's Precalculus and Discrete Mathematics book, which establishes the following on p. 169:<br /><b><br /></b><br /><blockquote class="tr_bq"><u style="font-weight: bold;">Reversible Steps Theorem</u></blockquote><blockquote class="tr_bq">Let \(f\), \(g\), and \(h\) be functions. Then, for all \(x\) in the intersection of the domains of functions \(f\), \(g\), and \(h\),</blockquote><blockquote class="tr_bq"><ol><li>\(f(x)=g(x) \Leftrightarrow f(x)+h(x)=g(x)+h(x)\)</li><li>\(f(x)=g(x) \Leftrightarrow f(x)\cdot h(x)=g(x)\cdot h(x)\) <i>[We'll actually come back to this one in a bit -- there's a slight issue with it.]</i></li><li>If \(h\) is 1-1, then for all \(x\) in the domains of \(f\) and \(g\) for which \(f(x)\) and \(g(x)\) are in the domain of \(h\), \[f(x)=g(x) \Leftrightarrow h(f(x))=h(g(x)).\]</li></ol></blockquote><br />Later on p. 318, the book says:<br /><br /><blockquote class="tr_bq"><b><i>"...there is no new or special logic for proving identities. Identities are equations and all the logic that was discussed with equation-solving applies to them."</i></b></blockquote><br />Yes, that whole "math isn't just a bunch of arbitrary rules" thing applies here too.<br /><br />The second important point, which you may have noticed while looking at the statement of the Reversible Steps Theorem, is that the implied <i>domain</i> of an identity matters a great deal. When you're proving a trig identity, you are trying to establish that it is true for all inputs that are in the <i>domain</i> of both sides. Most textbooks at least pay lip service to this fact, even though they don't follow it to its logical conclusion.<br /><br />To illustrate why domain is so important, consider this example:<br /><br />\[\dfrac{\cos{x}}{1-\sin{x}} = \dfrac{1+\sin{x}}{\cos{x}}\]<br />To verify this identity, I'm going to do something that may give you a visceral reaction: I'm going to "cross-multiply". Or, more properly, I'm going to multiply both sides by the expression \((1 - \sin x)\cos x\). I claim that this is a perfectly valid step to take, and what's more, it makes the rest of the proof downright easy by reducing to everyone's favorite Pythagorean identity:<br /><br />\[<br />\begin{align*}<br />(\cos{x})(\cos{x}) &= (1+\sin{x})(1-\sin{x})\\<br />\cos^2{x} &= 1-\sin^2{x}\\<br />\sin^2{x} + \cos^2{x} &= 1 \quad\blacksquare<br />\end{align*}<br />\]<br />"But wait," you ask, "what if \(x=\pi/2\)? Then you're multiplying both sides by zero, and that's certainly not reversible!"<br /><br />True. But if \(x=\pi/2\), then the denominators of both sides of the equation are zero, so the identity isn't even true in the first place. For any value of \(x\) that does <i>not</i> yield a zero in either denominator, though, multiplying both sides of an equation by that value is a reversible operation and therefore completely valid.<br /><br />Now, this isn't to say that multiplying both sides of an equation by a function can't lead to problems --- for example, if \(h(x)=0\) (as in the zero <i>function</i>), then \(f(x)\cdot h(x)=g(x)\cdot h(x)\) no matter what. This can even lead to problems in more subtle cases: suppose \(f\) and \(g\) are equal everywhere but a single point \(a\); for example, perhaps \(f(a)=1\) and \(g(a)=2\). If it just so happens that \(h(a)=0\), then \(f\cdot h\) and \(g\cdot h\) will be equal <i>as functions</i>, even though \(f\) and \(g\) are not themselves equal.<br /><br />The real issue here can be explained via a quick foray into higher mathematics. Functions form what's called a <u style="font-weight: bold;">ring</u> -- basically meaning you can add, subtract, and multiply them, and these operations have all the nice properties we'd expect. But being able to preserve that if-and-only-if relationship when multiplying a function by both sides of an equation requires a special kind of ring called an <u style="font-weight: bold;">integral domain</u>, which means that it's impossible to multiply two nonzero functions together and get a zero function.<br /><br />Unfortunately, functions in general don't form an integral domain --- not even continuous functions, or differentiable functions, or even <i>infinitely</i> differentiable functions do! But if we move up to the <i>complex numbers</i> (where everything works better!), then the set of <u style="font-weight: bold;">analytic</u> functions --- functions that can be written as power series (infinite polynomials) on an open domain --- <i>is </i>an integral domain. And most of the functions that precalculus students encounter generally turn out to be analytic**: polynomial, rational, exponential, logarithmic, trigonometric, and even inverse trigonometric. This means that when proving trigonometric identities, multiplying both sides by the same function is a "safe" operation.<br /><br />So in sum, when proving trigonometric identities, as long as you're careful to only use reversible steps (what a great time to spiral back to one-to-one functions, by the way!), you are welcome to apply all the same algebraic operations that you would when solving equations, and the chain of equalities you establish will prove the identity. Even "cross-multiplying" is fair game, because any input that would make the denominator zero would invalidate the identity anyway.*** Since trigonometric functions are generally "safe" (analytic), we're guaranteed to never run into any issues.<br /><br />Now, none of this is to say that there isn't intrinsic merit to learning how to prove an identity by working from one side to the other. Algebraic "tricks" --- like multiplying by an expression over itself (\(1\) in disguise!) to conveniently simplify certain expressions --- are important tools for students to have under their belts, especially when they encounter limits and integrals next year in calculus.<br /><br />What we need to do, then, is encourage our students to come up with multiple solution methods, and perhaps present working from one side to the other as an added challenge to build their mathematical muscles. And if students are going to work on both sides of an equation at once, then we need to hold them to high standards and make them <i>explicitly</i> state in their proofs that all the steps they have taken are reversible! If they're unsure on whether or not a step is valid, have them investigate it until they're convinced one way or the other.<br /><br />If we're artificially limiting our students by claiming that only one solution method is correct, we're sending the wrong message about what mathematics really is. Instead, celebrating and cultivating our students' creativity is the best way to prepare them for problem-solving in the real world.<br /><br />--<br /><br /><i>* Rather, I <b>would</b> say it baffles me, but actually I'm quite used to seeing textbooks treat mathematical topics as disparate and unconnected, like how a number of Precalculus books teach vectors in one chapter and matrices in the next, yet never once mentione how they are so beautifully tied together via transformations.</i><br /><i><br /></i><i>** Except perhaps at a few points. The more correct term for rational functions and certain trigonometric functions is actually <u style="font-weight: bold;">meromorphic</u>, which describes functions that are analytic everywhere except a discrete set of points, called the <u style="font-weight: bold;">poles</u> of the function, where the function blows up to infinity because of division by zero.</i><br /><br /><i>*** If you extend the domains of the trig functions to allow for division by zero, you do need to be more careful. <a href="https://1dividedby0.com/">Not because there's anything intrinsically wrong with dividing by zero</a>, but because \(0\cdot\infty\) is an indeterminate expression and causes problems that algebra simply can't handle.</i>Bill Shillitohttp://www.blogger.com/profile/17774101901445053590noreply@blogger.com0tag:blogger.com,1999:blog-3748625510348961342.post-53575738391182138112019-01-02T20:23:00.002-08:002019-01-02T21:42:13.891-08:00When Math is in Jeopardy!<i>(Forgive the somewhat dramatic title, but it was too good to pass up.)</i><br /><div><br /></div><div>I absolutely love game shows. It's always fun to imagine you're there on stage ... or at least to yell at the people on TV when they don't know a question that was just <i>so easy</i>!</div><div><br /></div><div>One of my favorite game shows, of course, is Jeopardy --- it has an air of intellectualism about it, with such an eclectic collection of topics. I don't even mind that I'm admittedly pretty terrible at it... I always find myself thinking, "Man, if they just had some <i style="font-weight: bold;">math</i> questions, I could knock those out of the park!"</div><div><br /></div><div>Well, a couple of days ago, as we were getting ready to celebrate the advent of 2019, I was elated to see that there was in fact a math question! (Or math answer, rather. You know what, I'm just going to call them "clue" and "response" to avoid confusion.)</div><div><br /></div><div>The category was "Ends in 'ITE'", and the <strike>question</strike> clue was as follows:</div><blockquote class="tr_bq"><b>"In math, when the number of elements in a set is countable, it's this type of set."</b></blockquote><table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"><tbody><tr><td style="text-align: center;"><img alt="Image may contain: text" height="271" src="https://scontent.fatl1-1.fna.fbcdn.net/v/t1.0-9/48921855_10103114694184809_3192423146435117056_o.jpg?_nc_cat=109&_nc_ht=scontent.fatl1-1.fna&oh=e7356f79462113f7836aa29a92d2f373&oe=5CC2D3F5" style="margin-left: auto; margin-right: auto;" width="400" /></td></tr><tr><td class="tr-caption" style="text-align: center;"><i>Thanks to @GoGoGadgetGabe on Twitter for the screenshot!</i></td></tr></tbody></table>As I was trying to think of what <strike>answer</strike> response would end in those three letters, a contestant buzzed in and said:<br /><div><blockquote class="tr_bq"><b>"What is 'finite'?"</b></blockquote>Alex Trebek notified the contestant that they were correct.<br /><br />Meanwhile, I was speechless because I knew they weren't.<br /><br />Well, at least not completely correct. See, in math, words have very specific definitions in order to describe very specific phenomena. In this case, the word "countable" is used in set theory to describe a set --- a collection of things --- that can be put into one-to-one correspondence with a subset of the natural numbers, \(\mathbb{N}=\{0,1,2,3,\ldots\}\).* So, while finite sets are indeed countable, there are also infinite countable sets --- the integers, the even numbers, and the rational numbers are all well-known examples. (That last one still amazes me --- in some sense, there are exactly as many fractions as there are whole numbers, even though it seems like the former should outnumber the latter!)<br /><br />That means that the statement "When the number of elements in a set is countable, it's a finite set" is actually incorrect.<br /><br />Naturally, I took to the internet to voice my displeasure with how my favorite subject was represented on national television. When talking with a friend of mine (who knows more about game shows than I ever will), I learned that there had been two other cases recently where math-centric Jeopardy questions had issues with them in the past couple of months.<br /><ul><li>In one case, the clue was, <b>"If \(x^2+2=18\), then \(x\) equals this."</b> The response that was judged to be correct was <b>"What is </b>'<b>\(4\)'?"</b>. Any algebra teacher reading this right now is shaking their head, because that answer won't get you full credit on any test. There are two possible values of \(x\): \(4\) and \(-4\).</li><li>In another case, the response to the clue was supposed to be <b>"What is the commutative property?"</b>. Nobody got it correct (which makes me sad in itself), but when Alex Trebek read the correct response out loud, he said it as "COM-myoo-TAY-tive" instead of "com-MYOO-tuh-TIVE".</li></ul>Now, in isolation, any one of these would be only a minor annoyance. After all, there are plenty of clues on Jeopardy that have multiple possible correct responses, and contestants aren't expected to give all correct responses, but rather just one.<br /><br />But the fact that questions about mathematics appear so infrequently on the show compared to topics such as history, combined with the fact that these kinds of details are not attended to, seems to send a message that mathematics is considered to be not as important, not worth researching fully in the spirit of the subject.<br /><br />We already live in a culture in which any time I tell somebody I teach math, the inevitable response is "Oh, haha, I was never any good at math." Somehow people seem proud to admit and even proclaim this. I'm willing to wager (maybe even make this a true Daily Double) that those people would be much more reluctant to say something like "Oh, haha, I was never any good at reading." There's a pervasive attitude that mathematics is a torturous and frivolous subject, devoid of the <a href="https://medium.com/q-e-d/stop-selling-math-for-its-usefulness-d9143e80d78d">awe-inspiring beauty and sheer fun</a> that those who embrace the subject know it to have.<br /><br />With that said, I'd like to challenge the writers of Jeopardy --- and perhaps other game shows as well --- to make a conscious effort not only to ask more questions about mathematics, but to take care to do them well, perhaps even consulting one or more mathematicians to make sure the precision and nuance of the subject are properly represented. (I know math teachers who have come up with versions of game shows for their classes with only mathematically-oriented questions... the students love it!)<br /><br />It doesn't have to be something like "\(1\times 2\times 3\times 4\times 5\)"** either. There's such a rich amount of material to pull from --- why not ask questions about, say, fractals? Or famous mathematicians? Or even unsolved problems (and those who eventually solved them)? I would be giddy to see something like,<br /><blockquote class="tr_bq"><b>"Shot and killed in a duel when he was only twenty, this mathematician spent the last night of his life writing down what he'd discovered about quintic equations."</b></blockquote>Doesn't that make you want to find out what the story was, why he felt fifth-degree equations were so important that he just had to share it, knowing he would soon die? Mathematics is full of stories like this, and perhaps letting people know those stories exist, that there's more to math than doing arithmetic problems, might change how people view the subject.<br /><br />Of course, if winning prize money is what you like about game shows, there's always the million-dollar <a href="http://www.claymath.org/millennium-problems/millennium-prize-problems">Millennium Prize Problems</a>...<br /><br />--<br /><br /><i>* You might be thinking, "But zero isn't a natural number!" As it turns out, there's no real consensus on whether zero is considered a natural number. Some mathematicians choose to include it, while others don't. To some extent it depends on your field of study --- for example, number theorists may be more likely to disinclude zero because it doesn't play very nicely with things like prime factorizations, while computer scientists are used to counting from zero instead of one. Peano actually started off his axiomatization of the natural numbers with one being the starting point, but then changed his mind later and started with zero!</i><br /><i><br /></i><i>** This was actually a clue on a Kid's Jeopardy episode, under the category "Non-Common Core Math". That will eventually be the impetus for another blog post in the future on the way we currently view mathematics teaching.</i></div>Bill Shillitohttp://www.blogger.com/profile/17774101901445053590noreply@blogger.com2tag:blogger.com,1999:blog-3748625510348961342.post-21107092009774900532017-10-21T17:23:00.000-07:002017-10-21T17:27:48.187-07:00Discreet Discrete CalculusOver the past week, I went to the Georgia Mathematics Conference (GMC) at Rock Eagle, held by the Georgia Council of Teachers of Mathematics (GCTM). The GMC is one of the events I look forward to most every year --- tons of math educators and advocates sharing lessons, techniques, and ideas about how to best teach math to students from kindergarten through college. I always enjoy sharing my own perspectives as well (even when they do get a bit bizarre!)<br /><br />This time, I got to share the results of a lesson that I guinea-pigged on my Honors Precalculus class last year, where they explored the relationships between polynomial sequences, common differences, and partial sums. The presentation from the GMC uses the techniques we looked at to develop the formula for the sum of the first \(n\) perfect squares:<br /><br />\[1^2+2^2+3^2+\cdots+n^2=\frac{n(n+1)(2n+1)}{6}\]<br /><br />At the GMC, we did go a bit further than my class did --- they didn't do the full development of discrete calculus --- but some knowledge of where the ideas lead is never a bad thing, and a different class at a different school may even be able to go further!<br /><br /><b><a href="https://drive.google.com/open?id=0B-uVGGkZosoPM0kwMllTVzJPQmc">Here is the PowerPoint from the presentation.</a></b> If you find it useful, or have any questions, please don't hesitate to leave a comment!<br /><br /><i>(I recommend downloading the PPTX file and viewing the slide show in PowerPoint, rather than using Google's online viewer.)</i>Bill Shillitohttp://www.blogger.com/profile/17774101901445053590noreply@blogger.com3tag:blogger.com,1999:blog-3748625510348961342.post-14626193589859870922017-09-15T14:12:00.000-07:002017-09-15T21:38:25.756-07:00I'm Done with Two Column Proofs.Wow.<br /><br />It really has been a while since I've posted here, hasn't it?<br /><br />But I suppose having a mini-crisis in my geometry class that has forced me to reject our textbook's philosophy on what "proofs" should be is a good enough reason to resurrect this blog. I finally have the time this year, and I feel the need to share my probably overly opinionated beliefs about math education.<br /><br />This is the first year I have tried to integrate proofs into our school's geometry curriculum across the board. (In the past, proofs were only discussed in honors classes, but I felt somehow that reasoning through how you knew something was true was important for everyone.) I was trying to justify the two-column format, as much as I hate it, as a way to "scaffold" student thinking --- yay educational buzzwords! But when I actually did it, it got exactly the reaction that I knew it would --- it just served to overly obfuscate the material and utterly drain the life out of it. I realized I should have stuck to my guns and listened to the likes of <a href="https://www.maa.org/external_archive/devlin/LockhartsLament.pdf">Paul Lockhart</a> and <a href="https://mathwithbaddrawings.com/2013/10/16/two-column-proofs-that-two-column-proofs-are-terrible/">Ben Orlin</a>.<br /><br />So, after some reflection and course correction, here's the email I just sent my students.<br /><br />---<br /><blockquote class="tr_bq">Hello mathematicians. I have a rather bizarre request. </blockquote><blockquote class="tr_bq">Don't do your geometry homework this weekend. </blockquote><blockquote class="tr_bq">Yes, you read that right. Don't. </blockquote><blockquote class="tr_bq">Let me explain.</blockquote><blockquote class="tr_bq">We've spent the past couple of days looking at "proofs" in geometry. The reason I say "proofs" in quotations is that, in all honesty, I don't believe the two-column proofs that our book does are are all that useful. I have actually been long opposed to them, but against my better judgment, decided to give them a shot anyway and make them sound reasonable. But you know what they say ... if you put lipstick on a chazir*, it's still a chazir. (They do say that, right?) The thing is, that style of proof just ends up sounding like an overly repetitive magical incantation rather than an actual logical argument --- as some of you pointed out in class today. I truly do value that honesty, by the way, and I hope you continue to be that honest with me. </blockquote><blockquote class="tr_bq">Here is what I actually will expect of you going forward. It's quite simple: </blockquote><blockquote class="tr_bq">I expect you to be able to tell me how you know something is true, and back it up with evidence. </blockquote><blockquote class="tr_bq">That's it. </blockquote><blockquote class="tr_bq">It may be a big-picture kind of question, or it may be telling me how we get from step A to step B, but when it really comes down to it, it's all just "here's why I know this is true, based on this evidence". It doesn't have to be some stilted-sounding name like the "Congruent Supplements Theorem" either --- just explain it in your own words. That doesn't mean that any explanation is correct --- it still has to be valid mathematical reasoning. You can't tell me that two segments on a page are congruent because they're drawn in the same color, or something silly like that. But it doesn't have to be in some prescribed way --- just as long as you show me you really do understand it. </blockquote><blockquote class="tr_bq">With that in mind, by the way, I'm also not going to be giving you a quiz on Monday, either. Instead, we're going to focus on how to make arguments that are a lot more convincing than just saying the same thing in different words. I think you'll find that Monday's class will make a lot more sense than the past few classes combined. </blockquote><blockquote class="tr_bq">So, relax, take a much-deserved Shabbat, and when we come back, I hope to invite you to see geometry the way I see it --- not as a set of arbitrary rules, but as something both logical and beautiful. </blockquote><blockquote class="tr_bq">Shabbat Shalom.</blockquote><br /><i>* I teach at a Jewish private school. "Chazir" is Hebrew for "pig", which has the added bonus of being non-Kosher. Two-column proofs are treif... at least in the context of introductory geometry.</i><br /><i><br /></i><i>P.S. I am not saying that two-column proofs NEVER have a place in mathematics. I am merely saying that introductory geometry, when kids are still getting used to much of geometry as a subject, is not the proper place to introduce the building of an axiomatic system. Save that for later courses for the students who choose to become STEM majors.</i>Bill Shillitohttp://www.blogger.com/profile/17774101901445053590noreply@blogger.com1tag:blogger.com,1999:blog-3748625510348961342.post-69831562498495885932015-08-14T20:22:00.000-07:002015-08-14T20:22:40.357-07:00A Real-Life Paradox: The Banach-Tarski Burrito<b><i>Who knew the Axiom of Choice could help me decide whether to get guacamole for an extra $1.95?</i></b><br /><br />A couple of weeks ago, the popular YouTube channel <a href="https://www.youtube.com/user/Vsauce">Vsauce</a> released a video that tackles what it details as “one of the strangest theorems in modern mathematics”: the <a href="https://en.wikipedia.org/wiki/Banach%E2%80%93Tarski_paradox">Banach-Tarski Paradox</a>. In the video, Michael Stevens explains how a single sphere can be decomposed into peculiar-looking sets, after which those sets can be recombined to form two spheres, each perfectly identical to the original in every way. If you haven’t had a chance to watch the video, go ahead and do so here:<br /><br /><div class="separator" style="clear: both; text-align: center;"><iframe width="320" height="266" class="YOUTUBE-iframe-video" data-thumbnail-src="https://i.ytimg.com/vi/s86-Z-CbaHA/0.jpg" src="https://www.youtube.com/embed/s86-Z-CbaHA?feature=player_embedded" frameborder="0" allowfullscreen></iframe></div><br />Although this seems like a purely theoretical abstraction of mathematics, the video leaves us wondering if perhaps there could be a real-world application of such a bizarre phenomenon. Stevens asks, “is [the Banach-Tarski paradox] a place where math and physics separate? We still don’t know … The Banach-Tarski Paradox could actually happen in our real world … some scientists think it may be physically valid.”<br /><br />Well, my friends, I would like to make the bold claim that I have indeed discovered a physical manifestation of this paradox.<br /><br />And it happened a few years ago at my local <a href="http://www.chipotle.com/">Chipotle</a>.<br /><br />Let me start off by saying this: I <b><i>love</i> </b>Chipotle. It’s a particularly good day for me when I walk in and get my burrito with brown rice, fajita veggies, steak, hot salsa, cheese, pico de gallo, corn, sour cream, guacamole (yes I know it’s extra, just put it on my burrito already!), and a bit of lettuce. No chips, Coke, and about a half hour later I’m one happily stuffed math teacher.<br /><br />The only thing that I don’t like about Chipotle is that the construction of said burritos often ends up failing at the most crucial step – the rolling into one coherent, tasty package. Given the sheer amount of food that gets crammed into a Chipotle burrito, it’s unsurprising that they eventually lose their structural integrity and burst, somewhat defeating the purpose of ordering a burrito in the first place.<br /><br />If you have ever felt the pain of seeing your glorious Mexican monstrosity explode with toppings like something out of an Alien movie because of an unlucky burrito-roller, you have probably been offered the opportunity to “double-wrap” your burrito for no extra charge, giving it an extra layer of tortilla to ensure the safe deliverance of guacamole-and-assorted-other-ingredients into your hungry maw.<br /><br />Now, being a mathematically-minded kind of guy, I asked the employee who made me this generous offer:<br /><br />“Well, could I just get my ingredients split between two tortillas instead?”<br /><br />The destroyer-of-burritos gave that look that you always get from anybody who works at a business that bandies about words like “company policy” when they realize they have to deny a customer’s request even in the face of logic, and said:<br /><br />“If you do that, we’ll have to charge you for two burritos.”<br /><br />I was dumbfounded.<br /><br />“Wait … so you’re saying that if you put a second tortilla around my burrito, you’ll charge me for one burrito, but if you rearrange the <i>exact same ingredients</i>, you’ll charge me for two?”<br /><br />“Yes sir – company policy.”<br /><br />Utterly defeated, I begrudgingly accepted the offer to give my burrito its extra layer of protection, doing my best to smile at the girl who probably knew as well as I did the sheer absurdity of the words that had come out of her mouth. I paid the cashier, let out an audible “oof” as I lifted the noticeably heavy paper bag covered with trendy lettering, and exited the store.<br /><br />When I arrived home, I took what looked like an aluminum foil-wrapped football out of the bag (which was a great source of amusement for my housemates), laid it out on the kitchen table, and decided to dismantle the burrito myself and arrange it into two much more manageable Mexican morsels. I wondered whether I should have done this juggling of ingredients right there at Chipotle, just to see whether the staff’s heads would explode. <br /><br />It was in that moment, with my head still throbbing from the madness of the entire experience, that I began to realize what had just happened. How <i>was</i> it possible that a given mass of food could cost one amount one moment and another amount the next? I immediately began to deconstruct my burrito, laying out the extra tortilla onto a plate and carefully making sure that precisely one-half of the ingredients – especially the guacamole – found their way into their new home. As I carefully re-wrapped both tortillas, my suspicions were confirmed. Sitting right in front of me were two delicious burritos, each identical in price to my original.<br /><br />I had discovered the Banach-Tarski Burrito.Bill Shillitohttp://www.blogger.com/profile/17774101901445053590noreply@blogger.com3tag:blogger.com,1999:blog-3748625510348961342.post-2272902034172907032015-04-24T17:47:00.000-07:002015-04-28T14:00:36.302-07:00A Radical New Look for Logarithms<i>"A good notation has a subtlety and suggestiveness which at times make it almost seem like a live teacher."</i> — Bertrand Russell<br /><br /><i>"We could, of course, use any notation we want; do not laugh at notations; invent them, they are powerful. In fact, mathematics is, to a large extent, invention of better notations."</i> — Richard Feynman<br /><br />"<i>By relieving the brain of all unnecessary work, a good notation sets it free to concentrate on more advanced problems, and in effect increases the mental power of the race.</i>" — Alfred North Whitehead<br /><br /><br />Notation is perhaps one of the most important aspects of mathematics. The right choice of notation can make a concept clear as day; the wrong choice can make extracting its meaning hopeless. Of course, one great thing about notation is that even if there's a poor choice of notation out there (such as \(\left[x\right]\) or \(\pi\)), often someone comes along and creates a better one (such as \(\lfloor x\rfloor\) for the <a href="http://www.solidangl.es/2014/09/throw-outdated-notation-to-floor.html">floor function</a> or multiples of <a href="http://tauday.com/tau-manifesto">tau</a>, \(\tau\approx 6.28318\), for radian measure of angles).<br /><br />Which brings me to one such poor choice of notation, one that I believe needs fixing: the rather asymmetrical notation of powers, roots, and logarithms.<br /><br />Here we have three very closely related concepts — both roots and logarithms are ways to invert exponentiation, the former returning the base and the latter returning the exponent. And yet their notation couldn't be more different:<br />\[2^3=8\\<br />\sqrt[3]{8}=2\\<br />\log_2{8}=3\]This always struck me as annoyingly inelegant. Wouldn't it be nice if these notations bore at least some resemblance to each other?<br /><br />After giving it some thought, I believe I have found a possible solution. As an alternative to writing \(\log_2 {8}\), I propose the following notation:<br /><br /><div class="separator" style="clear: both; text-align: center;"><a href="http://4.bp.blogspot.com/-vyIGboQ3DkY/VTrVNLnYilI/AAAAAAAACmA/eAkvCZAFl3Y/s1600/new%2Blog%2Bnotation.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" src="http://4.bp.blogspot.com/-vyIGboQ3DkY/VTrVNLnYilI/AAAAAAAACmA/eAkvCZAFl3Y/s1600/new%2Blog%2Bnotation.png" /></a></div><br />This notation makes use of a reflected radical symbol, such that the base of the logarithm is written in a similar manner to the index of a radical but below the "point" (the pointy part of the radical symbol), and the argument of the logarithm is written "inside". The use of this notation has a number of advantages:<br /><br /><ol><li>The symmetry between the normal radical for roots and the reflected radical for logarithms highlights both their similarities and their differences — each one "undoes" an exponential expression, but each one gives a different part of the expression (the base and the exponent, respectively.)</li><li>The radical symbol can be looked at as a modified lowercase letter "r". (This may actually be the origin of the symbol, where the "r" stands for <i>radix</i>, the Latin word for "root".) In a similar way, the new symbol for logarithms resembles a capital "L".</li><li>The placement of the "small number" and the "point" can take on a secondary spatial meaning: </li><ul><li>The "small number" represents a piece of information we <u>know</u> about an exponential expression, and its placement indicates <u>which part</u> we know.</li><ul><li>For a root, the "small number" is on top, so <i>we know the <u>exponent</u></i>.</li><li>For a logarithm, the "small number" is on bottom, so <i>we know the <u>base</u></i>.</li></ul><li>The symbol seems to "point" to the piece of information that we are <u>looking for</u>.</li><ul><li>For a root, the "point" is pointing downward, so <i>we are looking for the <u>base</u></i>.</li><li>For a logarithm, the "point" is pointing upward, so <i>we are looking for the <u>exponent</u></i>.</li></ul><li>Looking at the image above, the new notation seems to say "We know the <i>base</i> is 2, so <i>what's the exponent</i> that will get us to 8?"</li><li>Similarly, the expression \(\sqrt[3]{8}\) now can be interpreted as saying "We know the <i>exponent</i> is 3, so <i>what's the base</i> that will get us to 8?"</li></ul></ol><div><div><br /></div><div>This notation would obviously not make much of a difference for seasoned mathematicians who are perfectly comfortable with the \(\log\) and \(\ln\) functions. But from a pedagogical standpoint, the reflected radical, with its multi-layered meaning and auto-mnemonic properties, could help students become more comfortable with a concept that many look at as just meaningless manipulation of symbols.<br /><br />When I first came up with this reflected-radical notation, I had originally imagined that it should <i>replace</i> the current notation. However, after some feedback from various people and some further consideration, I think a better course of action would be to have this notation be used <i>alongside</i> the current notation, much in the way that we have multiple notations for other concepts in math (such as the many ways to write derivatives). However, I would suggest that, if it were to become commonplace*, this notation would be best to use when first introducing the concept in schools. The current notation isn't <i>wrong</i> per se — it's just not very evocative of the underlying concept. Anything that can better elucidate that concept can't be a bad thing when it comes to students learning mathematics!</div><div><br /></div><div>It may seem like a <b>radical </b>idea.</div><div>But it's a <b>logical</b> one.<br /><br /><br /><i>* Of course, for this notation to become commonplace, somebody would need to figure out how to replicate it in LaTeX. Any takers?</i></div></div>Bill Shillitohttp://www.blogger.com/profile/17774101901445053590noreply@blogger.com11tag:blogger.com,1999:blog-3748625510348961342.post-68697028658403006552014-09-15T20:31:00.000-07:002014-09-18T05:29:00.808-07:00Infinity is my favorite number.<br /><br />Yes, you read that right.<br /><br />I've recently been embroiled in a lovely debate on <a href="http://numberphile.com/">Numberphile</a>'s video, "Infinity is bigger than you think", in which Dr. James Grime starts off: "We're going to break a rule. We're breaking one of the rules of Numberphile. We're talking about something that isn't a number. We're going to talk about infinity."<br /><br /><div class="separator" style="clear: both; text-align: center;"><iframe allowfullscreen='allowfullscreen' webkitallowfullscreen='webkitallowfullscreen' mozallowfullscreen='mozallowfullscreen' width='320' height='266' src='https://www.youtube.com/embed/elvOZm0d4H0?feature=player_embedded' frameborder='0' /></div><br /><br />I, too, was a longtime believer of what high school students all over are told: "Infinity is not a number; infinity is a concept." As my studies of mathematics progressed, however, I began to see that perhaps the things I had always taken for granted were not as black-and-white as they had seemed. There was a lot more nuance to mathematics than I had ever realized, and learning those nuances opened up an entire new level of understanding, unlocking all sorts of links between concepts that had previously seemed worlds apart. So it's no wonder that "Infinity Is Not A Number" (which I will occasionally abbreviate as "IINAN") was one of the first claims to which I took a fine-tooth comb. What I learned changed my stance on infinity and firmly cemented it as my favorite number - not just concept, but honest-to-god <i>number</i>.*<br /><br />The most common argument made IINAN proponents involves the curious property that \(\infty +1=\infty\). This, they say, leads to all sorts of contradictions, because all one has to do is simply subtract \(\infty\) from both sides:<br />\[<br />\infty+1=\infty\\<br />\underline{-\infty\ \ \ \ \ \ \ \ -\infty}\ \ \\<br />\ \ \ \ \ \ 1=0\]Oh no! We know that the statement \(1=0\) is obviously false, so there must be a false assumption somewhere. Many IINAN defenders claim that the false assumption was that we tried to treat \(\infty\) as a number. But that's not actually where the problem with infinity lies.<br /><br />The problem is that we tried to do algebra with it.<br /><br />For mathematicians, the most convenient place to do algebra is in a structure called a <b>field</b>. If you're already familiar with what a field is, great, but if not, you can think of a field as a number system in which the age-old operations of addition, subtraction, multiplication, and division <span style="color: #545454; font-family: arial, sans-serif; font-size: x-small;"><span style="line-height: 18.2000007629395px;">—</span></span> the four operations that my father often notes are the only ones he ever needs when I talk about the kinds of math I teach <span style="color: #545454; font-family: arial, sans-serif; font-size: x-small;"><span style="line-height: 18.2000007629395px;">—</span></span> work exactly as we'd like them to. The fields with which we are most familiar are the rational numbers (\(\mathbb{Q}\)), the badly-named so-called "real" numbers (\(\mathbb{R}\)), and often the complex numbers (\(\mathbb{C}\)). One basic thing about a field is that the <b>subtraction property of equality</b> holds: For any numbers \(a\), \(b\), and \(c\) in our field, if \(a=b\), then \(a-c=b-c\).<br /><br />What about \(\infty\) though? When we attempted to use \(\infty\) in an algebra problem, we got back complete garbage. And we know that the subtraction property of equality should hold for <span style="font-style: italic;">any</span> numbers in a field. What this means, then, is that \(\infty\) <i>is not part of that field</i> (or any field as far as I'm aware). So, when someone says "Infinity is not a number", what they <i>really</i> mean is <b>"Infinity is not a </b><i style="font-weight: bold;">real </i><b>number."</b><i style="font-weight: bold;"> </i>(It's not a complex number, either, for that matter.) It doesn't follow the same rules that the real numbers do.<br /><br />But that doesn't mean it's not a number at all.<br /><br />We've seen this sort of thing happen before. The Greek mathematician Diophantus, when faced the equation \(4x+20=0\), called its solution of \(-5\) "absurd" — yet now students learn about negative numbers as early as elementary school, and we barely blink an eye at their use in everyday life. Square roots of negative numbers seemed equally preposterous to the Italian mathematician Gerolamo Cardano, and the French mathematician René Descartes called them "imaginary", a term that we're unfortunately stuck with today. But imaginary numbers — and the complex numbers we build from them — are a vital part of physics, from alternating currents to quantum mechanics.<br /><br />So what makes infinity any different from \(i\)?<br /><br />Sure, it seems bizarre that a number plus one could equal itself. But it's equally bizarre that the square of a number could be negative. And sure, we can get a contradiction if we do certain things to infinity. But that happens with \(i\) as well! If we attempt to use the identity \(\sqrt{a}\cdot\sqrt{b}=\sqrt{ab}\), we can arrive at a similar contradiction:<br />\[\sqrt{-1}\cdot\sqrt{-1}=\sqrt{-1\cdot-1}\\<br />\ \ \ i\cdot i=\sqrt{1}\\<br />-1=1\ \]When this equation fails, you don't see mathematicians clamoring that "\(i\) isn't a number"! Instead, the response is that the original equation doesn't work like we thought it did when we extend our real number system to include the complex numbers — instead, the square root function takes on a new life as a multi-valued function. There's that nuance again! For the same reason, infinity makes us look closer at something as simple as subtraction, at which point we find that \(\infty-\infty\) is an <b>indeterminate form</b>, something that we need the tools of calculus to properly deal with.<br /><br />The truth is, mathematicians have been treating \(\infty\) as a number** for quite some time now.<br /><br />In <b>real analysis</b>, which was developed to give the techniques of calculus a rigorous footing, points labelled \(+\infty\) and \(-\infty\) can be added to either end of the real number line to give what we call the <b>extended real number line</b>, often denoted \(\overline{\mathbb{R}}\) or \(\left[-\infty,+\infty\right]\). The extended real number line is useful in describing concepts in measure theory and integration, and it has algebraic rules of its own, though analysts are still careful to mention that these two extra points are not <i>real</i> numbers. What's more, the extended real line is not a <i>field</i>, because it doesn't satisfy all the nice properties that a field does. (But that just makes us appreciate working in a field that much more!)<br /><br /><b>Projective geometry</b> gives us a different sort of infinity, what I like to call an "unsigned infinity", one that is obtained by letting \(-\infty\) and \(+\infty\) overlap and creating what is known as the <b>real projective line</b>. And <b>complex analysis</b>, which extends calculus to the complex plane,<b> </b>takes it even further, letting <i>all</i> the different infinities in all directions overlap to create a sort of "complex infinity", sometimes written \(\tilde{\infty}\), sitting atop the <b>Riemann sphere</b>. What I particularly like about these projective infinities is that, using them, <i>you can actually divide by zero!</i> ***<br /><br />So, since there are actually a number of different kinds of infinity that can be referred to, I would say that, more specifically, <i>complex infinity is my favorite number</i>.<br /><br />The tough thing about this situation is that the concept of "number" is a very difficult one to precisely and universally define — similar to how linguists still struggle to come up with a universal definition of "word". By trying to come up with such a description, you end up either including things that you don't want to be numbers (such as matrices) or excluding things that you do want to be numbers (such as complex numbers). The best we can really do is keep an open mind about what a "number" is.<br /><br />After all, there's infinitely many of them already — so there's bound to be new ones we haven't seen yet sooner or later.<br /><br /><div class="separator" style="clear: both; text-align: center;"><a href="http://3.bp.blogspot.com/-GCL0xZc_dCI/VBeqZtjo_nI/AAAAAAAAATM/HSzchObC2l4/s1600/Infinity.jpg" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" src="http://3.bp.blogspot.com/-GCL0xZc_dCI/VBeqZtjo_nI/AAAAAAAAATM/HSzchObC2l4/s1600/Infinity.jpg" height="179" width="320" /></a></div><br /><br />∎<br /><br />*<i> I'm not saying that infinity isn't a concept. When it really comes down to it, every number is a concept. That's the beauty of having abstracted the number "two" as an adjective, as in "two sheep", to "two" as a noun.</i><br /><i><br /></i>**<i> There's an argument to be made that <b>treating</b> something like a number doesn't mean it <b>is</b> a number. But at some point, the semantic distinction between these two becomes somewhat blurred.</i><br /><i><br /></i>***<i> Don't worry, I'll make a post about how to legitimately divide by zero in the near future!</i>Bill Shillitohttp://www.blogger.com/profile/17774101901445053590noreply@blogger.com4tag:blogger.com,1999:blog-3748625510348961342.post-62208688449458893182014-09-08T06:27:00.001-07:002014-09-08T06:27:30.494-07:00Throw Outdated Notation to the FloorThere's no excuse for bad notation.<br /><br />Well, there used to be. In older times, mathematical notation and terminology were anything but standard - especially when the mathematics behind them were just coming into existence. What Newton called fluxions, Leibniz called derivatives. Even the now-ubiquitous \(\pi\) was one of many symbols used to represent the constant \(3.14159...\) by various mathematicians - and even then they weren't aware they only had half of the most fundamental circle constant.* But as the human species has progressed, so has our ability to communicate with each other, allowing us to collaborate and spread ideas more quickly and more extensively than ever before. It's no wonder that with the acceleration in communications technology, mathematicians have replaced clunky symbols with more elegant and consistent notations.<br /><br />The only problem is that the math textbooks haven't quite caught up.<br /><br />Let's take a look at a function that causes glazed-over eyes for many high schoolers every time it shows up in an Algebra II or Precalculus book: the <b>greatest integer function</b>, usually notated as \(\left[x\right]\) or occasionally \(\left[\!\left[x\right]\!\right]\). This function is defined as "the greatest integer that is less than or equal to \(x\)." For example:<br /><br />\[\left[5\right]=5\\\left[5.1\right]=5\\\left[5.9\right]=5\\\left[-5.1\right]=-6\]<br />The bracket notation was invented by Carl Friedrich Gauss, who used it in his 1808 proof of the law of quadratic reciprocity.**<br /><br />If you've taught high school students this function, you're probably familiar with the looks of confusion that come shortly after writing down the definition. "So wait, do we want something less than the number or greater than the number?" You could break down that definition piece by piece and hope your students can follow the roundabout logic, but perhaps you've found it's easier to link it to something that students are more familiar with - <b>rounding</b>. To find \(\left[x\right]\), all we need to do is round \(x\) <i>down </i>to the nearest integer (meaning to the left on the real number line). And of course, if \(x\) is <i>already</i> an integer, then \(\left[x\right]=x\), since no rounding needs to happen.<br /><br />So now it's a bit easier to describe \(\left[x\right]\), but the notation seems rather arbitrary, doesn't it? We already use square brackets for other things - mainly for grouping terms together in large expressions and to denote closed intervals. Writing \(\left[\!\left[x\right]\!\right]\) is at least unique, but even more clunky. And that's what Kenneth Iverson, the Canadian computer scientist who invented APL in 1962, must have thought as well. In his book <i>A Programming Language</i>, he gave the function a new name - the <b>floor function</b> - and a new notation:<br /><br />\[\lfloor x\rfloor\]<br />Now <i>there's </i>some solid notation! The name and the bottom half-brackets suggest exactly what to do: round \(x\) <i>down </i>to the nearest integer, just as described earlier. Take a look:<br /><br />\[\lfloor5\rfloor=5\\\lfloor5.1\rfloor=5\\\lfloor5.9\rfloor=5\\\lfloor-5.1\rfloor=-6\]<br />I showed this to my own students, most of whom vaguely remembered the "greatest integer function" and only half of whom knew which direction it went. It stuck. The light bulbs went off all around the room, and the comments were to the effect of "well, <i>that</i> makes a lot more sense!" and "why didn't they teach it this way in the first place?"<br /><br />But it gets better. Mathematics is all about patterns and symmetry. One of the students asked, "so if there's a floor function, is there a ceiling function?" Yes there is. The <b>ceiling function</b> is defined as "the least integer that is greater than or equal to \(x\)", and does exactly what you'd expect - it rounds the number <i>up</i> to the nearest integer. Can you guess what the notation is?<br /><br />\[\lceil x\rceil\]<br />And can you guess the answers to the following problems?<br /><br />\[\lceil5\rceil=?\\\lceil5.1\rceil=?\\\lceil5.9\rceil=?\\\lceil-5.1\rceil=?\]<br />If you guessed \(5\), \(6\), \(6\), and \(-5\), you're correct. See how easy math can be when the notation is evocative of the concept behind it?<br /><br />At this point, you may be wondering, as you should: why aren't textbooks aren't using this notation, given its obvious pedagogical advantage? If this notation were only invented in the past few years or so, it might be excusable that the publishers haven't caught up yet. But come on. It's been over 50 years. And with the way that textbook companies churn out new editions as often as they can, you'd think that one reason to do so would be to keep their math and their teaching up-to-date.<br /><br />...right?<br /><br />It seems that many people are under the unfortunate impression that math, unlike science or social studies, is already set in stone and nothing new really ever comes out of it. This impression couldn't be further from the truth. Mathematics is an ever-growing and ever-evolving body of knowledge.<br /><br />And our potential to understand it better - and teach it better - hasn't hit the ceiling yet.<br /><br /><br /><br /><br /><i>To read further about the floor and ceiling functions, visit the <span id="goog_2018768175"></span><a href="http://en.wikipedia.org/wiki/Floor_and_ceiling_functions">Wikipedia article</a> <span id="goog_2018768176"></span>on the subject.</i><br /><br /><i>* <a href="http://www.tauday.com/tau-manifesto">http://www.tauday.com/tau-manifesto</a></i><br /><br />** <i>The law of quadratic reciprocity is a theorem in number theory. Gauss thought it was so profound and beautiful that he occasionally referred to it as the "golden theorem".</i>Bill Shillitohttp://www.blogger.com/profile/17774101901445053590noreply@blogger.com8tag:blogger.com,1999:blog-3748625510348961342.post-41784786176470201052014-07-31T15:48:00.000-07:002014-07-31T19:37:28.487-07:00Poorly Executed Mnemonics Definitely Addle StudentsIf you've read my past two posts, you know by now that PEMDAS* (Parentheses, Exponents, Multiplication, Division, Addition, Subtraction) is one of my most hated mnemonics. There are two main reasons:<br /><ul><li>It's misleading.</li><li>It's unnecessary.</li></ul>First of all, why is PEMDAS misleading?<br /><br />Let's start with the "P" (Parentheses). People claim that PEMDAS <i>is</i> "the order of operations." This is already problematic because parentheses aren't really a mathematical operation.** Operations <i>do</i> things. Parentheses don't actually <i>do</i> anything - they just group things together. This distinction may seem like more of a technicality, but it actually brings to light the main issue: parentheses aren't the important thing, but the idea of grouping in general. There's lots of ways to group expressions.<br /><ul><li>You can group expressions using a fraction bar. \[\frac{1+2}{3+4}\]</li><li>You can group expressions under a radical. \[\sqrt{3^2+4^2}\]</li><li>You can even group expressions inside an exponent! \[2^{4+1}\](How are you supposed to "do" exponents before addition if there's addition <i>in</i> the exponent and you don't have parentheses to tell you what to do?)</li></ul><div>So in terms of grouping, PEMDAS is at best incomplete.</div><br />Though the "E" (Exponents) is pretty much unambiguous, the entire rest of the mnemonic causes problems. By putting the "MDAS" in linear order, a number of students get the idea that all Multiplication should be done before any Division, and that all Addition should be done before any Subtraction. Thus you get students who will make the following mistakes:\[4-1+2\\=4-3\\=1(?!)\]\[6\div2\times3\\=6\div6\\=1(?!)\]<br />What's even scarier is that PEMDAS has become so ingrained in our math education culture that <i>some teachers actually teach it this blatantly incorrect way</i>. Don't believe me? Take a look at this video from TED-Ed:<br /><br /><div class="separator" style="clear: both; text-align: center;"><iframe allowfullscreen='allowfullscreen' webkitallowfullscreen='webkitallowfullscreen' mozallowfullscreen='mozallowfullscreen' width='320' height='266' src='https://www.youtube.com/embed/H6syI3xiBBg?feature=player_embedded' frameborder='0' /></div><br />Try to tell me that doesn't lead you to believe that MDAS is done in linear order. And we wonder why kids have trouble.<br /><br />Of course, many teachers are careful to explain how the order of operations are <i>supposed </i>to work - that multiplication and division (which is just multiplication by the multiplicative inverse) are done in order from left to right as they appear. Likewise, subtraction is just addition by the additive inverse, so addition and subtraction are done left to right as well. Some teachers write the mnemonic as PE(MD)(AS) or in some other sort of arrangement to emphasize this fact. Others further extend the ridiculous mnemonic-for-a-mnemonic to say "Please Excuse My Dear Aunt Sally ... and Let her Rest".<br /><br />But now why is PEMDAS unnecessary?<br /><br />There is a way to bypass all of this mnemonic madness and teach the order of operations in a way that actually makes sense. How?<br /><br />By teaching <i>why</i> it works that way.<br /><br />When I've asked fellow teachers why order of operations the way it is, those who have been able to answer often gave something to the effect of "well, we needed to decide on <i>some</i> kind of convention to deal with possible ambiguity, so we decided on what we have today." This is half correct - getting rid of ambiguity is very much important. But it wasn't an arbitrary decision. It's not like we could have just as equally decided that addition and subtraction come first, then multiplication and division, and then exponents. There's a very good reason that the operations fall naturally in the order that they do.<br /><br />Think back to elementary school when all you knew about was addition and subtraction. Eventually you ran into expressions that looked like this:\[3+3+3+3+3+3+3\]You didn't want to write so many 3's, so you were introduced to a shorthand to write this expression. Since there were seven 3's being added together, you learned you could instead write:\[3\times7\]Thus you learned that <i>multiplication is repeated addition</i>.<br /><br />Fast forward a few years, when you had multiplication and division under your belt. Now you saw expressions like this instead:\[4\times4\times4\times4\times4\]Again, you were introduced to a shorthand to keep from having to write all those 4's. Since there were five 4's being multiplied together, you wrote:\[4^5\]Thus you learned that <i>exponentiation is repeated multiplication</i>.<br /><br />Now we come to an expression like this.\[5^2+4\times3\]What do we do first? Well, remembering that <i>exponentiation is repeated multiplication</i>, we rewrite our exponent to say what it really means.\[5\times5+4\times3\]Next, remembering that <i>multiplication is repeated addition</i>, we rewrite our multiplication in even more basic terms.\[5+5+5+5+5+4+4+4\]Now the expression is a cinch to evaluate - anyone can add! The value just comes out to 37. But, more remarkably, what we've just done is uncovered the reason <i>why</i> the order of operations is as it is:<br /><b style="font-style: italic;"><br /></b><b style="font-style: italic;">The most compact shorthand is evaluated first.</b><br /><br />Once students understand this, they won't need to actually write out the additions explicitly - they'll just evaluate things in the order they should be handled. But they'll know <i>why</i> to do it.<br /><i><b><br /></b></i>With this in mind, I propose a better way to teach order of operations. Students need only remember two things.<br /><ol><li><b><i>Pay attention to grouping.</i></b></li><li><b><i>Shorthand comes first.</i></b></li></ol><div>If we must use an acronym, instead of PEMDAS, let's use something like ... say ... GEMA. (Grouping, Exponentiation, Multiplication, Addition.) But if we do use GEMA or something similar, we shouldn't deprive students of the understanding that comes from knowing <i>why</i> the order of operations works.<br /><br /><br /><br /><i>* In other countries, variations on PEMDAS are used, such as BODMAS or BIDMAS. The "B" stands for Brackets, another word for parentheses, and "O" and "I" stand for Orders and Indices, respectively, which are both alternate words for exponents. Note that in BODMAS and BIDMAS, the "D" and "M" are interchanged - think about how much confusion that could cause for students!</i><br /><i><br /></i><i>** Thanks to Quintopia for pointing this out: Parentheses in the context of computer science CAN in fact be thought of as operators which let the computer know to make a call to a subroutine. It would be even harder to make this into an acronym ... unless it's a recursive acronym in which the G stands for GEMA, similar to how WINE stands for WINE Is Not an Emulator!</i></div>Bill Shillitohttp://www.blogger.com/profile/17774101901445053590noreply@blogger.com10tag:blogger.com,1999:blog-3748625510348961342.post-56485535321738134582014-07-17T07:03:00.000-07:002014-07-17T18:10:33.338-07:00The Implications of Being ImplicitIn my previous post, I presented three "tricky" (read: "inane") math problems from the Internet, the last of which was the following:<br /><br /><div class="separator" style="clear: both; text-align: center;"><a href="http://3.bp.blogspot.com/-pRfKJkGGzlw/U7h5BBs_Q9I/AAAAAAAAAQo/7UDFIv6L-Bo/s1600/multiplyordivide.jpg" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" src="http://3.bp.blogspot.com/-pRfKJkGGzlw/U7h5BBs_Q9I/AAAAAAAAAQo/7UDFIv6L-Bo/s1600/multiplyordivide.jpg" height="192" width="320" /></a></div><br />I promised I'd come back to this one because it merits special discussion. Now it's time to do exactly that, as this one (as well as its many variations) has piqued my ire every time I've seen it.<br /><br />Most people who have had at least a basic prealgebra class tend to agree that the \(1+2\) in parentheses should be evaluated first, reducing the problem to \[6\div 2(3).\]<br />After that the battlefield gets bloody.<br /><br />Do we do the multiplication first, or do we do the division first?<br /><br />Some people say the multiplication needs to come first, valiantly shouting "PEMDAS" as their battle cry, arguing that since "M" comes before "D", the answer is \(1\).<br /><br />Those who have more experience with order of operations and don't just rely on a silly (and wrong) mnemonic say that the division comes first, since multiplication and division are really the same under the hood - after all, division is equivalent to multiplication by the reciprocal - and by convention* are performed from left to right in the order they appear. For these more seasoned warriors, the answer is "obviously" \(9\).<br /><br />Those in the latter camp definitely are applying better mathematical reasoning than those in the former. But do they have "the" correct answer? If you're like me, though you know that multiplication and division are <i>supposed</i> to happen in order from left to right, there's just something about that \(2(3)\) that catches your eye, that makes you <i>feel</i> like for some reason it "should" come first.<br /><br />And that's why we need to talk about <b>implicit </b>(or <b>implied</b>)<b> multiplication</b>.<br /><br />When we first learn multiplication, we write it with a cross (\(\times\)). But once variables like \(x\) start to come into play, we have to find new, less confusing ways to write multiplication. So instead of writing \(a\times b\), we have a few options.<br /><ul><li>We can use a dot: \[a\cdot b\]</li><li>We can use parentheses: \[a(b)\]</li><li>Or as long as both factors aren't numerals, we can just concatenate (attach) them: \[ab\]</li></ul>There's actually a subtle difference between multiplication with a cross or a dot and multiplication with parentheses or concatentation. The former two are called <b>explicit multiplication</b>, because we've explicitly indicated our operation using a symbol. The latter two are called <b>implicit </b><b>multiplication</b> - we know the intended operation is multiplication because no explicit symbol was provided.<br /><br />Why does this matter? As it turns out, in some conventions, implicit multiplication may actually take precedence over explicit multiplication and therefore division! For example, the "<a href="http://d22izw7byeupn1.cloudfront.net/files/styleguide-pr.pdf#page=23">Style and Notation Guide</a>" for <i>Physical Review</i>, an American scientific journal, specifies that implicit multiplication should come before division when submitting manuscripts**. Of course there are other conventions in which this is not the case, but the point to understand here is that multiple conventions do exist.<br /><br />What's more, we can't even turn to our trusty calculators to tell us which way is "the" correct way, because different calculators may follow different conventions!<br /><table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"><tbody><tr><td style="text-align: center;"><a href="http://1.bp.blogspot.com/-xwr84YKoo6E/U8fOj-gtPvI/AAAAAAAAARQ/uWo0Uv5gHvs/s1600/casio.jpg" imageanchor="1" style="margin-left: auto; margin-right: auto;"><img border="0" src="http://1.bp.blogspot.com/-xwr84YKoo6E/U8fOj-gtPvI/AAAAAAAAARQ/uWo0Uv5gHvs/s1600/casio.jpg" height="258" width="320" /></a></td></tr><tr><td class="tr-caption" style="text-align: center;">Two Casio calculators</td></tr></tbody></table><br /><table align="center" cellpadding="0" cellspacing="0" class="tr-caption-container" style="margin-left: auto; margin-right: auto; text-align: center;"><tbody><tr><td style="text-align: center;"><img border="0" src="http://1.bp.blogspot.com/-6JYVvbxVwiI/U8fOjwnrQYI/AAAAAAAAARU/nQK_geBxI3k/s1600/ti.jpg" height="240" style="margin-left: auto; margin-right: auto;" width="320" /></td></tr><tr><td class="tr-caption" style="text-align: center;">Two TI calculators</td></tr></tbody></table><div class="separator" style="clear: both; text-align: center;"></div><div class="separator" style="clear: both; text-align: center;"><br /></div><div class="separator" style="clear: both; text-align: left;">(Try this on your own calculator and see which convention it uses!)</div><div class="separator" style="clear: both; text-align: left;"><br /></div><div class="separator" style="clear: both; text-align: left;">I'd like to posit one more reason that the \(2(3)\) may <i>feel</i> like it "should" come first. One unfortunate side effect of trying to use existing punctuation when possible to represent mathematics is that certain symbols become overloaded - the same symbols can represent different things. In this case, the notation \(2(3)\) for multiplication bears a very strong resemblance to the notation \(f(3)\) for function evaluation! If the question were \[6\div f(1+2),\] even though we have no idea what function \(f\) is, there's no question that it would be evaluated before the division took place! This may be a possible reason that the implicit-trumps-explicit convention is followed in some circles.</div><br />The inevitable conclusion is that there is no single correct answer - it all depends on what convention you're using. At this point you may be ready to throw your hands up in despair. But there is hope. The best way to solve this kind of problem is ... you guessed it ... to use better notation in the first place! (I mean who uses the obelus (\(\div\)) anymore past 5th grade anyway?)<br /><ul><li>If you mean for multiplication to be done first, then say so! \[\frac{6}{2(1+2)}=1\]</li><li>If you mean for division to be done first, then say so! \[\frac{6}{2}(1+2)=9\]</li></ul>And if you're using a calculator, it never hurts to have too many parentheses.***<br /><br />It all comes back to the point of the previous article, which I will make explicit one more time: Math isn't about symbols. Math is about ideas. If your symbols don't unambiguously convey your ideas, then use better symbols.<br /><br /><br /><br /><i>* The left-to-right convention is probably so because those who established the convention spoke the sorts of European languages for which \(6\div 2\) would be vocalized in that order - not always the case if you've ever heard how fractions are read out loud in Japanese or Korean!</i><br /><br /><i>** The guide may seem like it's claiming that all multiplication should come before division, but this is because they don't use explicit multiplication at all except in specific contexts such as indicating dimensions and performing operations on vectors.</i><br /><br /><i>*** If you have a newer calculator, you may have a nifty fraction template that you can use to clear up confusion even further! If you're using a TI-84+, and you've got the newest operating system on your calculator, try hitting [ALPHA] and then [Y=]. If you see a little menu come up, choose "n/d" and voilà - you can now do fractions without getting lost in a sea of parentheses!</i>Bill Shillitohttp://www.blogger.com/profile/17774101901445053590noreply@blogger.com2tag:blogger.com,1999:blog-3748625510348961342.post-80149003868457686052014-07-05T16:42:00.000-07:002014-07-05T16:42:01.394-07:0099% of People Get This Wrong!If you've been around social media recently, you've no doubt seen the influx of math questions whose sole purpose is supposedly to test whether people know their order of operations.<br /><br />Sometimes, it's a simple question of whether people know not to just blindly apply the operations in order from left to right (the results of which are often worrisome).<br /><br /><div class="separator" style="clear: both; text-align: center;"><a href="http://1.bp.blogspot.com/-_7QMAQlIVGU/U7h5FbxYLNI/AAAAAAAAAQs/RIZekNCocwY/s1600/original-question.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" src="http://1.bp.blogspot.com/-_7QMAQlIVGU/U7h5FbxYLNI/AAAAAAAAAQs/RIZekNCocwY/s1600/original-question.png" height="196" width="400" /></a></div><div class="separator" style="clear: both; text-align: center;"><br /></div><div class="separator" style="clear: both; text-align: center;"><br /></div><div class="separator" style="clear: both; text-align: left;">Sometimes, the spacing is modified to try to trick you into grouping things incorrectly.</div><div class="separator" style="clear: both; text-align: left;"><br /></div><div class="separator" style="clear: both; text-align: center;"><a href="http://4.bp.blogspot.com/-v_-02pJV8I4/U7h4q266MnI/AAAAAAAAAQc/1MANERDFRxY/s1600/5+5x5+5.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" src="http://4.bp.blogspot.com/-v_-02pJV8I4/U7h4q266MnI/AAAAAAAAAQc/1MANERDFRxY/s1600/5+5x5+5.png" height="200" width="166" /></a></div><div class="separator" style="clear: both; text-align: center;"><br /></div><div class="separator" style="clear: both; text-align: left;">But the worst offenders of all are the ones that combine single-line division, either with an obelus (÷) or a slash (/), with implied multiplication, i.e. multiplication without explicitly writing a dot (⋅) or cross (×) symbol.</div><div class="separator" style="clear: both; text-align: left;"><br /></div><div class="separator" style="clear: both; text-align: center;"><a href="http://2.bp.blogspot.com/-pRfKJkGGzlw/U7h5BBs_Q9I/AAAAAAAAAQk/54YdQ5PtAFs/s1600/multiplyordivide.jpg" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" src="http://2.bp.blogspot.com/-pRfKJkGGzlw/U7h5BBs_Q9I/AAAAAAAAAQk/54YdQ5PtAFs/s1600/multiplyordivide.jpg" height="192" width="320" /></a></div><div class="separator" style="clear: both; text-align: center;"><br /></div><div class="separator" style="clear: both; text-align: left;">Most of the time, these pictures are accompanied by a caption such as "99% of People Get This Wrong!" in an attempt to amp up the clickbait factor. And you know what? They're right. But not for the reason you might think.</div><div class="separator" style="clear: both; text-align: left;"><br /></div><div class="separator" style="clear: both; text-align: left;">While the first question isn't too bad, the second and third questions commit a major faux pas in mathematics: introducing ambiguity. Even as a math teacher, I had to give the second one a double-take because of the deceptive spacing. And the ambiguity of the third example is so profound that I'll be dedicating a separate blog post to that problem specifically.</div><div class="separator" style="clear: both; text-align: left;"><br /></div><div class="separator" style="clear: both; text-align: left;">You can cite "order of operations" or "PEMDAS"* all you want - and many people do, with an air of intellectual superiority - but in doing so you're missing the point. The point of establishing the order of operations is to <i>reduce </i>ambiguity. And if there's still a possibility of ambiguity after that, well, that's what we have parentheses for! By deliberately trying to deceive people, those who create and share these images aren't showing how clever they are, since they're actually doing the exact <i>opposite </i>of the thing they're supposedly trying to test people on.</div><div class="separator" style="clear: both; text-align: left;"><br /></div><div class="separator" style="clear: both; text-align: left;">I stress this to my students all the time: Math isn't about symbols. It's about ideas. If the symbols on your page aren't clearly conveying those ideas to the reader, then you need to use better symbols. (And words don't hurt either.)</div><div class="separator" style="clear: both; text-align: left;"><br /></div><div class="separator" style="clear: both; text-align: left;">So why do 99%** of people get these questions wrong?<br /><br />Because they bother to answer them at all.</div><div class="separator" style="clear: both; text-align: left;"><br /></div><div class="separator" style="clear: both; text-align: left;"><br /></div><div class="separator" style="clear: both; text-align: left;"><br /></div><div class="separator" style="clear: both; text-align: left;"><i>* Please don't cite PEMDAS. It's a horrible mnemonic, it's not even correct, and it too will be the subject of a future post.</i></div><div class="separator" style="clear: both;"><i>** 99% of statistics are made up on the spot.</i></div><div><br /></div>Bill Shillitohttp://www.blogger.com/profile/17774101901445053590noreply@blogger.com8tag:blogger.com,1999:blog-3748625510348961342.post-19916448443802781182014-07-01T13:08:00.002-07:002014-07-01T13:23:33.898-07:00Welcome!Hello everyone, and welcome to <b>SolidAngl.es</b>! This website is where I'll be posting my thoughts, observations, tips, and ideas with regard to mathematics, education, and their intersection. Discussion is very much welcome on the comments of each blog post - I want to hear what you all have to say on these issues as well!<br /><br />In the About section of this site, you can find information about how and why I teach, and in the Resume section, you can see where I've been and what I've done thus far.<br /><br />(There's also a Music section if you're interested. I don't imagine I'll be blogging much about music, but who knows - mathematics and music are quite related!)<br /><br />If you have any topics in mathematics and/or education you think should discussed on here, feel free to contact me using the Contact form above.<br /><br />Subscription links can be found to the right, so you can keep up with the latest posts.<br /><br />Enjoy your stay!Bill Shillitohttp://www.blogger.com/profile/17774101901445053590noreply@blogger.com3