## A Personal Note: Hacking Traffic Lights

This is a brief followup to my previous post “A Personal Note: Mysterious Accident”

It is possible to hack traffic lights and there has been some published research into how to do it for some traffic light systems.  Here are some links to articles and videos on the subject, mostly from 2014:

http://thehackernews.com/2014/08/hacking-traffic-lights-is-amazingly_20.html

https://www.wired.com/2014/04/traffic-lights-hacking/

https://www.schneier.com/blog/archives/2014/08/hacking_traffic.html

https://www.technologyreview.com/s/530216/researchers-hack-into-michigans-traffic-lights/

Talk by Cesar Cerrudo at DEFCON 22 on Hacking Traffic Lights

Someone with sufficient physical access to the traffic lights could always modify the hardware even if a computer-style “hack” was impossible.

I was driving a 1995 Nissan 200SX SE-R with minimal built-in electronics by modern car standards.  It would be difficult to hack my car without physical access and it was either with me, in a brightly lit parking lot at my office, or in a secured parking garage at my apartment building.

Just to be clear I am not saying my accident was caused by hacking of the traffic lights, only that it is possible.  As noted in my previous post, there are other possible explanations: an accidental failure of the traffic lights or a remarkable mental lapse on my part.  None of the three explanations seems likely to me.

(C) 2017 John F. McGowan, Ph.D.

John F. McGowan, Ph.D. solves problems using mathematics and mathematical software, including developing gesture recognition for touch devices, video compression and speech recognition technologies. He has extensive experience developing software in C, C++, MATLAB, Python, Visual Basic and many other programming languages. He has been a Visiting Scholar at HP Labs developing computer vision algorithms and software for mobile devices. He has worked as a contractor at NASA Ames Research Center involved in the research and development of image and video processing algorithms and technology. He has published articles on the origin and evolution of life, the exploration of Mars (anticipating the discovery of methane on Mars), and cheap access to space. He has a Ph.D. in physics from the University of Illinois at Urbana-Champaign and a B.S. in physics from the California Institute of Technology (Caltech).

## A Personal Note: Mysterious Accident

I had a serious accident on July 19, 2017 at about 6:30 AM in the morning making a left turn from the off ramp of Central Expressway East onto North Wolfe Road in Sunnyvale, California.  Fortunately no one was seriously injured but I could easily have been killed and my car was totaled.  I have some whiplash and may have had a very mild concussion.

The accident is something of a mystery.  I distinctly remember stopping at the red light at the off ramp from Central Expressway and Wolfe Road.  Then the light turned green and I made a left turn across the three lanes of South Wolfe Road.  I make this turn almost every morning.  I was awake, alert.  I was not texting or doing anything reckless.  I am generally a cautious driver; people have complained about it.

I had a glancing collision with an SUV coming down North Wolfe that should have had a red light but seemingly did not.  The lights from both directions are extremely obvious at this intersection.  I was extremely surprised both by the SUV and to see the light that should have stopped the SUV was green when I pulled over.  I don’t think the SUV ran a red light although it is hard to be absolutely certain.

Three Unlikely Explanations

I seem to be left with three unlikely explanations.  Despite my distinct memory and natural caution, I somehow made a left turn against multiple bright red lights directly in front of me and in my field of view.

(OR) The lights changed improperly giving both the other driver and myself green lights at the same time due to an electrical or electronic failure.

(OR) The lights changed improperly giving both the other driver and myself green lights at the same time because someone tampered with the lights somehow, presumably either a dangerous prank or attempted homicide.

None of these possibilities seems particularly probable, but it seems that the explanation for the accident must be one of them.

Mental glitches of this type may happen but must be very rare.  I was wide awake, alert, have made the turn many times and am familiar with the intersection.  I am a cautious driver.  The red lights should have been easily and clearly visible.  I was in good health with no noticeable cognitive issues in the last several years.  I remember seeing the red lights and  the light changing green and very distinctly no red lights as I made the left turn.

There is a lot of construction at and near the intersection of Wolfe Road and Central Expressway in Sunnyvale that seems associated with an office building under construction in the block bordered by Wolfe, Central, Arques, and Commercial Street.  The new office building faces Wolfe and Arques.

There is a lot of work going on right at the intersection of Central Expressway, technically the on/off ramps, and Wolfe Road.

There were a lot of signs for construction in and around the intersection between Wolfe and the Central Expressway on/off ramps when the accident occurred although I did not see any workmen (it was about 6:30 AM and there do not seem to have been any witnesses).

The question arises whether the construction could have caused a glitch in the lights.  To be sure, the lights could have some very rare, bizarre failure mode not due to the construction as well but surely this is less likely than a problem related to the construction.

Then, finally, there is the possibility of deliberate tampering with the traffic lights.  This could just be some random, extremely dangerous “prank” or an attempt to kill, harm, or frighten either the other driver or myself (I did not and do not know the other driver).  It is a bit concerning coming twelve days after my personal storage locker was broken into (July 7, 2017).

It is difficult to know what to make of deliberate tampering with the traffic lights.  Many years ago I worked briefly for a company that shall remain nameless where I developed very serious concerns about the company and the people connected to it after I took the job.  I didn’t hang around.  They would be my leading suspects if there was tampering but it has been a long time.  They would be risking the spotlight of unwanted public attention if they tried to kill me.

I am left with a mystery.  None of the three explanations — a serious mental glitch on my part, an improbable electronic failure of the traffic lights, or deliberate tampering with the traffic lights by a person or persons unknown — seems likely, but presumably one of them happened.  Previous accidents that I have had have been minor and it was always clear what had happened.  🙁

(C) 2017 John F. McGowan, Ph.D.

John F. McGowan, Ph.D. solves problems using mathematics and mathematical software, including developing gesture recognition for touch devices, video compression and speech recognition technologies. He has extensive experience developing software in C, C++, MATLAB, Python, Visual Basic and many other programming languages. He has been a Visiting Scholar at HP Labs developing computer vision algorithms and software for mobile devices. He has worked as a contractor at NASA Ames Research Center involved in the research and development of image and video processing algorithms and technology. He has published articles on the origin and evolution of life, the exploration of Mars (anticipating the discovery of methane on Mars), and cheap access to space. He has a Ph.D. in physics from the University of Illinois at Urbana-Champaign and a B.S. in physics from the California Institute of Technology (Caltech).

## The Problems with the “Math Teaches You to Think” Argument

An Unconvincing Argument

One of the most common arguments for learning math (or computer programming or chess or <insert your favorite subject here>)  is that math teaches you to think.  This argument has a long history of failing to convince skeptical students and adults especially where more advanced mathematics such as algebra and calculus is concerned.

The “math teaches you to think” argument has several problems.  Almost any intellectual activity including learning many sports teaches you to think.  Reading Shakespeare teaches you to think.  Playing Dungeons and Dragons teaches you to think.  What is so special about math?

Math teaches ways of thinking about quantitative problems that can be very powerful.  As I have argued in a previous post Why Should You Learn Mathematics? mathematics is genuinely needed to make informed decisions about pharmaceuticals and medical treatments, finance and real estate, important public policy issues such as global warming, and other specialized but important areas.  The need for mathematics skills and knowledge beyond the basic arithmetic level is growing rapidly due to the proliferation of, use, and misuse of statistics and mathematical modeling in recent years.

Book Smarts Versus Street Smarts

However, most math courses and even statistics courses such as AP Statistics teach ways of thinking that do not work well or even work at all for many “real world” problems, social interactions, and human society.

This is not a new problem.  One of Aesop’s Fables (circa 620 — 524 BC) is The Astronomer which tells the tale of an astronomer who falls into a well while looking up at the stars.   The ancient mathematics of the Greeks, Sumerians, and others had its roots in ancient astronomy and astrology.

Why does mathematical thinking often fail in the “real world?”  Most mathematics education other than statistics teaches that there is one right answer which can be found by precise logical and mathematical steps.  Two plus two is four and that is it.  The Pythagorean Theorem is proven step by step by rigorous logic starting with Euclid’s Postulates and Definitions.  There is no ambiguity and no uncertainty and no emotion.

If a student tries to apply this type of rigorous, exact thinking to social interactions, human society, even walking across a field where underbrush has obscured a well as in Aesop’s Fable of the Astronomer, the student will often fail.  Indeed, the results can be disastrous as in the fable.

In fact, at the K-12 level and even college, liberal arts such as English literature, history, debate, the law do a much better job than math in teaching students the reality that in many situations there are many possible interpretations.  Liberal arts deals with people and even the most advanced mathematics has failed to duplicate the human mind.

In dealing with other people, we can’t read their minds.  We have to guess (estimate) what they are thinking to predict what they may do in the future.  We are often wrong.  Mathematical models of human behavior generally don’t predict human behavior reliably.  Your intuition from personal experience, learning history, and other generally non-quantitative sources is often better.

The problem is not restricted to human beings and human society.  When navigating in a room or open field, some objects will be obscured by other objects or we won’t happen to be looking at them.  Whether we realize it or not, we are making estimates — educated guesses — about physical reality.  A bush might be just a bush or it might hide a dangerous well that one can fall into.

The Limits of Standard Statistics Courses

It is true that statistics courses such as AP Statistics and/or more advanced college and post-graduate statistics addresses these problems to some degree: unlike basic arithmetic, algebra, and calculus.  The famous Bayes Theorem gives a mathematical framework for estimating the probability that a hypothesis is true given the data/observations/evidence.  It allows us to make quantitative comparisons between competing hypotheses: just a bush versus a bush hiding a dangerous well.

However, many students at the K-12 level and even college get no exposure to statistics or very little.  How many students understand Bayes Theorem?  More importantly, there are significant unknowns in the interpretation and proper application of Bayes Theorem to the real world.  How many students or even practicing statisticians properly understand the complex debates over Bayes Theorem, Bayesian versus frequentist versus several other kinds of statistics?

All or nearly all statistics that most students learn is based explicitly or implicitly on the assumption of independent identically distributed random variables.  These are cases like flipping a “fair” coin where the probability of the outcome is the same every time and is not influenced by the previous outcomes.  Every time someone flips a “fair” coin there is the same fifty percent chance of heads and the same fifty percent chance of tails.  The coin flips are independent.  It does not matter whether the previous flip was heads or tails.  The coin flips are identically distributed.  The probability of heads or tails is always the same.

The assumption of independent identically distributed is accurate or very nearly accurate for flipping coins, most “fair” games of chance used as examples in statistics courses, radioactive decay, and some other natural phenomena.  It is generally not true for human beings and human society.  Human beings learn from experience and change over time.  Various physical things in the real world also change over time.

Although statistical thinking is closer to the “real world” than many other commonly taught forms of mathematics, it still in practice deviates substantially from everyday experience.

Teaching Students When to Think Mathematically

Claims that math (or computer programming or chess or <insert your favorite subject here>) teaches thinking should be qualified with what kind of thinking is taught, what are its strengths and weaknesses, and what problems is it good for solving.

(C) 2017 John F. McGowan, Ph.D.

John F. McGowan, Ph.D. solves problems using mathematics and mathematical software, including developing gesture recognition for touch devices, video compression and speech recognition technologies. He has extensive experience developing software in C, C++, MATLAB, Python, Visual Basic and many other programming languages. He has been a Visiting Scholar at HP Labs developing computer vision algorithms and software for mobile devices. He has worked as a contractor at NASA Ames Research Center involved in the research and development of image and video processing algorithms and technology. He has published articles on the origin and evolution of life, the exploration of Mars (anticipating the discovery of methane on Mars), and cheap access to space. He has a Ph.D. in physics from the University of Illinois at Urbana-Champaign and a B.S. in physics from the California Institute of Technology (Caltech).

Credits

The image of a Latin proof of the Pythagorean Theorem with diagrams is from Wikimedia Commons and is in the public domain.  The original source is a manuscript from 1200 A.D.

We received a very official looking letter today from Corporate Compliance Services, seeming to be from a federal government agency without actually stating this.  The document asked us to send a check or money order for $84 to Corporate Compliance Services for US federal government required posters for our workplace. This is a scam. The United States Department of Labor does indeed require that employers display posters informing employees of their rights under various laws such as the Fair Labor Standards Act (FLSA), but these are available for free from the US Department of Labor. The United States Department of Labor web site has a simple, easy-to-use eLaws Advisor to help you determine which posters are required for your business and to then locate and download the required posters for free. http://webapps.dol.gov/elaws/posters.htm Some Internet Links about Corporate Compliance Services 2-Minute Tip: Corporate Compliance Scam – Don’t Fall for It https://www.bbb.org/south-east-florida/business-reviews/compliance-consulting/corporate-compliance-services-in-pembroke-pines-fl-90114312/Alerts-and-Actions (C) 2017 John F. McGowan, Ph.D. About the author John F. McGowan, Ph.D. solves problems using mathematics and mathematical software, including developing gesture recognition for touch devices, video compression and speech recognition technologies. He has extensive experience developing software in C, C++, MATLAB, Python, Visual Basic and many other programming languages. He has been a Visiting Scholar at HP Labs developing computer vision algorithms and software for mobile devices. He has worked as a contractor at NASA Ames Research Center involved in the research and development of image and video processing algorithms and technology. He has published articles on the origin and evolution of life, the exploration of Mars (anticipating the discovery of methane on Mars), and cheap access to space. He has a Ph.D. in physics from the University of Illinois at Urbana-Champaign and a B.S. in physics from the California Institute of Technology (Caltech). ## Why Should You Learn Mathematics? Why should you learn mathematics? By mathematics, I am not referring to basic arithmetic: addition, subtraction, multiplication, division, and raising a number to a power — for example for an interest calculation in personal finance. There is little debate that in the modern world the vast majority of people need to know basic arithmetic to buy and sell goods and services and perform many other common tasks. By mathematics I mean more advanced mathematics such as algebra, geometry, trigonometry, calculus, linear algebra, and college level statistics. I am not referring to highly specialized advanced areas of mathematics such as number theory or differential geometry generally taught after the sophomore year in college or in graduate school. I am following the language of Andrew Hacker in his book The Math Myth: And Other STEM Delusions in which he argues the algebra requirement should be eliminated in high schools, community colleges, and universities except for degrees that genuinely require mathematics. Hacker draws a distinction between arithmetic which is clearly needed by all and mathematics such as algebra which few use professionally. A number of educators such as Eloy Ortiz Oakley, the chancellor of California’s community colleges, have embraced a similar view, even arguing that abolishing the algebra requirement is a civil rights issue since some minority groups fail the algebra requirement at higher rates than white students. Yes, he did say it is a civil rights issue: The second thing I’d say is yes, this is a civil rights issue, but this is also something that plagues all Americans — particularly low-income Americans. If you think about all the underemployed or unemployed Americans in this country who cannot connect to a job in this economy — which is unforgiving of those students who don’t have a credential — the biggest barrier for them is this algebra requirement. It’s what has kept them from achieving a credential. (emphasis added) Eloy Ortiz Oakley on NPR (Say Goodbye to X + Y: Should Community Colleges Abolish Algebra? July 19, 2017) At present, few jobs, including the much ballyhooed software development jobs, require more than basic arithmetic as defined above. For example, the famous code.org “What Most Schools Don’t Teach” video on coding features numerous software industry luminaries assuring the audience how easy software development is and how little math is involved. Notably Bill Gates at one minute and forty-eight seconds says: “addition, subtraction…that’s about it.” Bill Gates assessment of the math required in software development today is largely true unless you are one of the few percent of software developers working on highly mathematical software: video codecs, speech recognition engines, gesture recognition algorithms, computer graphics for games and video special effects, GPS, Deep Learning, FDA drug approvals, and other exotic areas. Thus, the question arises why people who do not use mathematics professionally ought to learn mathematics. I am not addressing the question of whether there should be a requirement to pass algebra to graduate high school or for a college degree such a veterinary degree where there is no professional need for mathematics. The question is whether people who do not need mathematics professionally should still learn mathematics — whether it is required or not. People should learn mathematics because they need mathematics to make informed decisions about their health care, their finances, public policy issues that affect them such as global warming, and engineering issues such as the safety of buildings, aircraft, and automobiles — even though they don’t use mathematics professionally. The need to understand mathematics to make informed decisions is increasing rapidly with the proliferation of “big data” and “data science” in recent years: the use and misuse of statistics and mathematical modeling on the large, rapidly expanding quantities of data now being collected with extremely powerful computers, high speed wired and wireless networks, cheap data storage capacity, and inexpensive miniature sensors. Health and Medicine An advanced knowledge of statistics is required to evaluate the safety and effectiveness of drugs, vaccines, medical treatments and devices including widely used prescription drugs. A study by the Mayo Clinic in 2013 found that nearly 7 in 10 (70%) of Americans take at least one prescription drug. Another study published in the Journal of the American Medical Association (JAMA) in 2015 estimated about 59% of Americans are taking a prescription drug. Taking a prescription drug can be a life and death decision as the horrific case of the deadly pain reliever Vioxx discussed below illustrates. The United States and the European Union have required randomized clinical trials and detailed sophisticated statistical analyses to evaluate the safety and effectiveness of drugs, medical devices, and treatments for many decades. Generally, these analyses are performed by medical and pharmaceutical companies who have an obvious conflict of interest. At present, doctors and patients often find themselves outmatched in evaluating the claims for the safety and effectiveness of drugs, both new and old. In the United States, at least thirty-five FDA approved drugs have been withdrawn due to serious safety problems, generally killing or contributing to the deaths of patients taking the drugs. The FDA has instituted an FDA Adverse Events Reporting System (FDAERS) for doctors and other medical professionals to report deaths and serious health problems such as hospitalization suspected of being caused by adverse reactions to drugs. In 2014, 123,927 deaths were reported to the FDAERS and 807,270 serious health problems. Of course, suspicion is not proof and a report does not necessarily mean the reported drug was the cause of the adverse event. Vioxx (generic name rofecoxib) was a pain-killer marketed by the giant pharmaceutical company Merck (NYSE:MRK) between May of 1999 when it was approved by the United States Food and Drug Administration (FDA) and September of 2004 when it was withdrawn from the market. Vioxx was marketed as a “super-aspirin,” allegedly safer and implicitly more effective than aspirin and much more expensive, primarily to elderly patients with arthritis or other chronic pain. Vioxx was a “blockbuster” drug with sales peaking at about$2.5 billion in 2003 1 and about 20 million users 2. Vioxx probably killed between 20,000 and 100,000 patients between 1999 and 2004 3.

Faulty blood clotting is thought to be the main cause of most heart attacks and strokes. Unlike aspirin, which lowers the probability of blood coagulation (clotting) and therefore heart attacks and strokes, Vioxx increased the probability of blood clotting and the probability of strokes and heart attacks by about two to five times.

Remarkably, Merck proposed and the FDA approved Phase III clinical trials of Vioxx with too few patients to show that Vioxx was actually safer than the putative 3.8 deaths per 10,000 patients rate (16,500 deaths per year according to a controversial study used to promote Vioxx) from aspirin and other non-steroidal anti-inflammatory drugs (NSAIDs) such as ibuprofen (the active ingredient in Advil and Motrin), naproxen (the active ingredient in Aleve), and others.

The FDA guideline, Guideline for Industry: The Extent of Population Exposure to Assess Clinical Safety: For Drugs Intended for Long-Term Treatment of Non-Life-Threatening Conditions (March 1995), only required enough patients in the clinical trials to reliably detect a risk of about 0.5 percent (50 deaths per 10,000) of death in patients treated for six months or less (roughly equivalent to one percent death rate for one year assuming a constant risk level) and about 3 percent (300 deaths per 10,000) for one year (recommending about 1,500 patients for six months or less and about 100 patients for at least one year without supporting statistical power computations and assumptions in the guideline document).

The implicit death rate detection threshold in the FDA guideline was well above the risk from aspirin and other NSAIDs and at the upper end of the rate of cardiovascular “events” caused by Vioxx. FDA did not tighten these requirements for Vioxx even though the only good reason for the drug was improved safety compared to aspirin and other NSAIDs. In general, the randomized clinical trials required by the FDA for drug approval have too few patients – insufficient statistical power in statistics terminology – to detect these rare but deadly events 4.

To this day, most doctors and patients lack the statistical skills and knowledge to evaluate the safety level that can be inferred from the FDA required clinical trials.  There are many other advanced statistical issues in evaluating the safety and effectiveness of drugs, vaccines, medical treatments, and devices.

Finance and Real Estate

Mathematical models have spread far and wide in finance and real estate, often behind the scenes invisible to casual investors.  A particularly visible example is Zillow’s ZEstimate of the value of homes, consulted by home buyers and sellers every day.  Zillow is arguably the leading online real estate company.  In March 2014, Zillow had over one billion page views, beating competitors Trulia.com and Realtor.com by a wide margin; Zillow has since acquired Trulia.

According to a 2013 Gallup poll, sixty-two percent (62%) of Americans say they own their home.  According to a May 2014 study by the Consumer Financial Protection Bureau, about eighty percent (80%) of Americans 65 and older own their home.  Homes are a large fraction of personal wealth and retirement savings for a large percentage of Americans.

Zillow’s algorithm for valuing homes is proprietary and Zillow does not disclose the details and/or the source code.  Zillow hedges by calling the estimate an “estimate” or a “starting point.”  It is not an appraisal.

However, Zillow is large and widely used, claiming estimates for about 110 million homes in the United States.   That is almost the total number of homes in the United States.  There is the question whether it is so large and influential that it can effectively set the market price.

Zillow makes money by selling advertising to realty agents.  Potential home buyers don’t pay for the estimates.  Home sellers and potential home sellers don’t pay directly for the estimates either.  This raises the question whether the advertising business model might have an incentive for a systematic bias in the estimates.  One could argue that a lower valuation would speed sales and increase commissions for agents.

Zillow was recently sued in Illinois over the ZEstimate by a homeowner — real estate lawyer Barbara Andersen 🙂 — claiming the estimate undervalued her home and made it difficult therefore to sell the home.  The suit argues that the estimate is in fact an appraisal, despite claims to the contrary by Zillow, and therefore subject to Illinois state regulations regarding appraisals.  Andersen has reportedly dropped this suit and expanded to a class-action lawsuit by home builders in Chicago again alleging that the ZEstimate is an appraisal and undervalues homes.

The accuracy of Zillow’s estimate has been questioned by others over the years including home owners, real estate agents, brokers, and the Los Angeles Times.  Complaints often seem to involve alleged undervaluation of homes.

On the other hand, Zillow CEO Spencer Rascoff’s Seattle home reportedly sold for $1.05 million on Feb. 29, 2016, 40 percent less than the Zestimate of$1.75 million shown on its property page a day later (March 1, 2016).  🙂

As in the example of Vioxx and other FDA drug approvals, it is actually a substantial statistical analysis project to independently evaluate the accuracy of Zillow’s estimates.  What do you do if Zillow substantially undervalues your home when you need to sell it?

Murky mathematical models of the value of mortgage backed securities played a central role in the financial crash in 2008.  In this case, the models were hidden behind the scenes and invisible to casual home buyers or other investors.  Even if you are aware of these models, how do you properly evaluate their effect on your investment decisions?

Public Policy

Misleading and incorrect statistics have a long history in public policy and government.  Darrell Huff’s classic How to Lie With Statistics (1954) is mostly concerned with misleading and false polls, statistics, and claims from American politics in the 1930’s and 1940’s.  It remains in print, popular and relevant today.  Increasingly however political controversies involve often opaque computerized mathematical models rather than the relatively simple counting statistics debunked in Huff’s classic book.

Huff’s classic and the false or misleading counting statistics in it generally required only basic arithmetic to understand.  Modern political controversies such as Value Added Models for teacher evaluation and the global climate models used in the global warming controversy go far beyond basic arithmetic and simple counting statistics.

The Misuse of Statistics and Mathematics

Precisely because many people are intimidated by mathematics and had difficulty with high school or college mathematics classes including failing the courses, statistics and mathematics are often used to exploit and defraud people.  Often the victims are the poor, marginalized, and poorly educated.  Mathematician Cathy O’Neil gives many examples of this in her recent book Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy (2016).

The misuse of statistics and mathematics is not limited to poor victims.  Bernie Madoff successfully conned large numbers of wealthy, highly educated investors in both the United States and Europe using the arcane mathematics of options as a smokescreen.  These sophisticated investors were often unable to perform the sort of mathematical analysis that would have exposed the fraud.

Rich and poor alike need to know mathematics to protect themselves from this frequent and growing misuse of statistics and mathematics.

Algebra and College Level Statistics

The misleading and false counting statistics lampooned by Darrell Huff in How to Lie With Statistics does not require algebra or calculus to understand. In contrast, the college level statistics often encountered in more complex issues today does require a mastery of algebra and sometimes calculus.

For example, one of the most common probability distributions encountered in real data and mathematical models is the Gaussian, better known as the Normal Distribution or Bell Curve. This is the common expression for the Gaussian in algebraic notation.

${P(x) = \frac{1}{{\sigma \sqrt {2\pi } }}e^{{{ - \left( {x - \mu } \right)^2 } \mathord{\left/ {\vphantom {{ - \left( {x - \mu } \right)^2 } {2\sigma ^2 }}} \right. \kern-\nulldelimiterspace} {2\sigma ^2 }}}}$

$x$ is the position of the data point. $\mu$ is the mean of the distribution. If I have a data set obeying the Normal Distribution, most of the data points will be near the mean $\mu$ and fewer further away.  $\sigma$ is the standard deviation — loosely the width — of the distribution. $\pi$ is the ratio of the circumference of a circle to the diameter. $e$ is Euler’s number (about 2.718281828459045).

This is a histogram of simulated data following the Normal Distribution/Bell Curve/Gaussian with a mean $\mu$ of zero (0.0) and a standard deviation $\sigma$ of one (1.0):

To truly understand the Normal Distribution you need to know Euler’s number e and algebraic notation and symbolic manipulation. It is very hard to express the Normal Distribution with English words or basic arithmetic. The Normal Distribution is just one example of the use of algebra in college level statistics.  In fact, an understanding of calculus is needed to have a solid understanding and mastery of college level statistics.

Conclusion

People should learn mathematics — meaning subjects beyond basic arithmetic such as algebra, geometry, trigonometry, calculus, linear algebra, and college level statistics — to make informed decisions about their health care, personal finances and retirement savings, important public policy issues such as teacher evaluation and public education, and other key issues such as evaluating the safety of buildings, airplanes, and automobiles.

There is no doubt that many people experience considerable difficulty learning mathematics whether due to poor teaching, inadequate learning materials or methods, or other causes.  There is and has been heated debate over the reasons.  These difficulties are not an argument for not learning mathematics.  Rather they are an argument for finding better methods to learn and teach mathematics to everyone.

End Notes

1 How did Vioxx debacle happen?” By Rita Rubin, USA Today, October 12, 2004 The move was a stunning denouement for a blockbuster drug that had been marketed in more than 80 countries with worldwide sales totaling $2.5 billion in 2003. 2 Several estimates of the number of patients killed and seriously harmed by Vioxx were made. Dr. David Graham’s November 2004 Testimony to the US Senate Finance Committee gives several estimates including his own. 3 A “blockbuster” drug is pharmaceutical industry jargon for a drug with at least$1 billion in annual sales. Like Vioxx, it need not be a “wonder drug” that cures or treats a fatal or very serious disease or condition.

Sonal Singh and Yoon K Loke

Trials 2012 13:138

Received: 9 February 2012 Accepted: 30 July 2012 Published: 20 August 2012

The premarketing clinical trials required for approval of a drug primarily guard against type 1 error. RCTs are usually statistically underpowered to detect the specific harm either by recruitment of a low-risk population or low intensity of ascertainment of events. The lack of statistical significance should not be used as proof of clinical safety in an underpowered clinical trial.

Credits

The image of an ancient mathematician or engineer with calipers, often identified as Euclid or Archimedes, is from The School of Athens fresco by Raphael by way of Wikimedia Commons.  It is in the public domain.

(C) 2017 John F. McGowan, Ph.D.

John F. McGowan, Ph.D. solves problems using mathematics and mathematical software, including developing gesture recognition for touch devices, video compression and speech recognition technologies. He has extensive experience developing software in C, C++, MATLAB, Python, Visual Basic and many other programming languages. He has been a Visiting Scholar at HP Labs developing computer vision algorithms and software for mobile devices. He has worked as a contractor at NASA Ames Research Center involved in the research and development of image and video processing algorithms and technology. He has published articles on the origin and evolution of life, the exploration of Mars (anticipating the discovery of methane on Mars), and cheap access to space. He has a Ph.D. in physics from the University of Illinois at Urbana-Champaign and a B.S. in physics from the California Institute of Technology (Caltech).

## Escaping the Distraction Economy: Seven Steps to Reduce Distractions

A few months ago, starting my new business, I noticed that I was spending large amounts of valuable time on valueless distractions — notifications, e-mails, and other disruptions from social media, web sites, mailing lists, other people, and other causes.  This is a problem that had been growing over the last few years in part with the growth of social media and increased smartphone use.

I stopped and spent some time figuring out how to reduce the many distractions to a low level.  Here are seven steps to reduce distractions.

Seven Steps to Reduce Distractions

2. Unsubscribe from Mailing Lists
3. Unfollow Distracting Friends and Colleagues
4. Remove Distracting Social Media Apps from Your Smartphone
5. Block Notifications from Distracting Apps You Can’t Remove
6. Get an Office or Work from Home
7. Read Deep Work by Cal Newport

By default, social media sites like Facebook and many other mobile/web sites/apps/services send large numbers of promotional and other e-mails to subscribers e-mail accounts and push notifications to subscriber smartphones.  In most of these services, it is possible to turn off most or even all e-mail and smartphone notifications in the Settings for your account.  Sometimes this is in Options or Account Options or some other section title, but Settings seems to be the most common at present.

This screenshot shows my Notifications settings on Facebook after aggressively turning off most of them:

E-Mail notifications are controlled by a section further down in the list — not shown in the screenshot above.  Here is a screenshot with the E-Mail section:

Clicking on Email gives:

Note that I select the “Only notifications about your account, security and privacy” option and other options to minimize the number of emails that I receive.

Most social media services, web sites, and apps have similar options to turn off all or most non-essential notifications, both push notifications and emails, buried in the Settings, Options, or a similar section.

Unsubscribe from Mailing Lists

Most modern mailings lists have an unsubscribe link buried in the email messages, usually at the bottom.  This is required by law.  In my experience, around eighty percent of these actually work.  The unsubscribe links usually work for large high profile web sites and services such as Facebook, LinkedIn, etc.

With older mailing lists, you may need to hunt further, e-mail the administrator etc.

If unsubscribe does not work, you can block the mailing list using your SPAM filter or message rules in your mail program.

One problem with social media is that you may get inundated with distracting or annoying posts from friends and colleagues.  Nonetheless, you may not want to offend them by unfriending or disconnecting.  Facebook and some other social media sites enable you to unfollow connections.  You remain connected but you no longer see their posts in your feed.

Remove Distracting Social Media Apps from Your Smartphone

I simply removed the Facebook, Facebook Messenger, Twitter, LinkedIn, and a few other social media apps from my smartphone.  This completely eliminates a large number of distracting and annoying notifications, saves space, and reduces the ability of your social media or other service to track you and invade your privacy.

Block Notifications from Distracting Apps You Can’t Remove

In my case, Facebook, Twitter, LinkedIn and several other apps were useless fluff on my smartphone.  I have a few apps that generate unnecessary notifications but that I would like to keep on my phone.  In this case, I blocked the notifications in the Settings (I have an iPhone).

Get an Office or Work from Home

With the cult of “collaboration” and open offices, distractions have proliferated at work and it has become increasingly difficult for many workers to work productively or enjoyably.  If possible, get an office with walls and a door you can close when you need to focus on your work or consider working from home.

Working from home has its own distractions including close proximity to a kitchen with food, small children, large children, spouses, books, toys, and other items that can distract you.  Setting aside a room or a a dedicated office area with a divider can help with this.

In addition, working from home often makes it difficult to separate home/family/personal life from work/office life.

In my case, I am self-employed and opted to rent an office with walls and a door.  This has worked well so far.

Obviously, these options depend on your situation and resources.

Read Deep Work by Cal Newport

Read Deep Work: Rules for Focused Success in a Distracted World by Cal Newport.  Cal Newport is an Associate Professor of Computer Science at Georgetown University who also writes self-help books about how to work more productively. I found Deep Work insightful with many useful thoughts and suggestions.

Cal Newport has also given a number of talks and interviews on his ideas available online including:

Quit Social Media (TEDxTysons on YouTube)

Deep Work: How to Focus and Resist Distractions (with Cal Newport)  — The College Info Geek Podcast

Interview: Deep Work with Cal Newport (Brian Johnson)

There are several more available if you poke around with Google or another search engine.

Overall, Deep Work is a great book and the author has many great points.  Like many self-help books, it tends to hyperbole and uses very exceptional people like Theodore Roosevelt as inspirational examples.  The definition of “Deep Work” could be more precise.  Pick the ideas that work for you.

Conclusion

These steps are not a complete cure for the tidal wave of distractions in modern society, most associated with the Internet, social media, and smartphones.  However, taken together they can reduce the level of distractions in your life substantially, making it easier to focus on both work and your personal life productively.

(C) 2017 John F. McGowan, Ph.D.

John F. McGowan, Ph.D. solves problems using mathematics and mathematical software, including developing gesture recognition for touch devices, video compression and speech recognition technologies. He has extensive experience developing software in C, C++, MATLAB, Python, Visual Basic and many other programming languages. He has been a Visiting Scholar at HP Labs developing computer vision algorithms and software for mobile devices. He has worked as a contractor at NASA Ames Research Center involved in the research and development of image and video processing algorithms and technology. He has published articles on the origin and evolution of life, the exploration of Mars (anticipating the discovery of methane on Mars), and cheap access to space. He has a Ph.D. in physics from the University of Illinois at Urbana-Champaign and a B.S. in physics from the California Institute of Technology (Caltech).

## A Murder Over Mars: Did Johannes Kepler Poison Tycho Brahe?

An Unexpected Death

On October 13, 1601 the famous astronomer and astrologer Tycho Brahe (1546-1601) — new friend, confidant, and adviser to the Holy Roman Emperor Rudolf II, one of the most powerful men in Europe — became unexpectedly and gravely ill at a banquet in the Imperial capital of Prague.  Tycho was a colorful, athletic and brilliant Danish nobleman who had defied convention by taking a commoner as his wife and by pursuing the study of astronomy instead of the more common pastimes of his fellow nobles in Denmark which he pointedly disdained as frivolous and unimportant.

Tycho suffered horribly for about a week, seemed to begin recovering, and then died unexpectedly on October 24, 1601.   Tycho was noted for his good health and vigor.  His death came as a surprise to his family, friends, and colleagues.

The Imperial court was a hotbed of intrigue and filled with peculiar and often ambitious men who frequently worked with highly toxic chemicals: astrologers, alchemists and magicians in the employ of the Holy Roman Emperor Rudolf II (1552-1612) who hoped to unlock the secrets of the universe by funding a research program that might be called the Manhattan Project of its time.

From the start there were rumors Tycho had been poisoned and some suspected his assistant the young, equally brilliant mathematician and astronomer Johannes Kepler, remembered today as the author of Kepler’s Three Laws of Planetary Motion.

Tycho was buried with great pomp and circumstance in Prague on November 4, 1601 with some of his friends and colleagues making pointed aspersions in Kepler’s direction during the ceremonies.  In time, his beloved wife Kirsten Barbara Jørgensdatter was buried beside him.

Tycho and Kepler

By one year earlier (1600), Tycho had accumulated over a lifetime by far the most accurate measurements of the positions of the planets over time, especially the planet Mars thought by astrologers and kings to influence the occurrence and outcomes of wars and conflict. After years of lavish royal patronage in Denmark, Tycho had a falling out with the new king and fled to the mostly German-speaking Holy Roman Empire of Rudolf II.  Here with funding from Rudolf II he hoped to analyze his data and confirm his own novel theory of the solar system, the known universe at the time, in which the Earth was the center with the Sun and Moon orbiting the Earth and all the other planets orbiting the Sun.

Tycho hired the brilliant young up-and-coming astronomer and mathematician Johannes Kepler (1571-1630) to analyze his data. Kepler hoped to use Tycho’s data to confirm his own Theory of Everything based on the hot new Sun-centered theory of Nicolaus Copernicus (1473-1543).

Tycho and Kepler had a stormy working relationship until Tycho’s untimely death in 1601 which left Kepler with the access to Tycho’s data that he desired.  In the chaos accompanying Tycho’s death, Kepler quietly walked off with Brahe’s notebooks containing his data on Mars.  The ensuing controversy with Brahe’s family was eventually resolved more-or-less amicably, but in the mean time Kepler had the data he had sought.

In one of the great ironies of scientific history, Kepler proceeded to discover that his pet theory, the other variants of Copernicus’s Sun-centered system, the traditional Earth-centered system of Klaudius Ptolemy, and Tycho’s hybrid Earth-Sun centered system were all wrong, although he was never able to fully accept that the data ruled out his system as well.

The models were all mathematically equivalent although they had different physical interpretations.  All incorrectly assumed that the motions of Mars and the other planets were built up from uniform circular motion, the infamous epicycles.

Kepler’s analysis of Tycho’s data on the planet Mars took about five years — including his work in 1600 and 1601 when he had limited access to the full data set.  In 1605, while taking a break during the Easter Holiday, Kepler had his Eureka moment.  He realized that the orbit of Mars was elliptical with the Sun at one focus of the ellipse and that the speed of the planet varied inversely with distance from the Sun so that the plane swept out the same area in the same time.  These two insights are now known as Kepler’s First and Second Laws, and they ensured the fame of both Brahe and Kepler to the present day.

Did Kepler Murder Tycho?

In 1991, soon after the end of the Cold War, the National Museum in Prague gave a somewhat peculiar goodwill gift to Denmark, a small box with a six centimeter long sample of Tycho Brahe’s mustache hair, acquired years earlier when Tycho’s crypt was refurbished in 1901.  Tycho and his wife’s skeletons were examined and then reburied in 1901, but a few samples were taken and given to the National Museum in Prague.

The gift reopened the old question of whether Kepler or someone else had poisoned Tycho in 1601.  Kepler had been a seemingly deeply religious man who had given up a comfortable teaching job in Graz rather than abandon his Lutheran faith and convert to Catholicism.  He was later excommunicated from the Lutheran Church for publicly rejecting the Lutheran doctrine of ubiquity with no apparent gain to himself.  This latter doctrine was an esoteric theological issue nonetheless of paramount importance in the conflict between the Lutherans, Calvinists, and Catholics that would soon lead to the horrific Thirty Years War (1618-1648).

This same stubbornness in holding to his views and perhaps the jealousy of his colleagues had led Kepler into bitter clashes with Tycho and others during his career.  Could such a man have committed murder for the lucrative position of Imperial Mathematician in Rudolf II’s court, fame, or even a fanatical desire to extend human knowledge whatever the cost?

The hair from Tycho’s mustache was examined using modern forensic techniques by Bent Kaempe, Director of the Department of Forensic Chemistry at the Institute of Forensic Medicine at the University of Copenhagen — one of the leading toxicologists in Europe, and potentially lethal levels of mercury detected.  Kaempe concluded that:

Tycho Brahe’s uremia can probably be traced to mercury poisoning, most likely due to Brahe’s experiments with his elixir 11-12 days before his death.

Mercury and mercury compounds, some extremely toxic, were widely used in alchemy.  Tycho himself used a mercury compound at low doses, potentially deadly at higher doses, for his health — following the alchemical ideas of Paracelsus (1493-1541): the elixir mentioned by Kaempe.

Some experts argued the mercury measurements demonstrated that  Tycho had been poisoned and murdered with a mercury compound.  Others suggested that the mercury was due to the embalming process or some other contamination of Tycho’s remains.

In 2004, journalists Joshua and Anne-Lee Gilder published a book Heavenly Intrigue: Johannes Kepler, Tycho Brahe, and the Murder Behind One of History’s Greatest Scientific Discoveries popularizing the theory that Kepler poisoned Tycho with mercuric chloride — based on the mercury measurements from Tycho’s mustache hair.

The controversy led to the exhumation of Tycho’s skeleton in 2010 in an attempt to settle the issue.  The analysis of Tycho’s remains seemingly ruled out lethal levels of mercury as the cause of death in 2012 and  seems to have been generally consistent with natural causes, a bladder infection.

The Limits of Forensic Science

After over four-hundred years, it seems unlikely that we will ever know for sure if Tycho Brahe was poisoned and, if so, by whom.  Even today, people in their fifties die unexpectedly from heart attacks and other causes at rates substantially higher than people in their twenties, thirties, and forties.  Medicine was very limited in Tycho’s time — often more dangerous than doing nothing in fact.  Modern sanitation measures were almost non-existent even at an Imperial court.

On the other hand, Rudolf II had recruited and gathered around himself in Prague some of the most brilliant, highly educated, ambitious, and strange men of his time, many experts like Tycho in toxic chemicals used in alchemy and medicine.  Many were probably familiar with plants and herbs available in Renaissance Europe, some of which could have been used as deadly poisons as well.  He offered these men enormous wealth at a time when most people in Europe lived in dire poverty.

Kepler’s own mother was accused of and convicted of witchcraft.  She was specifically accused of poisoning another woman with a magic potion.  Kepler himself was a highly educated and brilliant man. It is quite conceivable that he could have known much about poisons, perhaps even ones unknown or rarely used today.  He had close access to Tycho, his boss.

The mercury measurements of Tycho’s mustache hair is one of many examples of overconfidence in forensic science.  This overconfidence is often an explicit or implicit claim that forensic techniques — if done right — can give an absolutely certain or almost certain (for example, the one in many trillion odds often quoted for DNA matches in criminal cases) answer.

This false certainty is a claim made by governments, prosecutors, scientists who should know better, and many others.  It is heavily promoted in the popular media with television shows like CSI (2000-2015), Numb3rs (2005-2010), Quincy (1976-1983), blockbuster movies like Silence of the Lambs (1991), and many others.

Numerous cases in recent years have demonstrated the uncertainty of forensic methods in the real world.  These include the Brandon Mayfield case for fingerprint analysis, the questionable use of DNA “profiling” in the Amanda Knox murder case in Italy, the failure of DNA analysis in the Jaidyn Leskie case in Australia, and many more.

In the case of DNA, the astronomical DNA match odds frequently quoted by prosecutors are highly misleading because they do not include a valid statistical model for the probability of contamination of the samples in the field, at the crime scene by investigators, or at the forensic laboratory where the DNA match is performed.  Almost certainly the odds of a false match due to some sort of contamination scenario are much higher than the one in several trillion odds often cited by prosecutors.

Contamination in the field is a likely explanation for the mercury levels in Tycho’s mustache.  The mercury may have come from the embalming process.  Perhaps Tycho or someone near him somehow spilled some mercury compound on his mustache while he was ill and dying.  Tycho worked with and used mercury compounds frequently and they were likely present in his home.

The reality is that there is limited data in many crimes and possible crimes like Tycho’s death.  There are usually many interpretations possible for that data, some more likely than others, some improbable but not impossible.  In many cases, we don’t even know the prior probability of those interpretations.  In the case of Tycho, we don’t know the probability that his mustache was contaminated with mercury by embalming or some other cause.

Mathematically, we now know there are an infinite number of mathematical models that can match any finite set of data with a desired level of accuracy.  In an early example of this, Kepler was able to show that the traditional Ptolemaic Earth-centered model of the Solar System, the hot new Copernican Sun-centered model, and the hybrid model of Tycho were mathematically equivalent and matched the data equally well — predicting the future position of Mars in the Zodiac to about one percent accuracy, a few degrees.

Most of this infinity of mathematical models matching a finite data set are extremely complicated.  We typically throw out the more complicated models to get a finite set of arguably plausible choices.

Historically, this method of selecting mathematical models has proven remarkably powerful in physics and astronomy.  Kepler discovered that a simple mathematical model of non-uniform elliptical motion explained the seemingly extremely complex motions of Mars and the other planets.  Newton, Maxwell, Einstein, and others have repeated this success with other data and phenomena including gravitation, electromagnetism, and radioactive decay.

This infinity of possible mathematical models for a finite data set is the mathematical explanation for the many possible interpretations of the data from a crime scene.  Even if we exclude extremely complicated and implausible models and interpretations a priori, we are still typically left with a number of possibilities, notably including contamination scenarios with DNA and other forensic methods such as the measurements of the mercury in Tycho’s mustache hair.

The figure above illustrates the many models problem in a forensic context.  It shows a simulation with four simulated data points.  These data points could be, for example, the strength of a DNA signal or the mercury level in different parts of Tycho’s body.  The high point on the right is the mercury level in his mustache hair.  The problem is the level could be lower in his body — too low to cause death.  The other points could be measurements of the mercury level in various bones from his skeleton.

We don’t actually have the soft tissues where the putative poison would have done its deadly work.  These have decayed away.  The analyst must therefore infer the mercury level in those tissues over four hundred years ago from measurements today.  The red line represents, for example, the threshold for a lethal level of mercury in Tycho’s body.  Thus, depending on which model is chosen, mercury did or did not kill Tycho.  In reality the forensic analysis is often much more complex and difficult to perform than this simple simulated example and illustration.

In conclusion, the certainty or near certainty often claimed or implied in many forensic analyses is frequently illusory.

(C) 2017 John F. McGowan, Ph.D.

Credits

The image of the planet Mars is from the NASA Jet Propulsion Laboratory (JPL) and is in the public domain.  It is a mosaic of images taken by observation satellites in orbit around Mars.  It shows the giant Valles Marineris canyon on Mars front and center.  It is one of the most popular images of Mars.

The image of Tycho Brahe is from Wikimedia Commons and is in the public domain.

The image of Johannes Kepler is from Wikimedia Commons and is in the public domain.

John F. McGowan, Ph.D. solves problems using mathematics and mathematical software, including developing gesture recognition for touch devices, video compression and speech recognition technologies. He has extensive experience developing software in C, C++, MATLAB, Python, Visual Basic and many other programming languages. He has been a Visiting Scholar at HP Labs developing computer vision algorithms and software for mobile devices. He has worked as a contractor at NASA Ames Research Center involved in the research and development of image and video processing algorithms and technology. He has published articles on the origin and evolution of life, the exploration of Mars (anticipating the discovery of methane on Mars), and cheap access to space. He has a Ph.D. in physics from the University of Illinois at Urbana-Champaign and a B.S. in physics from the California Institute of Technology (Caltech).

## Coding Bootcamps Closing

Hack Education has a good article on the recent closures of Dev Bootcamp and The Iron Yard, early high profile “coding bootcamps.”

“Why are Coding Bootcamps Going Out of Business?”

which has got some attention on Hacker News:  https://news.ycombinator.com/item?id=14831918

A couple comments on the article:

In my experience in the Silicon Valley, software developers/engineers/programmers almost always have at least a bachelor’s degree from an accredited non-profit university or college, mostly in a STEM (Science, Technology, Engineering, and Mathematics) field with CS (Computer Science) and EE (Electrical Engineering) the largest sub-groups.

I have personally never encountered a graduate from controversial for-profit schools like DeVry, University of Phoenix, etc. or a bootcamp.  Even developers with a solid work history but no bachelor’s degree seem to encounter a significant prejudice against them.

Yes, Bill Gates and Mark Zuckerberg dropped out of college and made it big in software, but they are rich kids who graduated from elite prep schools and then dropped out of Harvard.

The article has a brief line about a Haskell programmer making $250,000 in the Silicon Valley. It is not clear if the author actually knows of a case like this. If real, it is probably very unusual. Top software engineers seem to be bringing in a base salary of around$150,000 in the Silicon Valley:

http://spectrum.ieee.org/view-from-the-valley/at-work/tech-careers/a-snapshot-of-software-engineering-salaries-at-silicon-valley-startups

There is always the question of stock options and RSU’s (restricted stock units) and cash bonuses which can sometimes boost the base salary significantly.

Keep in mind the Silicon Valley/San Francisco Bay Area is very expensive with some of the highest home prices and apartment rental rates in the United States.  The salaries are still attractive but not nearly as large as they sound if you are from an inexpensive region like Texas.

The bottom line is to be very cautious about paying large sums of money for coding bootcamps or other non-traditional education.

(C) 2017 John F. McGowan, Ph.D.

John F. McGowan, Ph.D. solves problems using mathematics and mathematical software, including developing gesture recognition for touch devices, video compression and speech recognition technologies. He has extensive experience developing software in C, C++, MATLAB, Python, Visual Basic and many other programming languages. He has been a Visiting Scholar at HP Labs developing computer vision algorithms and software for mobile devices. He has worked as a contractor at NASA Ames Research Center involved in the research and development of image and video processing algorithms and technology. He has published articles on the origin and evolution of life, the exploration of Mars (anticipating the discovery of methane on Mars), and cheap access to space. He has a Ph.D. in physics from the University of Illinois at Urbana-Champaign and a B.S. in physics from the California Institute of Technology (Caltech).

## The Accuracy of Fingerprint Identification in Criminal Cases

In the wake of the Brandon Mayfield case (2004) which raised serious questions about the accuracy of fingerprint identification by the FBI, the National Academy of Sciences was asked to perform a scientific assessment of the accuracy and reliability of latent fingerprint identification in criminal cases.  Initial results were published in:

Proceedings of the National Academy of Sciences (PNAS)

Bradford T. Ulery, 7733–7738, doi: 10.1073/pnas.1018707108

Accuracy and reliability of forensic latent fingerprint decisions

Bradford T. Ulery (a), R. Austin Hicklin (a), JoAnn Buscaglia (b),1, and Maria Antonia Roberts (c)

Edited by Stephen E. Fienberg, Carnegie Mellon University, Pittsburgh, PA, and approved March 31, 2011 (received for review December 16, 2010)

ABSTRACT

The interpretation of forensic fingerprint evidence relies on the expertise of latent print examiners. The National Research Council of the National Academies and the legal and forensic sciences communities have called for research to measure the accuracy and reliability of latent print examiners’ decisions, a challenging and complex problem in need of systematic analysis. Our research is focused on the development of empirical approaches to studying this problem. Here, we report on the first large-scale study of the accuracy and reliability of latent print examiners’ decisions, in which 169 latent print examiners each compared approximately 100 pairs of latent and exemplar fingerprints from a pool of 744 pairs. The fingerprints were selected to include a range of attributes and quality encountered in forensic casework, and to be comparable to searches of an automated fingerprint identification system containing more than 58 million subjects. This study evaluated examiners on key decision points in the fingerprint examination process; procedures used operationally include additional safeguards designed to minimize errors. Five examiners made false positive errors for an overall false positive rate of 0.1%. Eighty-five percent of examiners made at least one false negative error for an overall false negative rate of 7.5%. Independent examination of the same comparisons by different participants (analogous to blind verification) was found to detect all false positive errors and the majority of false negative errors in this study. Examiners frequently differed on whether fingerprints were suitable for reaching a conclusion.

http://www.pnas.org/content/108/19/7733.full

Authors

(a) Noblis, 3150 Fairview Park Drive, Falls Church, VA 22042;

R. Austin Hicklin (a) Noblis, 3150 Fairview Park Drive, Falls Church, VA 22042;

JoAnn Buscaglia (b) Counterterrorism and Forensic Science Research Unit, Federal Bureau of Investigation Laboratory Division, 2501 Investigation Parkway, Quantico, VA 22135; and

Maria Antonia Roberts (c) Latent Print Support Unit, Federal Bureau of Investigation Laboratory Division, 2501 Investigation Parkway, Quantico, VA 22135

Whether a 0.1 percent false positive rate is “small” is a subjective value judgement. Would you drive across a bridge that had a 1 in 1000 (0.1 percent) chance of collapsing and killing you as you drove across it? No, probably not.

In addition, the 0.1 percent false positive rate is based on a small sample of less than 1000 test cases, 744 pairs of latent and exemplar fingerprints. The Federal fingerprint databases such as the ones used in the Brandon Mayfield case have millions of people in them and may eventually have all US citizens (over 300 million people) in them. How does this “small” rate extrapolate when a fingerprint is compared to every fingerprint in the US or the world?

One might wonder why such an assessment was not done a long time ago.

This is a report on the Brandon Mayfield case:

https://oig.justice.gov/special/s0601/exec.pdf

The National Research Council also published a detailed report Strengthening Forensic Science in the United States: A Path Forward in 2009 addressing the scientific issues raised by the Mayfield case and other questions about the scientific validity of forensic science methods.

Fingerprint identification: advances since the 2009 National Research Council report by Christophe Campod (Philos Trans R Soc Lond B Biol Sci. 2015 Aug 5; 370(1674): 20140259.
doi: 10.1098/rstb.2014.0259) has a summary of work on the issue since the 2009 National Research Council Report.

The bottom line is fingerprints are much more accurate than random chance but hardly infallible as used to be widely believed.

(C) 2017 John F. McGowan, Ph.D.

Credits

The fingerprint image is from the United States National Institute of Standards and Technology (NIST) by way of Wikimedia Commons and is in the public domain.

John F. McGowan, Ph.D. solves problems using mathematics and mathematical software, including developing gesture recognition for touch devices, video compression and speech recognition technologies. He has extensive experience developing software in C, C++, MATLAB, Python, Visual Basic and many other programming languages. He has been a Visiting Scholar at HP Labs developing computer vision algorithms and software for mobile devices. He has worked as a contractor at NASA Ames Research Center involved in the research and development of image and video processing algorithms and technology. He has published articles on the origin and evolution of life, the exploration of Mars (anticipating the discovery of methane on Mars), and cheap access to space. He has a Ph.D. in physics from the University of Illinois at Urbana-Champaign and a B.S. in physics from the California Institute of Technology (Caltech).

## Microsoft Layoffs and STEM Shortage Claims (2009-2017)

Microsoft and its former CEO and founder Bill Gates are prominent in claiming that there is a severe shortage of STEM (Science, Technology, Engineering, and Mathematics) workers in the United States.  Bill Gates testified on this claim to the House Committee on Science and Technology in 2008.

I know we all want the U.S. to continue to be the world’s center for innovation. But our position is at risk. There are many reasons for this but two stand out. First, U.S. companies face a severe shortfall of scientists and engineers with expertise to develop the next generation of breakthroughs. Second, we don’t invest enough as a nation in the basic research needed to drive long-term innovation.

Bill Gates

Remarkably, Microsoft appears to have laid off about 35,000 of these allegedly rare, difficult to find STEM workers since 2008, with even more planned layoffs announced a few weeks ago.

Microsoft Layoffs

In January 2009, Microsoft announced planned layoffs of 5,000 employees, about five (5) percent of its workforce over the next eighteen months.

In July 2014, Microsoft announced layoffs of 18,000 employees.  Most of these employees, reportedly about 12,500, were part of the Nokia mobile phone division, many in Finland.  In 2015 Finland students were ranked sixth (6th) worldwide in math and science compared to the United States twenty-eighth (28th).  In 2001 Finland was tops in the PISA international tests.  The engineers and other STEM workers laid off by Microsoft would have been educated in Finland’s schools in the early 00’s when Finland was at or near the top.

In July 2015, Microsoft announced layoffs of 7,800 employees, also mostly related to Nokia.

In May 2016, Microsoft announced layoffs of about 2,000 employees, including about 1300 from Nokia.

In July 2016, Microsoft announced layoffs of about 2,850 employees.

In July 2017 (a few weeks ago) Microsoft confirmed reports of planned layoffs without confirming reports that about 3,000 employees would lose their jobs, primarily in sales.

Thus, Microsoft appears to have laid off about 35,000 employees with more cuts likely in the coming year since Bill Gates testimony to the House Committee on Science and Technology.  Microsoft reported to the SEC that it had about 114,000 full time employees in 2016.

Stack and Rank

Up until 2013, Microsoft overtly practiced a stack and rank employment system where employees were graded on a curve compared to co-workers and “low performers” apparently laid off or fired.  This stack and rank system was the subject of a highly critical article in Vanity Fair by Kurt Eichenwald in July 2012 which probably contributed to the decision to shelve the system.  It is unclear how many allegedly difficult to find and replace STEM workers were laid off, fired or constructively discharged due to stack ranking.

Microsoft has been sued over allegedly using stack ranking to discriminate against female employees.

Microsoft like other industry leaders such as Google, Facebook, Apple, and Amazon is noted for being extremely picky about who it even interviews for jobs and for a grueling, highly demanding interview process.  Nonetheless, Microsoft appears to have had a policy of laying off a certain percentage of these highly qualified STEM workers every year despite repeatedly claiming to have great difficulty in finding these same STEM workers!

Conclusion

Microsoft is not alone in announcing sizable layoffs at the same time that it claims a STEM worker shortage.  Many other  large STEM worker employers do the same thing.  In an exchange on Bloomberg TV in August 2014 interviewer Alix Steel confronted industry funded “immigration reform” PAC FWD.us then chief Joe Green on the inconsistency between numerous layoff announcements and the shortage claims.  His answer was especially unconvincing and he soon resigned as chief of FWD.us probably at the behest of his friend and colleague Facebook CEO Mark Zuckerberg.

It is difficult to know what to make of this.  On a short term quarterly basis replacing a highly experienced and more expensive STEM worker with a less experienced, cheaper, more error prone STEM worker is likely to make the quarterly and sometimes annual earnings numbers look better.  However, there is a reason more experienced STEM workers are on average more expensive than less experienced STEM workers.  Some problems simply require more experience to solve; two less experienced STEM workers is not always equivalent to one more experienced STEM worker.

I personally don’t doubt that these bizarre hiring and employment practices have seriously negative consequences in the longer term.  Many of Kurt Eichenwald’s unnamed sources in his Vanity Fair article on Microsoft’s stack and rank employment system blamed the system for Microsoft’s faltering fortunes.  Would Microsoft not have been better off reassigning its highly skilled workers in Finland to new projects?

Nonetheless, despite the STEM shortage claims and despite what seems like common sense, many major STEM worker employers like Microsoft continue to lay off, fire, or constructively discharge large numbers of the qualified STEM workers they claim they want.

(C) 2017 John F. McGowan, Ph.D.

Credits

The picture of Bill Gates at the World Economic Forum 2012 in Davos, Switzerland is from the World Economic Forum by way of Wikimedia Commons.  It is licensed under the Creative Commons Attribution 2.0 Generic license.