David P. Doane
Oakland University
Journal of Statistics Education Volume 12, Number 1 (2004), ww2.amstat.org/publications/jse/v12n1/doane.html
Copyright © 2004 by David P. Doane, all rights reserved. This text may be freely shared among individuals, but it may not be republished in any medium without express written consent from the author and advance notification of the editor.
Key Words: Computerintensive methods; Decisions; Risk.
Many students doubt that statistical distributions are of practical value. Simulation makes it possible for students to tackle challenging, understandable projects that illustrate how distributions can be used to answer “whatif” questions of the type often posed by analysts. Course materials that have been developed over two years of classroom trials will be shared, including (1) overviews of distributions and simulation; (2) basic capabilities of @RISK software; (3) simulation spreadsheets suitable for analysis by teams; and (4) exercises to guide the teams. These revisable materials could also be used as inclass demonstrations. Concepts illustrated include expected value, ktiles (e.g., quartiles), empirical distributions, distribution parameters, and the law of large numbers. For those who don't have @Risk, spreadsheets are provided which demonstrate elementary riskmodeling concepts using only Excel. All materials can be downloaded from www.sba.oakland.edu/Faculty/DOANE/downloads.htm .
A few years ago, a former student (now a CFO) volunteered to talk to my introductory statistics students about how statistics helped his company. He candidly told the class that he had forgotten a lot of what he learned in statistics. But with a single example, he showed how the normal distribution had improved a management dialogue about the profit/loss implications of fluctuations in the price of a key raw material that had a strong effect on his company’s bottom line. From historical data, they determined that the price variation was roughly normal, and calculated the best 5% and worst 5% cashflow possibilities. Discussion became focused on probabilities of future cash flows. In effect, the normal distribution served as a behavioral tool to sharpen a debate.
Students were complimentary of his talk. One said that he never realized distributions were actually useful! That got me to thinking about the way I was covering distributions in my classes. Although I am a dataoriented teacher, when it came to distributions I was following the book, working problems, creating exam questions of the familiar sort, and moving on to more transparently practical topics such as confidence intervals. It was time to rethink my approach to distributions.
Most textbooks discuss distributions using formulas (for the mathematicallyinclined), graphs (for the visuallyoriented), and examples/scenarios/problems (for the practicalminded). The assumption is that one or more of these methods of presentation will appeal to the learner’s pedagogical style. Yet doubts may trouble the thoughtful student: (1) If a measurement is only taken once, does a distribution have meaning? This question is especially germane in business and economics, where replicated experiments are rare. (2) Are distributions useful in real business decisions? I believe this second doubt arises because students see mostly sanitized or contrived textbook illustrations. (3) Students who are familiar with seemingly nonstochastic accounting statements (balance sheets, income statements, cash flow projections) may feel that financial decisions do not require statistics at all (let alone distributions).
I concluded that I needed to seek new ways to help practicalminded students (the vast majority) attempt their own answers to these questions. My search focused on the potential roles of distributions in (1) exploring our assumptions about key variables, and (2) structuring business dialogue. Basically, I was led toward a focus on risk modeling and decisionmaking. I wanted to capitalize on students’ acceptance of computers and spreadsheets as tools of proven value in the “real world” and to encourage individual exploration. Although I am still exploring this pedagogical approach, I feel increasingly confident that it is the right direction. Although a few textbooks have begun moving this way, my personal ontogeny required me to create my own teaching materials to see what worked for the types of students I normally encounter.
The purpose of this paper is to share my ideas on using spreadsheet simulation to motivate teaching of distributions, to share classroom materials and projects that I have developed, to obtain feedback and suggestions from colleagues, and to stimulate dialogue on the simulation/decision modeling approach.
Participants in the MSMESB conferences or the annual JSM sessions on statistical education will realize that educators everywhere are using spreadsheets, collaborative projects, and case studies to help students learn statistics. It is difficult to quantify how many use simulation per se, although clearly many do, as is evident in Mills’ (2002) comprehensive review of the literature on simulation in teaching statistics and probability. One suspects that many interesting approaches do not receive widespread attention. Nonetheless, there is agreement on the importance of computers (Strasser and Ozgur 1995), projects (Ledolter 1995; Love 1998; Holcomb and Ruffer 2000), cases (Parr and Smith 1998; Brady and Allen 2002), active learning (Snee 1993; Gnanadesikan 1997; Smith 1998), and cooperative learning (e.g., Magel, 1998).
Simulation materials can range from simple spreadsheets (Ageel 2002) to an entire language (Eckstein and Riedmueller 2001). Management science educators have taken an especially active role in using spreadsheet simulation of distributions to teach students to solve business problems, using both Monte Carlo simulation (e.g., Evans 2000) and process simulation (e.g., Hill 2001). While management science classes build upon and reinforce existing statistical skills for business and industrial engineering students, they do not directly address the question of how students learn statistics. Our present purpose is limited to using simulation to teach distributions, a subject which seems to have received somewhat less attention, unless we adopt broad definitions to include Central Limit Theorem demonstrations and sampling distributions.
Rather than viewing distributions as descriptions of data, we can view distributions as a way of visualizing our assumptions. For example, many students will tell the instructor “I am a B student” (call it 80 on a typical grading scale). Yet grades clearly are stochastic. Any student realizes that an 80 average does not imply an 80 on every exam. A more precise statement might be, “My exam grades range from 70 to 90 but on average I get 80.” Figure 1 shows three simple ways of visualizing the situation.
Uniform  Normal 
It can be seen that all three of the distributions shown in Figure 1 have an area of 1 (for the normal, using a triangle approximation). After seeing these diagrams, the student will deny that all scores within the range a to b would be equiprobable, hence rejecting the uniform model U(a, b). Assuming a normal distribution and the 4sigma empirical rule (Browne 2002) we could set = R/4 to model the situation as N(80, 5). Of course, students do not think in terms of normality, but upon seeing the sketch, they will agree that the implied N(, ) is a credible model of what they expect when they take an exam. But my experience suggests that the triangular distribution T(a, b, c) or T(min, max, mode) best approximates the way nontechnical people think, particularly in business. Although the triangular distribution is not mentioned in most textbooks, it is attractive because (a) it is easy to visualize, (b) it is flexible, (c) it prevents extremes, (d) it can be skewed, and (e) areas can be found using the familiar formula A = ½bh. The symmetric triangular resembles a normal, except that it lacks extreme values (see Appendix A for a few basic facts about the triangular distribution).
Monte Carlo simulation variables can be either deterministic (nonrandom or fixed) or stochastic (random variables). If a variable is stochastic, we must hypothesize its distribution (normal, triangular, etc.). By allowing the stochastic variables to vary, we can study the behavior of the output variables that interest us, to establish their ranges and likelihood of occurrence. But we are mainly interested in the sensitivity of our output variables to variation in the stochastic input variables.
Component 
Explanation 
List
of deterministic factors F_{1}, F_{2},
..., F_{m} 
These
are quantities that are known or fixed, or whose behavior we choose not to model
(i.e., exogenous). 
List
of stochastic input variables V_{1}, V_{2},
..., V_{k} 
These
are quantities whose value cannot be known with certainty, and are assumed to
vary randomly. 
List
of output variables O_{1}, O_{2}, ..., O_{p} 
These
are stochastic quantities that are important to a decision problem, but whose
value depends on many things in the model and whose distribution is not
easily found. 
Assumed
distribution for each stochastic input variable 
These
are chosen from known statistical distributions, such as normal, Poisson,
triangular, and so on. 
A model
which specifies the rules or formulas that define the relationships among Fs
and Vs 
Formulas
including simple accounting identities such as Profit = Revenue  Cost or
End Balance = Begin Balance  Withdrawals + Deposits, or behavioral hypotheses such as Car Sales
= b_{0} + b_{1}(Personal Income After Taxes) + b_{2}(Net
Worth) + b_{3}(3month TBill Interest Rate). 
A simulation
method that produces random data from the specified distributions and
captures the results. 
This
is a programming language (such as VBA) although it may be embedded invisibly
in a spreadsheet with builtin functions like Excel’s RAND() or other addins. 
An interface
that summarizes the model’s inputs, outputs, and simulation results. 
Typically, spreadsheet
tables or graphs to summarize the outcomes of the simulation. 
Before assigning any independent projects, the teacher needs a simple classroom demonstration. This one uses a simple spreadsheet which can be downloaded along with the student projects. The Axolotl Corporation sells three products (A, B, and C). Prices are set competitively, and are assumed constant. The quantity demanded, however, varies widely from month to month. To prepare a revenue forecast, the firm sets up a simple simulation model of its input variables, as shown in Table 2. The output variable of interest is total revenue P_{A} Q_{A} + P_{B} Q_{B} + P_{C} Q_{C}.
Variable 
Type 
Product A 
Product B 
Product C 
Price 
Deterministic 
P_{A}
= 80 
P_{A}
= 150 
P_{A}
= 400 
Quantity 
Stochastic
input 
Normal Q_{A}
~ N(50, 10) m =50, s = 10 
Triangular Q_{B} ~ T(0, 40, 5) Min =0, Max = 40, Mode = 5 
Exponential Q_{C}
~ E(2.5) l = 2.5 
Revenue 
Stochastic output 
P_{A}
Q_{A} 
P_{B}
Q_{B} 
P_{C}
Q_{C} 
Variation in the quantity demanded would make it difficult to predict total revenue. We could predict its mean, based on the mean of each distribution, but what about its range? This demonstration shows that simulation reveals things that are not obvious.
Table 3 shows the results of a typical simulation run using 100 Monte Carlo iterations (i.e., each input cell value is sampled, results are computed, and the process is repeated 100 times). This simulation can be repeated another 100 times by clicking a button. Results will vary, but as long as the number of iterations is reasonably large, the conclusions will be similar.
Percentile 
Product A 
Product B 
Product C 
Total Revenue 
Min 
26 
1 
0 
4,180 
5% 
34 
4 
0 
4,745 
25% 
44 
8 
1 
5,943 
50% 
50 
14 
2 
7,000 
75% 
56 
20 
3 
8,340 
95% 
64 
32 
6 
10,022 
Max 
69 
35 
11 
10,780 





Sample Mean 
49.85 
15.11 
2.30 
7,335 
Theory Mean 
50 
15 
2.5 
7,250 
Note The mean
demand of the triangular is m_{B} = (a + b
+ c)/3 = (0 + 40 + 5)/3 = 15.
Inserting other demand means m_{A} = 50 and m_{C} = 2.5 we can find the theoretical mean of total
revenue as P_{A} m_{A} + P_{B} m_{B} + P_{C} m_{C} = (80)(50) + (150)(15) +
(400)(2.5) = 4,000 + 2,250 + 1,000 = 7,250.
This compares well with the sample result over 100 iterations. Although product A contributes the
most to total revenue at the mean, this may not be the case in a
particular simulation, because demand can fluctuate. 
These results suggest that Axolotl’s revenue could be as low as $4,180 or as high as $10,780, with a 50% chance (25th and 75th percentiles) of revenue between $5,943 and $8,340. Median revenue seems to be below the mean, suggesting that total revenue is rightskewed. That is e expected, since both the triangular (Product B) and exponential (Product C) are rightskewed distributions (as is also evident in the simulation results for demand for Products B and C). The symmetric normal distribution (Product A) is reflected in the simulation results, which lie well within the ± 3 limits. Figure 2 summarizes the simulation results visually. Notice that colorcoding is used in the spreadsheet and graphs to distinguish inputs (purple) from outputs (orange).
Product A  Product B 
Product C  Total 
Figure 2. Histograms for 100 Iterations of Revenue Simulation
Once the basic idea is in place, it is natural to have students explore what can happen when they use distributions for stochastic decision parameters in a new scenario. My design goals were to develop scenarios that (1) were entertaining, (2) involved familiar problems, (3) had a clear “bottom line,” (4) called for “whatif” analysis, and (5) were complex enough that the solutions weren’t obvious. In short, I set out to create “light” simulation scenarios with sufficient relevance to attract practicalminded students.
Initially, I selected @Risk as the simulation tool (although Crystal Ball would be equivalent) because I was able to negotiate a reasonable oneyear academic site license fee (about $500). My university’s Teaching and Learning Committee underwrote the first year’s cost, and my academic unit agreed to pick up the license thereafter if the experiment proved successful. I developed three scenarios using @Risk and tried them in the classroom. Realizing that there is a learning curve for @Risk (as well as an added cost) I later developed Excelonly versions, although at the sacrifice of some powerful features of @Risk.
The scenarios presented here are intended to be undertaken by student teams of two students in a onehour computer lab session under instructor supervision. I find that more than two students cannot effectively share a single computer and have trouble agreeing on the joint report under a tight deadline. Why use teams at all? First, because even if one partner is dominant, interaction is beneficial to both. Second, teams more closely approximate the workplace. Each inclass project can be done in 6090 minutes if the simulation tools are demonstrated during a prior lecture and scenario overviews and project instructions are handed out in advance. The onehour deadline forces closure, just as in the “real world” where a report must be done by the deadline even if a better one could be written. A little time pressure rewards teams who do not dawdle.
The three scenarios discussed here use only the normal N(, ) and triangular T(a,b,c) distributions because they are flexible yet easy to understand. Each scenario involves a problem faced by a hypothetical character named “Bob.”
We illustrate with Bob’s Stochastic Balance Sheet. The basic spreadsheet (Excel version) is shown in Figure 3. Where relevant, cell range names are used (e.g., Net Worth) to increase students’ familiarity with this useful alternative to cell references (e.g., H9). Comments have been added to cells that specify stochastic inputs (purple highlight) or stochastic outputs (orange highlight). The idea is that the output cells are “bottom line” variable of interest while input cells are the “drivers” of the output(s).
Figure 3. Bob's Stochastic Balance Sheet
The spreadsheet contains a description of the scenario, a set of questions, and a graph showing Bob’s net worth. The idea is that on a given day, Bob’s actual net worth depends on the market value of his assets. Some of his asset and liability values are deterministic (e.g., checking account, savings account, student loans) while the values of his car, beer can collection, and stocks (and hence his net worth) are stochastic. Table 4 shows the characteristics of the stochastic inputs (distribution, skewness, coefficient of variation).
Asset 
Distribution 
Comments 
Mustang 
T(8000,15000,10000) 
Rightskewed 
Beer can collection 
T(0,1000,50) 
Very rightskewed 
Garland stock 
N(15.38,3.15) 
Symmetric, large C.V. 
Oxnard stock 
N(26.87,2.02) 
Symmetric, small C.V. 
Scamco stock 
N(3.56,0.26) 
Symmetric, small C.V. 
The graph and highlighted cells will update with new simulated values each time F9 is pressed. The goal is to get the learner to visualize (1) what values could appear in each cell, and (2) what these distributions say about Bob’s reasoning. For example, Bob thinks that if he finds the right buyer, his Mustang could be worth up to $15,000. He is quite sure he won’t get less than $8,000. He figures that $10,000 is the most likely value. This way of thinking strikes many students as congruent with their own.
While the prices of Bob’s stocks vary without limits, as a practical matter even ± 3 is conservative. Following the logic of probability plots (D'Agostino and Stephens 1986), the range that is likely to be observed in a sample is from F^{1}(0.5/n) to F^{1}(1(0.5/n)) where F^{1} is the inverse of the cumulative distribution function. For n = 100, the likely sample extremes are from F^{1}(0.005) to F^{1}(0.995). For the Garland stock this implies values between 15.38 ± 2.576(3.15) or $7.27 to $23.91. @Risk offers a truncated normal, to prevent extreme values, but its theoretical properties are awkward.
In this Excelonly version of Bob’s Stochastic Balance Sheet, a second worksheet shows the results of 100 iterations of sampling. This provides a basis for answering the question posed (and suggests some new questions). There are histograms for each input and output variable. Every time F9 is pressed, 100 new iterations will occur. Students can observe that despite variation in the individual stochastic cells, there is stability in the mean, standard deviation, and ktiles of each stochastic variable.
Figure 4. Simulation Summary for Bob's Balance Sheet (Excel Version)
@Risk is an Excel addin software package with powerful simulation capabilities (see Winston 2001). The general idea of @Risk is to provide builtin functions to create random values in designated cells. When we paste an @Risk input function for the desired statistical distribution in a cell, its contents become stochastic, so that every time the spreadsheet is updated a new value will appear. For example, an input cell containing =RiskNormal(26,4) is a random variable with = 26 and = 4. All the common distributions are available from Excel’s InsertFunction menu. An output cell is calculated as usual except that =RiskOutput()+ is added in front of the cell’s contents (e.g., =RiskOutput + TotalAssets  TotalDebt where TotalAssets and TotalDebt are defined elsewhere in the spreadsheet (of course, you can also use cell references like C12 and H7 instead of cell names). Thus, @Risk automates tasks that otherwise would require VBA programming.
Each time F9 is pressed, the spreadsheet will show updated values of all input and output cells. In replication mode, we can do up to 10,000 replications while @Risk keeps track of all simulated values of the input and output cells. For those variables, empirical distributions can be displayed in tables (statistics, percentiles) or charts (histograms, cumulative distributions). Using sliders, you can reveal any desired percentile on the cumulative distribution. Tornado charts can be used to reveal sensitivities of output variables to all the input variables.
Figure 5. Buttons to Set Up the Simulation
The @Risk toolbar lies above the regular Excel toolbar, as illustrated in Figure 5. The @Risk setup screens and typical settings are shown in Figure 6. Experience suggests that 500 iterations will suffice for the distributions used in our scenarios.
Iterations  Sampling 
Macros  Monitor 
Figure 6. Suggested Simulation Settings
Various reports can be generated and placed either in a new workbook or in the active workbook. To begin, click the Start Simulation button. A screen like Figure 7 will appear. Select a variable by clicking its row tab (in this illustration, we chose Output H9Net Worth). You will get a menu of graphs that are available.
Figure 7. Selecting a Graph
By default, the middle 90% of the outcomes are shown. In Figure 8, the distribution of net worth is symmetric and platykurtic with a mean of $8,881, exceeding $10,000 only about 5% of the time.
Figure 8. Histogram of 500 Iterations of Net Worth
You can drag the vertical sliders to show different percentiles. This is easy, but inexact. To select
Figure 9. Distribution of Net Worth for 500 Iterations
Figure 10. Tornado Graph for Sensitivity Analysis
After using these materials three times in a required undergraduate class and once in an MBA class, I can report that students are attracted to this approach. One of the exercises asks the team to rate advantages and disadvantages of @Risk simulation. Based on their comments and my own observations, I conclude that the advantages of @Risk are its ease of use, power, and versatility. Drawbacks of @Risk are its high price, lack of general availability to students after they leave the university, and rarity of use in universities, partly due to administrative and licensing issues. These drawbacks have prompted some educators to develop their own Excelonly simulation addins (e.g., Eckstein and Riedmueller 2001), although they must still teach students to use the addin.
These are the motivations for my simple Excelonly versions of the simulation spreadsheets without addins, despite the loss of powerful features such as sensitivity analysis and slider bars for percentiles. Another limitation is that the Excelonly histograms are set up for the initial parameter values, so they will not look right if the parameters are changed. But an advantage of the Excelbased versions is that curious students can “see how it is done” by looking at the cells, histogram bins, and tables of results, whereas @Risk’s functions are “black boxes.” Without VBA programming, we can create the normal and triangular distributions used in the three scenarios, although at the cost of some unaesthetic columns of intermediate calculations. While VBA would be needed to create most discrete distributions (e.g., binomial, Poisson) Appendix B shows Excel functions to create random data from a few common distributions.
Another issue is that mainstream textbooks do not discuss risk modeling, @Risk, or simulation in general. These become extra topics which impose a burden on the instructor. Students are usually suspicious of material that is “not in the book,” fearing that their instructor is going to subject them to some kind of iconoclastic experiment. There now are textbooks on risk modeling (e.g., Albright, Winston, and Zappe 1999; Evans and Olsen 2001; Moore and Weatherford 2001). But mainstream practice evolves very slowly. It is noteworthy that some popular textbook authors have tried abolishing some distribution tables (e.g., binomial, Poisson), partly for brevity but also because Excel functions make tables obsolete. But can students take exams without tables?
Syllabus crowding is also an issue. With an alreadycrowded syllabus, it is not easy to convince faculty to spend 23 extra class days for risk modeling and lab projects. My colleagues have not jumped on the @Risk bandwagon for this explicit reason. There is also resistance to teaching yet another software package like @Risk. We already have our hands full with Excel and Minitab.
Should we reduce traditional treatment of distributions to yield curricular space for simulation and risk modeling? Can simulation and risk modeling cases somehow increase understanding enough to replace time currently spent on distributions? The answer may be “not at this point.” Perhaps if Excel had builtin simulation capabilities the answer could be “yes.” Faculty may become convinced if they see that students enjoy doing risk modeling projects and gain useful knowledge. Experts (e.g., Love and Hildebrand 2002) do recommend less lecture and more use of collaborative cases, so there is support for experimentation of this sort.
One contemporary thought question: would companies whose defined benefit pension funds in 2002 faced underfunding because stock market returns have been subnormal have found it useful to engage in @Risk simulation, using (say) triangular distributions of possible returns on their investment portfolios? If so, perhaps at least a contingency plan could have been created to anticipate cash flow problems.
Interested parties can download the 542K file JSE.zip directly or from my web site:
www.sba.oakland.edu/Faculty/DOANE/downloads.htm
Appendix C lists the files contained in JSE.zip. The Word files and @Risk scenarios can be modified to fit local situations. However, as explained earlier, care must be taken in modifying the Excelonly versions because of assigned variable names, fixed histogram scales, and complex cell links. Permission is given to use these materials in an academic setting, as long as they are not resold for profit and are properly attributed. As a courtesy, I would appreciate comments and suggestions from those who use them.
Ageel, M.I., (2002) “Spreadsheets as a Simulation Tool for Solving Probability Problems,” Teaching Statistics, 24(2), 5154.
Albright, S. Christian., Winston, Wayne. L., and Zappe, Christopher (1999), Data Analysis and DecisionMaking with Microsoft Excel, Pacific Grove, CA: BrooksCole.
Brady, James E., and Allen, Theodore T. (2002), “Case Study Based Instruction of DOE and SPC,” The American Statistician, 56(4), 312315.
Browne, Richard H. (2002), “Using the Sample Range as a Basis for Calculating Sample Size in Power Calculations,” The American Statistician, 55(4), 293298.
D’Agostino, Ralph B., and Stephens, Michael A. (1986), Goodness of Fit Techniques, New York: Marcel Dekker, Inc., pp. 2435.
delMas, Robert C., Garfield, Joan, and Chance, Beth L. (1999), “A Model of Classroom Research in Action: Developing Simulation Activities to Improve Students’ Statistical Reasoning,” Journal of Statistics Education [Online], 7(3). ww2.amstat.org/publications/jse/secure//v7n3/delmas.cfm
Eckstein, Jonathan, and Riedmueller, Steven T. (2001), “YASAI: Yet Another AddIn for Teaching Elementary Monte Carlo Simulation in Excel,” INFORMS Transactions on Education [Online], 2(2). ite.pubs.informs.org/Vol2No2/EcksteinRiedmueller/index.php
Evans, James R., and Olsen, D. L. (2001), Introduction to Simulation and Risk Analysis, 2nd Ed., PrenticeHall.
Evans, James R. (2000), “Spreadsheets as a Tool for Teaching Simulation,” INFORMS Transactions on Education [Online], 1(1). ite.pubs.informs.org/Vol1No1/Evans/index.php
Gnanadesikan, Mrudulla, Scheaffer, Richard L., Watkins, Ann E., and Witmer, Jeffrey A. (1997), “An ActivityBased Statistics Course,” Journal of Statistics Education[Online], 5(2). ww2.amstat.org/publications/jse/v5n2/gnanadesikan.html
Hill, Raymond R., (2001) “Process Simulation in Excel for a Quantitative Management Course,” INFORMS Transactions on Education [Online], 2(3). ite.pubs.informs.org/Vol2No3/Hill/index.php
Holcomb, John P., and Ruffer, Rochelle L. (2000), “Using a LongTerm Project Sequence in Introductory Statistics," The American Statistician, 54(1), 4953.
Ledolter, Johannes (1995), “Projects in Introductory Statistics Courses," The American Statistician, 49(4), 364367.
Love, Thomas E. (1998), “A ProjectDriven Second Course,” Journal of Statistics Education [Online], 6(1). ww2.amstat.org/publications/jse/v6n1/love.html
Love, Thomas E. and Hildebrand, David K. (2002), “Statistics Education and the MSMESB Conference,” The American Statistician, 56(2), 107111.
Magel, Rhonda C. (1998), “Using Cooperative Learning in a Large Introductory Statistics Class,” Journal of Statistics Education [Online], 6(3). ww2.amstat.org/publications/jse/v6n3/magel.html
Mills, Jamie D. (2002), “Using Computer Simulation Methods to Teach Statistics: A Review of the Literature,” Journal of Statistics Education [Online], 10(1). ww2.amstat.org/publications/jse/v10n1/mills.html
Moore, Jeffrey H.. and Weatherford, Lawrence R. (2001), Decision Modeling with Microsoft Excel, 6th Edition, PrenticeHall.
Parr, William C. and Smith, Marlene A., (1998) “Developing CaseBased Business Statistics Courses,” The American Statistician, 52(4), 330337.
Smith, Gary (1998), “Learning Statistics By Doing Statistics,” Journal of Statistics Education [Online], 6(3). ww2.amstat.org/publications/jse/v6n3/smith.html
Snee, R. D. (1993), “What’s Missing in Statistical Education?” The American Statistician, 47(2), 149154.
Strasser, Sandra E., and Ozgur, Ceyhun (1995) “Undergraduate Business Statistics: A Survey of Topics and Teaching Methods,” Interfaces, 25(3), 95103.
Winston, Wayne L. (2001), Simulation Modeling Using @RISK, Pacific Grove, CA: BrooksCole.
Random numbers are at the heart of any simulation. In general, if you know F(x), the cumulative distribution function (CDF) of your distribution, you generate a uniform U(0,1) random number R and then find F^{1}(R) where F^{1} is the inverse of the CDF. Essentially, what you have to do is to set F(x)=R and then solve for x, as illustrated in Figure 11. However, this is sometimes easier said than done, because finding may be tricky, especially for discrete distributions as in Figure 12. With VBA it is not difficult, and there are plenty of commercial packages that do it. But it is harder if you are a doityourself person who wants only to use the functions available within Excel. Table 13 shows Excel functions to create random data for a few common distributions.
Inverse Continuous CDF  Inverse Discrete CDF 
Figure 11. Inverse Continuous CDF  Figure 12. Inverse Discrete CDF 
Distribution 
What to Put in Excel
Cell 
Explanation 
Uniform
U(0, 1) 
=RAND() 
Builtin
Excel function 
Uniform
U(a, b) 
=$A$1+($A$2$A$1)*RAND() 
$A$1 is the minimum and $A$2 is the maximum (or use cell names like Xmin and Xmax). 
Normal
N(0, 1) 
Method A: =NORMSINV(RAND()) Method B:
=SQRT(LN(RAND()))*COS(2*PI()*RAND()) 
Excel’s
inverse normal function or doityourself BoxMuller transform. 
Normal
(m, s) 
Method A:
=NORMINV(RAND(),$A$1,$A$2) Method B:
=$A$1+$A$2*SQRT(2*LN(RAND()))*COS(2*PI()*RAND()) 
$A$1 is the mean and $A$2 is the standard deviation (or use cell names like Mu and Sigma). 
Exponential
E(l) 
=LN(RAND())*$A$1 
$A$1 is the mean Poisson arrival rate (or use cell name
like Lambda) 
Triangular
T(a, b, c) 
No
single cell formula but can be done in Excel with two cells. 
Better to use @Risk or Crystal Ball. 
Bernoulli B(p) 
=IF(RAND()<$A$1,1,0) 
$A$1 is the
probability of success (or use a cell name like ProbSuccess) 
www.sba.oakland.edu/Faculty/DOANE/downloads.htm
File Name 
Type 
Contents 
SimulationBasics.doc 
Word 
Brief handout on simulation and distributions (examples,
functions, Excel, @Risk) 
HowToUse@Risk.doc 
Word 
Optional handout on @Risk software (methods, iterations, output options) to get teams started if they use the @Risk versions of the cases. 
ProjectInstructions.doc 
Word 
Instructions to student teams for inclass projects using the case studies. Includes separate instructions for Excelonly and @Risk versions. 
ThreeProductExample.xls 
Excel 
Suggested inclass demonstration using only Excel. Illustrates normal, triangular, and exponential distributions using a familiar business example. 
@RiskVersionCase1.xls @RiskVersionCase2.xls @RiskVersionCase3.xls 
Excel 
Three Excel spreadsheet cases using @Risk functions, with questions to be answered by student teams. Requires @Risk software. Stochastic cells will show #NAME? until the @Risk AddIn is installed and @Risk is started. 
ExcelVersion Case1.xls ExcelVersionCase2.xls ExcelVersionCase3.xls 
Excel 
Three Excel spreadsheet cases containing formulas, questions to be answered by the student teams, and a simulation results worksheet. Initial stochastic cell values change when F9 is pressed. Fewer capabilities than the @Risk version, but works using only Excel without any addins. 
David P. Doane
Department of Decision and Information Sciences
Oakland University
Rochester, Michigan 483094493
USA
doane@oakland.edu
Volume 12 (2004)  Archive  Index  Data Archive  Information Service  Editorial Board  Guidelines for Authors  Guidelines for Data Contributors  Home Page  Contact JSE  ASA Publications