Home

Risk Simulation Basics - Real Options Valuation, Inc.

image

Contents

1. 7 Show only data deviates less than std deviation Statistic Precision level that used to calculate the error 7 gne Show the following statistic on histogram Mean 7 1st Quartile 7 Median 4th Quartile Figure 11 Forecast chart options Using Forecast Charts and Confidence Intervals In forecast charts you can determine the probability of occurrence called confidence intervals that 1s given two values what are the chances that the outcome will fall between these two values Figure 12 illustrates that there is a 90 probability that the final outcome in this case the level of income will be between 0 2781 and 1 3068 The two tailed confidence interval can be obtained by first selecting Two Tail as the type entering the desired certainty value e g 90 and hitting Tab on the keyboard The two computed values corresponding to the certainty value then will be displayed In this example there is a 5 probability that income will be below 0 2781 and another 5 probability that income will be above 1 3068 that is the two tailed confidence interval is a symmetrical interval centered on the median or 50th percentile value Thus both tails will have the same probability Sa Figure 12 Forecast chart two tailed confidence interval Alternatively a one tail probability can be computed Figure 13 shows a Left Tail selection at 95 confidence i e choose Left Tail as the type
2. m I Distribution Fitting Result D Gog Ex 0 02 100 Gamma 0 03 0 99 3 0 98 0 97 0 74 2 j 4 5 6 gt 8 5 Mean 100 67 Standard Deviation 10 40 Kolmogorov Smirnov Test Statistic Test Statistic 0 02 P Value 1 00 Mean Stdev Skewness Automatically Generate Assumption Figure 32 Distributional fitting result Statistical Summary Fitted Assumption 100 61 Fitted Distribution Mormal Mean 100 67 Sigma 10 40 Ealmoagoroy Smirnoy Statistic 0 02 P Walue Far Test Statistic 0 9996 Actual Theoretical Mean 100 61 100 67 Standard Deviation 10 31 10 40 Skewnezz 0 07 0 00 Excess Kurtosis D 13 0 00 Original Fitted Data 73 53 78 21 73 52 73 50 73 72 73 74 51 56 82 08 82 68 a2 7h 83 34 83 54 84 03 54 66 35 00 85 35 25 51 56 04 26 79 56 52 36 91 a7 02 ar03 T 45 ar 53 BT 55 55 05 25 45 55 51 83 35 30 13 30 54 30 68 90 96 31 25 381 43 81 56 31 34 92 06 92 36 92 41 92 45 92 70 92 30 42 84 93 21 93 26 33 48 33 73 33 75 33 71 33 82 54 00 34 15 34 51 34 57 34 54 34 563 34 35 35 57 35 62 35 71 35 78 35 23 35 37 96 20 56 24 96 40 35 43 35 47 96 31 35 88 97 00 37 0 37 21 37 23 37 48 37 7 3T TT 37 85 38 15 35 17 38 24 35 28 38 32 38 33 38 35 35 65 99 03 33 27 35 46 33 47 33 55 33 73 33 35 1000 08 100 24 100 36 100 42 100 44 100 48 100 43 100 83 101 17 101 28 101 34 101 45 101 46 101 55 101 73 101 74 101 81 102 29 102 55 102 58 102 60 12 70 103 17 103 21 103 22 103 32 103 34 103 45 103 65
3. Figure 51 Distributional Analysis Tool Binomial Distribution with 20 Trials Figure 51 shows the same binomial distribution but now the CDF is computed The CDF is simply the sum of the PDF values up to the point x For instance in Figure 51 we see that the probabilities of O 1 and 2 are 0 000001 0 000019 and 0 000181 whose sum is 0 000201 which is the value of the CDF at x 2 in Figure 52 Whereas the PDF computes the probabilities of getting 2 heads the CDF computes the probability of getting no more than 2 heads or probabilities of 0 1 and 2 heads Taking the complement 1 e 1 0 00021 obtains 0 999799 or 99 9799 provides the probability of getting at least 3 heads or more Distribution Analysis This tool generates the probability density function PDF cumulative distribution function CDF and the Inverse CDF ICDF of all the distributions in Risk Simulator including theoretical moments and probability chart Dib Trials 20 Probability 0 5 0 411901 0 588039 0 748278 0 868412 0 342341 Figure 52 Distributional Analysis Tool Binomial Distribution s CDF with 20 Trials Using this Distributional Analysis tool even more advanced distributions can be analyzed such as the gamma beta negative binomial and many others in Risk Simulator As further example of the tool s use in a continuous distribution and the ICDF functionality Figure 53 shows the standard normal distribution normal distribution w
4. Use these steps to define the forecasts Select the cell on which you wish to set an assumption e g cell G10 in the Basic Simulation Model example Click on Risk Simulator Set Output Forecast or click on the set forecast icon on the Risk Simulator icon toolbar Enter the relevant information and click OK Figure 7 illustrates the set forecast properties which include Forecast Name Specify the name of the forecast cell This is important because when you have a large model with multiple forecast cells naming the forecast cells individually allows you to access the right results quickly Do not underestimate the importance of this simple step Good modeling practice is to use short but precise assumption names Forecast Precision Instead of relying on a guesstimate of how many trials to run in your simulation you can set up precision and error controls When an error precision combination has been achieved in the simulation the simulation will pause and inform you of the precision achieved Thus the number of simulation trials is an automated process you do not have to guess the required number of trials to simulate Review the section on error and precision control for more specific details Show Forecast Window This property allows you to show or not show a particular forecast window The default is to always show a forecast chart Specify the name of the forecast il Forecast Properties Forecast Name Income
5. Assumption Name This optional area allows you to enter in unique names for the assumptions to help track what each of the assumptions represents Good modeling practice is to use short but precise assumption names Distribution Gallery This area to the left shows all of the different distributions available in the software To change the views right click anywhere in the gallery and select large icons small icons or list More than two dozen distributions are available Input Parameters Depending on the distribution selected the required relevant parameters are shown You may either enter the parameters directly or link them to specific cells in your worksheet Click on the link icon to link an input parameter to a worksheet cell Hard coding or typing the parameters is useful when the assumption parameters are assumed not to change Linking to worksheet cells is useful when the input parameters themselves need to be visible on the worksheets or can be changed as in a dynamic simulation where the input parameters themselves are linked to assumptions in the worksheets creating a multidimensional simulation or simulation of simulations Data Boundary Typically the average analyst does not use distributional or data boundaries truncation but they exist for truncating the distributional assumptions For instance if a normal distribution is selected the theoretical boundaries are between negative infinity and positive infinity Howe
6. E Revenue Positive Correlation Risk Simulator Forecast EI 3 E Revenue Negative Correlation Risk Simulator Forec o El 3 Trials Mean Coefficient of Variation Maximum Minimum Panaka Eror Frecision at 95 Confidence Figure 17 Correlation results Figure 18 illustrates the results after running a simulation extracting the raw data of the assumptions and computing the correlations between the variables The figure shows that the input assumptions are recovered in the simulation that is you enter 0 9 and 0 9 correlations and the resulting simulated values have the same correlations Clearly there will be minor differences from one simulation run to another but when enough trials are run the resulting recovered correlations approach those that were inputted Spearman s Nonlinear Rank Correlation on Raw Data Extracted from Simulation Quantity Price Negative Quantity Negative Price Positive Positive Correlation Correlation Correlation Correlation Correlation Correlation O76 145 0 90 TU 158 0 89 JGA an 464 975 ba rata 515 ayy 239 ore ora E 122 rad rog rg agi bay dat avd 33b 639 be 445 4 oo az 190 241 ar bod Bra bo ord 138 285 oud 59 458 ang 55 a Oo dre F r 62 528 555 Ba 106 op ote Figure 18 Correlations recovered Tornado and Sensitivity Tools in Simulation One of the powerful simulation tools is tornado analysis it captures the static impacts of each
7. Histogram Cumulative Probability 6000 Trials 600 500 0 80 gt 400 i c 0 60 Statistics to Bootstrap 2 300 me g a m 040 F Mean Average Deviation Skewness ean Income Model B Risk Simulator Forecast Distribution Median Maximum Kurtosis 100 Standard Deviation Minimum 25 Percentile Histogram Statistics Preferences Options K L E i digz Variance C Range C 75 Percentile l Model B Hist Cumulative Probab mu ncome ode istogram Lumulative Froba Number of Bootstrap Trials 1002 s E 600 Type Two Tail v m ey 500 0 80 400 300 200 i 100 m il E 72 93 72 98 72 103 72 108 72 113 3400 Type Two Tail infinity Infinity Certainty 10000 e e e Frequency e ho e e A e yyiqeqoag aayenun M Figure 34 Nonparametric bootstrap simulation Mean Risk Simulator Forecast Distribution Standard Deviation Risk Simulator Forecast Distribution tj Histogram Statistics Preferences Options Histogram Statistics Preferences Options Mean Histogram Cumulative Probability 1000 Trials Standard Deviation Histogram Cumulative Probability 1000 Trials 1 gt o c pu c pu t ra ypiqeqoag sAnejunz Frequency ypiqegqo4g eAnepnun 100 15 Type TwoTal B 999028 1000707 Certainty TwoTal w B 35300 36487 Certainty Skewness Risk Simulator Forecast Di
8. enter 95 as the certainty level and hit Tab on the keyboard This means that there is a 9596 probability that the income will be below 1 3068 1 e 9596 on the left tail of 1 3068 or a 5 probability that income will be above 1 3068 corresponding perfectly with the results seen in Figure 12 Figure 13 Forecast chart one tailed confidence interval In addition to evaluating the confidence interval i e given a probability level and finding the relevant income values you can determine the probability of a given income value Figure 14 For instance what is the probability that income will be less than 1 To do this select the Left Tail probability type enter into the value input box and hit Tab The corresponding certainty will then be computed in this case there is a 64 80 probability income will be below 1 Te Letra e EN E Cerny ATH Figure 14 Forecast chart left tail probability evaluation For the sake of completeness you can select the Right Tail probability type and enter the value 1 in the value input box and hit Tab Figure 15 The resulting probability indicates the right tail probability past the value 1 that is the probability of income exceeding 1 in this case we see that there is a 35 20 probability of income exceeding 1 I Income Risk Simulator Forecast Figure 15 Forecast chart right tail probability evaluation Note tha
9. 0 100 0 10 00 5 00 Figure 23 Spider chart Depa 44 mc 43 ATO B Price Tt eo Por ME vg a gm de psa 7 P GILIATILIES B Quantity Pos m E C es ix Discount Rate am percent rane x Price Erosion Sales Growth omai oo22 Depreciation 8 41 Interest 22 18 Amortization 27 3 3 Capex D Met Capital 0 150 100 50 D 50 100 150 200 250 300 350 Figure 24 Tornado chart Although the tornado chart is easier to read the spider chart is important to determine if there are any nonlinearities in the model For instance Figure 25 shows another spider chart where nonlinearities are fairly evident the lines on the graph are not straight but curved The example model used is Tornado and Sensitivity Charts Nonlinear which applies the Black Scholes option pricing model Such nonlinearities cannot be ascertained from a tornado chart and may be important information in the model or may provide decision makers important insight into the model s dynamics For instance in this Black Scholes model the fact that stock price and strike price are nonlinearly related to the option value is important to know This characteristic implies that option value will not increase or decrease proportionally to the changes in stock or strike price and that there might be some interactions between these two prices as well as other variables As another example an engineering model depicting nonlinearitie
10. 0918 0 0082 Null Hypothesis The data is normally distributed 108 00 0 02 0 42 0 0977 0 0223 114 00 0 02 0 14 0 1038 0 0362 Conclusion The sample data is normally distributed at 127 00 0 02 0 16 0 1180 0 0420 the 1 alpha level 153 00 0 02 0 18 0 1504 0 0296 177 00 0 02 0 20 0 1851 0 0149 186 00 0 02 0 22 0 1994 0 0206 188 00 0 02 0 24 0 2026 0 0374 198 00 0 02 0 26 0 2193 0 0407 222 00 0 02 0 28 0 2625 0 0175 231 00 0 02 0 30 0 2797 0 0203 240 00 0 02 0 32 0 2975 0 0225 246 00 0 02 0 34 0 3096 0 0304 251 00 0 02 0 36 0 3199 0 0401 265 00 0 02 0 38 0 3494 0 0306 280 00 0 02 0 40 0 3820 0 0180 285 00 0 02 0 42 0 3931 0 0269 286 00 0 04 0 46 0 3953 0 0647 291 00 0 02 0 48 0 4065 0 0735 303 00 0 02 0 50 0 4336 0 0664 311 00 0 02 0 52 0 4519 0 0681 Figure 48 Sample Statistical Analysis Tool Report Normality Test Stochastic Process Parameter Estimations Statistical Summary A stochastic process is a sequence of events or paths generated by probabilistic laws That is random events can occur over time but are governed by specific statistical and probabilistic rules The main stochastic processes include Random Walk or Brownian Motion Mean Reversion and Jump Diffusion These processes can be used to forecast a multitude of variables that seemingly follow random trends but yet are restricted by probabilistic laws The process generating equation is known in advance but the actual results generated is unknown The Random Walk Brownian M
11. 1 e to see if the differences between the means and variances of two different forecasts that occur are based on random chance or if they are in fact statistically significantly different from one another This analysis 1s related to bootstrap simulation with several differences Classical hypothesis testing uses mathematical models and is based on theoretical distributions This means that the precision and power of the test 1s higher than bootstrap simulation s empirically based method of simulating a simulation and letting the data tell the story However the classical hypothesis test is applicable only for testing means and variances of two distributions and by extension standard deviations to see if they are statistically identical or different In contrast nonparametric bootstrap simulation can be used to test for any distributional statistics making it more useful the drawback is its lower testing power Risk Simulator provides both techniques from which to choose Procedure e Runa simulation e Select Risk Simulator Tools Hypothesis Testing e Select the two forecasts to test select the type of hypothesis test you wish to run and click OK Figure 36 MODEL A MODEL B E Revenue 200 00 Revenue 200 00 Hypothesis testing is used to determine if two or more forecast Cost 100 00 Cost 100 00 distributions have the same mean and variance i e if they are Income 100 00 Income 100 00 statistically different fr
12. A C is not a precedent as it is only an intermediate calculated value The range and number of values perturbed is user specified and can be set to test extreme values rather than smaller perturbations around the expected values In certain circumstances extreme values may have a larger smaller or unbalanced impact e g nonlinearities may occur where increasing or decreasing economies of scale and scope creep occurs for larger or smaller values of a variable and only a wider range will capture this nonlinear impact A tornado chart lists all the inputs that drive the model starting from the input variable that has the most effect on the results The chart is obtained by perturbing each precedent input at some consistent range e g 10 from the base case one at a time and comparing their results to the base case A spider chart looks like a spider with a central body and its many legs protruding The positively sloped lines indicate a positive relationship while a negatively sloped line indicates a negative relationship Further spider charts can be used to visualize linear and nonlinear relationships The tornado and spider charts help identify the critical success factors of an output cell in order to identify the inputs to simulate The identified critical variables that are uncertain are the ones that should be simulated Do not waste time simulating variables that are neither uncertain nor have little impact on the results Result Ba
13. Bot eone amp Risk New Edit SetInput Set Output Copy Paste Remove Run Step Reset ARIMA Maximum Nonlinear Regression Stochastic Time Series Run Set Set Set Simulator Profile Profile Assumption Forecast Likelihood Extrapolation Analysis Processes Analysis Optimization Objective Decision Constraint Menu Profile Assumptions Forecasts Editing Simulation Run Forecasting Optimization Next icon Icon Rotates SECOND TOOLBAR FORECASTING AND OPTIMIZATION Optimization analysis Run optimization set objective set among decision variables and set constraints different i Forecasting Models and Analytics Rotates icon toolbars in Risk Simulator 3 d c G Microsoft Excel 8 y Home Insert Page Layout Formulas Data Review View Developer Risk Simulator j J 3 T L m n r r t y e t4 sg qd F xXx gd a Bate ux odd GE gi oh s Risk Create Data Data Diagnostic Distribution Distribution Fitting Fitting Edit Hypothesis Nonparametric Sensitivity Scenario Statistical Tornado Example Help User Next Simulator Report Extraction Open Import Tool Analysis Designer Single Multiple Correlations Testing Bootstrap Analysis Analysis Analysis Analysis Models Manual icon Menu Analytical Tools Help Icon THIRD TOOLBAR ANALYTICAL TOOLS ji Rotates EXAMPLES AND HELP FILES x Example models Help among ApaMficaLTools files and User Manual different icon toolbars in Risk Simulator Figure 2B R
14. Error Level Name nple Second Forecast Enabled Yes Cell E 13 Forecast Precision Precision Level mE Error Level Name ample Third Forecast Enabled Yes Cell SESI4 Forecast Precision Precision Level Error Level Correlation Matrix Simulation Example Profile Number of Trials 1000 Stop Simulation on Error No Random Seed 123456 Enable Correlations Yes Name e Second Assumption Enabled Yes Cell SESO Dynamic Simulation No Range Minimum infinity Maximum Infinity Distribution Triangular Minimum 10 Most Likely 0 Maximum 10 0 00 9AA 59 197 Number of Datapoints Mean Median Standard Deviation Variance Average Deviation Maximum Minimum Range Skewness Kurtosis 25 Percentile 7396 Percentile Error Precision at 93 Number of Datapoints Mean Median Standard Deviation Variance Average Deviation Maximum Minimum Range Skewness Kurtosis 2596 Percentile 7595 Percentile Error Precision at 9596 Number of Datapoints Mean Median Standard Deviation Variance Average Deviation Maximum Minimum Range Skewness Kurtosis 2596 Percentile 7595 Percentile Error Precision at 9596 1000 100 0400 99 8427 9 8334 96 6903 7 0397 134 5452 66 9132 67 6320 0 1121 0 1401 93 3563 106 3153 0 0061 1000 0 0806 0 0755 41171 16 9506 3 3389 9 3923 9 7671 19 1594 0 0494 0 5394 2 8924 2 8015 3 1644 1000 0 2861 0 2621 0 1593 0 0254 0 1305 0
15. IE is used ta identify critical success factors of a model before running simulations Product Avg Prices Product B Avg Prices Product C Avg Prices Review the precedents below and make any necessary changes Ug 88 Seduta Sae ata worksheet Base Upside Downside Test Points Product B Sale Qua DCF Model Cs Er Product C Gale Qua Market Risk DCF Model O 15 10 00 10 00 Total Revenues Product 4 Avg DCF Model 10 10 00 10 00 Direct Cost of Goods Product B Avg OCF Model 12 25 10 00 10 00 an Gross Profit Product C Ava DCF Model 15 15 10 0095 10 00 Operating Expenses Product 4 Sale DEF Model S 10 009 10 009 A8 Sales General and Product B Sale DCF Model 3o 10 0092 10 0092 n5 Operating Incor Pront E Sale DCF Model 20 10 009 10 0095 7H Depreciation Depreciation DCF Model 10 10 0095 10 00 m Amortization Amortization DCF Model 3 10 0059 10 00 po EBIT Interest Payments OCF Model 2 10 00 10 00 7H Interest Payments Options DD em 3 Show All Variables ZH Taxes Net Income Show Top DK I Noncash Depreciat Moncash Change in Met Working Capital 0 00 Noncash Capital Expenditures Free Cash Flow Investment Outlay 180000 Financial Analysis Present Value of Free Cash Flow 528 24 440 60 367 26 305 97 254 62 Present Value of Investment Outlay 1 800 00 0 00 0 00 0 00 0 00 Net Cash Flows 1 271 75 bobo p485 7L bb 445 33 Figure 20 Running a tornado analysi
16. Regression Analysis lt Set Objective lt Create Report e FO Maximum Likelihood Stochastic Processes DSet Decision Fitting Single SetInput Set Output Copy i ste o Run Step Reset Run More Next Euer gio iem eund Forecast ZE Nonlinear Extrapolation Time Series Analysis Optimization amp Set Constraint License icon New Simulation Profile s Forecasts Editing Simulation Run Forecasting Optimization Analytical Tools Icon 1 Edit Simulation Profile l Y Change Simulation Profile Create Report Set Input Assumption Set Output Forecast Risk Simulator Icon Toolbar AO Copy Parameter Remove Parameter LC Edit Correlations Risk Simulator Menu Items Step Simulation w B Run Simulation i gt Reset Simulation I Example Models lilk Forecasting Optimization hd Yv Y Tools Options License About Risk Simulator eS User Manual Figure 1B Risk Simulator menu and icon toolbar in Excel 2007 Edit Profile Set Input Copy Paste Delete Multiple Nonlinear Stochastic Hypothesis ee One ic AEsuUmpion i i Regression ea aan Processes Testing oo Help pr P y 4 x M Re AaB M XX PN adgaewagd p Run g Reset time Series Distribution m Tornado Set G lput 1 Fitting D 3 Hew Simulation Simulation Step Simulation Analysis i e Analysis Forecast E O
17. are simple inputs then the testing range will be a simple perturbation based on the range chosen e g the default is 10 Each precedent variable can be perturbed at different percentages if required A wider range is important as it 1s better able to test extreme values rather than smaller perturbations around the expected values In certain circumstances extreme values may have a larger smaller or unbalanced impact e g nonlinearities may occur where increasing or decreasing economies of scale and scope creep in for larger or smaller values of a variable and only a wider range will capture this nonlinear impact Discounted Cash Flow Model Base Year 2005 Sum PV Net Benefits 1 896 63 Market Risk Adjusted Discount Rate 15 0058 Sum PV investments 1 800 00 Frivate Risk Discount Rate 3 0058 Met Present Value 96 63 Annuallzed Sales Growth Hate 2 00 Interna Rate of Retum 18 8058 Price Erosion Rate 5 5 Return on Investment a 2 75 Fffective Tax Rate 40 00 2008 2006 2007 2008 2008 Product A Avg Price Unit Product B Avg PricerUnit Product C Avg Price Unit Product A Sale Quantity 000s Product B Sale Quantity 000s Product C Sale Quantity 000s 20 00 Total Revenues Direct Cost of Goods Sold Gross Profit Operating Expenses Sales General and Admin Costs Operating Income EBITDA Depreciation Amortization EBIT Interest Payments EBT 822 82 Taxes Net Income Moncash Depreciation Amortization 13 00 13 00 13 00 13 00
18. do so Click on Start Control Panel Classic View on the left panel User Accounts Turn User Account Control On or Off and uncheck the option Use User Account Control UAC and restart the computer When restarting the computer you will get a message that UAC 1s turned off You can turn this message off by going to the Control Panel Security Center Change the Way Security Center Alerts Me Don t Notify Me and Don t Display the Icon The sections that follow provide step by step instructions for using the software As the software is continually updated and improved the examples in this book might be slightly different from the latest version downloaded from the Internet Ej Zl File Edit wiew Insert Format Tools Data Window Help Simulation Adobe PDF id M Ga x n Mew Simulation Profile L Edit Simulation Profile Change Simulation Profile Set Input 4ssumption Set Oukput Forecast Copy Parameter Paste Parameter Remove Parameter Edit Correlations Run Simulation Step Simulation Reset Simulation Forecasting H Optimization Tools 1 Options License About Risk Simulator Help Figure 1A Risk Simulator menu and icon toolbar in Excel XP and Excel 2003 ir M Microsoft Excel 000000 iiia Home Insert Page Layout ulas Data Review View Dev Risk Simulator lt _ Risk Simulator Tak e 7x el B a a ele E a gt
19. drug was applied and so forth Data Extraction Saving Simulation Results and Generating Reports Raw data of a simulation can be extracted very easily using Risk Simulator s Data Extraction routine Both assumptions and forecasts can be extracted but a simulation must be run first The extracted data can then be used for a variety of other analyses and the data can be extracted to different formats for use in spreadsheets databases and other software products Procedure e Open or create a model define assumptions and forecasts and run the simulation e Select Risk Simulator Tools Data Extraction e Select the assumptions and or forecasts you wish to extract the data from and click OK The simulated data can be extracted to an Excel worksheet a flat text file for easy import into other software applications or as risksim files which can be reopened as Risk Simulator forecast charts at a later date Finally you can create a simulation report of all the assumptions and forecasts in the model by going to Risk Simulator Create Report A sample report is shown in Figure 37 General Assumptions Name aple First Assumption Enabled Yes Cell E 8 Dynamic Simulation No Range Minimum Infinity Maximum Infinity Distribution Normal Mean 100 Standard Deviation 10 0 00 R21 77RA 925A 107 44 122 31 137 1 Forecasts Name Sample First Forecast Enabled Yes Cell SESI2 Forecast Precision Precision Level
20. examples However if correlations are added between the assumptions Figure 27 shows a very different picture Notice for instance price erosion had little impact on NPV but when some of the input assumptions are correlated the interaction that exists between these correlated variables makes price erosion have more impact Note that tornado analysis cannot capture these correlated dynamic relationships Only after a simulation is run will such relationships become evident in a sensitivity analysis A tornado chart s pre simulation critical success factors therefore sometimes will be different from a sensitivity chart s post simulation critical success factor The post simulation critical success factors should be the ones that are of interest as these more readily capture the interactions of the model precedents Nonlinear Rank Correlation Net 0 35 C Quantity EEE 75 A Price EE 921 B Price EEE 0 22 C Price CONI 0 17 Tax Rate 0 05 Price Erosion 0 05 Sales Growth 0 0 0 1 2 0 3 0 4 O 5 0 6 Figure 26 Sensitivity chart without correlations P 0 26 B Price 1 22 C Price ee Price Erosion per 18 Tax Rate 0 03 Sales Growth 0 0 0 1 0 2 0 3 0 4 0 5 0 5 Figure 27 Sensitivity chart with correlations Procedure Use these steps to create a sensitivity analysis e Open or create a model define assumptions and forecasts and run the simulation the example here uses the Tornado and Sensitiv
21. implies that the series follows a low order moving average process Partial correlation PAC k measures the correlation of values that are k periods apart after removing the correlation from the intervening lags If the pattern of autocorrelation can be captured by an autoregression of order less than k then the partial autocorrelation at lag k will be close to zero Ljung Box Q statistics and their p values at lag k has the null hypothesis that there 1s no autocorrelation up to order k The dotted lines in the plots of the autocorrelations are the approximate two standard error bounds If the autocorrelation is within these bounds it is not significantly different from zero at the 5 significance level Autocorrelation measures the relationship to the past of the dependent Y variable to itself Distributive Lags in contrast are time lag relationships between the dependent Y variable and different independent X variables For instance the movement and direction of mortgage rates tend to follow the Federal Funds Rate but at a time lag typically 1 to 3 months Sometimes time lags follow cycles and seasonality e g ice cream sales tend to peak during the summer months and are hence related to last summer s sales 12 months in the past The distributive lag analysis Figure 40 shows how the dependent variable is related to each of the independent variables at various time lags when all lags are considered simultaneously to determine which time
22. say that the values are statistically different In addition if a model s resulting skewness is 0 19 is this forecast distribution negatively skewed or is it statistically close enough to zero to state that this distribution is symmetrical and not skewed Thus if we bootstrapped this forecast 100 times i e run a 1 000 trial simulation for 100 times and collect the 100 skewness coefficients the skewness distribution would indicate how far zero is away from 0 19 If the 90 confidence on the bootstrapped skewness distribution contains the value zero then we can state that on a 90 confidence level this distribution 1s symmetrical and not skewed and the value 0 19 is statistically close enough to zero Otherwise if zero falls outside of this 90 confidence area then this distribution is negatively skewed The same analysis can be applied to excess kurtosis and other statistics Essentially bootstrap simulation is a hypothesis testing tool Classical methods used in the past relied on mathematical formulas to describe the accuracy of sample statistics These methods assume that the distribution of a sample statistic approaches a normal distribution making the calculation of the statistic s standard error or confidence interval relatively easy However when a statistic s sampling distribution is not normally distributed or easily found these classical methods are difficult to use In contrast bootstrapping analyzes sample st
23. success drivers of a project the results of the model depend on these critical success drivers These variables are the ones that should be simulated Do not waste time simulating variables that are neither uncertain nor have little impact on the results Tornado charts assist in identifying these critical success drivers quickly and easily Following this example it might be that price and quantity should be simulated assuming if the required investment and effective tax rate are both known in advance and unchanging Ps Base Value 86 6261638553219 Input Changes Precedent Cell Investment B Quantity C Price C Quantity Discount Rate Price Erosion sales Growth Lepreciation Interest Amortization Cage Met Capital Output owns ide 275 63 279 73 343 15 71 23 18 3053 40 15 4805 138 24 115 80 90 59 95 08 3709 95 16 96 63 96 63 tut Uinside 03 37 26 47 158 83 176 55 170 07 152 72 153 11 145 20 57 03 76 54 102 68 98 17 96 16 897 09 96 63 96 63 inout Downside 3e0 00 1 520 00 87 980 00 87 600 00 36 00 3 00 11 03 45 00 31 50 13 54 18 00 13 50 4 50 1 8008 900 1 50 2 70 0 00 000 inout Unside 44 00 131 00 13 48 55 00 38 50 15 57 22 00 16 5086 5 50 2 2006 11 00 220 330 0 00 0 00 Base Case ate Figure 22 Sensitivity table Spider Chart Investment EL E 2 Depreciation 50
24. tax rates tend not to fluctuate too much and also because tax rate is a straight percentage value of the income before taxes where other precedent variables have a larger effect on NPV This example proves that it is important to perform sensitivity analysis after a simulation run to ascertain if there are any interactions in the model and if the effects of certain variables still hold The second chart Figure 30 illustrates the percent variation explained that is of the fluctuations in the forecast how much of the variation can be explained by each of the assumptions after accounting for all the interactions among variables Notice that the sum of all variations explained is usually close to 100 sometimes other elements impact the model but they cannot be captured here directly and if correlations exist the sum may sometimes exceed 100 due to the interaction effects that are cumulative 0 22 C Price ee Tax Rate E Price Erosion 0 03 Sales Growth 0 0 0 1 02 O35 O 5 0 6 A Quantity 9 52 B Price 464 C Price 3 02 Tax Rate 0 26 Price Erosion 0 11 sales Growth 0 0 0 1 0 1 0 2 0 2 0 3 0 3 04 Figure 30 Contribution to variance chart Notes Tornado analysis is performed before a simulation run while sensitivity analysis is performed after a simulation run Spider charts in tornado analysis can consider nonlinearities while rank correlation charts in sensitivity analysis can account for nonl
25. trials Enter the number of simulation trials required Running 1 000 trials means that 1 000 different iterations of outcomes based on the input assumptions will be generated You can change this number as desired but the input has to be positive integers The default number of runs 1s 1 000 trials Pause on simulation error If checked the simulation stops every time an error is encountered in the Excel model that is if your model encounters a computational error e g some input values in a simulation trial may yield a divide by zero error in a spreadsheet cell the simulation stops This feature is important to help audit your model to make sure there are no computational errors in your Excel model However if you are sure the model works there is no need for you to check this preference Turn on correlations If checked correlations between paired input assumptions will be computed Otherwise correlations will all be set to zero and a simulation is run assuming no cross correlations between input assumptions Applying correlations will yield more accurate results if correlations do indeed exist and will tend to yield a lower forecast confidence if negative correlations exist Specify random number sequence By definition a simulation yields slightly different results every time it is run by virtue of the random number generation routine in Monte Carlo simulation This is a theoretical fact in all random number generators However
26. variable on the outcome of the model that 1s the tool automatically perturbs each variable in the model a preset amount captures the fluctuation on the model s forecast or final result and lists the resulting perturbations ranked from the most significant to the least Figures 19 through 24 illustrate the application of a tornado analysis For instance Figure 19 is a sample discounted cash flow model where the input assumptions in the model are shown The question is What are the critical success drivers that affect the model s output the most That is what really drives the net present value of 96 63 or which input variable impacts this value the most The tornado chart tool can be obtained through Simulation Tools Tornado Analysis To follow along the first example open the Tornado and Sensitivity Charts Linear file in the examples folder Figure 19 shows this sample model where cell G6 containing the net present value is chosen as the target result to be analyzed The target cell s precedents in the model are used in creating the tornado chart Precedents are all the input and intermediate variables that affect the outcome of the model For instance if the model consists of A B C and where C D E then B D and E are the precedents for A C is not a precedent as it 1s only an intermediate calculated value Figure 20 shows the testing range of each precedent variable used to estimate the target result If the precedent variables
27. zy Click here to link the m name of the forecast to Forecast Precision a cell Precision Level 75 Confidence Optional Specify the E precision confidence Error Level of Mean and error levels D value ofthe Mean Options Specify if you want this W Show Forecast Window forecast to be visible turned on by default Cancel Figure 7 Set output forecast 4 Run Simulation If everything looks right click on Risk Simulator Run Simulation or click on the Run icon on the Risk Simulator toolbar and the simulation will proceed You may also reset a simulation after it has run to rerun it Risk Simulator Reset Simulation or the Reset icon on the toolbar or to pause it during a run Also the step function Risk Simulator Step Simulation or the Step icon on the toolbar allows you to simulate a single trial one at a time which is useful for educating others on simulation i e you can show that at each trial all values in the assumption cells are replaced and the entire model is recalculated each time 5 Interpreting the Forecast Results The final step in Monte Carlo simulation is to interpret the resulting forecast charts Figures 8 to 15 show the forecast chart and the statistics generated after running the simulation Typically these sections on the forecast window are important in interpreting the results of a simulation Forecast Chart The forecast chart shown in Figure 8 is a probability histogram that shows the fr
28. 0 1886 for the positive correlation model and 0 0717 for the negative correlation model That is for simple models with positive relationships e g additions and multiplications negative correlations tend to reduce the average spread of the distribution and create a tighter and more concentrated forecast distribution as compared to positive correlations with larger average spreads However the mean remains relatively stable This implies that correlations do little to change the expected value of projects but can reduce or increase a project s risk Recall in financial theory that negatively correlated variables projects or assets when combined in a portfolio tend to create a diversification effect where the overall risk 1s reduced Therefore we see a smaller standard deviation for the negatively correlated model In a positively related model e g A B C or A x B C a negative correlation reduces the risk standard deviation and all other second moments of the distribution of the result C whereas a positive correlation between the inputs A and B will increase the overall risk The opposite is true for a negatively related model e g A B C or A B C where a positive correlation between the inputs will reduce the risk and a negative correlation increases the risk In more complex models as 1s often the case in real life situations the effects will be unknown a priori and can be determined only after a simulation 1s run
29. 103 65 103 72 103 81 103 30 103 99 104 46 104 57 104 76 106 20 105 44 105 50 105 52 105 68 105 66 105 57 105 30 105 30 106 29 106 35 106 69 107 01 107 68 107 70 107 33 108 17 108 20 108 34 103 42 108 43 108 43 108 70 105 15 103 22 103 35 103 52 103 75 110 04 110 16 110 25 110 54 111 05 111 06 111 44 111 76 111 50 111 55 12 07 112 13 12 29 12 32 12 42 12 48 112 35 112 92 113 50 113 53 113 63 113 70 114 13 114 14 114 21 114 31 114 35 115 40 115 58 115 66 116 55 116 98 117 60 118 67 113 24 113 52 124 14 124 16 124 39 132 30 Figure 33 Distributional fitting report Bootstrap Simulation Bootstrap simulation is a simple technique that estimates the reliability or accuracy of forecast statistics or other sample raw data Bootstrap simulation can be used to answer a lot of confidence and precision based questions in simulation For instance suppose an identical model with identical assumptions and forecasts but without any random seeds is run by 100 different people the results will clearly be slightly different The question is if we collected all the statistics from these 100 people how will the mean be distributed or the median or the skewness or excess kurtosis Suppose one person has a mean value of say 1 50 while another has 1 52 Are these two values statistically significantly different from one another or are they statistically similar and the slight difference is due entirely to random chance What about 1 53 So how far is far enough to
30. 13 00 Moncash Change in Met Working Capital 0 00 0 00 0 00 0 00 0 00 Mancash Capital Expenditures 0 00 0 00 0 00 0 00 0 00 Free Cash Flow 528 24 506 69 485 70 465 25 445 33 Investment Outlay Financial Analysis Present Value of Free Cash Flow 528 24 440 60 367 26 305 91 254 62 Present Value of Investment Outlay 7 800 00 0 00 0 00 0 00 0 00 Net Cash Flows 1 271 75 pa0b bu p455 70 465 25 445 33 Figure 19 Sample discounted cash flow model Procedure Use these steps to create a tornado analysis 2 a 4 5 Select the single output cell 1 e a cell with a function or equation in an Excel model e g cell G6 is selected in our example Select Risk Simulator Tools Tornado Analysis Review the precedents and rename them as appropriate renaming the precedents to shorter names allows a more visually pleasing tornado and spider chart and click OK Alternatively click on Use Cell Address to apply cell locations as the variable names A B 0 E F Discounted Cash Flow Model Dase Year z005 Sum PY Met Benefits 7 896 63 Market Risk Adjusted Discount Hate 15 0058 Sum PY Investments 57 800 00 Prvate Aisk Discount hate 5 c8 Met Present Value 96 63 Annualized Sales Ta Price Erosion Rate E Tornado Analysis on Effective Tax Rate Tornado analysis creates static perturbations I amp each precedent is perturbed one at a time to identify the impact to the results
31. 4 0 8355 0 0545 0 6828 0 7354 0 5093 0 3500 x4 0 0061 0 6739 0 7932 0 7719 0 6748 0 8627 0 5586 0 9046 0 5726 0 6304 0 4812 0 5707 X5 0 1591 0 2032 0 4123 0 5599 0 6416 0 3447 0 9190 0 9740 0 5185 0 2855 0 1489 0 7794 Figure 40 Autocorrelation and Distributive Lag Results Another requirement in running a regression model is the assumption of normality and sphericity of the error term If the assumption of normality is violated or outliers are present then the linear regression goodness of fit test may not be the most powerful or informative test available and this could mean the difference between detecting a linear fit or not If the errors are not independent and not normally distributed it may indicate that the data might be autocorrelated or suffer from nonlinearities or other more destructive errors Independence of the errors can also be detected in the heteroskedasticity tests Figure 41 The Normality test on the errors performed is a nonparametric test which makes no assumptions about the specific shape of the population from which the sample is drawn allowing for smaller sample data sets to be analyzed This test evaluates the null hypothesis of whether the sample errors were drawn from a normally distributed population versus an alternate hypothesis that the data sample is not normally distributed If the calculated D Statistic is greater than or equal to the D Critical values at various significance values then reject the null hy
32. 60 D E 40 g zx li Alii Caa Cora BB2 B4 152 54 337 38 837 38 Type A Infinity il Infinity Certainty 2 100 00 as p Financial Analysis Present Value of Free Cash Flow 528 24 5440 60 367 26 305 01 254 62 Present Value of Investment Outlay 1 800 00 0 00 0 00 0 00 0 00 Net Cash Flows 1 271 76 506 69 485 70 455 25 445 33 Figure 28 Running sensitivity analysis Results Interpretation The results of the sensitivity analysis comprise a report and two key charts The first is a nonlinear rank correlation chart Figure 29 that ranks from highest to lowest the assumption forecast correlation pairs These correlations are nonlinear and nonparametric making them free of any distributional requirements 1 e an assumption with a Weibull distribution can be compared to another with a beta distribution The results from this chart are fairly similar to that of the tornado analysis seen previously of course without the capital investment value which we decided was a known value and hence was not simulated with one special exception Tax rate was relegated to a much lower position in the sensitivity analysis chart Figure 29 as compared to the tornado chart Figure 24 This is because by itself tax rate will have a significant impact Once the other variables are interacting in the model however it appears that tax rate has less of a dominant effect This 1s because tax rate has a smaller distribution as historical
33. 728 14630 4008 30927 22322 3711 3136 185 600 3 2 142 oe 290 346 328 354 256 320 197 CO Data is from a single variable Data comprises multiple variables in columns Figure 44 Running the Statistical Analysis Tool ES Statistical A nalyses Select the analyses to run iw Descriptive Statistics Distributional Fitting t Continuous Histagram and Charts Hypothesis Testing Hypothesized Mean t Discrete Pp 0 Nonlinear Extrapolation Forecast Periods Normality Test i Stochastic Process Parameter Estimation Periadicity Annual Iw Time series Autocorrelation iw Time series Forecasting Seasonality Periads Cycle Forecast Periods 4 iw Trend Line Projection Forecast Periods Cancel Figure 45 Statistical Tests Descriptive Statistics Analysis of Statistics Almost all distributions can be described within 4 moments some distributions require one moment while others require two moments and so forth Descriptive statistics quantitatively capture these moments The first moment describes the location of a distribution i e mean median and mode and is interpreted as the expected value expected returns or the average value of occurrences The Arithmetic Mean calculates the average of all occurrences by summing up all ofthe data points and dividing them by the number of points The Geometric Mean is calculated by taking the power root of the produ
34. 8358 0 0126 0 8232 0 5797 0 2064 0 4590 0 3935 0 0345 Name Enabled Cell Dynamic Simulation Range Minimum Maximum Distribution Aipha Beta 0 0 nnn n48 89 62 109 62 iple Third Assumption Yes SESIO No infinity Infinity Beta 2 5 EIL DE 128 62 Aie 12 a4nemiun c E x q X c c b JIABAO a4nemiun Sample First Assumption ssumption ssumption Sample First Assumption 7 00 Sample Second Assumption 0 00 7 00 Sample Third Assumption 0 00 0 00 7 00 Figure 37 Sample Simulation Report Regression and Forecasting Diagnostic Tool This advanced analytical tool in Risk Simulator is used to determine the econometric properties of your data The diagnostics include checking the data for heteroskedasticity nonlinearity outliers specification errors micronumerosity stationarity and stochastic properties normality and sphericity of the errors and multicollinearity Each test is described in more detail in their respective reports in the model Procedure amp Open the example model Risk Simulator Examples Regression Diagnostics and go to the Time Series Data worksheet and select the data including the variable names cells C5 H55 amp Click on Risk Simulator Tools Diagnostic Tool amp Check the data and select the Dependent Variable Y from the drop down menu Click OK when finished Figure 38 Multiple Regression Analysis Data Set ES dee Si Variabl
35. Excel as an add in This software is compatible and often used with the Real Options SLS software used in Part II of this book also developed by the author Standalone software applications in C are also available for implementation into other existing proprietary software or databases The different functions or modules in both software applications are briefly described next The Appendix provides a more detailed list of all the functions tools and models The Simulation Module allows you to o Run simulations in your existing Excel based models o Generate and extract simulation forecasts distributions of results o Perform distributional fitting automatically finding the best fitting statistical distribution o Compute correlations maintain relationships among simulated random variables o lHidentify sensitivities creating tornado and sensitivity charts o Test statistical hypotheses finding statistical differences between pairs of forecasts o Run bootstrap simulation testing the robustness of result statistics o Run custom and nonparametric simulations simulations using historical data without specifying any distributions or their parameters for forecasting without data or applying expert opinion forecasts The Forecasting Module can be used to generate o Automatic time series forecasts with and without seasonality and trend o Automatic ARIMA automatically generate the best fitting ARIMA forecasts o Basic Econo
36. Risk Simulation The Basics of Quantitative Risk Analysis and Simulation Short Examples Series using Risk Simulator I heal Options Valuation For m infor rma atio Hos visit www r cao T onsvaluation com onta ct S u admin i eines valuation com Introduction to Risk Simulator This section also provides the novice risk analyst an introduction to the Risk Simulator software for performing Monte Carlo simulation where a 30 day trial version of the software is included in the book s DVD This section starts off by illustrating what Risk Simulator does and what steps are taken in a Monte Carlo simulation as well as some of the more basic elements in a simulation analysis It then continues with how to interpret the results from a simulation and ends with a discussion of correlating variables in a simulation as well as applying precision and error control Software versions with new enhancements are released continually Please review the software s user manual and the software download site www realoptionsvaluation com for more up to date details on using the latest version of the software See Modeling Risk Applying Monte Carlo Simulation Real Options Analysis Stochastic Forecasting and Portfolio Optimization Wiley 2007 also by the author for more technical details on using Risk Simulator Risk Simulator is a Monte Carlo simulation forecasting and optimization software It 1s written in Microsoft NET C and functions with
37. ased and the fitted slope and intercept estimates will not be meaningful Over a restricted range of independent or dependent variables nonlinear models may be well approximated by linear models this is in fact the basis of linear interpolation but for accurate prediction a model appropriate to the data should be selected A nonlinear transformation should first be applied to the data before running a regression One simple approach is to take the natural logarithm of the independent variable other approaches include taking the square root or raising the independent variable to the second or third power and run a regression or forecast using the nonlinearly transformed data Diagnostic Results Heteroskedasticity Micronumerosity Outliers Nonlinearity W Test Hypothesis Test Approximation Matural Matural Number of Monlinear Test Hypothesis Test Variable p value result result Lower Bound Upper Bound Potential Outliers p value result no problems Bb 671 70 2 variable x1 12543 Homoskedastic no problems 21377 95 b47 13 03 3 2458 linear Variable X2 0 3371 Homoskedastic na problems ria 145 83 2 0 0335 nonlinear Variable X3 0 3649 Homaoskedastic no problems i 15 69 3 0 0305 nonlinear variable x4 0 3066 Homoskedastic no problems 295 96 625 21 4 0 9298 linear Variable X5 0 2495 Homoskedastic no problems 3 35 Hg 3 D 2727 linear Figure 39 Results from Tests of Outliers Heteroskedasticity Micronumerosity and Nonlinearity Another typica
38. atistics empirically by sampling the data repeatedly and creating distributions of the different statistics from each sampling The classical methods of hypothesis testing are available in Risk Simulator and are explained in the next section Classical methods provide higher power in their tests but rely on normality assumptions and can be used only to test the mean and variance of a distribution as compared to bootstrap simulation which provides lower power but is nonparametric and distribution free and can be used to test any distributional statistic Procedure e Runa simulation with assumptions and forecasts e Select Risk Simulator Tools Nonparametric Bootstrap e Select only one forecast to bootstrap select the statistic s to bootstrap and enter the number of bootstrap trials and click OK Figure 34 fr Nonparametric Bootstrap X Nonparametric bootstrap simulation is a distribution free technique used to estimate the reliability or accuracy of MODEL A MODEL B forecast statistics I e to compute the forecast a intervals of each of the statistics Revenue 200 00 Revenue 200 00 Please select a forecast to run the non parametric bootstrap Cost 1 00 00 Cost 1 O00 00 Forecast Name Worksheet Income 100 00 Income 100 00 Income Model A Simulation Model C Income Model B Simulation Model Income Model A Risk Simulator Forecast Distribution aa Histogram Statistics Preferences Options Income Model A
39. be considered as severe multicollinearity A VIF exceeding 10 0 indicates destructive multicollinearity Figure 43 Correlation Matrix CORRELATION na x3 Ad AS A1 0 333 0 959 0 247 0 237 M2 TOO 0 348 0 319 0 120 M3 COAN 0 185 0 227 Md 1000 0 790 Variance Inflation Factor VIF X2 A3 x4 x5 X1 1312 12 46 1 06 1 065 X2 MEN 114 131 1 01 X3 MEA 1 04 1 05 x4 M 1 08 Figure 43 Multicollinearity Errors The Correlation Matrix lists the Pearson s Product Moment Correlations commonly referred to as the Pearson s R between variable pairs The correlation coefficient ranges between 1 0 and 1 0 inclusive The sign indicates the direction of association between the variables while the coefficient indicates the magnitude or strength of association The Pearson s R only measures a linear relationship and is less effective in measuring non linear relationships To test whether the correlations are significant a two tailed hypothesis test is performed and the resulting p values are listed above P values less than 0 10 0 05 and 0 01 are highlighted in blue to indicate statistical significance In other words a p value for a correlation pair that is less than a given significance value is statistically significantly different from zero indicating that there is significant a linear relationship between the two variables The Pearson s Product Moment Correlation Coefficient R between two variables x and y 1s related to COV
40. cance alpha level of the test and the number of data points Power decreases as the residual variance increases decreases as the significance level is decreased 1 e as the test is made more stringent increases as the variation in observed independent variable increases and increases as the number of data points increases Values may not be identically distributed because of the presence of outliers Outliers are anomalous values in the data Outliers may have a strong influence over the fitted slope and intercept giving a poor fit to the bulk of the data points Outliers tend to increase the estimate of residual variance lowering the chance of rejecting the null hypothesis 1 e creating higher prediction errors They may be due to recording errors which may be correctable or they may be due to the dependent variable values not all being sampled from the same population Apparent outliers may also be due to the dependent variable values being from the same but non normal population However a point may be an unusual value in either an independent or dependent variable without necessarily being an outlier in the scatter plot In regression analysis the fitted line can be highly sensitive to outliers In other words least squares regression is not resistant to outliers thus neither is the fitted slope estimate A point vertically removed from the other points can cause the fitted line to pass close to it instead of following the general
41. cts of all the data points and requires them to all be positive The Geometric Mean is more accurate for percentages or rates that fluctuate significantly For example you can use Geometric Mean to calculate average growth rate given compound interest with variable rates The Trimmed Mean calculates the arithmetic average ofthe data set after the extreme outliers have been trimmed As averages are prone to significant bias when outliers exist the Trimmed Mean reduces such bias in skewed distributions The Standard Error of the Mean calculates the error surrounding the sample mean The larger the sample size the smaller the error such that for an infinitely large sample size the error approaches zero indicating that the population parameter has been estimated Due to sampling errors the 9596 Confidence Interval for the Mean is provided Based on an analysis ofthe sample data points the actual population mean should fall between these Lower and Upper Intervals for the Mean Median is the data point where 5096 of all data points fall above this value and 5096 below this value Among the three first moment statistics the median is least susceptible to outliers A symmetrical distribution has the Median equal to the Arithmetic Mean A skewed distribution exists when the Median is far away from the Mean The Mode measures the most frequently occurring data point Minimum is the smallest value in the data set while Maximum is the largest value Range is the diff
42. e X1 Variable X2 Variable X3 Variable X4 Variable X5 Variable Y 52 18308 780 B 3b 1145 1 445 15066 323 365 fred B14 100484 x Wee This tool i used to diagnose forecasting problems in a set of multiple variables 397 4008 Dependent Variable Dependent Variable r l TEA 305947 427 773772 variable X2 Variable X3 VariaE 153 3711 16308 185 4 041 79 6 441 7136 1145 600 0 55 1 E 524 s0508 18066 E 3 665 32 3 326 20006 fen 142 2 351 45 1 FEIN 16995 100464 452 29 76 190 8 206 13035 16728 290 3 294 31 8 205 12973 14630 346 3 287 578 4 obo 16309 005 320 0 666 340 8 Gp 5227 30927 354 12 938 257 6 498 18235 az 22322 266 6 478 111 9 451 414857 lif i gt 465 14213 177 23619 188 3106 i i d h 450 24917 188 Ex d 74 3 5 6 106 38 2 195 0 799 Es 6 9 246 Bd45 183 1 575 20 5 d EE pams 417 1 202 10 5 SS bo 7128 23 1 109 ESNA TE Figure 38 Running the Data Diagnostic Tool A common violation in forecasting and regression analysis is heteroskedasticity that is the variance of the errors increases over time see Figure 39 for test results using the diagnostic tool Visually the width of the vertical data fluctuations increases or fans out over time and typically the coefficient of determination R squared coefficient drops significantly when heteroskedasticity exists If the variance of the dependent variable is not constant then the error s variance will not be constant Un
43. e a distributional fitting example The next discussion uses the Data Fitting file in the examples folder Procedure Use these steps to perform a distributional fitting model e Open a spreadsheet with existing data for fitting e g use the Data Fitting example file e Select the data you wish to fit not including the variable name Data should be in a single column with multiple rows e Select Risk Simulator Tools Distributional Fitting Single Variable e Select the specific distributions you wish to fit to or keep the default where all distributions are selected and click OK Figure 31 e Review the results of the fit choose the relevant distribution you want and click OK Figure 32 Student s T Triangular Uniform 4 56 783 36 S3 30 48 1 204 Ff 55 08 oO 24 745 67 S2 0d o 36 2198 85 45 81 I Single Fit Distribution fitting takes existing raw data and statistically finds the best fithing distribution L amp by optimizing the parameters of each distribution and performing statistical hypotheses tests Distribution Type 5 Fit to Continuous Distributions C5 Fit to Discrete Distributions Select Distributions bo Fit Cauchy Distribution ChiSquare Distribution Exponential Distribution F Distribution Gamma Distribution Gumbel M asinum Distribution vt Figure 31 Single variable distributional fitting Results Interpretation The null hypothesis H being tested is such that the
44. e this t test Figure 47 Sample Statistical Analysis Tool Report Hypothesis Testing of One Variable Test for Normality The Normality test is a form of nonparametric test which makes no assumptions aboutthe specific shape ofthe population from which the sample is drawn allowing for smaller sample data sets to be analyzed This test evaluates the null hypothesis of whether the data sample was drawn from a normally distributed population versus an alternate hypothesis that the data sample is not normally distributed If the calculated p value is less than or equal to the alpha significance value then reject the null hypothesis and acceptthe alternate hypothesis Otherwise if the p value is higher than the alpha significance value do not reject the null hypothesis This test relies on two cumulative frequencies one derived from the sample data set the second from a theoretical distribution based on the mean and standard deviation of the sample data An alternative to this test is the Chi Square test for normality The Chi Square test requires more data points to run compared to the Normality test used here Test Result Data Relative Observed Expected O E Data Average 331 92 Frequency Standard Deviation 172 91 47 00 0 02 0 02 0 0497 0 0297 D Statistic 0 0859 68 00 0 02 0 04 0 0635 0 0235 D Critical at 1 0 1150 87 00 0 02 0 06 0 0783 0 0183 D Critical at 5 0 1237 96 00 0 02 0 08 0 0862 0 0062 D Critical at 10 0 1473 102 00 0 02 0 10 0
45. elation coefficient Risk Simulator then applies its own algorithms to convert them into Spearman s rank correlation thereby simplifying the process Applying Correlations in Risk Simulator Correlations can be applied in Risk Simulator in several ways e When defining assumptions simply enter the correlations into the correlation grid in the Distribution Gallery e With existing data run the Multi Variable Distribution Fitting tool to perform distributional fitting and to obtain the correlation matrix between pairwise variables If a simulation profile exists the assumptions fitted automatically will contain the relevant correlation values e With the use of a direct input correlation matrix click on Risk Simulator Edit Correlations to view and edit the correlation matrix used in the simulation Note that the correlation matrix must be positive definite that is the correlation must be mathematically valid For instance suppose you are trying to correlate three variables grades of graduate students in a particular year the number of beers they consume a week and the number of hours they study a week You would assume that these correlation relationships exist Grades and Beer The more they drink the lower the grades no show on exams Grades and Study The more they study the higher the grades Beer and Study The more they drink the less they study drunk and partying all the time However if you input a negative c
46. equency counts of values occurring and the total number of trials simulated The vertical bars show the frequency of a particular x value occurring out of the total number of trials while the cumulative frequency smooth line shows the total probabilities of all values at and below x occurring in the forecast Forecast Statistics The forecast statistics shown in Figure 9 summarizes the distribution of the forecast values in terms of the four moments of a distribution You can rotate between the histogram and statistics tab by depressing the space bar Income Risk Simulator Forecast LE e J ET P TE Figure 8 Forecast chart E Income Risk Simulator Forecast Number of Trials Mean Median Standard Deviation Variance Coefficient of Variation Minimum Range Skewness Kurtosis 253 Percentile 754 Percentile Percentage Eror Precision at 355 Confidence Figure 9 Forecast statistics Preferences The preferences tab in the forecast chart Figure 10 allows you to change the look and feel of the charts For instance if Always Show Window On Top is selected the forecast charts will always be visible regardless of what other software is running on your computer The Semitransparent When Inactive is a powerful option used to compare or overlay multiple forecast charts at once e g enable this option on several forecast charts and drag them on top of one another to visually see the similarities or d
47. erence between the Maximum and Minimum values The second moment measures a distribution s spread or width and is frequently described using measures such as Standard Deviations Variances Quartiles and Inter Quartile Ranges Standard Deviation indicates the average deviation of all data points from their mean Itis a popular measure as is associated with risk higher standard deviations mean a wider distribution higher risk or wider dispersion of data points around the mean and its units are identical to original data sets The Sample Standard Deviation differs from the Population Standard Deviation in that the former uses a degree of freedom correction to account for small sample sizes Also Lower and Upper Confidence Intervals are provided for the Standard Deviation and the true population standard deviation falls within this interval If your data set covers every element of the population use the Population Standard Deviation instead The two Variance measures are simply the squared values ofthe standard deviations The Coefficient of Variability is the standard deviation of the sample divided by the sample mean proving a unit free measure of dispersion that can be compared across different distributions you can now compare distributions of values denominated in millions of dollars with one in billions of dollars or meters and kilograms etc The First Quartile measures the 25th percentile of the data points when arranged from its smallest to la
48. fitted distribution is the same distribution as the population from which the sample data to be fitted comes Thus if the computed p value is lower than a critical alpha level typically 0 10 or 0 05 then the distribution is the wrong distribution Conversely the higher the p value the better the distribution fits the data Roughly you can think of p value as a percentage explained that 1s if the p value is 1 00 Figure 32 then setting a normal distribution with a mean of 100 67 and a standard deviation of 10 40 explains close to 10096 of the variation in the data indicating an especially good fit The data was from a 1 000 trial simulation in Risk Simulator based on a normal distribution with a mean of 100 and a standard deviation of 10 Because only 1 000 trials were simulated the resulting distribution is fairly close to the specified distributional parameters and in this case about a 10046 precision Both the results Figure 32 and the report Figure 33 show the test statistic p value theoretical statistics based on the selected distribution empirical statistics based on the raw data the original data to maintain a record of the data used and the assumption complete with the relevant distributional parameters i e if you selected the option to automatically generate assumption and if a simulation profile already exists The results also rank all the selected distributions and how well they fit the data re 7E a7 eee
49. h this model profiling process Start a new simulation profile by performing these steps e Start Excel and create a new or open an existing model You can use the Basic Simulation Model example to follow along Risk Simulator Examples Basic Simulation Model e Click on Risk Simulator New Simulation Profile e Enter a title for your simulation including all other pertinent information Figure 3 M E Simulation Properties Entched5dived number of simulation trials default is 1 000 Title First Example Simulation Configurations Enter a relevant title for this simulation Humber of trials 1000 z Select if you want the simulation to stop Select if you want Pause simulation on error cnoncmori A i encountered default is considered urthe Turn on correlations unchecked simulation default is Specify random number sequence checked Select and enter a seed 999 value if you want the simulation to follow a specified random number sequence default is unchecked Figure 3 New simulation profile The elements in the new simulation profile dialog shown in Figure 3 include Title Specifying a simulation profile name or title allows you to create multiple simulation profiles in a single Excel model By so doing you can save different simulation scenario profiles within the same model without having to delete existing assumptions and change them each time a new simulation scenario 1s required Number of
50. ifferences Histogram Resolution allows you to change the number of bins of the histogram anywhere from 5 bins to 100 bins Also the Update Data Interval section allows you to control how fast the simulation runs versus how often the forecast chart is updated That is if you wish to see the forecast chart updated at almost every trial this will slow down the simulation as more memory is being allocated to updating the chart versus running the simulation This is merely a user preference and in no way changes the results of the simulation just the speed of completing the simulation Display 7 Always Show Window On Top 7 Semitransparent When Inactive Histogram Resolution Faster U Higher Simulation I I I I I I I I I I I I I I I I I I I Resolution Data Update Interval Faster ier Faster Update I I I I I I I I I I I I I I Simulation Figure 10 Forecast chart preferences Options This forecast chart option allows you to show all the forecast data or to filter in or out values that fall within some specified interval or within some standard deviation that you choose Also you can set the precision level here for this specific forecast to show the error levels in the Statistics view See the section on precision and error control for more details Data Filter amp Show all data Show only data between limit Hn nty and infinity
51. inear and distributional free conditions Distributional Fitting Single Variable and Multiple Variables Another powerful simulation tool is distributional fitting that is which distribution does an analyst or engineer use for a particular input variable in a model What are the relevant distributional parameters If no historical data exist then the analyst must make assumptions about the variables in question One approach is to use the Delphi method where a group of experts are tasked with estimating the behavior of each variable For instance a group of mechanical engineers can be tasked with evaluating the extreme possibilities of the diameter of a spring coil through rigorous experimentation or guesstimates These values can be used as the variable s input parameters e g uniform distribution with extreme values between 0 5 and 1 2 When testing is not possible e g market share and revenue growth rate management still can make estimates of potential outcomes and provide the best case most likely case and worst case scenarios whereupon a triangular or custom distribution can be created However if reliable historical data are available distributional fitting can be accomplished Assuming that historical patterns hold and that history tends to repeat itself historical data can be used to find the best fitting distribution with their relevant parameters to better define the variables to be simulated Figures 31 through 33 illustrat
52. isk Simulator icon toolbars in Excel 2007 Running a Monte Carlo Simulation Typically to run a simulation in your existing Excel model you must perform these five steps 1 Start a new or open an existing simulation profile Define input assumptions in the relevant cells Define output forecasts in the relevant cells Run the simulation Uu ea V PM Interpret the results If desired and for practice open the example file called Basic Simulation Model and follow along the examples on creating a simulation The example file can be found on the start menu at Start Real Options Valuation Risk Simulator Examples I Starting a New Simulation Profile To start a new simulation you must first create a simulation profile A simulation profile contains a complete set of instructions on how you would like to run a simulation it contains all the assumptions forecasts simulation run preferences and so forth Having profiles facilitate creating multiple scenarios of simulations that is using the same exact model several profiles can be created each with its own specific simulation assumptions forecasts properties and requirements The same analyst can create different test scenarios using different distributional assumptions and inputs or multiple users can test their own assumptions and inputs on the same model Instead of having to make duplicates of the model the same model can be used and different simulations can be run throug
53. ith a mean of zero and standard deviation of one where we apply the ICDF to find the value of x that corresponds to the cumulative probability of 97 50 CDF That is a one tail CDF of 97 50 is equivalent to a two tail 95 confidence interval there is a 2 50 probability in the right tail and 2 50 in the left tail leaving 95 in the center or confidence interval area which is equivalent to a 97 50 area for one tail The result is the familiar Z Score of 1 96 Therefore using this Distributional Analysis tool the standardized scores for other distributions the exact and cumulative probabilities of other distributions can all be obtained quickly and easily Distribution Analysis This tool generates the probability density function PDF cumulative distribution function CDF and the Inverse CDF ICDF of all the distributions in Risk Simulator including theoretical moments and probability chart Distribution Mu Sigma Type Formatting Single Value Probability Range of Values Lower Bound Upper Bound Step Size Figure 53 Distributional Analysis Tool Normal Distribution s ICDF and Z Score
54. ity Charts Linear file e Select Risk Simulator Tools Sensitivity Analysis e Select the forecast of choice to analyze and click OK Figure 28 Note that sensitivity analysis cannot be run unless assumptions and forecasts have been defined and a simulation has been run Discounted Cash Flow Model Dase ear 2005 Gum PY Met Benetits 7 896 63 Market Risk Adjusted Discount Hate 15 0058 aum PY Investments 1 800 00 Pnyate Hisk Discount Rate 5 00 Met Present Value 96 63 Annualized Sales Growth Rate 2 05s internal Rate of Return 18 80565 Price Erosion Rate 5 00 Return on Investment 5 3 05 Effective Tax Rate 40 00 2005 F Sensitivity Analysis ProductaA Avg PricefUnit 10 00 0 00 Sensiivi lus ue MT al ensitivity analysis creates dynamic perturbations e multiple Product B Avg PE 1225 assumptions are perturbed simultaneously to identity the impact Product C Avg PricefUnit 15 15 to the results Itis used to identify critical success factors of the ProductA Sale Quantity 0003 50 00 forecast Praduct B Sale Quantity 000s 35 00 Please select the forecasts to run sensitivity analysis Product C Sale Quantity CO00s 20 00 Total Revenues 1 231 75 Fateci I zmz Worksheet Direct Cost of Goods Sold 184 76 DCF Model Histogram Statistics Preferences Options Het Present Value Histagram Cumulative Probability 120 PI a 100 ann i j BU ln C E
55. l and other commodity prices using a simple regression model because these variables are highly uncertain and volatile and does not follow a predefined static rule of behavior in other words the process is not stationary Stationarity is checked here using the Runs Test while another visual clue is found in the Autocorrelation report the ACF tends to decay slowly A stochastic process is a sequence of events or paths generated by probabilistic laws That is random events can occur over time but are governed by specific statistical and probabilistic rules The main stochastic processes include Random Walk or Brownian Motion Mean Reversion and Jump Diffusion These processes can be used to forecast a multitude of variables that seemingly follow random trends but restricted by probabilistic laws The process generating equation is known in advance but the actual results generated is unknown Figure 42 The Random Walk Brownian Motion process can be used to forecast stock prices prices of commodities and other stochastic time series data given a drift or growth rate and volatility around the drift path The Mean Reversion process can be used to reduce the fluctuations of the Random Walk process by allowing the path to target a long term value making it useful for forecasting time series variables that have a long term rate such as interest rates and inflation rates these are long term target rates by regulatory authorities or the market The Ju
56. l issue when forecasting time series data is whether the independent variable values are truly independent of each other or are they dependent Dependent variable values collected over a time series may be autocorrelated For serially correlated dependent variable values the estimates of the slope and intercept will be unbiased but the estimates of their forecast and variances will not be reliable and hence the validity of certain statistical goodness of fit tests will be flawed For instance interest rates inflation rates sales revenues and many other time series data are typically autocorrelated where the value in the current period is related to the value in a previous period and so forth clearly the inflation rate in March is related to February s level which in turn is related to January s level and so forth Ignoring such blatant relationships will yield biased and less accurate forecasts In such events an autocorrelated regression model or an ARIMA model may be better suited Risk Simulator Forecasting ARIMA Finally the autocorrelation functions of a series that is nonstationary tend to decay slowly see Nonstationary report in the model If autocorrelation AC is nonzero it means that the series is first order serially correlated If AC K dies off more or less geometrically with increasing lag it implies that the series follows a low order autoregressive process If AC k drops to zero after a small number of lags it
57. lags are statistically significant and should be considered Autocorrelation Time Lag AC PAC LowerBound Upper Bound Q Stat Prob 1 0 0580 0 0580 0 2828 0 2828 0 1786 0 6726 2 0 1213 0 1251 0 2828 0 2828 0 9754 0 6140 3 0 0590 0 0756 0 2828 0 2828 1 1679 0 7607 4 0 2423 0 2232 0 2828 0 2828 4 4865 0 3442 5 0 0067 0 0078 0 2828 0 2828 4 4890 0 4814 6 0 2654 0 2345 0 2828 0 2828 8 6516 0 1941 d 7 0 0814 0 0939 0 2828 0 2828 9 0524 0 2489 I I 8 0 0634 0 0442 0 2828 0 2828 9 3012 0 3175 I I g 0 0204 0 0673 0 2828 0 2828 9 3276 0 4076 I I 10 0 0190 0 0865 0 2828 0 2828 9 3512 0 4991 I I 11 0 1035 0 0790 0 2828 0 2828 10 0648 0 5246 l 14 0 1658 0 0978 0 2828 0 2828 11 9466 0 4500 13 0 0524 0 0430 0 2828 0 2828 12 1394 0 5162 l 14 0 2050 0 2523 0 2828 0 2828 15 1738 0 3664 i I 15 0 1782 0 2089 0 2828 0 2828 17 5315 0 2881 i l 16 0 1022 0 2591 0 2828 0 2828 18 3296 0 3050 i 17 0 0861 0 0808 0 2828 0 2828 18 9141 0 3335 i l 18 0 0418 0 1987 0 2828 0 2828 19 0559 0 3884 19 0 0869 0 0821 0 2828 0 2828 19 6894 0 4135 20 0 0091 0 0269 0 2828 0 2828 19 6966 0 4770 Distributive Lags P Values of Distributive Lag Periods of Each Independent Variable Variable 1 2 3 4 B 5 7 g H 10 11 12 X1 0 8467 0 2045 0 3336 0 9105 0 9757 0 1020 0 9205 0 1267 0 5431 0 9110 0 7495 0 4016 X2 0 6077 0 9900 0 8422 0 2851 0 0638 0 0032 0 8007 0 1551 0 4823 0 1125 0 0519 0 4383 X3 0 7394 0 2396 0 2741 0 8372 0 3808 0 046
58. less the heteroskedasticity of the dependent variable is pronounced its effect will not be severe the least squares estimates will still be unbiased and the estimates of the slope and intercept will either be normally distributed if the errors are normally distributed or at least normally distributed asymptotically as the number of data points becomes large if the errors are not normally distributed The estimate for the variance of the slope and overall variance will be inaccurate but the inaccuracy is not likely to be substantial if the independent variable values are symmetric about their mean If the number of data points is small micronumerosity it may be difficult to detect assumption violations With small sample sizes assumption violations such as non normality or heteroskedasticity of variances are difficult to detect even when they are present With a small number of data points linear regression offers less protection against violation of assumptions With few data points it may be hard to determine how well the fitted line matches the data or whether a nonlinear function would be more appropriate Even if none of the test assumptions are violated a linear regression on a small number of data points may not have sufficient power to detect a significant difference between the slope and zero even if the slope is nonzero The power depends on the residual error the observed variation in the independent variable the selected signifi
59. linear trend of the rest of the data especially if the point is relatively far horizontally from the center of the data However great care should be taken when deciding if the outliers should be removed Although in most cases when outliers are removed the regression results look better a priori justification must first exist For instance if one is regressing the performance of a particular firm s stock returns outliers caused by downturns in the stock market should be included these are not truly outliers as they are inevitabilities in the business cycle Forgoing these outliers and using the regression equation to forecast one s retirement fund based on the firm s stocks will yield incorrect results at best In contrast suppose the outliers are caused by a single nonrecurring business condition e g merger and acquisition and such business structural changes are not forecast to recur then these outliers should be removed and the data cleansed prior to running a regression analysis The analysis here only identifies outliers and it is up to the user to determine if they should remain or be excluded Sometimes a nonlinear relationship between the dependent and independent variables is more appropriate than a linear relationship In such cases running a linear regression will not be optimal If the linear model is not the correct form then the slope and intercept estimates and the fitted values from the linear regression will be bi
60. metrics modified multivariate regression forecasts o Box Jenkins ARIMA econometric forecasts o GARCH Models forecasting and modeling volatility o J Curves exponential growth forecasts o Markov Chains market share and dynamics forecast o Multivariate regressions modeling linear and nonlinear relationships among variables o Nonlinear extrapolations curve fitting o S Curves logistic growth forecasts o Spline Curves interpolating and extrapolating missing values o Stochastic processes random walks mean reversions jump diffusion and mixed processes The Optimization Module is used for optimizing multiple decision variables subject to constraints to maximize or minimize an objective It and can be run as a static optimization as a dynamic optimization under uncertainty together with Monte Carlo simulation or as a stochastic optimization The software can handle linear and nonlinear optimizations with integer and continuous variables The Real Options Super Lattice Solver SLS is another standalone software that complements Risk Simulator used for solving simple to complex real options problems To install the software insert the accompanying CD ROM click on the nstall Risk Simulator link and follow the onscreen instructions You will need to be online to download the latest version of the software The software requires Windows 2000 XP Vista administrative privileges and Microsoft Net Framework 1 1 installed on the co
61. mp Diffusion process is useful for forecasting time series data when the variable can occasionally exhibit random jumps such as oil prices or price of electricity discrete exogenous event shocks can make prices jump up or down These processes can also be mixed and matched as required Statistical Summary The following are the estimated parameters for a stochastic process given the data provided It is up to you to determine if the probability of fit similar to a goodness of fit computation is sufficient to warrant the use of a stochastic process forecast and if so whether it is a random walk mean reversion or a jump diffusion model or combinations thereof In choosing the right stochastic process model you will have to rely on past experiences and a pron economic and financial expectations of what the underlying data set is best represented by These parameters can be entered into a stochastic process forecast Simulation Forecasting Stochastic Processes Periodic Drift Rate 1 489 Reversion Rate 283 89 Jump Rate 20 4196 Volatility 88 84 Long Term Value 327 72 Jump Size 237 89 Probability of stochastic mode fit 46 48 A high fit means a stochastic model is better than conventional models Runs 20 Standard Normal 1 7321 Positive 25 P Value 1 tai 0 0416 Negative 25 P Vl aiue 2 tai 0 0833 Expected Run 26 A low p value below 0 10 0 05 0 01 means that the sequence Js not random and hence suffers from stationarity p
62. mputer Most new computers come with Microsoft NET Framework 1 1 already preinstalled However if an error message pertaining to requiring NET Framework 1 1 occurs during the installation of Risk Simulator exit the installation Then install the relevant NET Framework software also included in the CD found in the DOT NET Framework folder Complete the NET installation restart the computer and then reinstall the Risk Simulator software Version 1 1 of the NET Framework is required even if your system has version 2 0 3 0 as they work independently of each other You may also download this software on the Download page of www realoptionsvaluation com Once installation is complete start Microsoft Excel If the installation was successful you should see an additional Simulation item on the menu bar in Excel and a new icon bar as shown in Figure 1 Figure 2 shows the icon toolbar in more detail You are now ready to start using the software Please note that Risk Simulator supports multiple languages e g English Chinese Japanese and Spanish and you can switch among languages by going to Risk Simulator Languages There is a default 30 day trial license file that comes with the software To obtain a full corporate license please contact the author s firm Real Options Valuation Inc at admin realoptionsvaluation com If you are using Windows Vista make sure to disable User Access Control before installing the software license To
63. nce levels If the null hypothesis is not rejected when the p values are high the means of the two forecast distributions are statistically similar to one another The same analysis is performed on variances of two forecasts at a time using the pairwise F test If the p values are small then the variances and standard deviations are statistically different from one another otherwise for large p values the variances are Statistically identical to one another The example file used was Hypothesis Testing and Bootstrap Simulation Notes The two variable t test with unequal variances the population variance of forecast 1 is expected to be different from the population variance of forecast 2 is appropriate when the forecast distributions are from different populations e g data collected from two different geographical locations or two different Operating business units and so forth The two variable t test with equal variances the population variance of forecast 1 is expected to be equal to the population variance of forecast 2 is appropriate when the forecast distributions are from similar populations e g data collected from two different engine designs with similar specifications and so forth The paired dependent two variable t test is appropriate when the forecast distributions are from exactly the same population and subjects e g data collected from the same group of patients before an experimental drug was used and after the
64. om one another or their differences are due to random chance Please select two forecasts to run hypothesis test Histogram Statistics Preferences Options Forecast Name Worksheet Cell Income Model 4 Simulation Model D10 Income Model A Histogram Cumulative Probability 5000 Trials Income Model B Simulation Model G10 600 D 1 00 500 o Z p 400 3 300 i 200 xj 100 n m Assumptions 87 Income Model B Histogram Cumulative Probability 5000 Trials Independent Samples With Unequal Variances D 600 1 00 Independent S amples With Equal Variances Type Two Tail 500 0 80 Dependent Sample Pairs 4 e e Frequency ho ww C C e e e e Ayiqeqoag eAnejnun db 72 93 72 98 72 103 72 108 72 113 300 Type Two Tail v je Infinity i Infinity Certainty 100 00 Figure 36 Hypothesis testing Results Interpretation A two tailed hypothesis test 1s performed on the null hypothesis H such that the population means of the two variables are statistically identical to one another The alternative hypothesis H is such that the population means are statistically different from one another If the calculated p values are less than or equal to 0 01 0 05 or 0 10 alpha test levels it means that the null hypothesis is rejected which implies that the forecast means are statistically significantly different at the 1 5 and 10 significa
65. opulation mean is statistically greater than the hypothesized mean when tested using the sample dataset Using a t test if the p value is less than a specified significance amount typically 0 10 0 05 or 0 01 this means that the population mean is statistically significantly greater than the hypothesized mean at 1096 5 and 1 significance value or 90 95 and 99 statistical confidence Conversely ifthe p value is higher than 0 10 0 05 or 0 01 the population mean is statistically similar or less than the hypothesized mean Left Tailed Hypothesis Test A left tailed hypothesis tests the null hypothesis Ho such that the population mean is statistically greater than or equal to the hypothesized mean The alternative hypothesis is thatthe real population mean is statistically less than the hypothesized mean when tested using the sample dataset Using a t test if the p value is less than a specified significance amount typically 0 10 0 05 or 0 01 this means that the population mean is statistically significantly less than the hypothesized mean at 1096 596 and 196 significance value or 9096 9596 and 9996 statistical confidence Conversely if the p value is higher than 0 10 0 05 or 0 01 the population mean is statistically similar or greater than the hypothesized mean and any differences are due ti random chance Because the t test is more conservative and does not require a known population standard deviation as in the Z test we only us
66. orrelation between Grades and Study and assuming that the correlation coefficients have high magnitudes the correlation matrix will be nonpositive definite It would defy logic correlation requirements and matrix mathematics However smaller coefficients sometimes still can work even with the bad logic When a nonpositive definite or bad correlation matrix is entered Risk Simulator automatically informs you of the error and offers to adjust these correlations to something that is semipositive definite while still maintaining the overall structure of the correlation relationship the same signs as well as the same relative strengths The Effects of Correlations in Monte Carlo Simulation Although the computations required to correlate variables in a simulation are complex the resulting effects are fairly clear Figure 16 shows a simple correlation model Correlation Effects Model in the example folder The calculation for revenue is simply price multiplied by quantity The same model is replicated for no correlations positive correlation 0 9 and negative correlation 0 9 between price and quantity Correlation Model Without Fositive Negative Correlation Correlation Correlation Price 2 00 2 00 2 00 Quantity 1 00 1 00 1 00 Revenue 2 00 2 00 2 00 Figure 16 Simple correlation model The resulting statistics are shown in Figure 17 Notice that the standard deviation of the model without correlations is 0 1450 compared to
67. otion process can be used to forecast stock prices prices of commodities and other stochastic time series data given a drift or growth rate and a volatility around the drift path The Mean Reversion process can be used to reduce the fluctuations of the Random Walk process by allowing the path to target a long term value making it useful for forecasting time series variables that have a long term rate such as interest rates and inflation rates these are long term target rates by regulatory authorities or the market The Jump Diffusion process is useful for forecasting time series data when the variable can occasionally exhibit random jumps such as oil prices or price of electricity discrete exogenous event shocks can make prices jump up or down Finally these three stochastic processes can be mixed and matched as required Statistical Summary The following are the estimated parameters for a stochastic process given the data provided It is up to you to determine ifthe probability of fit similar to a goodness of fit computation is sufficient to warrant the use of a stochastic process forecast and if so whether itis a random walk mean reversion or a jump diffusion model or combinations thereof In choosing the right stochastic process model you will have to rely on past experiences and a priori economic and financial expectations of what the underlying data set is best represented by These parameters can be entered into a stochastic process fo
68. pothesis and accept the alternate hypothesis the errors are not normally distributed Otherwise if the D Statistic is less than the D Critical value do not reject the null hypothesis the errors are normally distributed This test relies on two cumulative frequencies one derived from the sample data set the second from a theoretical distribution based on the mean and standard deviation of the sample data Test Result Errors P9f ve Observed Expected OE Regression Error Average B aO Frequency Standard Deviation of Errors 141 83 218 04 B 02 B O2 B O6 T2 0 fat O Statistic B 1035 202 253 Og b fle BOE 0 U356 D Critica at 1 o 1138 126 04 B O2 O06 B OSq8 0 03438 L Critica at 595 Uu 1225 Pr4dy Og b O8 U TOS 0 O28 L Critical at T0965 B 1458 162 13 B O2 B T OF 265 0 02523 Nul Hypothesis The errors are normally distributed T51 52 D O2 Od D T2727 0 OU 72 160 39 O02 Ufa B T2917 BU TOS Conclusion The errors are normally distributed at the 143 40 Oo O76 07526 OO rel Ts alpha level 138 92 ERE B T8 01637 OTE 133 81 B O2 B 20 Ogre B 0273 T20 76 b O2 D 22 Bo 1973 Gore 120 12 B O2 O 24d OF Se Dagia Figure 41 Test for Normality of Errors Sometimes certain types of time series data cannot be modeled using any other methods except for a stochastic process because the underlying events are stochastic in nature For instance you cannot adequately model and forecast stock prices interest rates price of oi
69. ptimization Profile Simulation i Honparameteric Bootstrap Figure 2A Risk Simulator icon toolbar in Excel XP and Excel 2003 N de c a Microsoft Excel em 93 Home o Insert i Page Layout Formulas Data Review f View Developer Risk Simulator a EER E EEA orale Next icon Icon Maximum Likelihood Stochastic Processes J Set Decision LE Fitting Single m among different icon toolbars in Risk Simulator Risk New Edit SetInput SetOutput Copy Paste Remove Run Step Reset prays 4 More Simulator Profile Profile Assumption Forecast Nonlinear Extrapolation Time Series Analysis Giese toe 5 Set Constraint 9 license Menu Profile Assumptions Forecasts T Simulation Run Forecasting Optimization Tools Risk Simulator Set a new input Copy Paste or Run a simulation with i Optimization analysis Run Most frequently Main Menu Items assumption or anew Remove an existing multiple trials run a Forecasting Models and Analytics optimization set objective used tools and output forecast assumption forecast single trial step or set decision variables and License Create a new or decision variable reset the model after set constraints Installation profile or edit the a simulation run existing active profile Shows all the Analytical Tools 5 do c amp Microsoft Excel S Home Insert Page Layout Formulas Data Review View Developer Risk Simulator L Bx xm Cr BPS ce amp
70. ptive statistics to testing for and calibrating the stochastic properties of the data Procedure amp Open the example model Risk Simulator Examples Statistical Analysis and go to the Data worksheet and select the data including the variable names cells C5 E55 amp Click on Risk Simulator Tools Statistical Analysis Figure 44 amp Check the data type whether the data selected is from a single variable or multiple variables arranged in rows In our example we assume that the data areas selected are from multiple variables Click OK when finished W Choose the statistical tests you wish to perform The suggestion and by default is to choose all the tests Click OK when finished Figure 45 Spend some time going through the reports generated to get a better understanding of the statistical tests performed sample reports are shown in Figures 46 49 Variable X1 Variable X2 variable X3 52 367 443 365 614 355 206 397 fb4 427 153 431 524 325 240 206 205 564 ab 496 481 460 177 188 455 108 Data Set 18308 1148 18058 7729 100484 15728 14630 4006 30927 22322 ale 313b 50508 20006 16996 13035 12973 15308 5227 15235 44487 414213 23513 9106 24917 38 2 185 BDD 3 2 Statistical Analysis This tool is used to describe and find statistical relationships in a set of raw data Selected Data Variable X1 Variable xz Variable x3 18306 1146 18068 P729 100484 15
71. recast Simulation Forecasting Stochastic Processes Annualized Drift Rate 5 86 Reversion Rate N A Jump Rate 16 33 Volatility 7 04 Long Term Value N A Jump Size 21 33 Probability of stochastic model fit 4 6396 Figure 49 Sample Statistical Analysis Tool Report Stochastic Parameter Estimation Distributional Analysis Tool This is a statistical probability tool in Risk Simulator that is rather useful in a variety of settings and can be used to compute the probability density function PDF which is also called the probability mass function PMF for discrete distributions we will use these terms interchangeably where given some distribution and its parameters we can determine the probability of occurrence given some outcome x In addition the cumulative distribution function CDF can also be computed which is the sum of the PDF values up to this x value Finally the inverse cumulative distribution function ICDF is used to compute the value x given the probability of occurrence This tool is accessible via Risk Simulator Tools Distributional Analysis As an example Figure 50 shows the computation of a binomial distribution i e a distribution with two outcomes such as the tossing of a coin where the outcome is either Heads or Tails with some prescribed probability of heads and tails Suppose we toss a coin two times and set the outcome Heads as a success we use the binomial distribution with Trials 2
72. rgest value The Third Quartile is the value of the 75th percentile data point Sometimes quartiles are used as the upper and lower ranges of a distribution as it truncates the data set to ignore outliers The Inter Quartile Range is the difference between the third and first quartiles and is often used to measure the width of the center of a distribution Skewness is the third momentin a distribution Skewness characterizes the degree of asymmetry of a distribution around its mean Positive skewness indicates a distribution with an asymmetric tail extending toward more positive values Negative skewness indicates a distribution with an asymmetric tail extending toward more negative values Kurtosis characterizes the relative peakedness or flatness of a distribution compared to the normal distribution It is the fourth moment in a distribution A positive Kurtosis value indicates a relatively peaked distribution A negative kurtosis indicates a relatively flat distribution The Kurtosis measured here has been centered to zero certain other kurtosis measures are centered around 3 0 While both are equally valid centering across zero makes the interpretation simpler A high positive Kurtosis indicates a peaked distribution around its center and leptokurtic or fattails This indicates a higher probability of extreme events e g catastrophic events terrorist attacks stock market crashes than is predicted in a normal distribution Summary Statistics Sta
73. ric simulation is simply randomly picking golf balls from a large basket with replacement where each golf ball is based on a historical data point Suppose there are 365 golf balls in the basket representing 365 historical data points Imagine if you will that the value of each golf ball picked at random is written on a large whiteboard The results of the 365 balls picked with replacement are written in the first column of the board with 365 rows of numbers Relevant statistics e g mean median mode standard deviation etc are calculated on these 365 rows The process is then repeated say 5 000 times The whiteboard will now be filled with 365 rows and 5 000 columns Hence 5 000 sets of statistics that is there will be 5 000 means 5 000 medians 5 000 modes 5 000 standard deviations and so forth are tabulated and their distributions shown The relevant statistics of the statistics are then tabulated where from these results you can ascertain how confident the simulated statistics are Finally bootstrap results are important because according to the Law of Large Numbers and Central Limit Theorem in statistics the mean of the sample means is an unbiased estimator and approaches the true population mean when the sample size increases Hypothesis Testing A hypothesis test is performed when testing the means and variances of two distributions to determine if they are statistically identical or statistically different from one another
74. roblems and an ARIMA model might be more appropriate Conversely higher p values Indicate randomness and stochastic process models might be appropriate Figure 42 Stochastic Process Parameter Estimation Multicollinearity exists when there is a linear relationship between the independent variables When this occurs the regression equation cannot be estimated at all In near collinearity situations the estimated regression equation will be biased and provide inaccurate results This situation is especially true when a step wise regression approach is used where the statistically significant independent variables will be thrown out of the regression mix earlier than expected resulting in a regression equation that is neither efficient nor accurate One quick test of the presence of multicollinearity in a multiple regression equation is that the R squared value is relatively high while the t statistics are relatively low Another quick test is to create a correlation matrix between the independent variables A high cross correlation indicates a potential for autocorrelation The rule of thumb is that a correlation with an absolute value greater than 0 75 is indicative of severe multicollinearity Another test for multicollinearity is the use of the Variance Inflation Factor VIF obtained by regressing each independent variable to all the other independent variables obtaining the R squared value and calculating the VIF A VIF exceeding 2 0 can
75. s Results Interpretation Figure 21 shows the resulting tornado analysis report which indicates that capital investment has the largest impact on net present value NPV followed by tax rate average sale price and quantity demanded of the product lines and so forth The report contains four distinct elements Statistical summary listing the procedure performed Sensitivity table Figure 22 shows the starting NPV base value of 96 63 and how each input is changed e g Investment is changed from 1 800 to 1 980 on the upside with a 10 swing and from 1 800 to 1 620 on the downside with a 10 swing The resulting upside and downside values on NPV are 83 37 and 276 63 with a total change of 360 making it the variable with the highest impact on NPV The precedent variables are ranked from the highest impact to the lowest impact The spider chart Figure 23 illustrates these effects graphically The y axis is the NPV target value while the x axis depicts the percentage change on each of the precedent value The central point is the base case value at 96 63 at 0 change from the base value of each precedent Positively sloped lines indicate a positive relationship or effect negatively sloped lines indicate a negative relationship e g investment is negatively sloped which means that the higher the investment level the lower the NPV The absolute value of the slope indicates the magnitude of the effect computed as
76. s might indicate that a particular part or component when subjected to a high enough force or tension will break Clearly it is important to understand such nonlinearities Mm Stock Price E trike Price m Dividend Yield 0 0 i 60 00 40 00 20 00 0 0036 20 00 40 00 60 00 Figure 25 Nonlinear spider chart Sensitivity Analysis A related feature is sensitivity analysis While tornado analysis tornado charts and spider charts applies static perturbations before a simulation run sensitivity analysis applies dynamic perturbations created after the simulation run Tornado and spider charts are the results of static perturbations meaning that each precedent or assumption variable is perturbed a preset amount one at a time and the fluctuations in the results are tabulated In contrast sensitivity charts are the results of dynamic perturbations in the sense that multiple assumptions are perturbed simultaneously and their interactions in the model and correlations among variables are captured in the fluctuations of the results Tornado charts therefore identify which variables drive the results the most and hence are suitable for simulation sensitivity charts identify the impact to the results when multiple interacting variables are simulated together in the model This effect is clearly illustrated in Figure 26 Notice that the ranking of critical success drivers is similar to the tornado chart in the previous
77. se Value 96 6261638553219 Input Changes Investment ea Output Output Effective Input Input Base Case Precedent Cell Downside Upside Range Downside Upside Value 276 63 83 37 360 00 1 620 00 1 980 00 1 800 00 219 73 26 47 f 44 00 40 00 3 43 189 83 F 11 00 16 71 176 55 13 48 23 18 170 07 55 00 30 58 162 72 j 38 50 40 15 153 11 16 67 48 05 145 20 i 22 00 138 24 57 03 16 50 116 80 76 64 i 5 50 90 59 102 69 j 2 2096 95 08 98 17 11 00 97 09 96 16 2 20 96 16 97 09 3 30 D Price Erosion 96 68 96 63 0 00 l i tax Bala 96 63 96 63 0 00 mvoctmers rv estmerimt Tax Rate A Quantity B Quantity C Price C Quantity 18 22 Discount Rate 0 165 NN 0 135 Price Erosion 0 055 II 0 045 Sales Growth o s 0 022 Depreciation sn Interest 22 18 Amortization 27 33 Capex Met Capital 0 Figure 21 Tornado analysis report Notes Remember that tornado analysis is a static sensitivity analysis applied on each input variable in the model that is each variable is perturbed individually and the resulting effects are tabulated This makes tornado analysis a key component to execute before running a simulation Caputiring and identifying the most important impact drivers in the modek is one of the very first steps in risk analysis The next step is to identify which of these important impact drivers are uncertain These uncertain impact drivers are the critical
78. stribution tj Kurtosis Risk Simulator Forecast Distribution tj Histogram Statistics Preferences Options Skewness Histogram Cumulative Probability 1000 Trials Kurtosis Histogram Cumulative Probability 1000 Trials OD M0 002 coeooo Frequency oo to m oo APEH aAqeinuns Frequency ea D oo Type TwoTal w B 0 0189 0 0952 Certainty Type TwoTal B 0 0514 0 1634 Certainty Figure 35 Bootstrap simulation results Results Interpretation Figure 35 illustrates some sample bootstrap results The example file used was Hypothesis Testing and Bootstrap Simulation For instance the 90 confidence for the skewness statistic is between 0 0189 and 0 0952 such that the value O falls within this confidence indicating that on a 90 confidence the skewness of this forecast is not statistically significantly different from zero or that this distribution can be considered as symmetrical and not skewed Conversely if the value 0 falls outside of this confidence then the opposite is true The distribution is skewed positively skewed if the forecast statistic is positive and negatively skewed if the forecast statistic is negative Notes The term bootstrap comes from the saying to pull oneself up by one s own bootstraps and is applicable because this method uses the distribution of statistics themselves to analyze the accuracy of the statistics Nonparamet
79. t the forecast window is resizable by clicking on and dragging the bottom right corner of the window Finally it is always advisable that before rerunning a simulation you reset the current simulation by selecting Risk Simulator Reset Simulation Correlations and Precision Control The correlation coefficient is a measure of the strength and direction of the relationship between two variables and can take on any values between 1 0 and 1 0 that is the correlation coefficient can be decomposed into its direction or sign positive or negative relationship between two variables and the magnitude or strength of the relationship the higher the absolute value of the correlation coefficient the stronger the relationship The correlation coefficient can be computed in several ways The first approach is to manually compute the correlation coefficient r of a pair of variables x and y using ae n X X 2 ynl D V2 x The second approach is to use Excel s CORREL function For instance if the 10 data points for x and y are listed in cells A1 B10 then the Excel function to use is CORREL A1 A10 BI BIO The third approach is to run Risk Simulator s Multi Variable Distributional Fitting Tool and the resulting correlation matrix will be computed and displayed It is important to note that correlation does not imply causation Two completely unrelated random variables might display some correlation but this does not imply an
80. the covariance cov measure where R 2 The benefit of dividing the covariance by the product M of the two variables standard deviation 5 1s that the resulting correlation coefficient is bounded between 1 0 and 1 0 inclusive This makes the correlation a good relative measure to compare among different variables particularly with different units and magnitude The Spearman rank based nonparametric correlation is also included below The Spearman s R is related to the Pearson s R in that the data is first ranked and then correlated The rank correlations provide a better estimate of the relationship between two variables when one or both of them 1s nonlinear It must be stressed that a significant correlation does not imply causation Associations between variables in no way imply that the change of one variable causes another variable to change When two variables that are moving independently of each other but in a related path they may be correlated but their relationship might be spurious e g a correlation between sunspots and the stock market might be strong but one can surmise that there is no causality and that this relationship is purely spurious Statistical Analysis Tool Another very powerful tool in Risk Simulator is the Statistical Analysis Tool which determines the Statistical properties of the data The diagnostics run include checking the data for various statistical properties from basic descri
81. the percentage change in the result given a percentage change in the precedent A steep line indicates a higher impact on the NPV y axis given a change in the precedent x axis The tornado chart Figure 24 illustrates the results in another graphical manner where the highest impacting precedent is listed first The x axis is the NPV value with the center of the chart being the base case condition Green lighter bars in the chart indicate a positive effect red darker bars indicates a negative effect Therefore for investments the red darker bar on the right side indicate a negative effect of investment on higher NPV in other words capital investment and NPV are negatively correlated The opposite is true for price and quantity of products A to C their green or lighter bars are on the right side of the chart Tornado and Spider Charts Statistical Summary One of the powerful simulation tools is the tornado chart it captures the static impacts of each variable on the outcome of the model That is the tool automatically perturbs each precedent variable in the model a user specified preset amount captures the fluctuation on the model s forecast or final result and lists the resulting perturbations ranked from the most significant to the least Precedents are all the input and intermediate variables that affect the outcome of the model For instance if the model consists of A B C where C D E then B D and E are the precedents for
82. tion is not known but the sampling distribution is assumed to be approximately normal the t test is used when the sample size is less than 30 but is also appropriate and in fact provides more conservative results with larger data sets This t test can be applied to three types of hypothesis tests a two tailed test a right tailed test and a left tailed test All three tests and their respective results are listed below for your reference Two Tailed Hypothesis Test A two tailed hypothesis tests the null hypothesis Ho such that the population mean is statistically identical to the hypothesized mean The alternative hypothesis is that the real population mean is statistically different from the hypothesized mean when tested using the sample dataset Using a t test if the computed p value is less than a specified significance amount typically 0 10 0 05 or 0 01 this means that the population mean is statistically significantly different than the hypothesized mean at 1096 5 and 1 significance value or at the 9096 95 and 99 statistical confidence Conversely if the p value is higher than 0 10 0 05 or 0 01 the population mean is statistically identical to the hypothesized mean and any differences are due to random chance Right Tailed Hypothesis Test A right tailed hypothesis tests the null hypothesis Ho such that the population mean is statistically less than or equal to the hypothesized mean The alternative hypothesis is that the real p
83. tions can be assigned only to cells without any equations or functions i e typed in numerical values that are inputs in a model whereas output forecasts can be assigned only to cells with equations and functions i e outputs of a model Recall that assumptions and forecasts cannot be set unless a simulation profile already exists Follow these steps to set new input assumptions in your model e Select the cell you wish to set an assumption on e g cell G8 in the Basic Simulation Model example e Click on Risk Simulator Set Input Assumption or click the Set Assumption icon in the Risk simulator icon toolbar e Select the relevant distribution you want enter the relevant distribution parameters and hit OK to insert the input assumption into your model Figure 5 IX Assumption Properties bod E e A Normal i Minimum 1 Ij Most Likely 2 zi 2 5 Ej Regular Input Percentile Input Bemoulli Enable Data Boundary Triangular Distribution The triangular distribution describes aj Minimum mt E finit situation where you know the minimum maximum and most likely values to occur Maximum Tinity E For example you could describe the E ic Simulat number af cars sold per week when past e Dynamic Simulations sales show the minimum maximum and _ Figure 5 Setting an input assumption Several key areas are worthy of mention in the Assumption Properties Figure 6 shows the different areas
84. tistics Variable X1 Observations 50 0000 Standard Deviation Sample 172 9140 Arithmetic Mean 331 9200 Standard Deviation Population 171 1761 Geometric Mean 281 3247 Lower Confidence Interval for Standard Deviation 148 6090 Trimmed Mean 325 1739 Upper Confidence Interval for Standard Deviation 207 7947 Standard Error of Arithmetic Mean 24 4537 Variance Sample 29899 2588 Lower Confidence Interval for Mean 283 0125 Variance Population 29301 2736 Upper Confidence Interval for Mean 380 8275 Coefficient of Variability 0 5210 Median 307 0000 First Quartile Q1 204 0000 Mode 47 0000 Third Quartile Q3 441 0000 Minimum 764 0000 Inter Quartile Range 237 0000 Maximum 717 0000 Skewness 0 4838 Range Kurtosis 0 0952 Figure 46 Sample Statistical Analysis Tool Report Hypothesis Test t Test on the Population Mean of One Variable Statistical Summary Statistics from Dataset Calculated Statistics Observations 50 t Statistic 13 5734 Sample Mean 331 92 P Value right tail 0 0000 Sample Standard Deviation 172 91 P Value left tailed 1 0000 P Value two tailed 0 0000 User Provided Statistics Null Hypothesis Ho u Hypothesized Mean Hypothesized Mean 0 00 Alternate Hypothesis Ha u gt Hypothesized Mean Notes denotes greater than for right tail Tess than for lef tail or not equal to for two tail hypothesis tests Hypothesis Testing Summary The one variable t test is appropriate when the population standard devia
85. tossing the coin twice and Probability 0 50 the probability of success of getting Heads Selecting the PDF and setting the range of values x as from O to 2 with a step size of 1 this means we are requesting the values 0 1 2 for x the resulting probabilities are provided in the table and graphically as well as the theoretical four moments of the distribution As the outcomes of the coin toss is Heads Heads Tails Tails Heads Tails and Tails Heads the probability of getting exactly no Heads is 25 one Heads is 50 and two Heads is 25 Distribution Analysis This tool generates the probability density function PDF cumulative distribution function CDF and the Inverse CDF ICDF of all the distributions in Risk Simulator including theoretical moments and probability chart Distribution Binomial Trials Fi Probability Figure 50 Distributional Analysis Tool Binomial Distribution with 2 Trials Similarly we can obtain the exact probabilities of tossing the coin say 20 times as seen in Figure 51 The results are presented both in table and graphical formats Distribution Analysis This tool generates the probability density function PDF cumulative distribution function CDF and the Inverse CDF ICDF of all the distributions in Risk Simulator including theoretical moments and probability chart Distribution Binomial Trials 20 Probability 0 5 0 014736 0 004621 0 001087 0 000131 0 000013 0 000001
86. ver in practice the simulated variable exists only within some smaller range This range can be entered to truncate the distribution appropriately Correlations Pairwise correlations can be assigned to input assumptions here If assumptions are required remember to check the Turn on Correlations preference by clicking on Risk Simulator Edit Simulation Profile See the discussion on correlations later in this chapter for more details about assigning correlations and the effects correlations will have on a model Short Descriptions These exist for each of the distributions in the gallery The short descriptions explain when a certain distribution is used as well as the input parameter requirements See the section in the appendix Understanding Probability Distributions in Modeling Risk Applying Monte Carlo Simulation Real Options Analysis Stochastic Forecasting and Portfolio Optimization Wiley 2006 also by the author for details about each distribution type available in the software Enter the assumption name here IX Assumption Properties bol a x Click on the link Normal A Triangular Assumption Name Revenue zh icons to link to any Uniform Ui Custom p cell in Excel Bemoulli il Bets E Ji Binomial FE Cauchy se Different views Chi Square illl Discrete Uniform m exist by right Exponential ATF 1 Enter the selected clicking in Pis A Gamma Geometric Most Likely distribution s Gallery Gumbel Ma
87. when making presentations sometimes you may require the same results For example during a live presentation you may like to shown the same results being generated as are in the report when you are sharing models with others you also may want the same results to be obtained every time If that 1s the case check this preference and enter in an initial seed number The seed number can be any positive integer Using the same initial seed value the same number of trials and the same input assumptions always will yield the same sequence of random numbers guaranteeing the same final set of results Note that once a new simulation profile has been created you can come back later and modify your selections In order to do this make sure that the current active profile is the profile you wish to modify otherwise click on Risk Simulator Change Simulation Profile select the profile you wish to change and click OK Figure 4 shows an example where there are multiple profiles and how to activate duplicate or delete a selected profile Then click on Risk Simulator Edit Simulation Profile and make the required changes E Change Active Simulation Jog Simulation Name Second Simulation E sample Third Simulation Example view simulation profiles in all workbooks Delete LIF Cancel rr Figure 4 Change active simulation 2 Defining Input Assumptions The next step is to set input assumptions in your model Note that assump
88. ximum J Gumbel Minimum 2 E required i Hypergeometric PJK Logistic parameters Lognormal Mi Negative Binomial Maximum Pareto ME Poisson 25 Li Rayleigh PT Regular Input Optional Alternate Weibull AP view of the input ercentile Input parameters iV Enable Correlation 7 Enable Data Boundary Optional Enter new Triangular Distribution Assumption Location Correlation or leave as is the The triangular distribution describes aj Minimum os distributional A short description situation where you know the minimum You can only edit correlations when boundaries of the distribution is maximum and most likely values to occur there is more than one assumption Maximum E available here For example you could describe the F E muc Edd Optional Select to number of cars sold per week when past Enable Dynamic Simulations enable sales show the minimum maximum and Cancel multidimensional simulation Use this area to add edit or remove any correlations among input assumptions Figure 6 Assumption properties Note If you are following along with the example continue by setting another assumption on cell G9 This time use the Uniform distribution with a minimum value of 0 9 and a maximum value of 1 1 Then proceed to defining the output forecasts in the next step 3 Defining Output Forecasts The next step is to define output forecasts in the model Forecasts can be defined only on output cells with equations or functions
89. y causation between the two e g sunspot activity and events in the stock market are correlated but there is no causation between the two There are two general types of correlations parametric and nonparametric correlations Pearson s correlation coefficient is the most common correlation measure and usually is referred to simply as the correlation coefficient However Pearson s correlation is a parametric measure which means that it requires both correlated variables to have an underlying normal distribution and that the relationship between the variables 1s linear When these conditions are violated which 1s often the case in Monte Carlo simulations the nonparametric counterparts become more important Spearman s rank correlation and Kendall s tau are the two nonparametric alternatives The Spearman correlation is used most commonly and is most appropriate when applied in the context of Monte Carlo simulation there is no dependence on normal distributions or linearity meaning that correlations between different variables with different distribution can be applied In order to compute the Spearman correlation first rank all the x and y variable values and then apply the Pearson s correlation computation Risk Simulator uses the more robust nonparametric Spearman s rank correlation However to simplify the simulation process and to be consistent with Excel s correlation function the correlation user inputs required are the Pearson s corr

Download Pdf Manuals

image

Related Search

Related Contents

  

Copyright © All rights reserved.
Failed to retrieve file