295
A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS-BASED APPROACH TO MEASURE THE SHORT-TERM RISK OF INITIAL PUBLIC OFFERINGS by Shabnam Sorkhi A thesis submitted in conformity with the requirements for the degree of Doctor of Philosophy Center for Management of Technology and Entrepreneurship Department of Chemical Engineering and Applied Chemistry University of Toronto © Copyright by Shabnam Sorkhi 2015

A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS-BASED

APPROACH TO MEASURE THE SHORT-TERM RISK OF

INITIAL PUBLIC OFFERINGS

by

Shabnam Sorkhi

A thesis submitted in conformity with the requirements for the degree of Doctor of Philosophy

Center for Management of Technology and Entrepreneurship Department of Chemical Engineering and Applied Chemistry

University of Toronto

© Copyright by Shabnam Sorkhi 2015

Page 2: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

ii

A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS-BASED

APPROACH TO MEASURE THE SHORT-TERM RISK OF

INITIAL PUBLIC OFFERINGS Shabnam Sorkhi

Doctor of Philosophy

Center for Management of Technology and Entrepreneurship Department of Chemical Engineering and Applied Chemistry

University of Toronto

2015

ABSTRACT

Initial public offerings (IPOs) are perhaps the most exhilarating events on stock

exchanges. Yet, the ‘ambiguity’ of the risk of IPOs overshadows the thrill and deters many

investors from possibly considering the IPOs. It is the insufficient accounting and market

history at the IPO stage that burdens their proper risk quantification. The main objective

pursued by this thesis is to offer a methodology for measuring the short-term risk of IPOs

which conforms to the mathematical principles of market risk analysis employed in the case

of public companies. Here, short-term risk is defined as the uncertainty associated with the

stock price of the IPO of interest (IPOI) 90 days subsequent to the issuing day and is

quantified as a value-at-risk (VaR) inferred from the probability density function of price on

day 90 (i.e., PDF90IPOI). This thesis develops a Bayesian framework where PDF90IPOI can be

estimated in a recursive and iterative process. In most IPO cases, there exist limited hard data,

Page 3: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

iii

yet, strong ‘prior’ belief (soft data). The Bayesian setting offers a unique risk quantification

approach which befits and serves these two characteristics of IPOs.

To obtain the data required for carrying out the risk analysis, this research relies upon the

‘closest comparable’ of IPOI. The ‘closest comparable’ would be a public firm whose pre-

IPO idiosyncratic financial data most resemble those of IPOI. Furthermore, it is expected to

have gone public in similar macro-economic and sector conditions. Concisely, the risk

quantification process involves two phases: In Phase I, a Data Envelopment Analysis-based

multi-dimensional similarity metric is developed to select the closest comparable. Phase I

indeed identifies the most suitable source of ‘prior’ knowledge and passes the output to Phase

II which encompasses the Bayesian process. Phase II is designed to formulate and refine the

‘prior’ evidence and then employ the achieved ‘posterior’ knowledge towards estimating

PDF90IPOI. This PDF90IPOI subsequently acts as the basis for VaR inferences. In the last stage of

the research, the proposed Bayesian VaR methodology is examined (backtested) using the

following two tests: test of uniform cumulative probability values and test of VaR break

frequency.

Page 4: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

iv

EXECUTIVE SUMMARY

This thesis proposes a methodology to quantify the short-term risk of investing in initial

public offerings (IPOs). Here, the risk of an IPO firm of interest (IPOI) is defined as a value-

at-risk measure computed based on the distribution of the IPOI’s stock price at the time point

of 90 trading days subsequent to the issuing day (i.e., PDF90IPOI). Limited accounting

information and/or short firm history at the IPO time poses a challenge in deriving the

parameters required to estimate PDF90IPOI. To mitigate this problem and to compensate for the

scarcity of data, this research incorporates relevant historical data from the past IPOs which

were the most comparable to the IPOI when they went public.

Finding comparables introduces another challenge to the risk analysis; selecting

comparables would require defining a ‘similarity’ metric (i.e., ‘distance’ metric) between the

IPOI and the respective candidates. In the multi-dimensional space of operating and financial

characteristics, each firm can be represented by one point. Yet, a simple ‘Euclidean’ distance

cannot be used to quantify the ‘similarity’ due to the heterogeneity of the dimensions and

units. Motivated by Simak’s (2000) [133] work, this thesis develops a methodology which uses

the framework of data envelopment analysis (DEA) to algorithmically and objectively select

comparables.

In working out the details of the method of comparable selection, this research sets forth a

novel model which augments the literature of ‘super-efficiency’ in DEA. This non-oriented

and non-radial model has been developed in the framework of ‘range directional’ models

(Silva Portela et al., 2004 [132]) and is unique by virtue of simultaneously holding all the

following characteristics: (i) it is unit invariant, (ii) it is translation invariant and has been

designed to maintain the integrity of any input or output array which is composed of positive

and negative elements; that is, the model does not decompose the respective mix into its

negative and non-negative components, and hence, precludes any complication imposed by

splitting an input or output vector, (iii) Its objective function is structured to minimize the

distance between the efficient DMU and the ‘reference’ point, in terms of both inputs and

Page 5: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

v

outputs; the reference point is the aggregate of the coordinates which serve as the basis of

super-efficiency measurements. In addition, the objective function is designed such that it

conforms to the conventional interpretation of the super-efficiency score, (iv) as pursued in

this thesis, by introducing further constraints, the model can be enhanced to accommodate

‘non-discretionary’ factors, as well.

The comparable selection step indicated above forms the first phase of this thesis (i.e.,

Phase I). Once the comparables of a given IPOI are identified, the next phase (i.e., Phase II)

employs the Bayesian framework to find the distribution of the IPOI’s stock price at the 90th

day of trading (i.e., PDF90IPOI). It is assumed that during this period, the IPOI’s price follows the

stochastic process of GBM. If the parameters underlying this stochastic process were given,

the methodology could solve for PDF90IPOI, which would be a log-normal distribution.

However, these parameters are not fixed (i.e., they are random). Therefore, this research draws

on their joint distribution to estimate PDF90IPOI. The joint distribution of the parameters for any

given IPOI is approximated by the posterior joint distribution of its closest comparable.

The distance function introduced in Phase I incorporates a parameter (𝛼𝛼) whose value is

objectively estimated by backtesting, using a recursive-cyclic algorithm. At each cycle, a value

is assigned to 𝛼𝛼. Next, Phase I and Phase II are applied to a comprehensive set of IPOIs (276

IPOIs in this thesis). Using the Bayesian framework, the algorithm subsequently solves for the

likelihood of the given 𝛼𝛼, conditioned on the realized prices on day 90 for the participating

IPOIs (276 IPOIs). Using different values of 𝛼𝛼, the above procedure is iterated in various

cycles. The optimum value of 𝛼𝛼 is selected as the one which leads to the maximum likelihood

across all the cycles.

Once the maximum likelihood estimate of 𝛼𝛼 is achieved (i.e., 𝛼𝛼 = 1.5), the

estimated PDF90IPOI’s using this value of 𝛼𝛼 undergo an extensive accuracy scrutiny, by means of

the following two tests: (i) test of uniform cumulative probability values, (ii) test of VaR break

frequency. The outcomes of both tests attest to the validity and predictive utility of the

proposed methodology and offer the yielded PDF90IPOI’s as reliable bases for deriving VaR

inferences.

Page 6: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

vi

ACKNOWLEDGMENTS

Over the long journey of building this thesis, I eagerly awaited a day that I would write

these lines where I can formally extend my most sincere gratitude to the people who helped

me succeed in this endeavour.

First and foremost, I wish to profoundly thank the members of my PhD committee: Prof.

Roy H. Kwon, Prof. Yuri A. Lawryshyn, and my PhD advisor, Prof. Joseph C. Paradi. It

would not have been possible to achieve this thesis without their extraordinary collaboration.

Their insightful suggestions helped me organize and systematically evolve my initial ideas in

the most efficient manner. I am also grateful to them for the time and effort they dedicated to

carefully review this thesis.

I must extend a special note of gratitude to my advisor, Prof. Joseph C. Paradi. Through his

courtesy and consideration, I had all the essential resources at my disposal. But beyond that,

he is an intelligent and forthright life mentor. I am thankful to him for having saved me from

learning many life lessons the intensive and unpleasant way.

I would like to genuinely thank Mr. Eric Kang, the assistant manager at the Finance Lab of

Rotman School of Management. His patience has no limit. He taught me the fundamentals of

many financial databases and never tired of addressing my most fastidious queries.

I wish to record my appreciation to Dr. Xiaopeng Yang with whom I was fortunate to share

office space. Our friendly brainstorming sessions sparked a nucleus idea around which the

framework of ‘Phase I’ gradually evolved. Above all, I will cherish our enjoyable

conversations, restaurant adventures, and his pastry treats.

I must express my gracious thanks to Ms. Leticia Gutierrez and Ms. Gorette Silva for

making my term at the department such a pleasant one. Their warm demeanor shall forever

last as a treasure in my memory.

Page 7: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

vii

Certainly, this acknowledgment would have been incomplete without extending heartfelt

thanks to my family. I wish to offer my deepest appreciation to all my family members,

particularly my parents, Mahin and Jafar. I was privileged to be born as the daughter of such

intellectual and affectionate parents who institutionalized the value of education in me at very

early stages of my life. I am forever in their debt for giving me life and for their subsequent

sacrifices and love.

This research was supported by Ontario Graduate Scholarship, Queen Elizabeth II

Graduate Scholarships in Science & Technology, and grants to the Center for Management of

Technology and Entrepreneurship from the Financial Services Industry.

Page 8: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

viii

TABLE OF CONTENTS

TABLE OF CONTENTS .................................................................................... VIII

LIST OF TABLES .............................................................................................. XIV

LIST OF APPENDIX TABLES ............................................................................. XV

LIST OF FIGURES ............................................................................................ XVI

LIST OF APPENDIX FIGURES........................................................................... XXI

LIST OF ABBREVIATIONS ............................................................................. XXIII

NOMENCLATURE ......................................................................................... XXVI

CHAPTER 1. INTRODUCTION ........................................................................... 1

1.1. Motivation .................................................................................................................1

1.2. Contributions and Challenges ....................................................................................2

1.3. Road Map ..................................................................................................................4

1.3.1. Phase I ................................................................................................................5

1.3.2. Phase II. ..............................................................................................................7

1.3.3. Calibration of the Distance Function .................................................................7

1.4. Organization of the Document ..................................................................................8

CHAPTER 2. LITERATURE REVIEW ON IPO ................................................. 10

2.1. Introduction .............................................................................................................10

2.2. Initial Public Offering Motives................................................................................10

2.3. IPO Timing ..............................................................................................................12

2.4. IPO Mechanisms......................................................................................................13

Page 9: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

ix

2.4.1. Book-Building (Firm Commitment) ................................................................13

2.4.2. Fixed Price Mechanism ....................................................................................13

2.4.3. Auctions ...........................................................................................................14

2.4.4. Prevailing Mechanism: Book-Building ............................................................14

2.5. Post-IPO Trading .....................................................................................................15

2.5.1. Difference between Offer Price and First-Day Opening Price ........................15

2.5.2. Post-IPO Underwriter Services ........................................................................16

2.6. Underpricing ............................................................................................................17

2.7. Post-IPO Long-Term Firm Performance .................................................................18

2.8. IPO Valuation ..........................................................................................................19

2.8.1. Discounted Free Cash Flows ............................................................................20

2.8.2. Comparable Multiples ......................................................................................22

2.8.3. Other Techniques .............................................................................................23

2.8.3.1. Asset-Oriented Techniques ........................................................................23

2.8.3.2. Real Options ...............................................................................................23

CHAPTER 3. LITERATURE REVIEW ON METHODOLOGY: COMPARABLE-

BASED METHODS ................................................................................................. 25

3.1. Introduction .............................................................................................................25

3.2. Comparable-Based Risk Analysis ...........................................................................25

3.2.1. Comparable-Based Method - More Reflective of Market Conditions .............26

3.2.2. Discounting Methods - Theoretically Sound but Practically Rigorous ............28

3.2.3. Discounting Methods - Added Complexity with Uncertain Accuracy Gain ...29

3.2.3.1. Evidence from Non-IPO Cases ..................................................................30

3.2.3.2. Evidence from French IPOs .......................................................................32

Page 10: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

x

3.2.4. Comparable-Based Method: Prevalent in the U.S. ..........................................33

3.3. Comparable Firm Selection .....................................................................................35

3.4. Factors Impacting the IPO Price ..............................................................................47

CHAPTER 4. LITERATURE REVIEW ON METHODOLOGY: DATA

ENVELOPMENT ANALYSIS ................................................................................... 53

4.1. Introduction .............................................................................................................53

4.2. Basic DEA Models ..................................................................................................55

4.2.1. CCR Model ......................................................................................................55

4.2.2. BCC Model ......................................................................................................60

4.2.3. Additive Model ................................................................................................61

4.2.4. Slacks-Based Measure of Efficiency ................................................................62

4.3. Negative Inputs or Outputs ......................................................................................64

4.3.1. Translating Data ...............................................................................................64

4.3.2. Treating Negative Outputs (Inputs) as Positive Inputs (Outputs) ....................67

4.3.3. Semi-Oriented Radial Measure ........................................................................68

4.3.4. Variant of Radial Measure ...............................................................................72

4.3.5. Range Directional Models ................................................................................74

4.4. Non-Discretionary Variables ...................................................................................79

CHAPTER 5. LITERATURE REVIEW ON METHODOLOGY: BAYESIAN

INFERENCE IN RISK ASSESSMENT ....................................................................... 88

5.1. Introduction .............................................................................................................88

5.2. Bayesian versus Frequentist Statistics .....................................................................90

5.3. Basics of Bayesian Inference ...................................................................................93

Page 11: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xi

5.4. Merits of the Bayesian Methodology ......................................................................96

5.5. Merits of the Bayesian Approach in Risk Assessment ..........................................101

CHAPTER 6. METHODOLOGY...................................................................... 108

6.1. Review of Objectives .............................................................................................108

6.2. Phase I: Comparable Selection ..............................................................................108

6.2.1. Pool of Candidates .........................................................................................114

6.2.2. Variable Selection ..........................................................................................115

6.2.3. The DEA Model .............................................................................................121

6.2.3.1. Negative Data ...........................................................................................122

6.2.3.2. Non-Discretionary Factors .......................................................................127

6.2.4. Efficient IPOI Treatment ................................................................................133

6.2.5. Algorithm of Phase I ......................................................................................138

6.3. Phase II: Assessment of Short-Term Risk .............................................................146

6.3.1. Stock Pricing Model .......................................................................................146

6.3.2. Distribution of Stock Price 90 Days after the Issuing Day ............................148

6.3.2.1. Estimating the Joint Distribution of the Parameters .................................151

6.3.2.1.1. Estimating the Posterior Distributions ....................................................................... 152

6.3.2.1.2. Estimating the Prior Distributions .............................................................................. 155

6.4. Calibrating the Distance Equation .........................................................................157

CHAPTER 7. DATA ....................................................................................... 161

7.1. Fundamental Financial Data ..................................................................................161

7.2. Sector Index and GDP ...........................................................................................164

CHAPTER 8. RESULTS ................................................................................. 166

Page 12: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xii

8.1. Results of Phase I: Comparable Selection .............................................................166

8.2. Results of Phase II: Assessment of Short-Term Risk ............................................175

8.2.1. Review of the Methodology of Phase II.........................................................175

8.2.2. Uniform Prior Joint Density ...........................................................................178

8.2.2.1. Estimating RΜ and RΣ ..............................................................................179

8.2.2.1.1. Maximum Likelihood Estimation Method ................................................................. 184

8.2.2.2. Estimating RΞ and RΩ. .............................................................................186

8.2.3. Calibrating the Distance Function ..................................................................191

8.2.4. Scrutiny of Estimated Probability Density Functions ....................................203

8.2.4.1. Test of Uniform Cumulative Probabilities ...............................................203

8.2.4.1.1. Non-Parametric Statistical Tests to Examine the Uniformity Assumption ................ 207

8.2.4.2. Test of VaR Break Frequency ..................................................................211

8.2.4.3. Test of Impact of Comparable Selection and Bayesian Updating on

Estimated PDFs .............................................................................................................212

CHAPTER 9. CONCLUSIONS AND FUTURE WORK ....................................... 219

9.1. In Conclusion: IPO Risk Analysis Unravelled Using the Bayesian Perspective ..219

9.2. Future Research Directions and Application Prospects ........................................225

9.2.1. Extension of Applications ..............................................................................225

9.2.1.1. Risk Measurement of Portfolios Containing IPOs ...................................225

9.2.1.2. Hedging the IPO Investment Risk ............................................................230

9.2.1.3. IPO VaR Decomposition ..........................................................................231

9.2.2. Future Research: Incorporating Information from Other Comparables .........231

Direction 1. Virtual Firm ..........................................................................................232

Direction 2. Compound Posterior Joint Density ......................................................234

Page 13: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xiii

REFERENCES .................................................................................................. 237

APPENDIX A. COMPLEMENTARY DETAILS ON THE CALIBRATION OF THE

DISTANCE FUNCTION......................................................................................... 253

A.1 Introduction ...........................................................................................................253

A.2 Supplementary Discussion on Section 8.2.4: Scrutiny of Estimated PDF90IPOI’s

under other Values of 𝛼𝛼 ........................................................................................................253

A.3 Behaviour of Log-Likelihood Functions of 𝛼𝛼 .......................................................260

Page 14: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xiv

LIST OF TABLES

Table 6-1. Input-Output Data Used by the CCR Model in an Input-Oriented Analysis .........110

Table 7-1. The table presents the number of remaining IPOs under each S&P sector after

completing the process of data mining. The third and fifth columns display the time

periods spanned by the respective IPOs. .........................................................................163

Table 7-2. Each cell displays the time period spanned by the respective sector and index in

COMPUSTAT. ................................................................................................................165

Table 8-1. For each nominal frequency (ῤ) on the leftmost column, the rightmost counterpart

presents the realized frequency which is computed as the quotient of the middle column

over the total number of the IPOIs (i.e., 276). .................................................................213

Table 8-2. This table can be regarded as the counterpart of Table 8-1. The latter reports the

realized VaR break frequencies under the ‘comprehensive’ methodology; whereas, this

table exhibits the outcomes of the test of VaR break frequency for the ‘simplified’

methodology. ...................................................................................................................217

Page 15: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xv

LIST OF APPENDIX TABLES

Table A-1. The table reports the individual K-S p-values computed under different

assumptions of 𝛼𝛼. For a given 𝛼𝛼, the null hypothesis can be stated as ‘the observed set of

276 CPVs has been drawn from a reference population with uniform density between 0

and 1’. Results indicate that in the cases of 0, 0.5, 1.9, and ≥ 3, there is no strong

evidence in favour of the null hypothesis, which suggests that the respective observed

sample set would be considered a rare event if one assumed that the null hypothesis is

valid. It can, therefore, be concluded that the individual sets of PDF90IPOI’s produced using

these 𝛼𝛼 values have failed to fulfil the objectives of the test of uniform CPVs, which

implies that no convincing evidence exists to support the standalone accuracy of each of

these sets. .........................................................................................................................254

Table A-2. Each cell presents the realized VaR break frequency under the corresponding 𝛼𝛼

and nominal frequency of VaR breaks. The exhibited frequencies can be interpreted in a

similar fashion to those displayed in Table 8-1 (see, Section 8.2.4.2). ...........................255

Page 16: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xvi

LIST OF FIGURES

Figure 1-1. This figure illustrates the main contributions of this thesis. .....................................3

Figure 6-1. This figure demonstrates the efficient hyperplanes determined by the constraints

presented in Equations (6.1) to (6.6). The shared feasible region lies above the efficient

hyperplanes. .....................................................................................................................112

Figure 6-2. This figure complements Figure 6-1 by adding the hyperplane of DMU A which is

represented by Equation (6.1). DMU A is an inefficient DMU, and thus, its hyperplane

lies below the hyperplanes of the efficient DMUs C, D, and E. The set of ‘efficient peers’

of DMU A comprises DMUs D and E. Note that the intersection of DMUs D and E is

where the hyperplane A has the closest proximity to the efficient frontier. DEA,

therefore, determines similar final weights for DMUs D, E, and A. ...............................113

Figure 6-3. A Snapshot of the Pool of Candidates for a Given IPOI in the ‘Health Care’ Sector.

.........................................................................................................................................116

Figure 6-4. General Layout of the DEA Model .......................................................................118

Figure 6-5. A Demonstration of Several Potential Improvement Directions for DMU E .......130

Figure 6-6. This flowchart presents the sequence of steps involved in Phase I of the analysis.

The MATLAB programming software is used to implement all the steps. The written

‘functions’ are classified in three different libraries: (i) Data Mining and Data Processing,

(ii) DEA Models, (iii) Analysis. In total, the libraries comprise approximately 7000 lines

of script. ...........................................................................................................................141

Figure 6-7. The outlier detection algorithm of De Sousa and Stošić (2005) [45] is implemented

as demonstrated by this flowchart. ..................................................................................143

Page 17: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xvii

Figure 6-8. This flowchart outlines the process of computing ‘threshold’, indicated in Figure

6-7. ...................................................................................................................................145

Figure 6-9. The figure presents the histogram of the ratio of ‘closing price of the first day to

offer price’ for 327 IPOs in the ‘health care’ sector. The IPOs span the time period of

1990-2012. .......................................................................................................................149

Figure 6-10. Using the data of the ‘health care’ sector (Sector 35 under GICS), this figure

provides an overview of the IPOIs and their ‘pools of candidates’. Each column presents

the IPOs used to carry out Phase I and Phase II for the respective IPOI. The reader is

referred to Section 6.2.1 for further information on the pool of candidates. The first 51

IPOs are referred to as ‘pioneers’. ...................................................................................153

Figure 6-11. This flowchart provides an overview of the process of calibrating the ‘distance

function’ and estimating the optimum value for 𝛼𝛼 (see, Equation (6.37)). .....................160

Figure 8-1. The bar plots visualize the number of comparables associated with the respective

IPOIs whose IDs are presented on the horizontal axis. ...................................................168

Figure 8-2. This figure depicts the efficient frontier for the data presented in Table 1.5 of

Cooper et al. (2007) [41]. Note that the output ‘Inpatients’ has been eliminated which

facilitates visualizing the frontier in 3 dimensions. It is illustrated how the shape of the

frontier changes by the exclusion of the efficient DMU G. The production possibility set,

which is capped by the efficient frontier, spans a smaller space subsequent to the removal

of the efficient DMU G. The segmentation of the frontier changes as well; the number of

efficient hyperplanes decreases in this example. .............................................................170

Figure 8-3. This figure demonstrates the ‘average-union’ ratio for individual IPOs. .............172

Figure 8-4. This figure provides an overview of the changes in the composition of the

comparables for the 52nd IPO. The horizontal axis represents the number of candidates

Page 18: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xviii

(i.e., the IPOs participating in the respective analysis of comparables). The vertical axis

exhibits the union of all the sets of comparables obtained for the 52nd IPO using the

candidate sets of different sizes. Under a given set of candidates, the selected

comparables are colored in red. .......................................................................................174

Figure 8-5. This figure visualizes the process of determining RΜ and RΣ. .............................181

Figure 8-6. The graph visualizes the estimated PDF90IPOI for the 276th IPOI (i.e., 327th IPO).

The dashed line marks the observed price on day 90 (𝑣𝑣90realized) normalized by the

respective inflation-adjusted offer price (𝑣𝑣offer). Its position is an indicative of the

probability that would be associated with the event 𝑣𝑣90realized/𝑣𝑣offer ≈ 1.015 if the

proposed methodology of this thesis is employed and if 𝛼𝛼 is set equal to 1.5. A similar

process is undertaken for each IPOI in order to achieve the value of the yielded PDF90IPOI

at the respective 𝑣𝑣90realized/𝑣𝑣offer. These values are recorded to be subsequently utilized in

determining the likelihood of 𝛼𝛼 = 1.5 following the steps described within the text. ....192

Figure 8-7. This graph exhibits the log-likelihood function of 𝛼𝛼. Each depicted point is

obtained by executing the following two main steps: Step 1 assigns a value to 𝛼𝛼 and

executes phase I and phase II for each of the 276 IPOIs which results in 276

individual PDF90IPOI’s under the postulated 𝛼𝛼. Step 2 computes the logarithm of the

product of 276 elements each representing the value of the PDF90IPOI of a given IPOI at the

respective realized 𝑉𝑉90. The likelihood function hits its maximum at 𝛼𝛼 ≅ 1.5. ..............194

Figure 8-8. The graph visualizes the two PDF90IPOI’s estimated for the 276th (last) IPOI under

the two assumptions of 𝛼𝛼 = 0 and 𝛼𝛼 = 1.5. A narrower PDF is achieved with the

maximum likelihood estimate of 𝛼𝛼 (𝛼𝛼 = 1.5), which is more desirable since it indicates

less variance and more certainty about the value of 𝑉𝑉90. .................................................197

Figure 8-9. Similar to Figure 8-8, this plot visualizes the two estimated PDF90IPOI’s for the 17th

IPOI when 𝛼𝛼 = 0 and 𝛼𝛼 = 1.5. The graph serves as an example of the case where the

Page 19: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xix

maximum likelihood estimate of the parameter (𝛼𝛼 = 1.5) has produced a narrower PDF;

yet, the mode of the PDF is comparatively remote from the realized price, marked by the

dashed line. Such an observation should not be viewed as an element of concern about

the accuracy of the PDF since the concept of maximum likelihood rests the basis of

accuracy comparison upon examining the assigned probability to the realized event, and

not upon the deviation of the realized event from the most likely value (mode) of the

PDF. .................................................................................................................................198

Figure 8-10. This graph exhibits the outcomes of the calibration process for varying sets of

engaged IPOs. As displayed on the horizontal axis, the set producing the first point on the

plot consists of the 52nd IPO, only. The second point on the figure is generated using the

set comprising the 52nd and 53rd IPOs. In a like fashion, the remaining points are obtained

by sequentially increasing the size of the set of participating IPOs by one. For each set,

the maximum likelihood estimate of 𝛼𝛼 is found through repeating the maximum

likelihood calibration process described in the text. As a case in point, consider the label

52-100, on the horizontal axis; it implies that the corresponding calibration process was

carried out on the basis of the PDF90IPOI’s estimated for the 52nd IPO to the 100th IPO, or

equivalently, the 1st IPOI to the 49th IPOI, under the individual 𝛼𝛼 values depicted on the

vertical axis. For this particular set of length 49, the likelihood function reaches its

maximum at the value of 𝛼𝛼 = 1.25; that is, the maximum likelihood estimate of 𝛼𝛼 is

equal to 1.25. This figure demonstrates that despite early fluctuations in the value of the

maximum likelihood estimate of 𝛼𝛼, the trend remains steady at 𝛼𝛼 = 1.5 once the number

of participating IPOs rises beyond a certain threshold (i.e., 148). ...................................202

Figure 8-11. The figure presents the histogram of the CPVs computed for the 276 IPOIs (i.e.,

52nd IPO to 327th IPO), using Equation (8.9). ..................................................................209

Figure 8-12. This figure is produced following the same steps used for Figure 8-11. The only

difference remains the underlying methodology which generates the CPVs. Figure 8-11

depicts the histogram of the CPVs computed using the ‘comprehensive’ methodology;

Page 20: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xx

whereas, here, the graph is formed on the basis of the CPVs resulted from the ‘simplified’

methodology. ...................................................................................................................216

Figure 9-1. The figure illustrates a one-input and one-output DEA example with a variable

returns to scale frontier of efficiency. Firm G represents the firm of interest whose

comparables are encircled in grey in the upper graph. The lower graph depicts the convex

hull of the set of comparables as well as the ‘virtual’ firm which resides on the boundary

of the convex hull and is designated by Firm K. .............................................................233

Page 21: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xxi

LIST OF APPENDIX FIGURES

Figure A-1. The histogram is built using 276 CPVs which are computed on the basis of the

respective set of 276 PDF90IPOI’s, estimated under the assumption of 𝛼𝛼 = 0. ...................257

Figure A-2. The histogram is built using 276 CPVs which are computed on the basis of the

respective set of 276 PDF90IPOI’s, estimated under the assumption of 𝛼𝛼 = 0.5. ................257

Figure A-3. The histogram is built using 276 CPVs which are computed on the basis of the

respective set of 276 PDF90IPOI’s, estimated under the assumption of 𝛼𝛼 = 1.25. ..............258

Figure A-4. The histogram is built using 276 CPVs which are computed on the basis of the

respective set of 276 PDF90IPOI’s, estimated under the assumption of 𝛼𝛼 = 1.9. ................258

Figure A-5. The histogram is built using 276 CPVs which are computed on the basis of the

respective set of 276 PDF90IPOI’s, estimated under the assumption of 𝛼𝛼 ≥ 1.9. Recall from

Section 8.2.3 that starting at the vicinity of 𝛼𝛼 = 3, the log-likelihood function stabilizes

which indicates that the second term (i.e., the 𝛼𝛼 term) has dominated the first term of the

similarity metric in all the 276 IPOI cases. Therefore, the individual sets of PDF90IPOI’s

estimated from this point onward remain unchanged, which consequently translates into

invariant sets of CPVs, and hence, identical CPV histograms. .......................................259

Figure A-6. The graph demonstrates the log-likelihood function of 𝛼𝛼 computed on the basis of

the subset consisting of the 52nd to 141st IPOs. This figure serves as an example of the

‘first type’ of behaviour described on Page 260. Similar behaviours have been recorded

for the two succeeding subsets (i.e., 52 to 142 and 52 to 143). .......................................262

Figure A-7. The graph demonstrates the log-likelihood function of 𝛼𝛼 computed on the basis of

the subset consisting of the 52nd to 144th IPOs. This figure serves as an example of the

Page 22: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xxii

‘second type’ of behaviour described on Page 260. A similar behaviour has been

observed for the set of the 52nd to 145th IPOs. .................................................................262

Figure A-8. The graph demonstrates the log-likelihood function of 𝛼𝛼 computed on the basis of

the subset consisting of the 52nd to 146th IPOs. This figure serves as an example of the

‘third type’ of behaviour described on Page 260. Similar behaviours have been recorded

for the three succeeding subsets (i.e., 52 to 147, 52 to 148, and 52 to 149). ...................263

Figure A-9. The graph exhibits the log-likelihood function of 𝛼𝛼 computed on the basis of the

subset consisting of the 52nd to 150th IPOs. More details regarding this plot are provided

on Page 261. .....................................................................................................................263

Page 23: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xxiii

LIST OF ABBREVIATIONS

BAM Bounded Adjusted Measure of Efficiency

BCC DEA Model of Banker, Charnes, and

Cooper

CCR DEA Model of Charnes, Cooper, and

Rhodes

CDF Cumulative Density Function

CDF-1 Inverse Cumulative Density Function

CFO Chief Financial Officer

CI Confidence Interval

CPI Consumer Price Index

CPV Cumulative Probability Value

CRLB Cramer-Rao Lower Bound

CRS Constant Returns to Scale

CUSIP Committee on Uniform Security

Identification Procedures

DCF Discounted Cash Flow

DDM Dividend Discount Model

DEA Data Envelopment Analysis

DMU Decision Making Unit

Page 24: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xxiv

DT Decision Tree

EBITDA Earnings Before Interest, Tax,

Depreciation, and Amortization

EVA Economic Value Added

EVS Enterprise Value-to-Sales Multiple

GBM Geometric Brownian Motion

GDP Gross Domestic Product

GICS Global Industry Classification Standard

i.i.d independent and identically distributed

IPO Initial Public Offering

IPOI IPO Firm of Interest

K-S Kolmogorov-Smirnov (Test)

LHS Left-Hand Side (of an Equation)

MLE Maximum Likelihood Estimation

MSBM Modified Slacks-Based Measure

ND Non-Discretionary

NPV Net Present Value

PB Price-to-Book Multiple

PDF Probability Density Function

𝐏𝐏𝐏𝐏𝐏𝐏𝟗𝟗𝟗𝟗𝐈𝐈𝐏𝐏𝐈𝐈𝐈𝐈 PDF of Stock Price on Day 90 Associated

with the IPO Firm under Scrutiny (IPOI)

Page 25: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xxv

𝐏𝐏𝐏𝐏𝐏𝐏𝟗𝟗𝟗𝟗𝐏𝐏𝐏𝐏𝐏𝐏𝐏𝐏𝐏𝐏𝐏𝐏𝐏𝐏𝐏𝐏𝐏𝐏 PDF of Portfolio Value 90 Days After Its

Formation

P/E Price-to-Earnings Multiple

PPE Property, Plant and Equipment

R&D Research and Development

RDM Range Directional Model of Silva Portela et

al. (2004) [132]

RE Retained Earnings

RHS Right-Hand Side (of an Equation)

RIM Residual Income Model

ROA Real Option Analysis

ROE Return on Equity

SBM Slacks-Based Measure of Efficiency

SDC Securities Data Company PlatinumTM

SDC Stochastic Differential Equation

SEC Securities and Exchange Commission

SN Standard Normal PDF

SORM Semi-Oriented Radial Measure

VaR Value at Risk

VRM Variant of Radial Measure

VRS Variable Returns to Scale

Page 26: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xxvi

WRDS Wharton Research Data Services

NOMENCLATURE

𝑫𝑫𝒄𝒄𝒄𝒄 Distance between the IPOI under Scrutiny, Identified by

the ID of 𝑏𝑏, and the Corresponding Comparable Firm 𝑐𝑐

𝒆𝒆 Vector of 1’s with a Dimension Matching Its Multiplier

𝒇𝒇𝑽𝑽𝟗𝟗𝟗𝟗(𝒗𝒗𝟗𝟗𝟗𝟗) Unconditional PDF of 𝑉𝑉90, Symbolizing PDF90IPOI

𝒇𝒇𝑽𝑽𝟗𝟗𝟗𝟗𝐏𝐏 �𝒗𝒗𝟗𝟗𝟗𝟗𝐏𝐏 � Unconditional PDF of 𝑉𝑉90P , Symbolizing PDF90Portfolio

𝒇𝒇𝚳𝚳,𝚺𝚺,𝚵𝚵,𝛀𝛀(𝝁𝝁,𝝈𝝈, 𝝃𝝃,𝝎𝝎) Joint PDF of the Bayesian Random Variables Μ, Σ, Ξ,

and Ω

𝒇𝒇𝚬𝚬𝐍𝐍(𝝐𝝐;𝟗𝟗,𝟏𝟏) Standard Normal PDF: 𝜖𝜖 represents a draw from the

standard normal random variable Ε.

𝒇𝒇𝑼𝑼𝐏𝐏𝐏𝐏𝐥𝐥𝐍𝐍(𝒖𝒖; 𝝃𝝃,𝝎𝝎)

Log-Normal PDF of the Random Variable 𝑈𝑈, with the

Corresponding Normal PDF N(𝜉𝜉,𝜔𝜔)

𝒈𝒈𝒙𝒙𝒊𝒊𝒊𝒊 Range of Possible Improvement for the 𝑖𝑖th Input of the 𝑗𝑗th

DMU, in the Range Directional Model of Silva Portela et

al. (2004) [132]

𝒈𝒈𝒚𝒚𝒓𝒓𝒊𝒊 Range of Possible Improvement for the 𝑟𝑟th Output of the

𝑗𝑗th DMU, in the Range Directional Model of Silva

Portela et al. (2004) [132]

𝓛𝓛𝑽𝑽𝟗𝟗𝟗𝟗(𝒗𝒗𝟗𝟗𝟗𝟗|𝝁𝝁,𝝈𝝈,𝒗𝒗𝟏𝟏) Conditional PDF of 𝑉𝑉90, Given the Event of (Μ ≈ 𝜇𝜇,

Σ ≈ 𝜎𝜎, 𝑉𝑉1 ≈ 𝑣𝑣1)

ῤ Nominal Frequency of VaR Breaks

Page 27: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xxvii

𝑹𝑹𝒙𝒙𝒊𝒊𝒊𝒊

Equivalent of 𝑔𝑔𝑥𝑥𝑖𝑖𝑖𝑖, Employed in the Modified Slacks-

Based Measure of Sharp et al. (2007) [131]

𝑹𝑹𝒚𝒚𝒓𝒓𝒊𝒊 Equivalent of 𝑔𝑔𝑦𝑦𝑟𝑟𝑟𝑟, Employed in the Modified Slacks-

Based Measure of Sharp et al. (2007) [131]

𝐑𝐑𝚳𝚳

An Interval Reflecting the Prior Evidence on the Range

of Typical Values for Short-Term Expected Daily Rate

of Return

𝐑𝐑𝚺𝚺 An Interval Reflecting the Prior Evidence on the Range

of Typical Values for Short-Term Daily Volatility

𝐑𝐑𝚵𝚵 An Interval Reflecting the Prior Evidence on the Ξ Range

𝐑𝐑𝛀𝛀 An Interval Reflecting the Prior Evidence on the Ω

Range

𝒔𝒔− Input Slack Vector of a Given DMU

𝒔𝒔+ Output Slack Vector of a Given DMU

𝒕𝒕 Time

𝑼𝑼

A Random Variable Denoting the Ratio of Closing Price

of the First Trading Day (𝑉𝑉1) to Offer Price (𝑣𝑣offer), for a

Given IPO

𝒖𝒖 A Realization of the Random Variable 𝑈𝑈

𝑽𝑽𝒕𝒕 A Random Variable Denoting the Stock Price at Time 𝑡𝑡

𝒗𝒗𝒕𝒕 A ‘Realized’ Stock Price at Time 𝑡𝑡

𝑽𝑽𝟏𝟏 A Random Variable Denoting the Closing Price of the

Page 28: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xxviii

First Trading Day, for a Given IPO

𝒗𝒗𝟏𝟏 A Realization of the Random Variable 𝑉𝑉1

𝑽𝑽𝟗𝟗𝟗𝟗 A Given IPOI’s Price at the Time Point of 90 Trading

Days after the Issuing Day

𝒗𝒗𝟗𝟗𝟗𝟗 A Realization of the Random Variable 𝑉𝑉90

𝑽𝑽𝟗𝟗𝟗𝟗𝐏𝐏

Price of a Portfolio Containing the Stocks of a Given

IPO, at the Time Point of 90 Trading Days after the

Issuing Day of the IPO

𝒗𝒗𝟗𝟗𝟗𝟗𝐏𝐏 A Realization of the Random Variable 𝑉𝑉90P

𝒗𝒗𝐏𝐏𝐏𝐏𝐏𝐏𝐨𝐨𝐏𝐏 Offer Price of a Given IPO

𝒘𝒘 Weight of an Asset in a Given Portfolio

𝒙𝒙 Input Vector of a Given DMU

𝑿𝑿 Matrix of DEA Inputs

𝒚𝒚 Output Vector of a Given DMU

𝒀𝒀 Matrix of DEA Outputs

𝒁𝒁 A Variable Following a Wiener Process

Greek Symbols

𝜶𝜶 Coefficient of the Second Term (i.e., the 𝜓𝜓 Term) in the

Distance Function (Equation (6.37))

𝜷𝜷 Radial Measure of Inefficiency in the VRM Model

Page 29: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xxix

𝜹𝜹 𝜆𝜆 Difference Indicator

𝚬𝚬 A Random Variable with a Standard Normal Distribution

𝝐𝝐 A Realization of the Random Variable Ε

𝜼𝜼 Number of Comparables of a Given IPOI

𝜽𝜽 Radial Efficiency Score of BCC or CCR

𝝀𝝀 Non-Negative Weight of a DMU in the Envelopment

Form of DEA

𝚳𝚳 A Bayesian Random Variable Denoting the Expected

Daily Rate of Return of a Given Stock

𝝁𝝁

A ‘Realized’ Expected Daily Rate of Return: It is a

constant representing a draw from the random variable

Μ.

𝝁𝝁�𝑳𝑳 Lower Bound of the Confidence Interval for 𝜇𝜇

𝝁𝝁�𝑼𝑼 Upper Bound of the Confidence Interval for 𝜇𝜇

𝚵𝚵

A Bayesian Random Variable Representing the Mean of

the Normal PDF Associated with the Log-Normal PDF

of 𝑈𝑈

𝝃𝝃 A Realization of the Random Variable Ξ

𝝃𝝃�𝑳𝑳 Lower Boundary of the Confidence Interval of 𝜉𝜉

𝝃𝝃�𝑼𝑼 Upper Boundary of the Confidence Interval of 𝜉𝜉

𝝆𝝆 A Slacks-Based Measure of Efficiency

𝚺𝚺 A Bayesian Random Variable Denoting the Daily

Page 30: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xxx

Volatility of a Given Stock

𝝈𝝈 A ‘Realized’ Daily Volatility: It is a constant

representing a draw from the random variable Σ.

𝝈𝝈�𝑳𝑳 Lower Boundary of the Confidence Interval of 𝜎𝜎

𝝈𝝈�𝑼𝑼 Upper Boundary of the Confidence Interval of 𝜎𝜎

𝝍𝝍 Efficiency Score of the DEA Optimization Problem

Developed in this Thesis

𝛀𝛀

A Bayesian Random Variable Representing the Standard

Deviation of the Normal PDF Associated with the Log-

Normal PDF of 𝑈𝑈

𝝎𝝎 A Realization of the Random Variable Ω

𝝎𝝎�𝑳𝑳 Lower Bound of the Confidence Interval of 𝜔𝜔

𝝎𝝎�𝑼𝑼 Upper Bound of the Confidence Interval of 𝜔𝜔

Functions and Operators

∈ Denotes Set Membership

● ∈ █ : ● is an element or a member of the set █.

|●| Absolute Value or Modulus of ●

Underbar Denoting the Vector Characteristic of

|||| ℓ2-Norm or Euclidean Norm of the Vector

⊙ Element-Wise Multiplication of Two Vectors

𝐝𝐝● Infinitesimal Change in the Variable ●

Page 31: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xxxi

𝚫𝚫● Non-Infinitesimal Change in the Variable ●

𝐈𝐈() An Indicator Function: It is assigned the value of 1 when

its condition, , is true and equals 0 if its argument

returns false.

ln(●) Natural Logarithm of ●

log(●) Logarithm of ● to Base 10

𝐂𝐂𝐏𝐏𝐏𝐏𝐒𝐒𝐍𝐍,Ƿ/𝟐𝟐−𝟏𝟏

Absolute Value of the Inverse Cumulative Density

Function (CDF−1) of the Standard Normal Distribution

(SN), at the Cumulative Probability of Ƿ/2

𝐂𝐂𝐏𝐏𝐏𝐏𝝌𝝌𝒏𝒏−𝟏𝟏𝟐𝟐 ,Ƿ/𝟐𝟐 −𝟏𝟏 Lower Ƿ/2 × 100 % Quantile of the Chi-Square

Distribution with the Degree of Freedom of (𝑛𝑛 − 1)

𝐂𝐂𝐏𝐏𝐏𝐏𝝌𝝌𝒏𝒏−𝟏𝟏𝟐𝟐 ,(𝟏𝟏−Ƿ/𝟐𝟐) −𝟏𝟏 Upper Ƿ/2 × 100 % Quantile of the Chi-Square

Distribution with the Degree of Freedom of (𝑛𝑛 − 1)

𝐍𝐍(,2) Normal Probability Density Function with Mean and

Variance 2

𝐏𝐏(❷|❶) Conditional Probability Statement: Probability of Event

❷ Given the Occurrence of Event ❶

𝐏𝐏�❷� Unconditional Probability of Event ❷

Notes on Notation

1. Random variables are denoted by uppercase letters.

2. A realization of a random variable is denoted by the corresponding

lowercase letter.

Page 32: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

xxxii

3. For an arbitrary random variable Τ, P(Τ ≈ 𝜏𝜏) can, more precisely, be

expressed as P(𝜏𝜏 < Τ ≤ 𝜏𝜏 + d𝜏𝜏).

Page 33: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

1

Chapter 1. INTRODUCTION

1.1. Motivation

Every year all over the world, Initial Public Offerings (IPOs) provide opportunity for many

significant private companies and relatively young firms to raise the capital required for future

growth, market expansion, and/or repayment of debt. Investing in IPOs is inherently risky. It

is, however, the ambiguity, not the quantity, of this risk that often hinders risk-seeking

investors from investing in IPOs. Hence, properly quantifying the risk associated with

investment in an IPO can attract many new investors to this market and let them contribute to

the expansion of smaller/younger firms.

A methodology for quantifying the risk of IPOs can benefit other parties engaged in an IPO

event, as well. In general, the ‘game’ of IPO involves two more ‘players’ besides investors:

the IPO company itself and the underwriter. Although not as distinctly recognized as in the

case of investors versus the IPO company, the incentives of the underwriter and the IPO

company are not perfectly aligned, either. Both parties seek to maximize the IPO proceeds;

yet, they face different sets of constraints. Under the ‘book-building’ (‘firm commitment’)

mechanism (see, Section 2.4), the IPO company would aim at the maximum attainable gain

with no litigation risk; whereas, the underwriter endeavours to maximize the proceeds within

the prescribed bounds of commitment to sell the entire stock inventory as well as enhancing

the expected value of its future share of the IPO market through maintaining its reputation

among investors and catering to potential future clients; that is, presently private firms which

may decide to go public. Evidently, all such constraints necessitate hedging against litigation

risk, as well. Thus, while the IPO company would intend to achieve zero ‘money left on the

table’, the underwriter might ‘underprice’ the issue, to a certain degree, to cause

oversubscription which would guarantee the inventory exhaustion (see, Sections 2.4 and 2.6).

The ‘intentional’ underpricing also serves as a strategy to induce truthful information

disclosure by the investors which would reduce the risk of a ‘broke’ issue or significant first-

Page 34: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 1. Introduction 1.2. Contributions and Challenges

2

day underpricing; the former scenario (i.e., broke issue) would tarnish the credibility of the

underwriter’s due diligence among investors, and the latter (i.e., substantial first-day

underpricing) would impair the underwriter’s reputation among potential future clients.

The IPO incentive lattice is undoubtedly more complex and goes beyond the simple

situation pictured above. Chapter 2 is devoted to further clarifying the aforementioned jargon

terms and presents the nuts and bolts of the process of going public. Within this game of

conflicting and aligning interests, a methodology that quantifies the short-term IPO payoff

would provide extra pieces of information, mitigating the risk of extreme short-term

overpricing or underpricing, and would lead to a new equilibrium state where the incentives of

all the players are commonly skewed towards minimizing the risk.

1.2. Contributions and Challenges

Figure 1-1 summarizes the main contributions of this thesis. As depicted, the main

objective is to quantify the short-term risk of IPOs. A comprehensive risk measure for an IPO

ought to be developed based on the distributions of the IPO’s future prices, estimated utilizing

the following two components of information: (i) firm-specific financial details, (ii) market

information. In conformity with the mathematical principles of market risk analysis employed

in the case of public companies, in this thesis, the short-term risk of an IPO of interest (IPOI)

is defined as a value-at-risk (VaR), to be ‘inferred’ from the corresponding ‘predictive’

unconditional probability density function (PDF) which models the uncertainty associated

with the stock price at a short-term post-IPO horizon (here, on the 90th day subsequent to the

issue day). Since VaR is, in essence, just a quantile, the ultimate objective pursued by this

research is, therefore, to estimate the underlying predictive PDF of the stock price of the IPOI

at the time point of 90 trading days after the issue day (PDF90IPOI). This problem, however, is a

fundamentally difficult one due to lack of market information and limited accounting history

at the IPO time, rendering estimation of the distribution of the ‘short-term’ payoff restricted.

To mitigate this problem, any proposed methodology needs to compensate for the paucity

of financial statements and absence of market history. To this end, this research relies on the

Page 35: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 1. Introduction 1.2. Contributions and Challenges

3

accounting information and market history of preceding IPOs which were ‘comparable’ to the

IPOI at their issuing times. Following Simak (2000) [133], this thesis develops a Data

Envelopment Analysis (DEA)-based high-dimensional similarity metric. This metric has the

capacity to objectively and algorithmically choose the ‘comparable’ IPOs preceding the IPOI.

As indicated, the comparables are identified using a DEA-based algorithm. One step

underlying this algorithm involves selecting comparables for an ‘efficient’ IPOI. The search

for a DEA model which can be utilized toward this end underscored the potential for a

noteworthy contribution to the literature of ‘super-efficiency’ models in DEA. This study

sought a model which would simultaneously fulfil all the characteristics listed below. Yet, to

Figure 1-1. This figure illustrates the main contributions of this thesis.

Page 36: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 1. Introduction 1.3. Road Map

4

the best of the author’s knowledge, no DEA study offers such a model. Hence, drawing on the

principles of ‘range directional’ models (Silva Portela et al., 2004 [132]), this thesis develops a

novel model which accomplishes all the following objectives (see, Section 6.2.4):

First. It provides a non-radial measure of super-efficiency for a ‘strongly’ efficient DMU.

Second. It is unit-invariant. Third. It is translation-invariant; in achieving this property, it

is crucial that the proposed model preserves the integrity of any input or output vector

which comprises a mix of positive and negative elements. Such a model would preclude

any complication imposed by decomposing a vector into its negative and non-negative

components. Moreover, only would the efficiency score yielded by such a model comply

with the role considered for it, in this thesis; that is, the efficiency score of a DMU acts as a

proxy for its distance from the ‘efficiency frontier’. Fourth. The objective function

minimizes the distance between the efficient DMU (IPOI) and its ‘reference point’, in terms

of both inputs and outputs. Furthermore, the objective function is structured such that the

achieved super-efficiency score can be interpreted in a similar fashion to the ones produced

by conventional super-efficiency models. Fifth. As pursued in this study, by means of

specifying further constraints, the model can be enhanced to cope with ‘non-discretionary’

factors, as well.

Once the comparables are identified, they are utilized in a recursive procedure which is

developed to estimate the distribution of the respective IPOI’s price at day 90 (i.e., PDF90IPOI).

This distribution would then serve as a basis for quantifying risk using a VaR measure.

1.3. Road Map

The process of estimating PDF90IPOI comprises two phases:

Phase I. In this phase, the closest comparable to a given IPOI is identified using the

aforementioned DEA-based similarity metric.

Page 37: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 1. Introduction 1.3. Road Map

5

Phase II. This phase offers a Bayesian framework where the knowledge supplied by Phase

I is formulated and ‘refined’ to be utilized in estimating PDF90IPOI, and subsequently,

quantifying the short-term market risk of the given IPOI.

Sections 1.3.1 and 1.3.2 briefly describe these two phases. Section 1.3.3 provides a short

overview of the steps involved in calibrating the distance function (or conversely, similarity

metric), developed in Phase I. In the final stage of the risk quantification process, the proposed

methodology is examined (backtested) to ensure that the outcomes are reliable and the

promised objectives are achieved. The following two tests are employed to this end:

(i) Test of uniform cumulative probability values (see, Section 8.2.4.1)

(ii) Test of VaR break frequency (see, Section 8.2.4.2)

1.3.1. Phase I

Phase I is focused on identifying the comparable IPOs to a given IPOI. In this research, the

‘pool of candidates’ for comparables consists of preceding IPOs in the same sector as the

IPOI. Since a given IPOI succeeds its comparables in time, one must account for the time

disparity when selecting the comparables. Comparison criteria should, therefore, be adjusted

for inflation. Furthermore, GDP and the sector-specific index can be incorporated to control

for economic productivity and the performance of the respective sector since these

environmental variables can accelerate or impede the course of actions undertaken by a firm

and, consequently, its success. Once adjusted for the time effect and the impact of

macroeconomic and sector factors, a public firm which most resembled the IPOI at its IPO

stage would be selected as the most similar comparable.

By confining the search space to past IPOs, the proposed methodology accounts for many

IPO-specific events and characteristics, leveling grounds of comparison. As a case in point,

consider issuers’ incentive to manipulate pre-IPO earnings using income-increasing

accounting choices. By limiting the search space to former IPOs, the methodology controls

for the impact of earnings management. In addition, growth prospects would be matched more

accurately; an IPOI and a public firm, which share similar financial characteristics, may not

Page 38: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 1. Introduction 1.3. Road Map

6

correspond in terms of growth potential. It is possible that the high growth regime has elapsed

for the public firm and it has reached a steady growth; whereas, the IPOI may be on the verge

of its high growth period.

Finding comparable firms could be a challenging task since it requires defining a similarity

metric (or conversely, a distance metric) between the IPOI and any preceding IPO. Any firm

can be pictured as a point in the multi-dimensional space of operating characteristics and

financial performance. In a heterogeneous space, where the dimensions are of different units

and quantities, a simple ‘Euclidean’ distance metric cannot represent the similarity between

the firms in the real world.

A DEA-based methodology is developed to algorithmically select the comparables. The

approach for selecting comparables was originally motivated by the work of Simak (2000)

[133]. The algorithm of Simak (2000) [133] is, however, extended and evolved into a new and

more comprehensive and analytically rigorous framework which not only addresses the

limitations of Simak’s (2000) [133] work but also provides enhanced features and functionality

and ensures a more reliable performance. Sections 3.3 and 6.2 present, in detail, how this

thesis develops a new DEA-based methodology for selecting comparables. The main

shortcomings of Simak’s (2000) work can broadly be listed as follows. The reader is referred

to Section 3.3 for a detailed discussion on the objective and limitations of Simak’s (2000) [133]

work.

(i) Simak (2000) [133] utilizes the basic input-oriented, radial model of CCR (CRS) or

BCC (VRS). In this thesis, this model is replaced with new non-oriented DEA

models which can cope with negative and ‘non-discretionary’ inputs and outputs.

(ii) Simak’s (2000) [133] methodology is incapable of identifying comparables for an

‘efficient’ firm of interest. This research develops an additional model to eliminate

this shortcoming.

(iii) Simak’s (2000) [133] approach is subjective and requires analyst input. The algorithm

suggested in this thesis is fully objective. To accomplish this goal, an additional

term, accompanied by a coefficient (parameter), is incorporated into Simak’s (2000)

Page 39: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 1. Introduction 1.3. Road Map

7

similarity metric. The new similarity function is then calibrated using an

optimization program which is based on the Bayesian framework. The calibration

process does not require any human intervention.

1.3.2. Phase II.

In Phase II, the distribution of the IPOI’s stock price at the time point of 90 days after the

issuing day (PDF90IPOI) is estimated. The steps involved in the process are detailed in Section

6.3. Concisely, the joint probability density function of the following four parameters is

required in order to solve for PDF90IPOI: expected rate of return of the stock price (𝜇𝜇), volatility

of the stock price (𝜎𝜎), mean (𝜉𝜉) and variance (𝜔𝜔) of the normal distribution associated with the

log-normal distribution describing the first closing price normalized by the respective offer

price. This joint density function is represented by 𝑓𝑓Μ,Σ,Ξ,Ω(𝜇𝜇,𝜎𝜎, 𝜉𝜉,ω) and cannot directly be

calculated for a given IPOI due to lack of market history. This research, therefore,

approximates it by the posterior joint distribution of the corresponding closest comparable.

Phase I supplies the ID of the closest comparable, and Phase II adopts the Bayesian framework

to construct the posterior distribution of the closest comparable through updating its prior

distribution, using the respective realized 90-day trajectory of stock prices.

Once the prior joint distribution of the four parameters (𝑓𝑓Μ,Σ,Ξ,Ω(𝜇𝜇,𝜎𝜎, 𝜉𝜉,ω)) is estimated

for each IPOI, the methodology can then solve for the distribution of the respective IPOI’s

stock price at day 90 (PDF90IPOI). This distribution can subsequently be used to quantify risk.

1.3.3. Calibration of the Distance Function

In addition to the aforementioned parameters (i.e., 𝜇𝜇, 𝜎𝜎, 𝜉𝜉, and ω), which can indeed be

considered as the parameters of the stock price model (see, Section 6.3.1), there exists one

more parameter that must be estimated. Denoted by 𝛼𝛼, it resides in the distance function

developed in Phase I and represents a coefficient (see, Sections 6.4 and 8.2.3). In order to

estimate this coefficient (i.e., 𝛼𝛼), the distance function is calibrated using a recursive-cyclic

algorithm. At each cycle, a value is assigned to 𝛼𝛼. Subsequently, Phase I and Phase II are

Page 40: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 1. Introduction 1.4. Organization of the Document

8

repeated for various IPOIs (276 IPOs in this thesis) given the assigned 𝛼𝛼. Using the Bayesian

framework, the method then computes the likelihood of the assumed 𝛼𝛼 conditioned on the

realized prices of day 90 for the participating IPOIs (276 IPOs). The above cycle is iterated

using different values of 𝛼𝛼. The optimum 𝛼𝛼 is selected as the one whose likelihood is

maximum across all the cycles.

1.4. Organization of the Document

The remainder of this document is organized as follows.

Chapter 2.

It provides background information on IPOs which is required to understand the challenges

associated with the risk assessment of IPOs.

Chapter 3.

It is composed of three main sections: The literature review presented in Section 3.2 offers

strong grounds for using a market-based comparables approach in quantifying the short-

term risk of IPOs. Section 3.3 reviews the existing methods for selecting comparable firms.

Section 3.4 surveys the literature in order to create an inventory of the financial factors

identified to be influential on the value and short-term performance of IPOs. These factors

will subsequently build the inputs and outputs of the DEA models.

Chapter 4.

The goal of Chapter 4 is to provide a comprehensive literature review on DEA as pertinent

to the methodology suggested in this thesis. Section 4.2 introduces basic DEA models, such

as CCR (CRS), BCC (VRS) and SBM. Subsequently, Sections 4.3 and 4.4 thoroughly

survey the advances in the fields of negative data and non-discretionary factors in DEA.

Page 41: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 1. Introduction 1.4. Organization of the Document

9

Chapter 5.

The focal intent of Chapter 5 is to review the merits of the Bayesian approach in risk

assessments. The Bayesian framework hosts a vast array of topics which extend beyond the

capacity of this thesis. The choice of the topics thus follows the specific objectives of this

thesis. Section 5.1 provides an introduction on the concept of VaR and enlightens why the

Bayesian framework would offer a comprehensive setting to tackle VaR problems. Sections

5.2 to 5.4 acquaint the reader with practical nuts and bolts as well as perils and promises of

the Bayesian and frequentist schools of thought. Section 5.5 is devoted to delineating the

promises of the Bayesian wisdom in risk management.

Chapter 6.

The proposed methods are presented in Chapter 6. Sections 6.2 and 6.3 provide a detailed

discussion on Phase I and Phase II. The method of calibrating the distance function is

explained in Section 6.4.

Chapter 7.

Chapter 7 introduces the data used for the study and provides information on the sources of

data. It describes the process of collecting data and matching 9-digit CUSIPs across

different resources. Chapter 7 also gives an overview of the steps involved in data cleaning,

combining, re-classifying, and adjusting.

Chapter 8.

The results of the proposed methodology (Chapter 6) are presented and examined (i.e.,

backtested) in Chapter 8.

Chapter 9.

This chapter summarizes the key conclusions and highlights potential future research

avenues.

Page 42: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

10

Chapter 2. LITERATURE REVIEW ON IPO

2.1. Introduction

This chapter provides a brief overview of the process of going public and the details

surrounding initial public offerings (IPOs). It is explained why companies decide to go public

and when the best time for an initial public offering would be (Sections 2.2 and 2.3). The

shares of an issuing company can be priced and allocated under different mechanisms. These

mechanisms are introduced in Section 2.4. Once the shares are priced and allocated in the

primary market, secondary market trading commences. Sections 2.5 to 2.7 focus on the details

of post-IPO trading, the ‘underpricing’ phenomenon, and the issuing company’s post-IPO

long-term performance. Section 2.8 concludes this chapter by describing the conventional IPO

valuation methods.

2.2. Initial Public Offering Motives

Initial Public Offering (IPO) is an event where a private company (an issuer) sells a portion

of its authorized shares to the public investors for the first time. IPO is one of the most

fundamental decisions a corporation may face in its life. Through the IPO decision, the issuer

commits to major responsibilities including shared ownership, increased monitoring, in

addition to direct one-time filing and offering costs. Thus, there must be strong motives that a

company decides to offer a fraction of its ownership to public.

Pagano et al. (1998) [104] studied the causes of going public by investigating the pre-IPO

characteristics of a private firm and the post-IPO performance of the public firm. They

examined a large sample of Italian private companies in the 1982-1992 period. The outcome

of their research indicates the following general motivations for initial public offerings:

(i) The main objective of going public is to attain a high market-to-book ratio. High

market-to-book ratios may be the indicative of high growth opportunities in the

Page 43: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 2. Literature Review on IPO 2.2. Initial Public Offering Motives

11

respective sector which demand high investments. High market-to-book ratios can

also originate from temporary mispricing in hot markets. Pagano et al. (1998) [104]

observed reductions in investments during the post-IPO period. This outcome

supports the second perception and implies that issuers may spot hot markets and

decide to go public.

(ii) The cost of debt decreases for independent IPOs (i.e., not carve-outs). It is not

certain whether the cheaper debt occurs due to the greater transparency of a public

firm’s business activities or due to increased bargaining power resulting from

increased number of alternative lending institutions.

(iii) According to the empirical analysis of Pagano et al. (1998) [104], portfolio

diversification does not seem to be a significant IPO inspiration. The high turnovers

of controlling shareholders in some firms after the three years from the IPO time,

however, support the cashing-out scenario in IPOs.

Unique to the Italian market, Pagano et al. (1998) [104] conclude that in general, Italian IPO

firms are greater in size and the IPO market in Italy is not a financing engine of many young

firms. On the contrary, the U.S-based research (e.g., Mikkelson et al., 1997 [99]) reveals that the

IPO market in the U.S. is a means for many young firms to raise capital for further growth.

Another significant incentive to go public is post-IPO acquisition plans. Brau and Fawcett

(2006) [31] investigate the reasons for going public by surveying 336 chief financial officers

(CFOs) whose companies had successfully completed or attempted but withdrawn IPOs during

the time period of January 2000 to December 2002, in the U.S. The outcome indicates that

acquisition is the primary objective. The IPO proceeding can be used to accomplish

acquisitions. Alternatively, the IPO firm can be a target since the post-IPO transparent

business activities and market value facilitate the acquisition.

From an overall perspective, the IPO motives can be categorized as endogenous and

exogenous factors. Examples of endogenous factors can be financial motives (e.g., cheaper

debt) and governance restructuring (e.g., venture capitalists’ exit). ‘Hot’ issues markets and

Page 44: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 2. Literature Review on IPO 2.3. IPO Timing

12

strategic motives (e.g., competitors going public) can exogenously trigger a private firm to go

public.

2.3. IPO Timing

IPO timing can be determined based on the development stage of the firm. After a period of

high growth, a firm may decide to finance the next stages of its projects by offering shares to

public. A firm may also gauge market conditions to determine when to go public. In a study

about the impact of the underwriter reputation on an IPO’s success, Lee (2011) [89] concludes

that an IPO’s success can partially be attributed to the correct timing of a reputable

underwriter.

During ‘hot’ issues markets, many firms decide to go public to exploit the favourable

market conditions and achieve high market-to-book ratios (Pagano et al., 1998 [104]). Hot issues

often occur when the increases from the offer prices to the aftermarket prices are above the

expected market risk premium for a specific cluster of stocks. The above-average short-term

returns prompt investors to funnel more wealth to the respective industry which in turn

motivates more issues. The cycle of self-fulfilling prophecy continues until the market

optimism disappears. New technologies becoming obsolete or reaching their growth ceilings,

unfavourable information spillovers about the newly public firms’ activities and/or unfulfilled

pre-IPO promises can be among the factors leading to ‘cold’ markets.

Hot issues create a temporary window of opportunity for the private firms within the

respective market to have successful offerings. Hence, it is important to identify key hot

market drivers. The hot issues markets have been examined and documented by many

researchers (e.g., Ibbotson, 1975 [76]; Ibbotson and Jaffe, 1975 [77]; Ritter, 1984 [118]; Draho,

2004 [48]; Ljungqvist et al., 2006 [92]). Different scenarios can lead to hot issues markets. These

scenarios span both the investor (demand) and issuer (supply) sides. Issuers learn from the

public offering of their competitors. An IPO completed successfully by a competitor motivates

other firms in the same industry to go public. On the investor side, a new technology and/or

industry can cause investors’ demand to surge. Furthermore, investors may approximate the

Page 45: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 2. Literature Review on IPO 2.4. IPO Mechanisms

13

expected performance of a new issue by the most recent successful IPOs. Such irrational

extrapolating may fuel hot markets as well.

2.4. IPO Mechanisms

The IPO mechanism outlines how the underwriter prepares a firm for its IPO and sells

shares to the investors. Draho (2004) [48] categorizes IPO mechanisms into three classes: firm

commitment or book-building, fixed-price offering, and auctions. Prior to delving into more

details, it is worthwhile to note that the primary focus of the IPO discussion in this document

is on the U.S. IPOs unless otherwise stated.

2.4.1. Book-Building (Firm Commitment)

Under the book-building mechanism, the underwriter is responsible for all the preliminary

legal steps for the public offering, share pricing and share allocation. The notion of ‘book-

building’ refers to the underwriter soliciting the investor demand through road trips and

marketing activities and building an order book accordingly. The terminology of ‘firm

commitment’ denotes the underwriter’s commitment to sell the inventory at the price

determined in the final prospectus (Draho, 2004 [48]).

2.4.2. Fixed Price Mechanism

The price is determined in advance if the offering is performed using the fixed price

mechanism. The fixed price offering does not incorporate the firm commitment and book-

building components of the book-building mechanism. The underwriter and issuer set the final

offer price without directly gauging the investor demand in advance. The shares are then

allocated based on the orders submitted. Under this mechanism, the underwriter does not

guarantee the sale of the entire inventory (Draho, 2004 [48]).

Page 46: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 2. Literature Review on IPO 2.4. IPO Mechanisms

14

2.4.3. Auctions

The most commonly used auction mechanisms are uniform price auction (or Dutch auction)

and discriminatory price auction (Draho, 2004 [48]). In the auction mechanism, the underwriter

is not directly involved in setting the offer price. The investors submit their orders and limit

prices. After collecting all the orders, based on the auction type, the offer price is determined.

If the auction is of the uniform type, the offer price is determined as the intersection of the

supply curve (which is fixed) and the demand curve. The shares are then allocated, at the offer

price, to the investors whose bids exceed the offer price. The orders of the investors with bids

at the offer price may partially be filled, based on the supply availability. In the discriminatory

price auction, the shares are allocated starting from the highest bid until no share remains.

2.4.4. Prevailing Mechanism: Book-Building

Despite the generally higher costs associated with the book-building mechanism, it is the

most commonly used mechanism in the U.S. (Ljungqvist et al., 2003 [91]). The book-building

mechanism can lead to the highest efficiency; that is, it can maximize the expected proceeds

(Biais and Faugeron-Crouzet, 2002 [26]). To reach the theoretical maximum, the underwriter

must be informed about different investors’ true pricing of the IPO and set the final offer price

accordingly. This cannot be achieved in practice. However, the mechanism of the book-

building allows the underwriter to acquire maximum information from the investors (Cornelli

and Goldreich, 2003 [43]).

The underwriters often determine a preliminary range for the share price, based on their

own valuation of the firm. They then collect information from the investors about how much a

share is worth for them (i.e., asking for ‘private’ information). The expected value of the

disclosed prices by investors determines the final IPO offer price. Since the investors have the

incentive to understate their true opinion about the share price, it is required that the

underwriters take some steps toward aligning their incentives with the investors’. The study of

Benveniste and Spindt (1989) [20] elaborates on some of these practices.

Page 47: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 2. Literature Review on IPO 2.5. Post-IPO Trading

15

Benveniste and Spindt (1989) [20] show that in the presence of information asymmetry, the

underwriter can solicit the investors’ honest opinion of the share price by allocating more

shares to those who indicate higher prices. If the majority of the investors report high prices

(i.e., the offering becomes ‘hot’), the underwriter may underprice the shares; that is, instead of

setting the final offer price to the high end of the preliminary offer range, the price is

decreased to some degree. The amount of this reduction varies among different offerings and

can be perceived as the cost of information acquisition. Such intentional underpricing is,

therefore, another practice to reward the truthful information disclosure when the overall

interest is high and discriminating in allocation does not apply. In practice, the private

information disclosure can be compensated using a combination of underpricing and

discrimination in share allocation.

As follows from the discussion above, under the book-building mechanism, the

underwriters do not discriminate in price, but only in allocating shares. The review of different

research indicates that under the book-building mechanism, institutional investors often

receive precedence for IPO shares in the primary market (e.g., Benveniste and Spindt, 1989

[20]; Ibbotson and Ritter, 1995 [78]; Cornelli and Goldreich, 2001 [42]; Ljungqvist and Wilhelm,

2002 [90]; Draho, 2004 [48]).

Note: Henceforth, the IPO mechanism is assumed to be book-building in this document.

2.5. Post-IPO Trading

Trading in the secondary market starts following the share allocation in the primary market,

at the IPO final offer price. The share prices are then determined by independent agents

competing in a public market.

2.5.1. Difference between Offer Price and First-Day Opening Price

Share allocation process and the number of shares offered remain the main reason for the

discrepancy between the offer price and the first-day opening price of an IPO. Interested

Page 48: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 2. Literature Review on IPO 2.5. Post-IPO Trading

16

investors who do not acquire any share or the desired fraction in the primary market (at the

IPO final offer price) can purchase them in the secondary market once the stocks hit the

market. Therefore, secondary market trading is a game between the investors with IPO shares

willing to sell and those who wish to receive a (larger) stake in the newly- public firm. Any

gain or loss from these trades remains with the selling parties and does not affect the issuer’s

IPO proceeds. It is the volume and types of orders placed with the brokers that determine the

opening price of the first day. A significant investor demand can result in (extraordinarily)

appreciated prices in the secondary market immediately following the primary market.

2.5.2. Post-IPO Underwriter Services

The underwriting contract often obligates the underwriter to price stabilizing activities in

the IPO aftermarket. What motivates the issuer to promise price stabilization is encouraging

investor participation in the primary market and increasing the IPO final offer price.

Underwriters obligate themselves to different types of price stabilizing activities such as

setting constraints and monitoring reselling for quick gains (i.e., flipping) in the aftermarket

(Draho, 2004 [48]). The standard approach to the price stabilization is using the ‘overallotment’

or ‘greenshoe’ option which can be explained as follows:

The underwriter borrows additional shares from the issuer and resells them in the primary

market, at the offer price. In the case of poor short-term performance (i.e., price decline in the

secondary market), the underwriter purchases back these additional shares. The practice is an

attempt to create a demand for the flipped shares and prevent price declines in the short-term.

If the prices increase in the immediate aftermarket, the exercise of the overallotment option

gives the underwriter the right to purchase the additional shares at the IPO final offer price

from the issuer. This implies that the underwriter does not have to cover the short position and

the size of the offer increases by the amount of additional shares sold in the primary market

(Aggarwal, 2000 [3]; Geddes, 2003 [57]).

Another important feature of the aftermarket trading is the ‘lock-up’ period set in advance

by the underwriter; that is, the pre-IPO shareholders are not allowed to sell their shares in the

Page 49: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 2. Literature Review on IPO 2.6. Underpricing

17

secondary market, for a predetermined period following the IPO. The purpose of the lock-up

period is to signal the insiders’ belief in the firm’s prospects (Ayayi, 2005 [14]).

2.6. Underpricing

In the IPO context, significant research has been devoted to underpricing (e.g., Hughes and

Thakor, 1992 [74]; Garfinkel, 1993 [56]; Arthurs et al., (2008) [11]). Understanding this

phenomenon is essential to correctly quantifying the risk associated with the post-IPO market

prices. Underpricing is defined as the excess of the short-term post-IPO market prices over the

IPO final offer price. It can be determined as the difference between the IPO final offer price

and the first-day closing price; however, the length of the period considered to estimate

underpricing can be longer (e.g., several weeks). Underpricing is often the consequence of a

two-step process: intentional underpricing arranged by the underwriter and/or issuer and post-

IPO market forces.

In the theories proposed to explain the underpricing phenomenon, the IPO is considered as

a game whose key players are the underwriting committee, issuer, and investors. The aligned

incentives or conflicting interests of these players can lead to underpricing. Most of the

proposed hypotheses focus on the intentional underpricing.

One of the most significant hypotheses of intentional underpricing perceives underpricing

as a cost that the underwriter and issuer endure to acquire the investors’ private information.

This hypothesis, suggested by Benveniste and Spindt (1989) [20], is discussed in detail in

Section 2.4.4. Agency conflicts between the underwriter and issuer can also lead to

underpricing. If information asymmetry exists between the two parties, the underwriter has the

incentive to further underprice the new issue. Greater underpricing by the underwriter reduces

the marketing effort required under the book-building mechanism. This theory is first

suggested by Baron and Holmström (1980) [18] and Baron (1982) [17]. Other theories have been

postulated to further enlighten the underpricing phenomenon. Ibbotson and Ritter (1995) [78] is

a comprehensive study which assembles these theories.

Page 50: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 2. Literature Review on IPO 2.7. Post-IPO Long-Term Firm Performance

18

The final underpricing incorporates a behavioural component besides intentional

underpricing. Interested investors with strong beliefs in the short- or long-term

outperformance of an IPO can cause the aftermarket prices to rise.

Due to the lack of empirical data, it is difficult to determine the contribution of intentional

underpricing. Roosenboom (2012) [121] is the first study which uses empirical evidence from

French IPOs to support the anecdotal claim about intentional underpricing. The researcher

obtained access to the underwriter reports of 228 IPOs during the 1990-1999 period. The

assessment of these reports reveals the percentage of intentional underpricing in the respective

firms. Roosenboom (2012) [121] documents an average intentional underpricing of 18.2%,

where reputable underwriters discount less.

In considering Roosenboom’s (2012) [121] research, it should be noted that there exists a

fundamental difference between the book-building mechanisms in France and the U.S. In

France, the underwriter solicits investor interest prior to the IPO registration. Thus, the

preliminary price incorporates much of the investor sentiment and has already been discounted

to compensate for the information revelation.

2.7. Post-IPO Long-Term Firm Performance

The preceding sections focused on the pre-IPO process and short-term post-IPO

performance. Understanding the post-IPO long-term performance is equally important.

According to Ritter (1991) [119], such information would impact the IPO investors’ long-term

trading strategies. Furthermore, it would reveal how much the short-term prices are influenced

by proper IPO timing (i.e., going public in hot markets) and behavioural components like

buying frenzy.

Some empirical evidence indicates poor post-IPO long-term performance for the equity of

the firms. Ritter (1991) [119] is the first academic study which documents this phenomenon

based on an empirical analysis. The study examines 1526 new issues during the period of 1975

to 1984 across different industries. These issuing firms are compared to their matching firms

Page 51: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 2. Literature Review on IPO 2.8. IPO Valuation

19

(i.e., public firms matched in terms of industry and market capitalization). The comparison

shows that on average, the issuing firms underperform the stock market in the three-year

period following their IPOs. The outcomes show greater underperformance for the firms going

public during hot markets. This evidence suggests that the short-term outperformance of these

firms can be attributed to the firms taking advantage of the temporary ‘window of opportunity’

to go public (see, Section 2.3).

The long-term underperformance can also be the consequence of the overvaluation of the

firms by investors at the time of going public. In the case of the firms with rapid pre-IPO

growth, investors may extend high pre-IPO growth rates over a long period in the future and,

hence, assign high values to the firms. Certo et al. (2009) [33] postulate that the long-term

underperformance can be the outcome of the sophisticated competitions to which a newly-

public firm is exposed. The firm’s strategic plans and resources to engage actively in the

competition define its long-term performance.

Studying a sample of 4753 IPOs during the 1970-1999 period, Loughran and Ritter (1995)

[93] also report smaller post-IPO 5-year average returns compared to a control group of non-

issuing firms. An interesting feature of the long-term underperformance is its persistence

despite being documented in different sources. Loughran and Ritter (1995) argue that the

pattern of short-term outperformance and long-term underperformance does not disappear

(i.e., arbitraged away) since there always exist optimistic investors with strong beliefs in the

fate of the IPO firms. These investors systematically overvalue the IPOs and it takes a long

period for them to realize the actual potential of the firms.

2.8. IPO Valuation

The first and perhaps the most critical step in the process of going public is the valuation of

the firm. According to Draho (2004) [48] and Roosenboom (2012) [121], underwriters use the

comparable multiples technique (see, Section 2.8.2) more frequently for this purpose. It is

indicated that the most commonly used multiples are price-to-earnings (P/E), price-to-cash

flow, price-to-sales, and enterprise value ratios.

Page 52: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 2. Literature Review on IPO 2.8. IPO Valuation

20

Underwriters, particularly in the U.S., do not often disclose their approach to valuing IPOs.

However, based on the evidence provided by Draho (2004) [48] and Roosenboom (2012) [121],

the methods described in Sections 2.8.1 to 2.8.3 can be considered standard.

2.8.1. Discounted Free Cash Flows

In theory, the value of equity can be determined as the sum of the present value of the

expected free cash flows that the shareholders would receive in the future. The expected free

cash flows can be discounted at the risk-adjusted discount rate (see, Ross et al., 2010 [122]).

Alternatively, one can adjust the expected cash flows for risk (i.e., determine certainty

equivalents) and use the risk-free rate to discount the certain cash flows.

The accuracy of the discounted cash flow (DCF) method is limited to the accuracy of the

cash flow projections. Due to the strong theoretical basis, this technique can yield correct

results if free cash flows are projected accurately. In practice, the cash flows are often

predicted by extrapolating the past events to the future and/or by assuming fixed (i.e., pre-

committed with no flexibility) and limited scenarios for the future operations and activities of

the firm. These somewhat arbitrary and highly subjective projections (common in practice)

would adversely impact the accuracy of the valuation outcomes.

The accuracy of the outcomes is also sensitive to the discount rate used to calculate the

present value of the expected free cash flows. In the case of IPOs, where there is no stock price

history, finding an appropriate risk-adjusted discount rate can be challenging. Often the

discount rate for IPOs is approximated by the industry cost of capital with some case-specific

adjustments (Draho, 2004 [48]).

In addition to the DCF method, this section can be generalized to include other discounting

techniques as listed below:

(i) Dividend Discount Model (DDM)

(ii) Residual Income Model (RIM)

Page 53: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 2. Literature Review on IPO 2.8. IPO Valuation

21

The dividend discount model (DDM) would be a suitable technique in the case of

companies with steady dividend payments. Considering that dividends form the only stream of

cash flows to the shareholders, DDM states that the current stock price can be computed by

summing the present values of all the future cash flows to the shareholders.

Although the DDM method avoids the challenges associated with free cash flow

projections, it is incapable of generating reliable outcomes for the firms without regular

dividend payments. Furthermore, as with the case of DCF, the accuracy of the outcomes

would be highly sensitive to the estimates of the dividend growth rate and the discount rate

(i.e., shareholders’ required rate of return).

The residual income model (RIM) is another frequently employed technique to estimate the

‘fair value’ of stocks. The technique is particularly proven to be useful in valuing firms with

no steady dividend payments or negative free cash flows. Essentially, RIM assumes that the

current stock price of a company is equal to its current book value plus the sum of present

values of all future residual incomes, where ‘residual income’ is defined as the difference

between the corresponding net income and shareholders’ expected dollar return (i.e., cost of

equity × shareholders’ equity).

The rationale behind RIM is to account for the equity cost in addition to the cost of debt

which is already incorporated into the figure of ‘net income’. Thus, the net income which is

adjusted for the equity suppliers’ expected return for the period would be a more

representative figure of a company’s profit. This approach would clarify if the company is

indeed profitable when its net income is adjusted for the risk taken by its shareholders.

RIM is indeed equivalent to DDM. The dividend-based framework can be transformed into

the residual income-based framework by theoretically substituting dividends by net income net

of retained earnings. Residual income with the above-indicated form is then obtained

following some algebraic reformulations detailed in Palepu and Healy (2008) [105]. Although

RIM and DDM are mathematically identical, the latter is only applicable to the companies

with predictable and steady dividends; whereas, by means of the aforementioned substitutions,

the former allows for the valuation of a wider variety of companies.

Page 54: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 2. Literature Review on IPO 2.8. IPO Valuation

22

As to the challenges associated with RIM, the projections of future residual incomes share

the same complications as the free cash flow predictions under the DCF method. Moreover,

finding an appropriate discount rate (i.e., cost of equity in this case) makes the valuation

practice challenging, especially in the lack of market information

2.8.2. Comparable Multiples

Models described in Section 2.8.1 are sometimes referred to as ‘absolute’ models. The term

implies that ‘value’ is estimated with reference to the fundamental financial information and

the models do not utilize the information available from other companies in the respective

sector to estimate value.

The method of comparable multiples, on the other hand, is considered a ‘relative’ model

implying that the model estimates value with reference to other companies. This technique

chooses a comparable firm or a group of comparable firms and uses their multiples (e.g.,

price-to-earnings and market-to-book ratios) to determine the value of the firm of interest. In

theory, the comparable firm should match the firm under scrutiny in terms of fundamental

characteristics such as net income and its growth rate, risk, and the product or service offered.

Section 3.3 discusses different approaches for selecting comparable firms in practice.

Some of the weaknesses associated with the comparable multiples method can be listed as

follows (Draho, 2004 [48]): (i) The method is limited to sufficiently diverse markets where

there exist several public firms comparable to the firm of interest, (ii) the accuracy of the

valuation outcome is affected by the accuracy of selecting the comparable firm(s), (iii) all the

assumptions and financial incentives underlying the price(s) of the comparable(s) would be

incorporated into the value estimates of the firm under scrutiny.

As indicated at the onset of this section, the method of comparable multiples is the most

frequently used IPO pricing technique and, for that matter, is the most commonly cited

technique in the IPO literature. Such interest in the comparable multiples technique is not

unique to IPOs. Comparables are extensively used to estimate diverse financial ratios. Section

Page 55: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 2. Literature Review on IPO 2.8. IPO Valuation

23

3.2 clarifies the implication of such a wide usage of comparables on developing a short-term

risk metric for IPOs.

2.8.3. Other Techniques

In addition to the aforementioned standard models, depending on the characteristics of the

IPO of interest, underwriters may choose to improve the valuation outcomes using specific

techniques. The remainder of this section provides a summary on two such methods: asset-

oriented methods and real options.

2.8.3.1. Asset-Oriented Techniques

Asset-oriented techniques are often used to provide tentative value estimates for asset-

intensive businesses, such as manufacturing businesses and financial services companies. The

value is often assumed to be the liquidation value of the business; that is, if the company

ceases its business activities, what its assets would be worth. To arrive at an estimate of the

value, the approach assesses the market worth of tangible assets, net of any liability. Examples

of tangible assets would be assets in the raw material inventory, finished goods, and fixed

assets (e.g., property, plant and equipment). Liquidation costs and other miscellaneous costs

associated with terminating a business, such as charges levied by the employees, should also

be taken into account.

One shortcoming of the above-indicated asset-oriented technique is the lack of explicit

accounting for intangible assets and the potential of the business to grow in the future and

create value from its tangible assets (Draho, 2004 [48]). Thus, asset-oriented methods have

limited applications in the IPO industry mainly comprising young businesses which would be

prone to undervaluation using this method.

2.8.3.2. Real Options

Real Option Analysis (ROA) is a framework which has been developed to analyze and

evaluate any arbitrary investment opportunity. ROA is based on the theory of financial option

pricing. However, in contrast to a securities option, a real option is an investment in a real

Page 56: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 2. Literature Review on IPO 2.8. IPO Valuation

24

asset, for example, a business project. The technique gives the right, but not the obligation, to

exercise a specific decision, such as to withdraw from or expand the project, at some future

time. Although the initial seed of ROA was planted with the early papers such as Black and

Scholes (1973) [27] and Merton (1973) [98], it was Stewart Myers who first used the terminology

‘real option’ in 1984 (Myers, 1984 [101]). He argued that the standard DCF method

underestimates the value of the real options attached to growing businesses.

ROA is an approach which combines strategic intuition and rigorous analytic computations.

ROA is perceived as a complement to the DCF technique and adds extra flexibility to the

valuation framework. The value of a project or a company (i.e., a portfolio of various projects)

can be described as a combination of its value with no flexibility and the value of its options.

Based on anecdotal evidence from U.S. underwriters (Draho, 2004 [48]) and firm evidence

from French IPOs (Roosenboom, 2012 [121]), the methodology of real option has not yet gained

a wide application in the IPO industry. The reason is that the technique is mathematically

intensive; yet, its contribution to accuracy remains unclear. ROA does not pose a suitable

solution to our problem, either. The reason can be described as follows: The main objective of

this research is to quantify the short-term risk of investing in IPOs, and the risk will be

estimated based on post-IPO market prices (see, Section 6.3). The reluctant and narrow

acceptance of the real options method by investors minimizes its impact on investors’

perception of future performance of companies, and consequently, their stock prices. Thus, the

seemingly insignificant impact of the real options method on market prices makes the

technique extraneous to this research.

Page 57: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

25

Chapter 3. LITERATURE REVIEW ON METHODOLOGY:

COMPARABLE-BASED METHODS

3.1. Introduction

This chapter centers on the literature of the ‘comparable-based’ methods. In Section 3.2, it

is established why it would be a prudent approach to base the risk assessment of IPOs on the

market information of the respective comparables. Section 3.3 covers the history of the

techniques involved in selecting comparables and extensively discusses their shortcomings.

Based on the survey conducted in this section, this thesis later proposes a new technique for

selecting comparables which addresses these limitations and augments the literature of

comparable firm selection (see, Chapter 6). The final section surveys the work focused on

identifying the factors impacting the IPO valuation and pricing. The scrutiny of such studies

assembles a set of factors which subsequently serve as the comparison criteria (dimensions of

comparison) between the IPOI and the candidates to act as its comparables.

3.2. Comparable-Based Risk Analysis

This section aims to establish why founding the risk analysis on comparables offers a

promising approach to quantify the market price-based risk of IPOs. By the end of this section,

it will become clear why top valuation specialists including bulge bracket investment banks

and ‘first team-ranked’ analysts frequently base their valuation on comparables. Furthermore,

evidence is provided that such specialists will perhaps be the key providers of direction to the

market. Thus, considering the prevalence of the comparable-based approach among analysts

and its potentially significant role on market prices, this approach can be suggested as the

building block of the risk quantification of IPOs (see, Chapter 6). To state it simply, if the

information derived from comparables impacts prices, then, a comprehensive risk metric ought

Page 58: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.2. Comparable-Based Risk Analysis

26

to be designed such that it extracts maximum information from the most representative group

of comparables.

The comparable-based valuation method is often contrasted with more theoretically-

acceptable discounting models. The wide interest in the former compared to the latter can be

justified from three different standpoints:

(i) The outcomes of the comparable-based technique are more reflective of market

characteristics and periodicities, behavioural components as well as investors’

general and specific perception of value (see, Section 3.2.1).

(ii) There are intrinsic limitations associated with the discounting (‘absolute’) models

(see, Section 3.2.2).

(iii) In terms of equity price projections, the increased complexity in case of discounting

models is often compensated with insignificant gain in accuracy (see, Section

3.2.3).

In the remainder of this discussion, Sections 3.2.1 to 3.2.3 detail each item on the above-

indicated list. Section 3.2.4 closes this discussion by providing evidence on the prevalence of

the comparable multiples method among U.S. bulge bracket investment banks and ‘first team-

ranked’ analysts whose reports often cause strong market reactions.

3.2.1. Comparable-Based Method - More Reflective of Market

Conditions

Discounting techniques such as DCF determine value primarily with reference to the ‘free’

cash generating potential of the business. These techniques do not utterly account for many

behavioral components, investors’ perception of value, or market conditions which impact

stock prices. Some of such details may be reflected in individual investors’ projections of

future free cash flows or discounting rates; e.g., in hot markets, investors tend to forecast

future free cash flows more optimistically. However, in general, the DCF framework does not

account for aggregate investor preferences, opinions and expectations, herding dynamics, etc.

Page 59: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.2. Comparable-Based Risk Analysis

27

As an example of the investors’ perception of value, consider R&D costs. Research has

shown that the R&D expenditures are positively correlated with the prices (e.g., Bartov et al.,

(2002) [19]). This implies that market assigns a positive value to the option of the firm to

commercialize the product (and expand) upon the success of the R&D project. Part of the

information asymmetry present at the IPO stage can indeed be attributed to how investors

value these options (Hall, 2005 [64]). Contrary to what has just been stated, under any

discounting method, R&D costs would be considered period expenses, and may even lead to

negative free cash flows for the period. In other words, within the DCF framework, not only

do R&D expenditures not contribute to the value, but they adversely impact the value.

Behavioral components, such as reacting to the underwriter’s reputation and the retention

ratio of the insiders’ shares, also cannot be directly incorporated into a cash flow-based

valuation framework. They have no influence on the firm’s operations and its capability to

generate free cash flow and, yet, they impact prices (e.g., Draho, 2004 [48]; Certo et al., 2009

[33]).

In case of such complicating market features, a market-based valuation approach would

lead to more descriptive outcomes. As a case in point, again, consider R&D costs.

Furthermore, consider the following two players: (i) a current IPO (i.e., IPO of interest) (ii) a

past IPO (i.e., a currently public firm whose IPO took place in the past). Given similar R&D

expenditures under similar aggregate (macro) market conditions at the issue time, it can be

assumed that the R&D information would be factored in for the current IPO in a similar

fashion to the past IPO. Underlying such assumption is the fact that other things being equal,

the average investor’s recognition of value drivers does not change critically over time.

It is worthwhile to note here that mathematically rigorous methods have been developed in

attempt to address some aspects of the above-indicated complications, surrounding market

value estimations. An example would be the variety of stochastic differential equation (SDC)-

based techniques which have been developed in attempt to capture part of the uncertainty

associated with a firm’s future free cash flows, and hence, improve the estimation of its

Page 60: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.2. Comparable-Based Risk Analysis

28

current stock price (e.g., Schwartz and Moon, 2000 [127]; 2001 [128]; Maya, 2004 [97]; Berk et al.,

2004 [23]; Wu et al., 2008 [139]). It should, however, be noted that even the most comprehensive

valuation technique, as long as it stays private and is not shared information among investors

or market makers, would not deliver accurate estimations of the market prices. Furthermore,

research has shown that such analyst-specific techniques do not supplant the standard

valuation techniques discussed in Section 2.8; top-tier underwriters and first-team ranked

analysts only marginally depend on these mathematically demanding techniques in estimating

market values (Roosenboom, 2007 [120]).

3.2.2. Discounting Methods - Theoretically Sound but Practically

Rigorous

As explained in Section 2.8.1, the DCF technique has a firm theoretical basis. It calculates

the value as the sum of the discounted expected future free cash flows. Therefore, the accuracy

of the outcomes of this method heavily relies on the quality of cash flow projections.

Forecasting cash flows is the most error-prone and challenging step of the DCF method.

Future free cash flows generated by a firm fluctuate over time due to the occurrence of various

events such as market moves, competition, or success/failure of the launch of a new product.

Some of these events may cause significant up or down jumps in the free cash flows. The free

cash flow can be considered as a combination of a deterministic dynamic component and a

stochastic component. As mentioned, in theory, the DCF technique computes the expected

values of the free cash flows at given points in time and discounts them back to the present.

This would require considering many different scenarios for the evolution of the free cash

flows. Subsequently, for each point in time, the technique should calculate the expected value

of the respective incremental free cash flows across different scenarios.

In practice, however, the expected free cash flows, under the traditional DCF technique, are

estimated in a somewhat arbitrary fashion, subjective to the analyst’s opinion and/or by

extrapolating the past trends to the future. Furthermore, under this technique, limited investing

Page 61: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.2. Comparable-Based Risk Analysis

29

and operating scenarios are assumed for the firm and the firm pre-commits to each scenario

with no flexibility for change conditional on future events.

The practical challenge of the DCF technique is not limited to forecasting future free cash

flows. As also detailed in Section 2.8.1, finding an appropriate discount rate, which can be

used to calculate the net present value of the future cash flows, poses another challenge. This

shortcoming is pronounced in the case of IPOs where the businesses have no stock price

history. Due to the mathematical sensitivity of the discounting methods to the growth rate and

discount rate, any minor error in the estimations of these two factors could lead to major

swings in the final estimates of value.

The aforementioned practical obstacles have impeded the broad adoption of the

theoretically-strong discounting methods by the analysts. Within the context of French IPOs,

Roosenboom (2007) [120] has shown that often it is in the volatile or high-growth markets

where underwriters assign greater weights to the outcomes of the DCF method. The reason can

be stated as follows: In high-growth markets, analysts can choose future growth rates of cash

such that the valuation outcomes closely follow market prices, and then, they will have a broad

latitude in defending their choices of growth rates. Furthermore, once volatility hits stock

prices, investors tend to rely more on the DCF method since as an ‘absolute’ method (see,

Section 2.8.2), it determines value primarily based on the operations of the firm and its

potential to generate cash in the future.

3.2.3. Discounting Methods - Added Complexity with Uncertain

Accuracy Gain

Despite its more theoretically firm foundation and computationally intensive process,

research shows that the DCF method has similar accuracy to the comparable multiples method

in predicting market prices.

Page 62: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.2. Comparable-Based Risk Analysis

30

3.2.3.1. Evidence from Non-IPO Cases

Using the DCF method, Kaplan and Ruback (1995) [82] valued a sample of 51 highly

leveraged transactions (HLTs) which took place in the 1983-1989 period. The study shows

that the outcomes of the DCF and comparable multiples method stay similar and they both

have comparable accuracy in predicting the market values of the transactions. The comparable

multiples method is centered upon market value to EBITDA multiples, each emerging from a

different definition of comparable company. The researchers argue that the comparable

multiples technique is capable of outputting more accurate results than the DCF method given

that the company under scrutiny and its comparable correspond and be proportionally

comparable in risk and size of cash flows. Furthermore, the choice of a multiple could impact

the accuracy of outcomes since the multiple is the reference connecting the relative values of

the two companies. The comparable multiples approach which meets all the above-indicated

criteria is expected to predict market values more accurately than the DCF method does since

it additionally accounts for market conditions and characteristics (see, Section 3.2.1).

In a different setting from Kaplan and Ruback’s (1995) [82] study and within the context of

businesses emerging from Chapter 11 bankruptcy, the research of Gilson et al. (2000) [59] also

attested to the approximately akin outputs of the DCF and comparable multiples techniques.

The objective of the study was to compare the following two figures for 63 ‘reorganized’

companies over the period of 1984-1993: (i) the market value post reorganization (ii) the value

obtained using the cash flow predictions that had been supplied as part of the process of

Chapter 11 bankruptcy. The study also estimates value using the comparable multiples

method. The multiple is selected to be the industry median of the ratio of total

capital/EBITDA. The performance (i.e., value) measure for the firm under scrutiny is assumed

to be the projected EBITDA for the first year after reorganization. The study concludes that

the DCF and comparable methods generate similar results and lead to similar level of accuracy

in predicting market values.

Page 63: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.2. Comparable-Based Risk Analysis

31

The comprehensive research conducted by Asquith et al. (2005) [13] further supports the

notion of indifference between the DCF model and the comparable multiples method. The

primary objective was to create a unique database which collects and categorizes the details

provided in analyst reports, such as price targets and valuation techniques. To this end, the

study manually scrutinized 1126 equity analyst reports released during the 1997-1999 period.

In order to be included in the study, analysts had to be members of Institutional Investor’s All-

American Research Team and ranked as ‘first-team’ at least one time. This implies that the

research only included top analysts whose buy-sell-hold recommendations would likely elicit

strong reactions by market participants.

The setting of Asquith et al.’s (2005) [13] research facilitated pursuing another goal which

was studying the association between the target price accuracy and the respective valuation

technique. The study reports that there is no significant association between the accuracy of

target price forecasts and the techniques used to arrive at these estimates, provided that the set

of techniques is limited to revenue or earnings multiples or DCF. Moreover, the choice of

valuation method was shown to be as ineffective in driving market reaction. Interestingly, the

research reports reduced accuracy in forecasted target prices if analysts use other techniques

such as Economic Value Added (EVA) or some analyst-specific techniques, uncommon

among analysts.

Worthy of note here is the definition of ‘accuracy’ in Asquith et al. (2005) [13]: if the

forecasted target price or a higher price is realized in the one-year period pertaining to the

forecasted target price, the target price is considered ‘accurate’. In this study, the analysts’

projected target prices were accurate in 54% of the cases. The projection accuracy can be

considered satisfactory for the rest of the cases where, on average, 84% of the target price was

observed in the market. Section 3.2.4 highlights more details from this study regarding the

distribution of valuation methods amongst financial analysts.

Page 64: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.2. Comparable-Based Risk Analysis

32

3.2.3.2. Evidence from French IPOs

The previous section (Section 3.2.3.1) demonstrated that in non-IPO cases, the DCF

technique and comparable multiples approach (particularly, earnings multiples) bear similar

accuracy levels. The research on IPO cases, on the other hand, is sparse; the underwriters,

particularly in the U.S., do not often disclose the valuation details, and information such as

cash flow forecasts and alike is seldom filed with SEC (Kaplan and Ruback, 1995 [82]). In the

IPO literature, only few studies could access the underwriters’ reports containing the details of

their valuation methodologies. Roosenboom (2007 [120], 2012 [121]) has perhaps conducted the

most comprehensive study in this regard.

The objective of Roosenboom’s (2012) [121] study was to document the methodologies used

by underwriters to compute ‘fair’ values of the IPOs. Moreover, the study elucidates the IPO

pricing process which shows all the modifications applied to the fair value estimate prior to

setting the final IPO offer price. To achieve the research goals, Roosenboom (2012) [121]

scrutinized the underwriter reports of 228 French IPOs which issued in the 1990-1999 period.

He reports the following three methods as the most frequently employed ones by underwriters:

(i) comparable multiples, (ii) dividend discount model (DDM), (iii) DCF. Underwriters may

assign a greater weight to a technique based on idiosyncratic characteristics of the firm of

interest or aggregate market conditions (Roosenboom, 2007 [120]; 2012 [121]); however, overall,

the comparable multiples technique, used in more than 87% of the cases, remains the most

frequently used technique. The bias and accuracy of these methods are compared below. It

should be noted that to prevent any sampling bias which can skew the outcomes in favor of

one method, any pairwise comparison between the techniques is conducted across only the

firms which employed both techniques.

Roosenboom (2012) [121] calculated ‘bias’ as the ratio of (value estimate - equilibrium

market value) to (equilibrium market value). He documented that all three techniques lead to,

on average, positively-biased outcomes; however, there is no significant difference between

the techniques in terms of bias. As with the case of ‘bias’, this study does not detect any

Page 65: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.2. Comparable-Based Risk Analysis

33

significant difference between the ‘accuracy’ of the three methods; accuracy is defined as the

absolute value of bias.

Concisely, despite the sparse public information on underwriters’ valuation details, the

scrutiny of the reports of 228 French IPOs provides firm evidence that, at least in the case of

French IPOs, the comparable multiples method and the computationally intensive DCF

method exhibit similar accuracy and bias in predicting equilibrium market values.

3.2.4. Comparable-Based Method: Prevalent in the U.S.

The comparable multiples method is the most commonly used valuation method in the U.S.

perhaps due to its more intuitive approach to valuation, yet, roughly identical outcomes to the

mathematically demanding DCF method. The remainder of this section is focused on the

studies which further support this fact.

According to the study conducted by DeAngelo (1990) [46], the comparable multiples model

is favored by bulge bracket investment banks such as CS First Boston (CSFB). In one of its

reports, CSFB documented the sensitivity of the DCF method to the terminal value estimates

and indicated that they favored the earnings multiple model over the DCF method.

Asquith et al.’s (2005) [13] study, introduced in Section 3.2.3.1, provides important

information regarding the distribution of the valuation methods amongst All-American

Research Team members, considered to be top U.S. analysts whose reports and buy-sell-hold

recommendations can lead to changes in the direction of the market. As indicated, the primary

objective of Asquith et al.’s (2005) [13] study was to create a unique database which collects

and categorizes the details provided in the analyst reports such as target prices and valuation

techniques. To accomplish this goal, the researchers scrutinized the equity reports of 1126

first-team ranked sell-side analysts. The reports covered the period of 1997 to 1999.

The assessment of the reports reveals that the earnings multiple method, employed by

99.1% of the analysts, is the most frequently used method; whereas, the usage of DCF or ‘any

of the other discounted cash flow approaches favored by finance textbooks and MBA

Page 66: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.2. Comparable-Based Risk Analysis

34

curriculums’ is not as dominant. Attesting to this fact is the figure of only 12.8% of the

analysts reporting the use of this method. The DCF method is employed more frequently if

downgrading is recommended. The research indicates that it is used in 20.8% of the

downgrade reports. Worthy of recalling here is the accuracy of target prices forecasted by

these analysts (refer to Section 3.2.3.1 for further details): The analysts forecasted target prices

‘accurately’ in 54% of the cases. The ‘accuracy’ of the remaining cases (i.e., 46%) can be

considered satisfactory since, on average, 84% of the target price was observed in the one-year

period corresponding to the target price forecast.

Houston et al. (2006) [70] is another study which indicates the prevalence of the comparable

multiples method among the U.S. analysts. This study was, however, conducted in a different

setting than Asquith et al.’s (2005) [13] research and focused on the initiating coverage reports

for IPO firms. One of the main objectives of this study was to investigate the methodologies

applied by the analysts in arriving at their target price estimates. Note that the target price of a

stock is an estimation of its price one year from the date the report is released. To perform this

analysis, the study scrutinized 454 reports initiating coverage on 248 firms whose IPO took

place during the 1996-2000 period. In order to be included in the study, a report had to be

published sometime during the 115-calendar day period following the respective firm’s issue

date. The research revealed that the DCF method was used in 10% of the reports. The

comparable multiples method was applied in a notably higher number of cases (i.e., 73% of

the reports) with earnings multiples being the most dominant. Furthermore, the research shows

that 87% of the firms were linked to at least one analyst report which used the comparable

multiples method to estimate the respective target prices. Houston et al. (2006) [70] also

examined the association between the characteristics of a firm and the valuation method

preferred by the analysts for forecasting target prices. The results indicate that the comparable

multiples method is preferred to other mathematically rigorous methods in case of young and

small businesses which pose many valuation challenges.

The studies presented in this section (Section 3.2) have documented that the method of

comparables is utilized frequently and yields estimates which are accuracy-wise comparable to

Page 67: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.3. Comparable Firm Selection

35

other analytically rigorous valuation techniques. The accuracy of the comparables approach in

predicting prices can be attributed to two different sources: (i) Value estimates derived with

reference to comparables incorporate market data (i.e., trends, insights, etc.) which are hard to

account for with other valuation methods (see, Section 3.2.1), (ii) The methodology is the

most frequently applied in the first-team ranked analysts’ reports. These reports initiate

significant market reaction and may impact the stock trends, skewing them toward the

recommended direction. Researchers have documented such impact of analyst reports on the

stock prices (e.g., Brav and Lehavy, 2003 [32]; Houston et al., 2006 [70]). Considering the

dominance and accuracy of the comparable-based approaches in predicting prices, they could

similarly offer a suitable ground for projecting future distributions of prices, and hence,

estimating the short-term risk. The process is detailed in Chapter 6.

3.3. Comparable Firm Selection

As established in the previous section, a comparable-based methodology is adopted in this

thesis to quantify the short-term risk of IPOs (also see, Chapter 1). One of the challenges of

this approach is identifying appropriate comparable firms to the firm under consideration. This

section reviews the existing methodologies for selecting comparables and notes any

shortcoming that may make the respective technique incompatible with the objectives of this

research.

A comparable firm, which is often selected from the industry of the firm of interest, is

expected to bear a resemblance to the firm of interest in terms of risk, future growth rates, and

earnings (Draho, 2004 [48]). The approaches adopted in the literature to find comparables are

quite diverse; yet, they fail to match all the above-indicated aspects, simultaneously. The

studies by Boatsman and Baskin (1981) [29] and Alford (1992) [4] are presented below as cases

in point.

Boatsman and Baskin (1981) [29] studied two different approaches for selecting a

comparable firm: (i) selecting a firm randomly from the corresponding industry, (ii) examining

Page 68: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.3. Comparable Firm Selection

36

the 10-year mean growth rate of earnings for individual firms from the same industry and

selecting a firm with the most similar mean. Using the P/E ratio, the second approach yielded

smaller price prediction errors.

Alford (1992) [4] examined three different approaches for choosing a comparable firm: (i)

based on industry using 4-digit SIC codes, (ii) based on risk which is proxied by size (i.e., total

assets), (iii) based on earnings growth using book return on equity (ROE) as its measure. The

study examined three different combinations of these approaches as well (i.e., industry and

size, industry and growth, and growth and size).

Using the P/E multiple to test the above-indicated approaches, Alford (1992) [4] concludes

that the industry approach and the combination of risk and earnings growth produce

comparable outcomes in terms of accuracy in predicting stock prices. The pairwise

combination of the industry approach with the approach of risk or earnings growth does not

significantly increase the accuracy of predictions. This implies that industry indeed

incorporates the impacts of risk and earnings growth. Selecting comparables using risk or

earnings growth individually does not provide effective solutions, either. Overall, the industry

approach is recommended in this study. Furthermore, it is noted that selecting comparable

firms based on industry yields more accurate valuation outcomes for larger than smaller firms.

Choosing comparables for an IPO firm is more complicated. The limited history available

at the IPO stage restricts use of factors such as long-term mean growth rate of earnings, used

by the studies presented above. The work of Kim and Ritter (1999) [87] was focused on valuing

IPOs and comparing the outcomes to the offer prices. The sample consisted of 190 IPOs from

the 1992-1993 period and the valuation technique was selected to be the comparable multiples

method. Two groups of firms were used as comparables: (i) firms that had recently gone

public in the same industry, (ii) firms selected by Renaissance Capital [117] which is a

‘boutique’ [87] research firm specializing in IPO valuation.

The first approach algorithmically chooses comparables based on the 4-digit SIC code.

Although this technique can only be as accurate as the SIC classification is, it hedges against

Page 69: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.3. Comparable Firm Selection

37

biases in the selection process. As an illustrative instance, consider hot issues markets where

underwriters have the incentive to manage the comparable firms selection in order to justify

their high estimations of the P/E multiple. The selection approach of Renaissance Capital, on

the other hand, is based on the information provided in the prospectuses about the firm’s direct

competitors. The prediction error is reduced when using the second approach.

Another study in the IPO context is conducted by Purnanandam and Swaminathan (2004)

[114]. Their primary objective was somewhat similar to Kim and Ritter’s (1999) [87]: they

investigated whether IPOs are overpriced with respect to the offer price. To this end, the study

compared the offer prices to the values obtained using price multiples and concluded that the

median IPO in their sample of more than 2000 IPOs (1980-1997) is overpriced. As to the

process of choosing comparable firms, Purnanandam and Swaminathan (2004) [114] adopted a

different approach than Kim and Ritter (1999) [87]. The comparable firm, which is a non-IPO

firm, is selected such that it matches the IPO of interest in terms of industry, sales, and

EBITDA profit margin. The study assumes that industry-matching would restrict the sample of

candidates to those firms with some potential of matching the IPO of interest with respect to

operating growth prospects, risk and profitability. Factors representing the operating

performance could be extended beyond sales and EBITDA profit margin. However, the

methodology suggested by Purnanandam and Swaminathan (2004) [114] cannot handle an

increased number of such factors; due to its sequential nature, as the set of factors is

broadened, the chance of finding a single comparable decreases. The matching methodology

can be described as follows:

The study classifies the firms based on the industry classification system of Fama and

French (1997) [50]. They used 4-digit SIC codes to cluster firms into 48 industries. Each

industry group is divided into three sales portfolios based on sales and, each sales portfolio, in

sequence, is divided into three margin portfolios based on EBITDA profit margin. The

comparable firm to the IPO of interest is then selected from the margin portfolio of the IPO of

interest and is defined to be the firm with the most similar level of sales to the IPO of interest.

Page 70: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.3. Comparable Firm Selection

38

Industry-based approaches approximate the performance of the firm of interest with the

overall characteristics of its industry. These approaches disregard firm-specific details such as

operating and financial characteristics. In theory, the comparable firm must resemble various

aspects of the firm under scrutiny such as the product or service, profitability, short-term

solvency, turnovers, and capital structure or financial leverage.

A comparable firm which satisfies all the above-indicated criteria seems unachievable in

practice. However, a sufficiently similar firm can be identified based on a multi-dimensional

comparison between important characteristics of the firm under scrutiny and other candidate

firms. The study of Purnanandam and Swaminathan (2004) [114] extends the basis of the

comparable selection process beyond using only industry multiples, by additionally accounting

for similarities in sales and EBITDA profit margin. However, as indicated above, due to the

sequential nature of their approach, it is impractical to base the comparison on a broad

category of factors. The review of literature returns two studies whose proposed algorithms

take into account multiple accounting and financial factors in choosing comparables: (i)

Bhojraj and Lee (2002) [25], (ii) Simak (2000) [133]. This section is terminated by concisely

describing the two studies.

In addition to developing an algorithm which identifies comparables primarily based on

multiple financial variables, Bhojraj and Lee’s (2002) [25] study stands apart from the prior

studies in two other respects:

(i) The previous studies selected the most suitable multiple as the one which best

explains the stock price of the firm of interest (e.g., Price-to-Earnings multiple). The

study by Bhojraj and Lee (2002) [25], on the other hand, chooses ex ante the

enterprise value-to-sales (EVS) and price-to-book (PB) ratios because these two

ratios are deemed as representatives of intrinsic value and can be obtained by

restructuring the residual income model and the dividend discount model,

respectively.

Page 71: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.3. Comparable Firm Selection

39

(ii) Despite prior studies which tested the comparable-based method by its capacity to

predict the stock price accurately, the study of Bhojraj and Lee (2002) [25] examined

the accuracy of the method based on its capacity to predict the EVS and PB

multiples. The reason is that this study is concerned with the intrinsic value of the

firm of interest and aims for developing a comparable-based methodology which

can be utilized to detect over- or under-priced stocks. Therefore, the stock price

itself does not present a suitable reference for gauging the accuracy of the method.

The methodology suggested by Bhojraj and Lee (2002) [25] identifies the comparables using

regression which clarifies why it can incorporate multiple accounting and financial factors.

Moreover, using regression in conjunction with choosing the multiples of EVS and PB as the

value indicators make the methodology applicable to the case of unprofitable firms with

negative earnings. Note that the widely used P/E multiple is not suitable in the case of negative

earnings. Bhojraj and Lee (2002) [25] adopted the following steps to identify peers. As will

become clear, the methodology is objective and, therefore, can be automated.

First, the cross-sectional variation of EVS or PB is modeled as a function of eight

explanatory variables, all pertaining to the same year but may be spanning different industries;

that is, data belong to firms from different industries. Considering that the PB and EVS ratios

can be obtained by restructuring the residual income and dividend discount models, the set of

explanatory variables comprises the ratios which either directly appear in these two models

(e.g., return on equity (ROE)) or act as the underlying factors of the ratios which directly

appear in the two models (e.g., return on net operating assets defined as operating profit

divided by net operating assets). In order to estimate a firms’ current or future EVS or PB

ratio, the study uses the regression equation resulted from the last year’s data. The estimated

EVS or PB is termed ‘warranted’ EVS or PB.

Second, Bhojraj and Lee (2002) [25] proposed five different ‘definitions’ for comparable

firms. In order to assess the significance of each of these definitions in providing the current or

up to three-year forward estimates of EVS or PB, the two ratios were regressed on a set of

Page 72: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.3. Comparable Firm Selection

40

explanatory variables. Each member of this set is the representative of one definition of

comparable firm. The members can be listed as follows: (i) ‘harmonic’ mean of the actual (not

warranted) EVS or PB ratios of the firms which are classified under the same industry as the

firm under scrutiny. In the study, a firm’s industry is determined based on its 2-digit SIC code.

The harmonic mean can be computed by taking the inverse of the mean of the inversed ratios

(e.g., EVS’s), (ii) harmonic mean of the EVS or PB ratios of the four firms whose market

capitalizations remain closest to the firm under scrutiny and whose industry is the same as the

firm of interest, (iii) the warranted EVS or PB ratio of the firm under scrutiny itself, (iv)

harmonic mean of the EVS (PB) ratios of the four firms with the closest warranted EVS (PB)

ratios to the EVS (PB) ratio of the firm under scrutiny, (v) harmonic mean of the EVS (PB)

ratios of the four firms whose warranted EVS (PB) ratios remain closest to EVS (PB) ratio of

the firm under scrutiny and whose industry is the same as the firm of interest as well.

The results from the second stage demonstrate that including warranted multiples (i.e.,

definitions (iii) to (v) discussed above) significantly improves the predictive power of the

regression equations. Bhojraj and Lee (2002) [25] thus concluded that using comparables which

are selected based on the proximity of the warranted multiples leads to more accurate

predictions of EVS or PB. It should, however, be noted that the outputs of the PB regression is

not as satisfactory as the EVS model.

Despite the merits of Bhojraj and Lee’s (2002) [25] approach, the following prominent

limitations are associated with their methodology:

(i) This method is regression-based. In consequence, it incorporates all the limitations

associated with the regression framework. As a case in point, regression is a

parametric approach which requires an ex ante specification of the relationship

between the dependent variable (or output) and explanatory variables (or inputs).

The accuracy of regression estimates is thus not only affected by the choice of

explanatory variables but by how they are related to the dependent variable. This

limitation is even more critical in the case of stock prices. Not only is the stock price

Page 73: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.3. Comparable Firm Selection

41

associated with a broad category of value drivers but the nature of this association is

complex and may be quite arbitrary. In such cases, non-parametric methods offer a

more concrete and integrative approach since they endogenously characterize the

link between the inputs and outputs and do not require a priori knowledge of the

functional relationship between them. Section 6.2 provides more details on selecting

comparables based on financial and accounting metrics and using a non-parametric

method called data envelopment analysis (DEA).

(ii) Outliers can substantially distort the regression fit. In order to mitigate the effect of

outliers, Bhojraj and Lee (2002) [25] excluded extreme low and high ends of the data.

While this step may improve the outcomes in case of the firms positioned close to

typical observations, it can adversely impact the comparable selection for extreme

firms as they may happen to match excluded outliers more closely.

(iii) The effectiveness of choosing comparables based on the proximity of EVS or PB

multiples is benchmarked against the accuracy of the EVS or PB forecasts yielded

by these comparables. Therefore, the capacity of the methodology to estimate prices

accurately remains indefinite. Hence, this approach does not fulfill the objectives of

this thesis. Considering that our research centers upon developing a risk measure,

which is essentially price-based, it is the price estimation efficacy that should

determine which comparable selection algorithm is indeed the most suitable.

(iv) In accordance with the preceding point, the accuracy of Bhojraj and Lee’s (2002) [25]

method is influenced by the choice of multiples. As discussed earlier, choosing

comparables based on warranted multiples leads to satisfactorily accurate

predictions only for EVS; the PB outcomes, although improved, remain far from

their market realizations.

(v) The methodology involves some details which render its use restricted in the case of

IPOs:

(a) The second regression described on Page 39 incorporates an explanatory

variable (ISEVS or ISPB) which can be computed as follows: The

Page 74: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.3. Comparable Firm Selection

42

market capitalizations of all the firms from the same industry as the firm

of interest are ranked and four firms with the closest market

capitalizations to that of the firm of interest are selected. The harmonic

mean of the EVS or PB ratios of these four firms represents a value for

ISEVS or ISPB, respectively. As inferred from the context, this

approach is inapplicable to IPOs as no market information exists for IPO

firms.

(b) Bhojraj and Lee’s (2002) [25] regression model requires data from two

consecutive years (i.e., last and current year) in order to estimate

warranted multiples for each firm under scrutiny. This requirement of

the model may not be fulfilled considering the data scarcity at the IPO

stage.

(vi) Bhojraj and Lee’s (2002) [25] first regression model includes consensus earnings

growth forecasts provided by I/B/E/S. Analysts’ forecasts are often subjective and

may be optimistically biased (Houston et al., 2006 [70]). Their use must, therefore, be

accompanied by a great caution; i.e., such forecasts should be scrutinized in light of

the analysts' motivations, particularly agency-driven incentives and local politics

(e.g., potential future business opportunities between the firm of interest and the

institution bound to the analyst).

(vii) The algorithm is multiples-based; that is, it draws on ratios to find comparables.

Basing the analysis on ratios as opposed to absolute values brings along the

disadvantages associated with ratios; e.g., by dividing two accounting factors, the

individual pieces of information contained within each figure are lost. Likewise, the

division may lead to the loss of economies of scale.

(viii) The cross-sectional regression which produces warranted multiples is calibrated

based on firms from different industries whose accounting data pertain to the same

year. This model is essentially regressing EVS or PB on its potential value drivers.

Calibrating the regression equation using cross-industry data would assign greater

Page 75: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.3. Comparable Firm Selection

43

weights to general value drivers and would disregard or assign less significance to

the value drivers specific to a particular industry. Negative earnings in the

biotechnology industry can be considered as an example for this point. Negative

earnings - arisen from, e.g., R&D expenditures - may positively contribute to value

as some investors tend to optimistically view them as investments. This cross-

industry calibration can consequently adversely impact the accuracy of comparable

selection. The reason is that it is eventually the warranted multiples that form the

basis for selecting comparables. These warranted multiples are generated based on

the regression coefficients that reflect the cross-industry general perspective on the

value drivers and disregard any industry-specific perspective.

As indicated above, Simak (2000) [133] is another study which also draws upon multiple

accounting criteria to choose comparables. Simak’s (2000) study - which appears to be

developed independently from Bhojraj and Lee (2002) [25] research - addresses the problem of

choosing comparables using a DEA model (see, Chapter 4 for more details on DEA). Since

DEA is a non-parametric approach, it resolves the regression-related issues existing in the

methodology proposed by Bhojraj and Lee (2002) [25]. Furthermore, Simak’s (2000) [133]

methodology was originally developed in the context of private firm valuation, and hence,

seems to be more compatible with the IPO framework. It should also be noted that Simak’s

(2000) [133] approach utilizes financial factors in ‘absolute’ format, as opposed to multiples,

which eliminates the ratio-related disadvantages pointed above (see, shortcoming (vii)).

Simak’s (2000) [133] methodology is, however, a semi-objective approach and cannot be

completely automated. Moreover, there exist other methodology-specific limitations which

will be discussed below once the technique is introduced.

As pointed out above, the objective of Simak’s (2000) [133] study is to find the market value

of a private firm. Therefore, after choosing the most comparable publicly traded firm to the

private firm under consideration, the market value of the private firm is approximated by the

market value of the comparable public firm.

Page 76: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.3. Comparable Firm Selection

44

In order to quantify ‘similarity’, Simak (2000) [133] developed a similarity metric using

DEA. The analysis is summarized below. The reader is referred to Chapter 4 for more details

on DEA and the respective parameters.

Step 1. Using the input-oriented DEA model of CCR or BCC, the ‘efficient frontier’ is

identified for a set consisting of the private firm and candidate public firms. The input

matrix includes variables such as total assets and total liabilities. The outputs can be

shareholders’ equity, net sales, net income and net cash flow. Based on the outcomes

of the DEA model, the firms with the same ‘reference set’ are clustered together.

Step 2. Simak (2000) [133] suggests a metric, referred to as ‘difference indicator’, for

gauging the difference between the private firm and other firms in the cluster. The

metric is defined as shown in the following equation:

𝛿𝛿𝑖𝑖𝑖𝑖 = ��𝜆𝜆𝑖𝑖𝑖𝑖 − 𝜆𝜆𝑖𝑖𝑖𝑖�2

(3.1)

Eq. 7.1 [133]

where, 𝛿𝛿 represents the difference indicator. The private firm and the corresponding

public firm are designated with the subscripts 𝑖𝑖 and 𝑗𝑗, respectively, and 𝑘𝑘 denotes the

efficient firms in the reference set and implies that the firms 𝑖𝑖 and 𝑗𝑗 belong to the same

cluster.

Step 3. The difference indicator is indeed a measure which shows the similarity in the

values of 𝜆𝜆. In other words, it shows how closely the ‘efficient targets’ of the two firms

are positioned on the efficient frontier. It is possible to find two firms located distant

from each other in space, yet, with efficient targets situated close to each other, leading

to a small value of difference indicator. This case indicates the necessity for another

dimension to complete the comparison. To this end, the firms which have been

identified to have similar 𝜆𝜆s, using the equation of difference indicator, are then

Page 77: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.3. Comparable Firm Selection

45

compared in terms of the radial ‘efficiency score’ (𝜃𝜃). Each firm’s efficiency score can

be considered as a measure of the firm’s distance from the frontier. The comparison of

𝜃𝜃s would, therefore, ascertain that the two firms are positioned close to each other in

space.

Anadol (2000) [7] put Simak’s (2000) [133] proposed methodology into test using the 1998

accounting data of 457 Canadian public companies from the manufacturing sector, as

classified under the SIC system. Different sets of analyses were conducted in her study. In one

set, 62 ‘inefficient’ companies from the SIC cluster of 3600s were assessed. The results

showed that in 49% of the cases, the market capitalization of the respective company falls

within the range yielded by the methodology. In 13% of the cases, the estimated market

capitalization was reported to be within the 25% of the realized market capitalization. The

reader is referred to the original study for further statistics.

In Anadol’s (2000) [7] thesis, the input-oriented BCC model was chosen (over CCR) to

conduct the analysis. BCC allows for variable returns to scale which is essential when the

analysis incorporates firms of various sizes. Moreover, the input-oriented BCC model can

handle negative outputs. Since in Anadol’s (2000) [7] dataset, only outputs (i.e., net sales and

net income) included negative numbers, the input-oriented BCC model suited the framework

of her analysis.

Despite the advantages listed above, the following limitations are associated with the

algorithm suggested by Simak (2000) [133]:

(i) The methodology is not applicable to the cases where the private firm under

scrutiny is ‘efficient’ itself. Anadol (2000) [7] mitigated this problem by proposing

an approach which only determines a lower bound for the market value of the

efficient private firm. The lower bound is computed as the weighted sum of the

market capitalizations of the inefficient public firms whose sets of efficient peer(s)

include the private firm of interest. The weights are defined as the normalized 𝜆𝜆s of

the inefficient public firms associated with the efficient private firm.

Page 78: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.3. Comparable Firm Selection

46

It should be noted that the complication associated with ‘efficient’ firms of

interest still remains untackled. Anadol’s (2000) [7] method only offers an

approximate solution with no reference to the theory. The reasons are specified

below:

In Simak’s (2000) [133] methodology, 𝜆𝜆’s and 𝜃𝜃 represent a new set of coordinates

which determine the position of a DMU in the new space of 𝜆𝜆’s and 𝜃𝜃. The

parameters 𝜆𝜆 and 𝜃𝜃 for each DMU are determined based on the ‘efficient’ peers.

Therefore, for a group of DMUs which do not share identical efficient peers, the 𝜃𝜃

and 𝜆𝜆 factors would not be comparable. This argument clarifies where Anadol’s

(2000) [7] approach conflicts with the theory of DEA: In determining the lower

bound for an ‘efficient’ firm of interest (FOIe), Anadol’s (2000) [7] methodology

draws on any ‘inefficient’ firm whose set of efficient peers includes (but is not

limited to) FOIe. In other words, these inefficient firms are not associated with an

identical set of efficient peers and hence, their 𝜆𝜆 (or 𝜃𝜃) values cannot be examined

against each other.

(ii) The BCC model is a radial model and the radial efficiency score yielded by this

model (i.e., 𝜃𝜃) does not incorporate all sources of inefficiency (see, Chapter 4).

Thus, in the presence of ‘slacks’, the parameter 𝜃𝜃 would not provide a correct

representation of the distance of the respective firm from the efficient frontier.

(iii) The study does not formulate the trade-off between 𝜃𝜃 and 𝜆𝜆. It is not clear how

much of the similarity in 𝜆𝜆s should be forgone to achieve similarity with respect to

𝜃𝜃s. This issue, which arises in Step 3 (see, Page 44), was handled manually and

subjectively by Simak (2000) [133] and Anadol (2000) [7]. In order to obtain an

absolutely objective measure for similarity, the compromise between 𝜃𝜃 and 𝜆𝜆 needs

to be modelled.

The algorithm of Simak (2000) [133] (examined by Anadol (2000) [7]), should be extended

and evolved into a new and more comprehensive and analytically rigorous framework which

Page 79: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.4. Factors Impacting the IPO Price

47

not only addresses the aforementioned limitations but also provides enhanced features and

functionality and ensures a more reliable performance. Section 6.2 details how this thesis

develops a new methodology for selecting comparables using DEA.

3.4. Factors Impacting the IPO Price

Prior to delving into the discussion of choosing comparables, it is essential to understand

which factors impact IPO offer prices as well as post-IPO short-term market prices. If two

firms are similar with respect to the factors deemed to impact investors’ opinion about prices,

they may experience comparable market performances as well. Hence, the outcome of this

discussion is required to underline the factors which would form the basis of the comparison

among different IPOs.

The fundamental challenge in pricing the shares of an IPO or a public company is that there

is no well-defined ‘true’ value for the firm. Even pricing a firm’s assets is not a trivial task.

Usually, every firm has some intangible assets like intellectual property, patents, copyrights,

trademarks, etc. which are not easy to price. The value of a firm (and hence its share price)

depends on the ‘perception’ of the investors about its future performance and profitability

derived based on the information available today.

The value drivers of an IPO firm can be classified in two broad categories, accounting

information and non-financial metrics (Draho, 2004 [48]). Accounting information is available

through financial statements. Non-financial metrics represent pieces of information which are

often hard to quantify or to be directly incorporated into the valuation models; yet, they impact

investors’ opinion about expected financial performance of issuers. The ‘Use of Proceed’

section of a prospectus, for example, provides insight into future business activities; whether

the issuer plans to use the IPO proceeds for acquisition or to invest in an R&D project or to

launch a new product, or to cover past debts would lead to different interpretations of the

firm’s growth prospects. The list can be extended to factors such as the share retention ratio by

Page 80: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.4. Factors Impacting the IPO Price

48

pre-IPO shareholders, underwriter reputation, possible risk factors indicated in the

prospectuses, etc.

Although focused on public firms, Ohlson (1995) [103] was one of the first few scholars who

attempted to model market value of a share as a function of accounting metrics (Draho, 2004

[48]). Assuming ‘clean surplus accounting’, Ohlson (1995) [103] built an autoregressive model to

express the market value of equity. The research indicates that the value of an equity today is

derived based on the book value and residual income information provided in the most recent

financial statements. Any difference between the realized and estimated equity prices is

attributed to ‘non-accounting information’ [48] impacting future residual incomes and is

incorporated into the noise term of the model.

Since the onset of internet industry in 1994 (Bartov et al., 2002 [19]), one possible

classification of the literature on stock valuation/pricing can be internet and non-internet. High

market value ratios (e.g., market-to-book), high revenue growth expectations, and negative

income at the IPO time (particularly in the late 1999) introduce new challenges to the

conventional valuation literature and require diligence.

Using 295 public internet companies from the period of January 1996 to June 2000, Jorion

and Talmor (2001) [81] studied the impact of financial and non-financial data on the market

value of internet companies. They employed different regression models to examine the

impact of explanatory variables of book value, earnings, and web traffic, on the dependent

variable of market value. All variables were assumed to be synchronous. The former two

independent variables represent accounting information and the third explanatory factor is

considered a non-accounting factor. These accounting and non-accounting explanatory factors

are decomposed into their underlying components for a more detailed scrutiny of the relevance

of the aforementioned independent factors to the market value of equity. The variable

earnings, as an example, is decomposed into gross profit, sales, general, and administrative

expenses (SGA) and R&D expenditures. To examine the impact of growth rates on the

explanatory power of the regression models, the study augments the list of explanatory

Page 81: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.4. Factors Impacting the IPO Price

49

variables by the addition of the following two variables: scaled growth in sales and scaled

growth in SGA over the most recent quarter.

The study identifies statistically significant relevance between the market value of equity

and the financial explanatory factors such as gross profit, scaled growth in sales and scaled

growth in SGA over the most recent quarter. However, the non-financial factor (i.e., web

traffic) expressed a more significant impact. The web factor was decomposed into the

following components: total number of times the page was viewed and total number of hours

the page was viewed. With further scrutiny, however, Jorion and Talmor (2001) [81] detected a

declining temporal trend in the impact of the non-financial factors on the market value of

equity and a growing temporal trend in the value relevance of the financial factors.

Considering that the time period of this study coincides with the dot-com bubble period (from

its birth to its burst), such trends indicate that as the frenzy of new economies abates and the

industry matures, the financial factors resume the significance they hold in the conventional

valuation framework.

Hand (2000) [65] reported comparable outcomes to Jorion and Talmor (2001) [81] in his

study of price-relevance of accounting information. He examined the quarterly financial

information of 167 ‘net’ public firms in the 1997-1999 period. His findings indicate a

significant positive relationship between the market value of equity and the following factors:

(i) book equity, (ii) positive net income, (iii) revenue, (iv) selling and marketing expenses or

R&D expenditures given negative net income. Among these factors, the market value of

equity is linearly increasing in the former (i.e., book equity) and log-linearly increasing in the

rest. It should, however, be noted that revenue which is a component of the net income was

found to be weakly related to the market value of equity. In short, it can be concluded that

regardless of the type of the relationship, accounting variables which are indicative of the

operational and structural characteristics of a firm remain the key value-drivers of stocks.

A subsequent more detailed study conducted by Hand (2001) [66] on R&D expenditures of

young biotechnology firms showed a significant positive market reaction to R&D expenditures

Page 82: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.4. Factors Impacting the IPO Price

50

in this industry. The younger the firm is and the faster it grows, the more pronounced this

positive impact is. The evidence from the biotechnology industry is promising as it indicates

that even the market valuation of young biotechnology firms, which have limited transparency

of income and sales, demonstrates significant dependency on financial metrics.

The discussion has so far focused on the key value drivers of stock prices in general which

also apply to the post-IPO market prices. The necessity of such discussion can be outlined as

follows: Any algorithm for identifying comparables would require a set of factors to act as the

basis of comparison. Considering that the comparables would eventually serve to estimate the

distributions of post-IPO market prices for the IPOI, it is essential to ensure that the

foundation of comparison bears significant relevance to post-IPO market prices. Only in this

case can the post-IPO market history of a comparable provide insight into the post-IPO market

status of the IPOI. It should, however, be noted that this study is focused on quantifying the

short-term risk of investing in IPOs. Thus, the time of concern remains close to the issue day.

It is, therefore, interesting to scrutinize the factors which may be influential on the offer price

of an IPO or its immediate post-IPO market prices such as the closing price of the first day. In

their seminal study focused on valuing IPOs using accounting data, Kim and Ritter (1999) [87]

addressed this avenue of the research.

One of the objectives of Kim and Ritter’s (1999) [87] study was to examine a series of

accounting ratios in order to identify those leading to more accurate predictions of IPO offer

price. They studied 190 domestic operating firms over the 1992-1993 period. Firms were

selected to have positive earnings per share (EPS) over the immediately preceding year of the

issue day and positive book value per share (BPSpreissue) prior to the issue day. Kim and Ritter

(1999) [87] adopted the comparable multiples methodology to predict IPO offer prices. Using

least squares linear regression, the study regressed multiples of an IPO (e.g., P/E, M/Bpostissue)

on the same multiples of the respective comparables. The study reported moderate prediction

accuracy for the following multiples: price-to-earnings, price-to-sales, enterprise value-to-

sales, enterprise value-to-operating cash flow, and market-to-book. It is, however, noted that

replacing earnings with the forecast of earnings would increase the prediction accuracy.

Page 83: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.4. Factors Impacting the IPO Price

51

Although Kim and Ritter’s (1999) [87] primary goal was to gauge the accuracy of the

comparable multiples technique in estimating offer prices, their results pose implications for

determining value drivers of equity market prices as well. Based on this study, it can be

concluded that earnings per share, sales per share, book value, and operating cash flow play a

material role in forming investors’ opinion about stock prices. It is worthy of note, however,

that due to the shortcomings associated with ratios, absolute accounting measures are preferred

over ratios (Certo et al., 2009 [33]). Loss of economies of scale and loss of information are

some of the common challenges related to ratios.

Based on his investigation of the pre-IPO underwriter reports of 228 French IPOs,

Roosenboom (2012) [121] provides the following list of the most commonly used multiples by

the underwriters in computing the estimates of ‘fair value’ of IPOs: price-to-earnings, price-to-

cash flow, price-to-sales, price-to-book, enterprise value-to-sales, enterprise value-to-earnings.

His list is in line with Kim and Ritter’s (1999) [87] study; Roosenboom (2012) [121], likewise,

highlights the potentially significant role of the following variables in determining value:

earnings per share, sales per share, book value, and operating cash flow.

In a different setting, Zhu (2000) [141] also recognized profit and revenue as the key factors

impacting the market performance of a company. The primary objective of his research was to

develop a methodology for gauging the financial performance of the Fortune 500 companies

by integrating the existing diverse measures, serving this purpose, into a single framework. He

accomplished this goal using DEA. His two-stage DEA model consists of two sub-models: (1)

profitability model, (2) stock marketability model. The inputs and outputs of each stage can be

taken to discern the key factors impacting a company’s profitability and market performance.

In the first stage, it is assumed that profit and revenue are primarily impacted by the following

three factors: (i) number of employees, (ii) assets, (iii) stockholder’s equity. In the second

model, profit and revenue, in turn, serve as the key factors impacting the market performance

proxied by (i) market value, (ii) total return to investors, (iii) earnings per share.

Page 84: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 3. Literature Review on Methodology: Comparable-Based Methods

3.4. Factors Impacting the IPO Price

52

As aimed at the onset of this section, a search was conducted to assemble key value

indicators from different studies. These factors which can serve as the comparison criteria in

choosing comparables can be summarized as follows:

(i) Revenue

(ii) Gross profit

(iii) Selling, general and administrative expenses

(iv) Net income

(v) Shareholders’ equity

(vi) Operating cash flow

Page 85: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

53

Chapter 4. LITERATURE REVIEW ON METHODOLOGY:

DATA ENVELOPMENT ANALYSIS

4.1. Introduction

This chapter reviews the literature on DEA as pertinent to this research. Section 4.2

provides a brief summary on the basic DEA models, such as CCR (CRS), BCC (VRS), and

SBM. This section aims to build the necessary background for discussing more advanced DEA

models which can cope with negative data and ‘non-discretionary’ factors. These two latter

lines of research are respectively unfolded in Sections 4.3 and 4.4. Section 4.3 presents an

overview of existing methods and recent advances in the field of negative data in DEA. The

outcome of this survey is subsequently utilized in selecting a suitable approach for handling

the mix of negative and positive data in this thesis. Section 4.4 scrutinizes the current models

for non-discretionary factors. This review reveals the present gaps and forms the basis for

addressing the existing methodological barriers. Prior to delving into the details of various

DEA models, the following paragraphs provide a general overview of DEA.

Established by Charnes, Cooper, and Rhodes (Charnes et al., 1978 [36]), Data Envelopment

Analysis (DEA) is a nonparametric linear-fractional programming technique which can be

used to evaluate the relative performance of decision making units (DMUs). The performance

of each DMU is represented by a set of inputs and outputs of the respective DMU. The DMUs

are comparable which implies that the same types of inputs and outputs are used across the

DMUs (Cooper et al., 2007 [41]).

DEA determines a frontier of efficient DMUs that ‘envelopes’ the inefficient DMUs. The

frontier which is called the ‘efficient frontier’ is determined endogenously using the set of all

DMUs whose performances are to be evaluated. DEA identifies a ‘reference set’ located on

the efficient frontier as a benchmark for each DMU. The performance of each DMU is

Page 86: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.1. Introduction

54

evaluated with respect to its reference set. As the outcome of the evaluation, DEA assigns a

performance degree (i.e., ‘degree of efficiency’) to each DMU (Cooper et al., 2007 [41]).

It should be noted that the source and amount of the inefficiency of a DMU are determined

by projecting it on the efficient frontier. The direction of this projection is based on the model

selected and/or the objectives of the analysis. The projection can be input-oriented, output-

oriented or can combine both orientations. The objective of an input-oriented model is to

minimize the inputs without compromising the given level of outputs (i.e., outputs remain at

least at the given levels). An output-oriented model aims to maximize outputs without

exceeding the given level of inputs. Some DEA models such as the additive model or the SBM

model (Sections 4.2.3 and 4.2.4) combine both orientations and simultaneously minimize

inputs and maximize outputs.

The shape of efficient frontier is determined based on the ‘returns to scale’ and convexity

assumptions for the ‘production possibility set’. Production possibility set is a set which

incorporates all the feasible activities (i.e., pairs of semi-positive inputs and outputs denoted

by (𝑥𝑥,𝑦𝑦)). As summarized by Bogetoft and Otto (2011) [30], the production possibility set is

defined to have

(i) constant returns to scale if (𝑥𝑥,𝑦𝑦) ∈ P and 𝛼𝛼 ≥ 0 → (𝛼𝛼𝑥𝑥,𝛼𝛼𝑦𝑦) ∈ P, where P

represents the production possibility set

(ii) decreasing returns to scale if (𝑥𝑥,𝑦𝑦) ∈ P and 0 ≤ 𝛼𝛼 ≤ 1 → (𝛼𝛼𝑥𝑥,𝛼𝛼𝑦𝑦) ∈ P

(iii) increasing returns to scale if (𝑥𝑥,𝑦𝑦) ∈ P and 𝛼𝛼 ≥ 1 → (𝛼𝛼𝑥𝑥,𝛼𝛼𝑦𝑦) ∈ P

(iv) variable returns to scale if 𝛼𝛼 = 1

The most commonly used DEA models can be listed as follows and are discussed next.

(i) Charnes, Cooper, and Rhodes (CCR) model (Charnes et al., 1978 [36])

(ii) Banker, Charnes, and Cooper (BCC) model (Banker et al., 1984 [15])

(iii) Basic additive model (Cooper et al., 2007 [41])

(iv) Slacks-based measure of efficiency (SBM) model (Tone, 2001 [136])

Page 87: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.2. Basic DEA Models

55

To clarify the essence of DEA, the CCR model (Section 4.2.1) is first presented in the form

of a fractional program. It is then converted to a linear program whose dual problem is

presented in this document. The dual problem is referred to as the envelopment form. The

BCC, additive, and SBM models are shown in the envelopment form as well. It should be

noted that the following sections provide a brief overview of the DEA models which form the

basis for the subsequent discussions. The reader is referred to the original studies for further

details.

4.2. Basic DEA Models

4.2.1. CCR Model

Efficiency is often evaluated using Farrell’s measure of efficiency (Farrell, 1957 [51]):

𝜃𝜃 =outputinput

(4.1)

In the case of multiple inputs and multiple outputs, Equation (4.1) can be modified as follows:

𝜃𝜃DMU =virtual outputvirtual input

=weighted sum of outputsweighted sum of inputs

(4.2)

If the weights are set in advance, it would not be clear what fraction of the inefficiency of a

DMU arises from its actual inefficiency and what fraction of it reflects the error of setting

weights. The CCR model developed by Charnes et al., 1978 [36] addresses this problem by

setting variable weights derived endogenously from the data of DMUs; that is, for each DMU,

Page 88: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.2. Basic DEA Models

56

the weights are set as unknowns. Equation (4.3) is then maximized subject to using the same

weights for bounding the efficiency scores of the remaining DMUs to values in the range of

[0,1] (see, Equations (4.4) to (4.6)). The same steps are repeated for each DMU. It is inferred

from this explanation that in addition to finding the weights endogenously, the procedure puts

the performance of each DMU in its best light by setting the corresponding weights such that

the efficiency score is the maximum for that DMU.

To illustrate the above-indicated steps, it is assumed that there are 𝑛𝑛 DMUs (𝑗𝑗 = 1, … ,𝑛𝑛)

which are to be evaluated, and 𝑚𝑚 and 𝑠𝑠 denote the number of inputs and outputs, respectively.

The inputs and outputs of DMU𝑖𝑖 are represented as semipositive vectors of (𝑥𝑥1𝑖𝑖 , … , 𝑥𝑥𝑚𝑚𝑖𝑖) and

(𝑦𝑦1𝑖𝑖 , … ,𝑦𝑦𝑠𝑠𝑖𝑖), respectively. The input and output weights are shown by the 𝑣𝑣 and 𝑢𝑢 vectors,

respectively. The radial efficiency of each DMU is determined using the input and output data

of all the DMUs and after running the optimization presented below:

𝐏𝐏𝐏𝐏𝐏𝐏

max𝑣𝑣,𝑢𝑢

𝜃𝜃𝑜𝑜 =𝑢𝑢1𝑦𝑦1𝑜𝑜 + 𝑢𝑢2𝑦𝑦2𝑜𝑜 + ⋯+ 𝑢𝑢𝑠𝑠𝑦𝑦𝑠𝑠𝑜𝑜𝑣𝑣1𝑥𝑥1𝑜𝑜 + 𝑣𝑣2𝑥𝑥2𝑜𝑜 + ⋯+ 𝑣𝑣𝑚𝑚𝑥𝑥𝑚𝑚𝑜𝑜

(4.3)

Eq. 2.3 [41]

Subject to 𝑢𝑢1𝑦𝑦1𝑖𝑖 + 𝑢𝑢2𝑦𝑦2𝑖𝑖 + ⋯+ 𝑢𝑢𝑠𝑠𝑦𝑦𝑠𝑠𝑖𝑖𝑣𝑣1𝑥𝑥1𝑖𝑖 + 𝑣𝑣2𝑥𝑥2𝑖𝑖 + ⋯+ 𝑣𝑣𝑚𝑚𝑥𝑥𝑚𝑚𝑖𝑖

≤ 1 for 𝑗𝑗 = 1, … ,𝑛𝑛 (4.4)

Eq. 2.4 [41]

𝑣𝑣1, 𝑣𝑣2, … , 𝑣𝑣𝑚𝑚 ≥ 0 (4.5)

Eq. 2.5 [41]

𝑢𝑢1,𝑢𝑢2, … ,𝑢𝑢𝑠𝑠 ≥ 0 (4.6)

Eq. 2.6 [41]

Page 89: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.2. Basic DEA Models

57

The efficiency scores of all the DMUs are computed by running 𝑛𝑛 such optimizations. At each

optimization, the DMU under scrutiny is denoted by DMU𝑜𝑜 and its efficiency score is

represented by 𝜃𝜃𝑜𝑜, where 0 < 𝜃𝜃 ≤ 1.

In a given optimization, if the efficiency score of a DMU is equal to 1 using the calculated

weights, that DMU belongs to the reference set of DMU𝑜𝑜. The efficiency score of DMU𝑜𝑜 (𝜃𝜃𝑜𝑜)

can be equal to 1 as well. A DMU is radially efficient if 𝜃𝜃 is equal to 1 and the model finds at

least one positive pair of weight vectors.

To make the computations less intensive, Equations (4.3) to (4.6) can be converted to a

linear programming problem which is called the ‘multiplier form’ in the literature (see,

Equations 2.7 to 2.11 of Cooper et al., 2007 [41]). The dual problem of the multiplier form

(called the ‘envelopment’ form) is shown below. Note that the production possibility set (P) in

the CCR model is assumed to have the constant returns to scale property discussed in the

preceding section.

Dual problem (Input oriented)

min𝜃𝜃,λ

𝜃𝜃 (4.7)

Eq. 3.6 [41]

Subject to 𝜃𝜃𝑥𝑥𝑜𝑜 − 𝑋𝑋𝜆𝜆 ≥ 0 (4.8)

Eq. 3.7 [41]

𝑌𝑌𝜆𝜆 ≥ 𝑦𝑦𝑜𝑜 (4.9)

Eq. 3.8 [41]

𝜆𝜆 ≥ 0 (4.10)

Eq. 3.9 [41]

Page 90: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.2. Basic DEA Models

58

Here, 𝜆𝜆 ∈ R𝑛𝑛 is a 𝑛𝑛-dimensional semipositive vector. The semipositive input and output

vectors of the DMU under scrutiny (DMU𝑜𝑜) are denoted by 𝑥𝑥𝑜𝑜 and 𝑦𝑦𝑜𝑜, respectively. The

optimal 𝜃𝜃 (𝜃𝜃∗) is a real number from the (0,1] range. The multipliers of 𝜆𝜆, 𝑋𝑋 and 𝑌𝑌, are

defined as matrices which incorporate the data of all the DMUs and can be represented as

follows:

𝑋𝑋 =

⎜⎛𝑥𝑥11𝑥𝑥21

⋮𝑥𝑥𝑚𝑚1

𝑥𝑥12𝑥𝑥22⋮

𝑥𝑥𝑚𝑚2

⋯⋯⋱⋯

𝑥𝑥1𝑛𝑛𝑥𝑥2𝑛𝑛⋮

𝑥𝑥𝑚𝑚𝑛𝑛⎠

⎟⎞

(4.11)

Eq. 2.1 [41]

𝑌𝑌 =

⎜⎛𝑦𝑦11𝑦𝑦21

⋮𝑦𝑦𝑠𝑠1

𝑦𝑦12𝑦𝑦22⋮𝑦𝑦𝑠𝑠2

⋯⋯⋱⋯

𝑦𝑦1𝑛𝑛𝑦𝑦2𝑛𝑛⋮𝑦𝑦𝑠𝑠𝑛𝑛⎠

⎟⎞

(4.12)

Eq. 2.2 [41]

The above-indicated dual problem minimizes the inputs while keeping the outputs at least

at the given levels. The reference set for DMU𝑜𝑜 is a set comprising the DMUs with non-zero

𝜆𝜆s. Equation (4.8) implies that the projection of DMU𝑜𝑜 on the efficient frontier can be found

by reducing all the inputs radially from the origin with the scale of 𝜃𝜃 without sacrificing any

output.

The dual problem just considered shows that reducing all the inputs with the same

proportion may put DMU𝑜𝑜 on a terminating segment of the envelope. The terminating

segments are the extensions of the envelope parallel to the axes in the 𝑚𝑚 + 𝑠𝑠 dimensional

space. A DMU on a terminating segment is the indicative of the remaining input excess or

output shortfall in that dimension.

Page 91: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.2. Basic DEA Models

59

To account for the ‘input excesses’ (𝑠𝑠− ∈ R𝑚𝑚) and ‘output shortfalls’ (𝑠𝑠+ ∈ R𝑠𝑠), the

following extra linear programming (Phase 2) is solved using the outcome of the dual problem

presented by Equations (4.7) to (4.10) (i.e., Phase 1):

max𝜆𝜆,𝑠𝑠+,𝑠𝑠−

𝜔𝜔 = 𝑒𝑒 𝑠𝑠+ + 𝑒𝑒 𝑠𝑠− (4.13)

Eq. 3.11 [41]

Subject to 𝑠𝑠− = 𝜃𝜃∗𝑥𝑥𝑜𝑜 − 𝑋𝑋𝜆𝜆 (4.14)

Eq. 3.12 [41]

𝑠𝑠+ = 𝑌𝑌𝜆𝜆 − 𝑦𝑦𝑜𝑜 (4.15)

Eq. 3.13 [41]

𝜆𝜆 ≥ 0, 𝑠𝑠− ≥ 0, 𝑠𝑠+ ≥ 0

where, 𝑒𝑒 is a vector of ones whose dimension matches its multiplier.

A DMU achieves the Pareto-Koopmans or ‘strong’ efficiency if the optimal solutions of

Phase 1 and Phase 2 satisfy:

(i) 𝜃𝜃∗ = 1 which is referred to as radial efficiency, technical efficiency, or weak

efficiency

(ii) 𝑠𝑠+ = 0 and 𝑠𝑠− = 0

If 𝜃𝜃 is equal to 1 with non-zero slacks, this indicates that all the inputs are reduced radially

from the origin with the same proportion; however, some inputs still show excesses which is

referred to as ‘mix inefficiency’. Accounting for these excesses changes the proportions of the

inputs.

Page 92: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.2. Basic DEA Models

60

Based on the preceding discussions, to improve the performance of the DUM under

scrutiny (DMU𝑜𝑜) using an input-oriented model, first the inputs are minimized radially and

then the input and output slacks are accounted for. Therefore, the projection of DMU𝑜𝑜 on the

frontier is represented by (𝑥𝑥�𝑜𝑜 ,𝑦𝑦�𝑜𝑜), where

𝑥𝑥�𝑜𝑜 = 𝜃𝜃∗𝑥𝑥𝑜𝑜 − 𝑠𝑠−∗ = 𝑋𝑋𝜆𝜆 (4.16)

𝑦𝑦�𝑜𝑜 = 𝑠𝑠+∗ + 𝑦𝑦𝑜𝑜 = 𝑌𝑌𝜆𝜆 (4.17)

Equations (4.16) and (4.17) imply that both the input and output coordinates of the

improved DMU can be expressed as the weighted sum of the respective inputs and outputs of

the DMUs in the reference set. Therefore, 𝜆𝜆 indeed represents the weights assigned to the

DMUs in the reference set.

As indicated, the CCR model can be output oriented which aims to maximize outputs

without increasing the input levels. This model can be derived in the similar fashion to the

CCR input-oriented model described above. More details can be found in Charnes et al. (1978)

[36] or Cooper et al (2007) [41].

4.2.2. BCC Model

As indicated, the CCR model assumes constant returns to scale. The assumption of constant

returns to scale may not hold in practice due to the existing limitations in the production

process in reality. For instance, if a factory owns limited land and facilities, increasing inputs

cannot be translated to increasing outputs indefinitely. To achieve a more realistic production

frontier, Banker, Charnes, and Cooper (Banker et al.,1984 [15]) posit a model whose production

frontier is a piecewise linear convex hull of the given DMUs and exhibits the properties of

variable returns to scale. The only difference between the BCC and CCR models is that the

Page 93: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.2. Basic DEA Models

61

BCC model adds the following equation as an extra constraint to Phase 1 of the dual problem

in the CCR model:

Due to this added constraint, the production frontier includes more efficient DMUs. Therefore,

a CCR-efficient DMU is expected to be BCC-efficient as well.

4.2.3. Additive Model

In using the CCR and BCC models, the analyst is required to select the orientation of the

model in advance. Additive models eliminate this requirement by combining the input and

output orientations. The envelopment form of the basic additive model is shown below:

max𝜆𝜆,𝑠𝑠+,𝑠𝑠−

𝑧𝑧 = 𝑒𝑒 𝑠𝑠+ + 𝑒𝑒 𝑠𝑠− (4.19)

Eq. 4.34 [41]

Subject to 𝑠𝑠− + 𝑋𝑋𝜆𝜆 = 𝑥𝑥𝑜𝑜 (4.20)

Eq. 4.35 [41]

𝑌𝑌𝜆𝜆 − 𝑠𝑠+ = 𝑦𝑦𝑜𝑜 (4.21)

Eq. 4.36 [41]

𝑒𝑒𝜆𝜆 = 1 (4.22)

Eq. 4.37 [41]

�𝜆𝜆𝑖𝑖

𝑛𝑛

𝑖𝑖=1

= 1, 𝜆𝜆𝑖𝑖 ≥ 0 for all 𝑗𝑗′s (4.18)

Page 94: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.2. Basic DEA Models

62

𝜆𝜆 ≥ 0, 𝑠𝑠− ≥ 0, 𝑠𝑠+ ≥ 0 (4.23)

Eq. 4.38 [41]

Based on Equation (4.22), the production possibility set is characterized by variable returns to

scale.

Under the additive model, a DMU is efficient if and only if 𝑠𝑠− = 0 and 𝑠𝑠+ = 0. The

coordinates of the improved version of the inefficient DMU under scrutiny (DMUo), denoted

by 𝑥𝑥�𝑜𝑜 and 𝑦𝑦�𝑜𝑜, can be expressed as follows:

𝑥𝑥�𝑜𝑜 = 𝑥𝑥𝑜𝑜 − 𝑠𝑠−∗ (4.24)

Eq. 4.44 [41]

𝑦𝑦�𝑜𝑜 = 𝑦𝑦𝑜𝑜 + 𝑠𝑠+∗ (4.25)

Eq. 4.45 [41]

The additive model accounts for the input and output inefficiencies directly, and the 𝑧𝑧 in

Equation (4.19) is an additive measure which demonstrates the combined input excesses and

output shortfalls. Despite the efficiency scores of CCR and BCC which only show the redial

efficiency, 𝑧𝑧 has the advantage of incorporating all the sources of inefficiency. It, however,

denotes an absolute measure and despite a relative measure such as 𝜃𝜃, does not reflect the

relative relationship between inputs and outputs.

4.2.4. Slacks-Based Measure of Efficiency

The additive model gauges efficiency in absolute measure which is not ‘dimension free’ or

‘unit invariant’ since changing the units of inputs and outputs impacts the outcomes. The

Page 95: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.2. Basic DEA Models

63

model shown in this section modifies the objective function (i.e., efficiency measure) of the

additive model so that it is unit invariant. The model was first suggested by Pastor et al. (1999)

[111], under the name ‘Enhanced Russell Graph Measure of Efficiency’. Shortly after, in an

independent study, Tone (2001) [136] developed the same model and referred to it as ‘slacks-

based measure of efficiency’ (SBM). Although the model is commonly designated by the

latter name, Enhanced Russell Graph Measure of Efficiency and SBM are indeed equivalent.

The efficiency measure of the model (𝜌𝜌) is defined as the ratio of mean ‘input mix

inefficiencies’ and mean ‘output mix inefficiencies’. It is a real number within the range [0,1]

(i.e., 𝜌𝜌 ∈ [0,1]) and can be demonstrated as follows:

min 𝜌𝜌 =𝜆𝜆, 𝑠𝑠−,𝑠𝑠+

�1 −1𝑚𝑚�

𝑠𝑠𝑖𝑖−

𝑥𝑥𝑖𝑖𝑜𝑜

𝑚𝑚

𝑖𝑖=1

� /�1 +1𝑠𝑠�

𝑠𝑠𝑟𝑟+

𝑦𝑦𝑟𝑟𝑜𝑜

𝑠𝑠

𝑟𝑟=1

= �1𝑚𝑚�

𝑥𝑥𝑖𝑖𝑜𝑜 − 𝑠𝑠𝑖𝑖−

𝑥𝑥𝑖𝑖𝑜𝑜

𝑚𝑚

𝑖𝑖=1

� /�1𝑠𝑠�

𝑦𝑦𝑟𝑟𝑜𝑜 + 𝑠𝑠𝑟𝑟+

𝑦𝑦𝑟𝑟𝑜𝑜

𝑠𝑠

𝑟𝑟=1

(4.26)

Eq. 4.48 [41]

Subject to 𝑠𝑠− + 𝑋𝑋𝜆𝜆 = 𝑥𝑥𝑜𝑜 (4.27)

Eq. 4.48 [41]

𝑌𝑌𝜆𝜆 − 𝑠𝑠+ = 𝑦𝑦𝑜𝑜 (4.28)

Eq. 4.48 [41]

𝜆𝜆 ≥ 0, 𝑠𝑠− ≥ 0, 𝑠𝑠+ ≥ 0 (4.29)

Eq. 4.48 [41]

Under the SBM model, a DMU is efficient if 𝜌𝜌 = 1 which implies zero slacks. Based on

Equations (4.27) to (4.29), the coordinates of the improved activity on the efficient frontier

remain the same as those of the additive model (i.e., Equations (4.24) and (4.25)).

Page 96: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.3. Negative Inputs or Outputs

64

4.3. Negative Inputs or Outputs

The discussion of the DEA models presented so far assumes semipositive data. However, as

is the case in this thesis, the vectors of inputs or outputs may consist of negative data or they

may comprise a mix of negative and positive numbers. As a case in point, consider profit; it is

often treated as an output since more of the variable is desired. Profit can become negative for

the respective period, and there is, therefore, a need for a method which can handle such cases.

The theoretical discussion concerning the topic of negative data in DEA can be considered

to have commenced in mid-1990s, with the work of Lovell and Pastor (1995) [95] and Pastor

(1996) [108]. Since then, a broad range of research projects has emerged, proposing various

techniques to tackle this aspect of DEA. The reader is encouraged to consult Pastor and Ruiz

(2007) [110] and Pastor and Aparicio (2015) [109] for a comprehensive survey on the

advancement of this crucial line of DEA research up until then. This section is devoted to

introducing and reviewing the existing seminal methodologies, in light of the specific

objectives pursued by this thesis. Later, in Section 6.2, it is presented how these methods are

tailored to suit the characteristics of the problem addressed in this document.

4.3.1. Translating Data

One solution to deal with negative inputs and/or outputs that may be encountered in

practice is ‘translating’ them to positive values. In DEA, shifting the inputs and/or outputs by

changing the origin (i.e., translating) results in a new problem. A DEA model is ‘translation

invariant’ if the optimal solution of the new problem remains the same as that of the original

problem (Cooper et al., 2007 [41]). Such model is also called ‘solution invariant’ (Seiford and

Zhu, 2002 [130]).

Within the DEA models introduced so far, the input (output)-oriented BCC model is

translation invariant if only outputs (inputs) are shifted. Another model that can be translation

invariant is the basic additive model which has the convexity constraint of 𝑒𝑒𝜆𝜆 = 1 (Ali and

Page 97: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.3. Negative Inputs or Outputs

65

Seiford, 1990 [5]). The constraint of 𝑒𝑒𝜆𝜆 = 1 is required since only in its presence do the

constraints of the additive model remain unaffected by the shift in the origin. The additive

model is translation invariant with respect to both inputs and outputs. However, it should be

recalled that the efficiency measure yielded by this model is an absolute and unit-sensitive

number. Lacking a unit-invariant efficiency measure with the capacity to show the relative

relationship between inputs and outputs makes this model incompatible with the objectives of

many studies.

Lovell and Pastor (1995) [95] proposed a model which overcomes the unit dependency

problem: normalized weighted additive model. The model was the first DEA model that

integrated both translation and unit invariance. The constraints of the normalized weighted

additive model are identical to the conventional additive model, introduced in Section 4.2.3;

however, the objective function is maximizing the weighted sum of slacks. For each input or

output slack, the weight is set to be the inverse of the sample standard deviation of the

respective input or output. It can immediately be perceived that the objective function is unit

invariant. It is translation invariant, as well since the standard deviation of a sample would not

be affected by translating the data.

Although the normalized weighted additive model of Lovell and Pastor (1995) [95] is both

translation and unit invariant, it suffers from the following shortcoming: It is hard to make

sense of the efficiency measure generated by the model. The figure of weighted sum of slacks,

with weights being assigned as indicated above, does not clarify the relationship between the

outputs and inputs; nor does it provide a direct implication for improving the performance of

an inefficient DMU or determining its distance from the efficiency frontier.

A variety of approaches have been used in the literature to transform data in order to make

them compatible with the existing standard DEA models. As an example, in order to eliminate

the problem of negative data, Lovell (1995) [94] mapped his output indicators onto the scale of

0 to 100. It should be noted that one cannot arbitrarily change the origin of the numbers since

the final outcomes would be sensitive to the amount of the transformation. This point is

Page 98: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.3. Negative Inputs or Outputs

66

illustrated below using the standard input-oriented BCC model, introduced in Section 4.2.2.

As indicated before, this model is not translation invariant with respect to inputs. Suppose, the

inputs and outputs are translated by adding the positive constants 𝜏𝜏 and 𝜉𝜉, respectively. These

constants are sufficiently large to turn all negative numbers into positive numbers. If the

untranslated input and output matrices are respectively denoted by 𝑋𝑋𝑂𝑂𝑂𝑂𝑂𝑂 and 𝑌𝑌𝑂𝑂𝑂𝑂𝑂𝑂, and their

translated counterparts are represented by 𝑋𝑋𝑁𝑁𝑁𝑁𝑁𝑁 and 𝑌𝑌𝑁𝑁𝑁𝑁𝑁𝑁, the latter can be expressed in terms

of the former as shown below:

𝑋𝑋𝑁𝑁𝑁𝑁𝑁𝑁 = 𝑋𝑋𝑂𝑂𝑂𝑂𝑂𝑂 + 𝜏𝜏 (4.30)

𝑌𝑌𝑁𝑁𝑁𝑁𝑁𝑁 = 𝑌𝑌𝑂𝑂𝑂𝑂𝑂𝑂 + 𝜉𝜉 (4.31)

Replacing 𝑋𝑋𝑂𝑂𝑂𝑂𝑂𝑂 and 𝑌𝑌𝑂𝑂𝑂𝑂𝑂𝑂 in the BCC model with the transformed data (i.e., 𝑋𝑋𝑁𝑁𝑁𝑁𝑁𝑁 and

𝑌𝑌𝑁𝑁𝑁𝑁𝑁𝑁 shown by Equations (4.30) and (4.31)) leads to the following equations:

min𝜃𝜃,λ

𝜃𝜃 (4.32)

Subject to 𝜃𝜃𝑥𝑥𝑜𝑜𝑁𝑁𝑁𝑁𝑁𝑁 − 𝜏𝜏𝜃𝜃 − 𝑋𝑋𝑁𝑁𝑁𝑁𝑁𝑁𝜆𝜆 + 𝜏𝜏𝑒𝑒𝜆𝜆 ≥ 0 (4.33)

𝑌𝑌𝑁𝑁𝑁𝑁𝑁𝑁𝜆𝜆 − 𝜉𝜉𝑒𝑒𝜆𝜆 ≥ 𝑦𝑦𝑜𝑜𝑁𝑁𝑁𝑁𝑁𝑁 − 𝜉𝜉 (4.34)

𝑒𝑒𝜆𝜆 = 1 (4.35)

𝜆𝜆 ≥ 0 (4.36)

Page 99: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.3. Negative Inputs or Outputs

67

Applying the convexity constraint (i.e., 𝑒𝑒𝜆𝜆 = 1), the model evolves into the following form:

min𝜃𝜃,λ

𝜃𝜃 (4.37)

Subject to 𝜃𝜃𝑥𝑥𝑜𝑜𝑁𝑁𝑁𝑁𝑁𝑁 − 𝑋𝑋𝑁𝑁𝑁𝑁𝑁𝑁𝜆𝜆 + 𝜏𝜏(1 − 𝜃𝜃) ≥ 0 (4.38)

𝑌𝑌𝑁𝑁𝑁𝑁𝑁𝑁𝜆𝜆 ≥ 𝑦𝑦𝑜𝑜𝑁𝑁𝑁𝑁𝑁𝑁 (4.39)

𝑒𝑒𝜆𝜆 = 1 (4.40)

𝜆𝜆 ≥ 0 (4.41)

This model is not solution invariant since despite the rest of the equations, Equation (4.38)

has not preserved its old format (i.e., Equation (4.8)); the data transformation has entered one

extra term into this equation (i.e., 𝜏𝜏(1 − 𝜃𝜃)) which changes the optimization results. In other

words, the 𝜆𝜆 and 𝜃𝜃 values do not remain intact as a result of changing the origin or translating

the data. Furthermore, Equation (4.38) shows that the final outcomes would be sensitive to the

size of the transformation (i.e., 𝜏𝜏 in this case).

4.3.2. Treating Negative Outputs (Inputs) as Positive Inputs

(Outputs)

Another possible, and perhaps the simplest, approach is to treat negative outputs as positive

inputs, and vice versa. This approach emerged as a possible solution for dealing with the cases

of undesirable outputs or desirable inputs (e.g., Koopman, 1951 [88]; Scheel, 2001 [126]; Seiford

and Zhu, 2002 [130]). In such cases, the efficient units are expected to achieve lower levels of

undesirable outputs like the amount of carbon dioxide produced by a process or higher levels

of desirable inputs, such as the magnitude of toxic chemicals fed into a filter for treatment.

Such settings contrast with the conventional DEA framework where the true production

Page 100: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.3. Negative Inputs or Outputs

68

process demands the outputs to be maximized and the inputs to be minimized. One approach

to address this problem is to design the DEA model such that it presents the undesirable

outputs as inputs. Similarly, the desirable inputs can be treated as outputs. The production

possibility set of this DEA model would be identical to that of the case which keeps the

negative equivalents of the undesirable outputs on the output side or similarly, continues to use

the desirable inputs as inputs but with negative signs (Scheel, 2001 [126]).

Drawing upon the above-described approach, one may treat negative outputs as positive

inputs, and vice versa. If a vector of input or output consists of a mix of positive and negative

numbers, this approach would require dividing the respective vector into two sub-vectors. One

of these sub-vectors would hold the positive numbers and fill in the place of negative values

with zero (or a very small positive value). Following the same format, the other one would

keep the absolute values of the negative elements and substitute zero (or a very small positive

value) in the place of positive numbers. The context would then dictate which sub-vector

needs to be maximized (minimized), and hence, should reside on the output (input) side.

This approach may be considered adequate if the entire vector of an input or output is

consistently negative. However, in cases where the vector comprises a mix of negative and

positive numbers, this approach would not be as reliant. The approach suffers from significant

theoretical shortcomings which are detailed in Section 6.2.3.1.

4.3.3. Semi-Oriented Radial Measure

The model of semi-oriented radial measure (SORM) suggested by Emrouznejad et al.

(2010) [49] is closely related to the approach discussed in the previous section which will be

referred to as ‘N-to-P’, henceforth, for the sake of brevity. In an identical manner to the N-to-P

approach, SORM splits the vector of input or output - which contains a mix of positive and

negative numbers - into two sub-vectors. The SORM model, however, handles the constraints

somewhat differently. The difference can best be explained using the mathematical

Page 101: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.3. Negative Inputs or Outputs

69

demonstration of the model. Below, the input-oriented SORM model is presented which

appears as ‘Model 8’ in the work of Emrouznejad et al. (2010) ) [49]:

Input-Oriented SORM Model (VRS)

min𝜃𝜃,λ

𝜃𝜃 (4.42)

Model 8 [49]

Subject to 𝑋𝑋+𝜆𝜆 ≤ 𝜃𝜃𝑥𝑥𝑜𝑜+ (4.43)

𝑋𝑋−1𝜆𝜆 ≤ 𝜃𝜃𝑥𝑥𝑜𝑜−1 (4.44)

𝑋𝑋−2𝜆𝜆 ≥ 𝜃𝜃𝑥𝑥𝑜𝑜−2 (4.45)

𝑌𝑌+𝜆𝜆 ≥ 𝑦𝑦𝑜𝑜+ (4.46)

𝑌𝑌−1𝜆𝜆 ≥ 𝑦𝑦𝑜𝑜−1 (4.47)

𝑌𝑌−2𝜆𝜆 ≤ 𝑦𝑦𝑜𝑜−2 (4.48)

𝑒𝑒𝜆𝜆 = 1 (4.49)

𝜆𝜆 ≥ 0 (4.50)

Note that the notation of the original model is modified to ensure that it agrees with the

notation of the previously introduced DEA models. The reader is referred to the original paper

for the output-oriented SORM model and further details.

Above, 𝑋𝑋− denotes a 𝑙𝑙 × 𝑛𝑛 matrix of inputs. The constant 𝑛𝑛 stands for the number of

DMUs and 𝑙𝑙 represents the number of inputs which contain a mix of positive and negative

values. Also, suppose 𝑋𝑋+ is a 𝑚𝑚 × 𝑛𝑛 matrix identical to matrix 𝑋𝑋 shown in Equation (4.11)

Page 102: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.3. Negative Inputs or Outputs

70

which only includes semi-positive inputs with 𝑚𝑚 representing the number of positive inputs.

The vector of 𝑥𝑥𝑜𝑜− (𝑥𝑥𝑜𝑜+) with the size of 𝑙𝑙 × 1 (𝑚𝑚 ×1) comprises the inputs corresponding to

DMUo selected from 𝑋𝑋− (𝑋𝑋+).

The matrices 𝑌𝑌− and 𝑌𝑌+, with the dimensions of 𝑘𝑘 × 𝑛𝑛 and 𝑠𝑠 × 𝑛𝑛, respectively, denote

equivalents of 𝑋𝑋− and 𝑋𝑋+ for the outputs. Similarly, the vectors of 𝑦𝑦𝑜𝑜− and 𝑦𝑦𝑜𝑜+, with the

lengths of 𝑘𝑘 and 𝑠𝑠, can be considered as the counterparts of 𝑥𝑥𝑜𝑜− and 𝑥𝑥𝑜𝑜+, shown above.

The matrix 𝑋𝑋−1 is constructed by keeping all the positive values in 𝑋𝑋− and substituting the

negative numbers with zeros. The matrix 𝑋𝑋−2, on the other hand, is formed by replacing all

the negative values in 𝑋𝑋− with the corresponding absolute values and converting the rest of the

indices to zero. The vectors of 𝑥𝑥𝑜𝑜−1 and 𝑥𝑥𝑜𝑜−

2 represent the columns corresponding to DMUo

in the matrices of 𝑋𝑋−1 and 𝑋𝑋−2, respectively. The matrices of 𝑌𝑌−1 and 𝑌𝑌−2 and the vectors of

𝑦𝑦𝑜𝑜−1 and 𝑦𝑦𝑜𝑜−

2 can be likewise interpreted for the outputs.

The rest of the notation remains the same as those presented under ‘Basic DEA Models’ in

the onset of this Chapter (see, Section 4.2).

It is readily observable that this model resembles the N-to-P approach which allocates

negative and positive numbers to individual sub-vectors (see, Section 4.3.2). There, however,

exists one important difference. In the N-to-P approach, once the sub-vectors are formed, the

problem is transformed to a new one where the vector of mixed negative and positive values is

omitted and is replaced with two new sub-vectors. The problem essentially gains new inputs

and outputs which are treated as ordinary inputs and outputs with no reference to their origins.

It is important to settle the fact that when the N-to-P approach is used, from the mathematical

point of view, any DEA model would not distinguish between the new sub-vector created on

the input side and the all-positive input vectors which were originally resided on the input

side. The same point holds for the outputs. In the SORM model, on the other hand, the sub-

Page 103: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.3. Negative Inputs or Outputs

71

vectors are not simply treated as ordinary semi-positive inputs or outputs. If this were the case,

Equation (4.45) would have the following format:

𝑋𝑋−2𝜆𝜆 ≥ 𝑥𝑥𝑜𝑜−2 (4.51)

as opposed to

𝑋𝑋−2𝜆𝜆 ≥ 𝜃𝜃𝑥𝑥𝑜𝑜−2 (4.45)

Equation (4.45) implies that the SORM model chooses to radially contract the absolute value

of a negative input and then, it searches for a target which is superior to this figure. The same

argument applies to the negative outputs in the output-oriented SORM.

Cheng et al. (2013) [38] have observed cases where the SORM model sets worse targets for

the inefficient units. The point just illustrated may be one cause for this matter. Equation

(4.45) indicates that in case of negative inputs where the model aims to maximize the absolute

values, the model may identify a target whose absolute value is greater than the contracted

value of the corresponding input of DMUo; yet, it is possible that this target does not exceed

the uncontracted value of the input. In other words, while Equation (4.45) holds, Equation

(4.51) may not always be true. When Equation (4.45) is not satisfied, the absolute value of the

target’s respective input would be less than the absolute value of the corresponding input used

by DMUo which implies that the target indeed consumes more input.

Considering the close resemblance of the SORM model to the N-to-P approach, it

incorporates all the shortcomings imposed by disaggregating the original input or output

vector. These limitations are discussed in details in Section 6.2.3.1.

Page 104: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.3. Negative Inputs or Outputs

72

4.3.4. Variant of Radial Measure

Put forth by Cheng et al. (2013) [38], the variant of the radial measure (VRM) indeed extends

the mathematics of conventional radial DEA models (i.e., CCR and BCC) so that they can

handle negative data, as well. The core of the model is to radially contract (expand) the

absolute values of the inputs (outputs) in order to measure the absolute value of the respective

possible improvement (i.e., input decrease or output increase). This would prevent the model

from identifying worse targets as may be the case with the methodology developed by

Emrouznejad et al. (2010) [49] (see, Section 4.3.3). It should be noted that the model proposed

by Cheng et al. (2013) [38] bears a close resemblance to the ‘generalised proportional distance

function’ suggested by Kerstens and De Woestyne (2011) [84]. The two studies, which seem to

be conducted independently and almost concurrently, have arrived at the same solution to the

following problem: How can the traditional radial DEA models be modified to accommodate

negative values?

In order to clarify the specific contribution of the model, this thesis presents the input-

oriented version of the model which satisfies the VRS condition. The reader is referred to the

original paper for the output-oriented model. Cheng et al. (2013) [38] substitutes the efficiency

score of BCC (i.e., 𝜃𝜃) with (1 − 𝛽𝛽), where 𝛽𝛽 can be considered a measure of inefficiency.

Such modification restructures the BCC model into the following form (Equation 7 in Cheng

et al., 2013 [38]):

max𝛽𝛽,λ

𝛽𝛽 (4.52)

Eq. 7 [38]

Subject to 𝛽𝛽𝑥𝑥𝑜𝑜 + 𝑋𝑋𝜆𝜆 ≤ 𝑥𝑥𝑜𝑜 (4.53)

𝑌𝑌𝜆𝜆 ≥ 𝑦𝑦𝑜𝑜 (4.54)

Page 105: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.3. Negative Inputs or Outputs

73

𝑒𝑒𝜆𝜆 = 1 (4.55)

𝜆𝜆 ≥ 0 (4.56)

As indicated, in the input-oriented model of Cheng et al. (2013) [38], the absolute values of the

inputs are radially reduced. Therefore, Equation (4.53), shown above, is replaced with the

following equation:

𝛽𝛽 �𝑥𝑥𝑜𝑜� + 𝑋𝑋𝜆𝜆 ≤ 𝑥𝑥𝑜𝑜

In its complete form, the model can be demonstrated as below:

Input-Oriented VRM Model (VRS)

max𝛽𝛽,λ

𝛽𝛽 (4.57)

Eq. 9 [38]

Subject to 𝛽𝛽 �𝑥𝑥𝑜𝑜� + 𝑋𝑋𝜆𝜆 ≤ 𝑥𝑥𝑜𝑜 (4.58)

𝑌𝑌𝜆𝜆 ≥ 𝑦𝑦𝑜𝑜 (4.59)

𝑒𝑒𝜆𝜆 = 1 (4.60)

𝜆𝜆 ≥ 0 (4.61)

In the case of positive inputs, this model is identical to the traditional BCC model. For

negative inputs, Equation (4.58) can be written as follows:

Page 106: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.3. Negative Inputs or Outputs

74

−𝛽𝛽𝑥𝑥𝑜𝑜 + 𝑋𝑋𝜆𝜆 ≤ 𝑥𝑥𝑜𝑜

→ 𝑋𝑋𝜆𝜆 ≤ (1 + 𝛽𝛽)𝑥𝑥𝑜𝑜 (4.62)

Bearing in mind that Equation (4.62) concerns negative numbers, it is readily observable that

if the target’s respective input (i.e., 𝑋𝑋𝜆𝜆) is smaller than a greater-than-one multiple of the

DMUo’s corresponding negative input (i.e., (1 + 𝛽𝛽)𝑥𝑥𝑜𝑜), it cannot exceed 𝑥𝑥𝑜𝑜, either.

With a similar argument, it can be proven that the output-oriented model always identifies

better-performing targets and it can, therefore, be concluded that the VRM model would

always lead to efficient targets with superior performance.

Although the VRM model is capable of providing correct directions for improving DMUo’s

performance, it suffers from a significant shortcoming: the model does not comply with the

‘monotonicity’ rule; that is, the efficiency score of a DMU is expected to monotonically

decrease with its distance from the corresponding target on the efficient frontier. Under the

VRM approach, it is possible that a DMU requiring a greater improvement to reach its

efficient target is identified as more efficient than a DMU which was located closer to its

efficient target. It should be noted that this issue only arises in the cases where input or output

vectors contain mixed signs. Otherwise, if input or output vectors are purely positive or

negative, the ‘monotonicity’ rule holds.

4.3.5. Range Directional Models

The range directional models and the SORM (Emrouznejad et al., 2010) [49] and VRM

(Cheng et al., 2013) [38] models, introduced above, can be considered two different lines of

methodologies which deal with negative data. The former (i.e., range directional models)

precede the latter (i.e., SORM and VRM models) in the timeline of the approaches developed

to handle negative data in DEA. Yet, they do not encompass the shortcomings associated with

the SORM and VRM models (see, Sections 4.3.3 and 4.3.4). In the cases of SORM and VRM,

researchers indeed aimed to pioneer a different line of methodologies which are capable of

Page 107: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.3. Negative Inputs or Outputs

75

coping with negative data, and yet, do not require the change of origin (as is the case with

directional distance models). However, further research is required to overcome the limitations

of SORM and VRM and to make them applicable to a broader range of situations. The range

directional models, on the other hand, offer a more comprehensive framework to base the

analysis on and to expand upon.

Generic directional distance functions were first studied by the economists in the following

research: Chung et al. (1997) [39] and Chambers et al., (1996 and 1998) [34,35]. Motivated by

these directional distance functions, Silva Portela et al. (2004) [132] developed a range

directional model to address cases where some data may be negative within the DEA context.

Their range directional model, referred to as RDM, is unit-invariant, and it can

straightforwardly be proven that if the convexity assumption holds, it is translation-invariant,

as well. The model is presented below. Except where noted, the notation is consistent with the

basic DEA models, introduced in Section 4.2.

A Non-Oriented Range Directional Model

max𝛽𝛽,λ

𝛽𝛽 (4.63)

Eq. 1 [132]

Subject to 𝑋𝑋𝜆𝜆 + 𝛽𝛽𝑔𝑔𝑥𝑥𝑜𝑜 ≤ 𝑥𝑥𝑜𝑜 (4.64)

𝑌𝑌𝜆𝜆 − 𝛽𝛽𝑔𝑔𝑦𝑦𝑜𝑜 ≥ 𝑦𝑦𝑜𝑜 (4.65)

𝑒𝑒𝜆𝜆 = 1 (4.66)

𝜆𝜆 ≥ 0 (4.67)

Page 108: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.3. Negative Inputs or Outputs

76

Silva Portela et al. (2004) [132] refer to the vectors of 𝑔𝑔𝑥𝑥𝑜𝑜 and 𝑔𝑔𝑦𝑦𝑜𝑜 as the ‘ranges of possible

improvement’ for DMUo. Each member of these vectors is defined as follows:

𝑔𝑔𝑥𝑥𝑖𝑖𝑜𝑜 = 𝑥𝑥𝑖𝑖𝑜𝑜 − min𝐽𝐽�𝑥𝑥𝑖𝑖𝑟𝑟�

(4.68)

𝑔𝑔𝑦𝑦𝑟𝑟𝑜𝑜 = max𝐽𝐽

�𝑦𝑦𝑟𝑟𝑟𝑟� − 𝑦𝑦𝑟𝑟𝑜𝑜 (4.69)

where, 𝑟𝑟 = 1, … , 𝑠𝑠

𝑖𝑖 = 1, … ,𝑚𝑚 𝑗𝑗 = 1, … , 𝐽𝐽

Consistent with the basic DEA models (see, Section 4.2), here, 𝑚𝑚 and 𝑠𝑠 denote the counts of

inputs and outputs, respectively. The number of the DMUs participating in the analysis is

represented by 𝐽𝐽.

In non-mathematical terms, the range for each specific input (output) of DMUo is defined as

the absolute value of the difference between that specific input (output) and the minimum

(maximum) value observed across all the DMUs for the respective input (output). The range

indeed shows the maximum improvement that DMUo can achieve. If for any input (output),

DMUo exhibits the minimum (maximum) value across all the DMUs, the respective range

would be zero for DMUo.

As can be inferred from the equations shown above, 𝛽𝛽 represents the fraction of the

maximum possible slack that is attained by DMUo at its optimum state. It can, therefore, be

considered an inefficiency score, and consequently, the RDM efficiency score is determined

by ‘1 − 𝛽𝛽’. To make even more sense, note the similarity between this model and the

reconfigured BCC model (Equations (4.52) to (4.56)) if 𝑔𝑔𝑥𝑥𝑜𝑜 and 𝑔𝑔𝑦𝑦𝑜𝑜 are respectively

substituted with the corresponding observed inputs and outputs of DMUo (i.e., 𝑥𝑥𝑜𝑜 and 𝑦𝑦𝑜𝑜). It

Page 109: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.3. Negative Inputs or Outputs

77

should be noted that the non-oriented model presented above can be converted to an input-

oriented (output-oriented) model by setting the coefficient of 𝛽𝛽 to ‘0’ in Equation (4.65)

(Equation (4.64)).

In short, the RDM model is a unit- and translation-invariant radial model with the

efficiency score in the range of 0 to 1. The essence of the RDM model is that it chooses the

intersection of the minimums of all the inputs and the maximums of all the outputs (called the

‘ideal point’) as the origin instead of the origin of ‘0’. To elaborate more, consider a case with

one input and one output where the minimum observed input and the maximum observed

output are respectively -2 and 10, across all the DMUs. The RDM model is designed such that

the point (-2,10), as opposed to (0,0), is considered as the reference of the efficiency

calculations.

Similar to other traditional radial DEA models like BCC and CCR, the radial efficiency

score yielded by RDM does not incorporate all sources of inefficiency. In other words, the

efficiency score of ‘1’ does not automatically translate into the Pareto-Koopmans or ‘strong’

efficiency (see, Section 4.2). This shortcoming sparks the idea of developing a purely slacks-

based measure of efficiency for RDM. The slacks-based model could be considered as the

equivalent of SBM (Tone, 2001 [136]) in the traditional DEA context. It would, therefore, offer

similar advantages over RDM as does SBM over BCC or CCR. Sharp et al. (2007) [131] has

addressed this gap by proposing the ‘modified slacks-based measure’ (MSBM).

The MSBM model differs from the SBM model in the efficiency score or objective

function. In the former’s efficiency score, slacks are scaled using ‘ranges of possible

improvement’ (i.e., 𝑔𝑔𝑥𝑥𝑖𝑖𝑜𝑜 and 𝑔𝑔𝑦𝑦𝑟𝑟𝑜𝑜); whereas, in the efficiency score of the latter model, slacks

are scaled using the realized value of the respective input or output of DMUo. The efficiency

score of MSBM diverges from that of SBM in one more aspect: the efficiency score yielded

by the conventional SBM assigns equal weights to the slacks; whereas, MSBM has the

capacity to incorporate user-defined weights into the efficiency score. The following equations

clarify the mathematical structure of MSBM:

Page 110: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.3. Negative Inputs or Outputs

78

Modified Slacks-Based Measure

min 𝜌𝜌 =𝜆𝜆, 𝑠𝑠−,𝑠𝑠+

�1 − ∑ 𝜔𝜔𝑖𝑖𝑠𝑠𝑖𝑖−𝑔𝑔𝑥𝑥𝑖𝑖𝑜𝑜

𝑚𝑚𝑖𝑖=1 �

�1 + ∑ 𝜗𝜗𝑟𝑟𝑠𝑠𝑟𝑟+𝑔𝑔𝑦𝑦𝑟𝑟𝑜𝑜

𝑠𝑠𝑟𝑟=1 �

(4.70)

Eq. 6 [131]

Subject to 𝑠𝑠− + 𝑋𝑋𝜆𝜆 = 𝑥𝑥𝑜𝑜 (4.71)

𝑌𝑌𝜆𝜆 − 𝑠𝑠+ = 𝑦𝑦𝑜𝑜 (4.72)

𝑒𝑒𝜆𝜆 = 1 (4.73)

𝑒𝑒𝜔𝜔 = 1 (4.74)

𝑒𝑒𝜗𝜗 = 1 (4.75)

𝜆𝜆 ≥ 0, 𝑠𝑠− ≥ 0, 𝑠𝑠+ ≥ 0

𝜔𝜔 ≥ 0, 𝜗𝜗 ≥ 0 (4.76)

Here, the variables represent the same factors as the previous models. The coefficients 𝜔𝜔 and

𝜗𝜗 denote the user-fed weights associated with the slacks. Note that the ranges 𝑔𝑔𝑥𝑥𝑖𝑖𝑜𝑜 and 𝑔𝑔𝑦𝑦𝑟𝑟𝑜𝑜

remain identical to the RDM ranges defined above (see, Equations (4.68) and (4.69)).

It is worthy of note that the efficiency score of RDM (i.e., ‘1 − 𝛽𝛽’) is greater than or equal

to the efficiency score of MSBM (see, Eq. 8 in Sharp et al., 2007 [131]). In intuitive terms, since

the MSBM model accounts for all sources of inefficiency using slacks, it generally yields

lower efficiency scores.

As final remarks on the topic of the range directional models, it is worthwhile to point out

that similar to the conventional setting of DEA, where an additive model is considered a close

cousin of SBM, in the context of range (directional) models, the following additive model falls

Page 111: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.4. Non-Discretionary Variables

79

close to MSBM in terms of the their capacity to fulfil the unit-invariance and translation-

invariance properties: Bounded Adjusted Measure of Efficiency or BAM (Cooper et al., 2011

[40]). It is a weighted additive model, where the objective function represents a weighted sum

of slacks. For a given input or output, the weight is set to be the inverse of the product of the

respective range of possible improvement and the total number of inputs and outputs. In

comparing BAM and MSBM, it should be noted that the figure of weighted sum of slacks in

BAM does not clarify the relationship between the outputs and inputs; nor does it provide a

direct implication for improving the performance of an inefficient DMU or determining its

distance from the efficiency frontier. However, these shortcomings can be considered

negligible in the scope of many projects and BAM can serve as an alternative to MSBM.

In conclusion, different approaches have been suggested to cope with negative inputs and

outputs in DEA. Considering the merits and limitations of each technique, discussed in details

above, the range directional models (i.e., RDM and MSBM) offer a more advanced framework

to handle negative data, particularly in the cases where the vector of an input or an output

comprises a mix of positive and negative values. To put into very brief terms, despite the other

DEA models presented in this section, RDM and MSBM hold the characteristic of translation

invariance while preserving the unit-invariance feature as well as the true production process.

Furthermore, the efficiency scores yielded by these models can readily be interpreted similar

to the efficiency scores of the conventional DEA models, such as BCC, CCR and SBM. The

reader is referred to the details presented above for a comprehensive discussion on the virtues

and flaws of the existing methods.

4.4. Non-Discretionary Variables

In the DEA research, there may exist variables which do impact the performance of

different DMUs; yet, they are ‘exogenously fixed’ (Cooper et al., 2007 [41]) and cannot be

varied at the discretion of the analyst or the management team. To contemplate the

performance analysis from all sides, such variables, referred to as non-discretionary (ND)

Page 112: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.4. Non-Discretionary Variables

80

variables, should be taken into account. As a case in point, in terms of sales performance, a

supermarket with few competitors, located at a densely-populated residential region would be

at advantage compared to one of its kind based at a sparsely-populated residential area

competing with more rivals. In such cases, the performance should be assessed in light of

these exogenous variables which cannot be modified or improved but do impact the conduct

and outcome of the operations.

The conventional DEA research does not directly associate the ND variables with the

efficiency score; that is, in a slacks-based model, as an example, the slacks of the ND variables

do not appear in the objective function (i.e., the efficiency score). In a radial model, as another

example, the problem does not pursue the radial improvement of the ND variables. This

restriction is justified by the viewpoint that it would not eventually be possible to enhance the

ND variables based on the suggestions provided by the efficiency scores. The DEA research,

however, accounts for the impact of the ND variables by allowing them to act as constraints in

the optimization problem. The ND variables, therefore, restrict the production possibility set,

and consequently, modify the efficiency score indirectly via their presence in the constraints.

To make it applicable to the cases with ND factors, Banker and Morey (1986) [16] modified

the traditional two-phase BCC as follows:

Non-Discretionary Input-Oriented BCC

min 𝜃𝜃 − 𝜀𝜀 ��𝑠𝑠𝑖𝑖−

𝑖𝑖∈D

+ �𝑠𝑠𝑟𝑟+𝑠𝑠

𝑟𝑟=1

(4.77)

Model 7.1 [41]

or Eq. 19 [16]

Subject to 𝜃𝜃𝑥𝑥𝑖𝑖𝑜𝑜 = �𝑥𝑥𝑖𝑖𝑖𝑖𝜆𝜆𝑖𝑖

𝑛𝑛

𝑖𝑖=1

+ 𝑠𝑠𝑖𝑖−, 𝑖𝑖 ∈ D (4.78)

Page 113: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.4. Non-Discretionary Variables

81

𝑥𝑥𝑖𝑖𝑜𝑜 = �𝑥𝑥𝑖𝑖𝑖𝑖𝜆𝜆𝑖𝑖

𝑛𝑛

𝑖𝑖=1

+ 𝑠𝑠𝑖𝑖−, 𝑖𝑖 ∈ ND (4.79)

𝑦𝑦𝑟𝑟𝑜𝑜 = �𝑦𝑦𝑟𝑟𝑖𝑖𝜆𝜆𝑖𝑖

𝑛𝑛

𝑖𝑖=1

− 𝑠𝑠𝑟𝑟+ (4.80)

�𝜆𝜆𝑖𝑖

𝑛𝑛

𝑖𝑖=1

= 1 (4.81)

𝜆𝜆𝑖𝑖 ≥ 0, 𝑗𝑗 = 1, … ,𝑛𝑛 (4.82)

𝑠𝑠𝑖𝑖− ≥ 0, 𝑖𝑖 = 1, … ,𝑚𝑚

𝑠𝑠𝑟𝑟+ ≥ 0, 𝑟𝑟 = 1, … , 𝑠𝑠

The model presented above is input-oriented. The output-oriented model with ND outputs

is readily attainable following the same notion. The model preserves the previous notation in

this document. The only difference is that the vector format has been expanded into

summation terms in order to better demonstrate the modifications applied to account for the

ND inputs. Moreover, the coefficient 𝜀𝜀 in the objective function indicates that slacks are

maximized in the second phase. To elaborate more, the first phase minimizes 𝜃𝜃 which is then

used as a constant in the second phase where the slacks are maximized. Note that the objective

function does not encompass the ND slacks. In addition, the problem does not allow the ND

inputs to be radially reduced.

With minor changes, Banker and Morey’s (1986) [16] model can be extended such that it

handles ND inputs and outputs simultaneously. To this end, Equation (4.77) can be modified

to only incorporate the slacks of discretionary outputs and Equation (4.80) can be split into

two equations, addressing the discretionary and non-discretionary outputs separately but

preserving the same format. Such extensions have also been documented by Golany and Roll

Page 114: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.4. Non-Discretionary Variables

82

(1993) [61]. The authors also modified Banker and Morey’s (1986) [16] model so that it can

quantify ‘partial discretion’. The binary approach of either discretionary or non-discretionary

may not be applicable in all frameworks. The working hours is one example provided by the

authors. Managers can exercise some control over the number of hours. They are, however,

obligated to comply with the profession’s standards and occupational guidelines. To express

this principle in mathematical terms, Golany and Roll (1993) [61] restructured Equations (4.77)

to (4.82) and introduced two new variables which account for the level of discretion in inputs

and outputs. The two variables vary over the range of 0 to 1, with 1 and 0 respectively

assigned to fully discretionary and non-discretionary factors. Further details in this regard are

beyond the scope of this research and the reader is referred to the original study for further

details.

If the ND slacks in Banker and Morey’s (1986) [16] models are forced to be zero (i.e. the ND

constraints are ‘tight’), the model is then called a ‘non-controllable’ model; otherwise, it is

referred to as a ‘non-discretionary’ model. To put a ND slack of zero into perspective,

consider the objective function of the multiplier (i.e., dual) equivalent of Banker and Morey’s

(1986) [16] model:

max𝑣𝑣,𝑢𝑢

�𝑢𝑢𝑟𝑟𝑦𝑦𝑟𝑟𝑜𝑜

𝑠𝑠

𝑟𝑟=1

−� 𝑣𝑣𝑖𝑖𝑥𝑥𝑖𝑖𝑜𝑜𝑖𝑖∈ND

(4.83)

Model 7.7 [41]

Based on the dual theorem of the linear programming, 𝑠𝑠𝑖𝑖− = 0 where 𝑖𝑖 ∈ ND, translates into 𝑣𝑣𝑖𝑖

not equal to ‘0’ and implies that the objective function would be reduced by the amount of

𝑣𝑣𝑖𝑖𝑥𝑥𝑖𝑖𝑜𝑜. To explain in non-technical terms, a reduction in the ND input would lead to an

enhanced performance. A case in point could be excessive rainfall, causing a serious loss of

crops. The excessive rainfall is considered a non-discretionary input whose increase would be

destructive (Cooper et al., 2007 [41]). A non-discretionary slack can assume a non-zero

Page 115: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.4. Non-Discretionary Variables

83

positive number, as well which implies that any excess in the respective ND input does not

cause any conflict with the performance.

Ruggiero (1996) [124] modified the models of Banker and Morey (1986) [16] in the following

manner: any DMU with more favourable non-discretionary inputs or outputs compared to

DMUo is excluded from the program of efficiency calculation for DMUo. This goal is achieved

by setting the 𝜆𝜆s of the corresponding DMUs to be zero. The objective of Ruggiero’s (1996)

[124] study - which was purely concerned with the calculation of technical efficiency - was to

build a model which sets the closest references for DMUo in terms of non-discretionary

factors. In other words, the study sought to identify, as efficient references, those DMUs that

operated under a similar or, perhaps, inferior environment. Ruggiero’s (1996) [124] approach is

supported by the viewpoint that it may not be feasible for DMUo to achieve the status of an

efficient target which operates under stronger exogenously fixed factors.

The efficiency estimation yielded by the methodology of Ruggiero (1996) [124], however,

deteriorates as the number of ND factors increases and the model tends to identify more

DMUs as efficient. Such sensitivity may be caused by the fact that with increasing number of

ND variables, more DMUs are possibly eliminated from the production possibility set which

makes DMUo more likely to be identified as efficient. Ruggiero (1998) [123] offered a three

stage DEA model to alleviate this shortcoming. Prior to unfolding that model, it should be

noted that the viewpoint pursued by Ruggiero (1996) [124] can also be achieved by the model of

Banker and Morey (1986) [16], if the ND constraints are ‘tight’ as explained above.

The first stage in the model of Ruggiero (1998) [123] involves using BCC with discretionary

factors to compute the efficiency score of the DMUs. In the second stage, the obtained

efficiency scores are linearly regressed on the exogenously fixed factors. This stage indeed

allocates a cross-sectional weight to each ND factor specifying its contribution to the yielded

efficiency scores in the first stage. For each DMU, the second stage finally computes a single

consolidated non-discretionary figure which is a weighted sum of all the ND factors with

Page 116: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.4. Non-Discretionary Variables

84

weights as yielded by the regression. Only DMUs whose weighted sum is inferior to that of

DMUo would be preserved in the optimization program of DMUo.

Further contributions coping with ND factors in the framework of radial DEA models can

be observed in the works of Yang and Paradi (2006) [140] and Muñiz et al. (2006) [100]. In line

with the goals pursued in Ruggiero (1996 and 1998) [124,123], the foremost objective of these

studies is to compute a measure of efficiency which is best deprived of the impact of

exogenously fixed factors, and hence, reflects the true technical efficiency of the DMU of

interest. To achieve this goal, the methodologies depend on a series of assumptions. In the

model of Ruggiero (1998) [123], for example, regression is used to assign cross-sectional

weights to the ND factors. Therefore, it is assumed a priori that the ND factors are linearly

related to the efficiency score. Such assumptions expose the analysis to new sources of

uncertainty.

This thesis acknowledges the benefits of these studies when the context is solely concerned

with calculating technical efficiency and providing guidance on which aspects of the business

can be improved. However, the goal of these models does not align with the specific

objectives of this thesis mainly due to the following reason: each DMU can indeed be

represented by a point in the multi-dimensional space of discretionary and non-discretionary

factors and in this thesis, DEA seeks to map these points into a new space defined by the

coordinates of 𝜆𝜆 and efficiency score (see, Section 3.3 for more details). As will be detailed in

Chapter 6, the efficiency score and 𝜆𝜆s can be utilized to gauge ‘similarity’ among a group of

firms. In this context, the efficiency score is considered one coordinate of the comparison,

representing the distance between the efficiency frontier and the firm of interest, in terms of

the discretionary variables. Furthermore, with respect to the ND factors, the firm of interest is

only compared to the similarly positioned firms. This outcome is ensured by the ND

constraints. The models discussed above cannot preserve the integrity of discretionary and

non-discretionary factors in the transformation from the old space to the new space and fail to

simultaneously account for the effect of the ND variables on the values of 𝜆𝜆 yielded by the

optimization program.

Page 117: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.4. Non-Discretionary Variables

85

In addition to the radial models, the slacks-based models have also been extended to cope

with the cases with ND inputs and outputs. Below, the additive model with this characteristic

is presented:

Additive Model with ND Factors

max 𝑒𝑒 𝑠𝑠+ + 𝑒𝑒 𝑠𝑠− (4.84)

Eq. 7.2 [41]

Subject to 𝑠𝑠− + 𝑋𝑋𝜆𝜆 = 𝑥𝑥𝑜𝑜 (4.85)

𝑌𝑌𝜆𝜆 − 𝑠𝑠+ = 𝑦𝑦𝑜𝑜 (4.86)

𝑠𝑠− ≤ 𝜋𝜋 ⊙ 𝑥𝑥𝑜𝑜 (4.87)

𝑠𝑠+ ≤ 𝛾𝛾 ⊙ 𝑦𝑦𝑜𝑜 (4.88)

𝜆𝜆 ≥ 0, 𝑠𝑠− ≥ 0, 𝑠𝑠+ ≥ 0 (4.89)

This model was originally developed by Charnes et al. (1987) [37] and similar to the model of

Golany and Roll (1993) [61], it accommodates ‘partial discretion’. Except for what is explained

below, the rest of the notation agrees with the previous models in this chapter.

The right-hand sides of Equations (4.87) and (4.88) represent the element-wise

multiplication of the respective two vectors. Each element of the 𝑚𝑚 × 1 vector 𝜋𝜋 can vary over

the range of 0 to 1, with 0 representing a fully non-discretionary input and 1 denoting a

discretionary input. The elements of the 𝑠𝑠 × 1 vector 𝛾𝛾 can adopt any value from the range of

[0,∞). The upper bound (i.e., 𝛾𝛾𝑟𝑟 → ∞) applies to the cases where the respective output is

discretionary and indeed implies that the corresponding fully discretionary slack is only

bounded to be non-negative. The lower bound, on the other hand, applies to ND outputs. In

Page 118: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.4. Non-Discretionary Variables

86

both cases of inputs and outputs, any number between the upper and lower bounds would

represent ‘partial discretion’ which would be determined based on the degree of discretion that

can be exercised with respect to the corresponding input or output. Note that the model assigns

‘0’ to the ND slacks which can be justified using the same rationale provided for the case of

non-controllable radial model discussed above.

Farzipoor Saen (2005) [52] extended the SBM model (see, Section 4.2.4) to accommodate

non-discretionary inputs and outputs. His model essentially appends the constraints of the

model of Charnes et al. (1987) [37] (i.e., Equations (4.85) to (4.89)) to the list of constraints of

the standard SBM model. The following equations exhibit the outcome:

SBM Model with ND Factors

min 𝜌𝜌 =𝜆𝜆, 𝑠𝑠−,𝑠𝑠+

�1 −1𝑚𝑚�

𝑠𝑠𝑖𝑖−

𝑥𝑥𝑖𝑖𝑜𝑜

𝑚𝑚

𝑖𝑖=1

� /�1 +1𝑠𝑠�

𝑠𝑠𝑟𝑟+

𝑦𝑦𝑟𝑟𝑜𝑜

𝑠𝑠

𝑟𝑟=1

� (4.90)

Eq. 4.48 [41]

Subject to 𝑠𝑠− + 𝑋𝑋𝜆𝜆 = 𝑥𝑥𝑜𝑜 (4.91)

𝑌𝑌𝜆𝜆 − 𝑠𝑠+ = 𝑦𝑦𝑜𝑜 (4.92)

𝑠𝑠− ≤ 𝜋𝜋 ⊙ 𝑥𝑥𝑜𝑜 (4.93)

𝑠𝑠+ ≤ 𝛾𝛾 ⊙ 𝑦𝑦𝑜𝑜 (4.94)

𝜆𝜆 ≥ 0, 𝑠𝑠− ≥ 0, 𝑠𝑠+ ≥ 0 (4.95)

As with the additive model of Charnes et al. (1987) [37], the ND slacks adopt the value of ‘0’

in this model. Therefore, the objective function is not directly affected by the ND slacks.

However, note that 𝑚𝑚 and 𝑠𝑠 denote the total number of inputs and outputs, respectively. Thus,

the number of ND factors does impact the weights assigned to the discretionary slacks in the

Page 119: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 4. Literature Review on Methodology: Data Envelopment Analysis

4.4. Non-Discretionary Variables

87

objective function of Farzipoor Saen (2005) [52] and makes the efficiency score positively

biased. To correct for this bias, Equation (4.90) can be corrected by replacing 𝑚𝑚 and 𝑟𝑟 with the

corresponding numbers of discretionary inputs and outputs.

Based on the scrutiny of the sequence of the DEA models for the ND variables, it can be

concluded that the production process is as affected by the ND factors as it is by the

discretionary factors (Ray, 1988 [116]). It is, therefore, cogent to incorporate the ND constraints

into the DEA framework in a similar manner to the discretionary variables. Any model

addressing the impact of ND factors should, however, be designed such that it meets the

individual goals of the problem. As an example, if a radial model is used to provide managers

with some guidance on which aspects of the respective business can be improved, it cannot be

designed to radially improve the ND factors. The reason is that the ND variables are

exogenously fixed and cannot often be controlled by the management team. In this research,

the ND constraints will primarily be used for classification purposes and to ensure that the

firm of interest is compared to the firms operating in a worse or alike environment. This thesis

guarantees this outcome by focusing on the slacks of the ND constraints. Section 6.2.3.2

details the contribution of this thesis to the literature of the ND factors in DEA.

Page 120: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

88

Chapter 5. LITERATURE REVIEW ON METHODOLOGY:

BAYESIAN INFERENCE IN RISK ASSESSMENT

5.1. Introduction

In 1986, J.P. Morgan [62], the world-renowned investment bank, brought forth a

revolutionary, yet practical and conceptually simple, measure in risk analysis – Value at Risk

(VaR). It is a single statistical figure and an explicit quantity that best summarizes the total

risk borne by a party holding a given portfolio. Since its emergence, VaR has remained a

popular risk measure by virtue of its simplicity and the ease of interpretation of its outcome

which can effortlessly be communicated to managers and stakeholders. Furthermore, the

authenticity of the measure contributes to its perpetuity. Its reliability is reflected in its

recognition as a crucial factor for determining the ‘capital adequacy’ of a bank or any other

financial institution. Regulatory bodies rely on the VaR analysis to decide on the amount of

equity per risk-weighted assets that must be held by a bank or financial institution to ensure

sufficient funds for safe operations and cushioning the hypothetical future problems.

In essence, VaR is a ‘quantile’ of a pdf representing gain or loss of an investment over a

given time period (e.g., between the present and a ‘target day’). It is worth noting that the

underlying pdf is not restricted to be a gain-loss one; since VaR is a quantile, its definition can

straightforwardly be extended to include other types of pdf as well. In layman’s terms, VaR

expresses the minimum loss, in absolute numbers, that an investment may suffer, with the

probability ƥ, over a given period of time (e.g., next N days). Phrased otherwise, with the

probability ƥ, in the next N days, the loss ascends to an amount equal to, or more than, a

threshold, referred to as VaR. As can be perceived, VaR is always accompanied by a

probabilistic statement about how likely it is for the investor to incur the loss of $VaR, or

more, in the next N days. Time horizon (or target date) and probability (or level of confidence)

are indeed two key explanatory variables that assemble to define VaR, for a given portfolio.

Page 121: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.1. Introduction

89

The confidence level is often chosen to be 99% or 95% in practice (i.e., ƥ = 1% or 5%,

respectively). The concept of VaR has been well preached in the literature and the following

studies fall among many which center on this topic: Dowd (1999) [47], Khindanova and Rachev

(2000) [85], Khindanova et al. (2001) [86], and Jorion (2006) [80].

The simplicity of interpretation and ‘intuitivity’ of the VaR approach should not be

mistaken for a lack of sophistication. Underlying any VaR statement is a pdf whose estimation

poses the major challenge for the approach. The literature of finance is abundant with a variety

of techniques for calculating VaR among which two approaches stand out as the prevalently

applied ones: historical simulation and model building (Siu et al., 2004 [134]). As suggested by

the terminology, the former approach is data-driven; that is, it relies on historical data to build

the empirical density required for the VaR estimation. The conspicuous shortcoming of this

method is its requirement of a long history to generate reliable and accurate outcomes. The

latter approach is a parametric one which concludes on the form of the gain-loss distribution a

priori. Given sufficiently simple initial assumptions, this method can lead to closed-form

formulae for VaR. A case in point would be the study conducted by Holton (1996) [68] which

rests on two assumptions: (i) gain-loss follows a normal distribution, (ii) the respective

portfolio is represented as a linear combination of its assets.

Since the gain-loss density is determined a priori in the model building approach of the

VaR analysis, it puts forth the challenge of how these models are developed and configured to

provide insight regarding the future scenarios of risk. Risk analysis is essentially a subjective

framework since it is impacted by how an analyst views and interprets uncertainties of the

financial market and how prior human beliefs influence the choice of the ‘right’ model

(Holton, 1997 [69]). The Bayesian methodology offers a formal means of incorporating

subjectivity and prior knowledge in the risk analysis. Concisely, in the Bayesian framework,

‘probability’ is interpreted as the subjective degree of belief in a proposition, which can be

contrasted against the rival ‘frequentist’ interpretation, expressing the probability of an event

as its long-run frequency of occurrence (see, Section 5.2 for more details on the two schools of

thought).

Page 122: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.2. Bayesian versus Frequentist Statistics

90

As the rest of this chapter unfolds, the Bayesian framework provides risk analysts with the

flexibility of integrating their subjective probability judgments with data-based objective

inferences. The Bayesian algorithm ‘learns’ from new evidence and sequentially ‘updates’ the

prior probabilities as new data accumulate. The Bayesian philosophy is mostly praised for

incorporating an analyst’s degree of uncertainty about the model of interest and its parameters,

in addition to the uncertainties pertinent to the random fluctuations in the market – the only

type of uncertainty quantified in the classical setting. In general, the VaR risk analysis within

the Bayesian context involves the following steps:

(i) Identify risk factors.

(ii) Layout a model relating the portfolio value to the risk factors.

(iii) Design probabilistic models, such as probability densities, for the risk factors, and

refine (update) them in light of new information.

(iv) Compute the predictive density of the portfolio value by integrating (averaging)

over the values of the risk factors.

(v) Estimate the portfolio VaR based on the predictive density

These steps will become clearer as more insights are gleaned from the Bayesian

methodology. The remainder of this chapter acquaints the reader with practical nuts and bolts

as well as perils and promises of the Bayesian and frequentist perspectives. The Bayesian

framework hosts a vast array of topics which extend beyond the capacity of this thesis. The

choice of the topics thus follows the specific objectives of this thesis and aligns with the

application of the Bayesian framework in this document.

5.2. Bayesian versus Frequentist Statistics

In the history of statistics, two main schools of thought have emerged regarding the

interpretation of probability. Although the roots of the Bayesian approach can be traced back

to the eighteenth century, the rejuvenation of the technique began in the second half of the

twentieth century due to advances in computational power. The debate of Bayesian vs.

Page 123: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.2. Bayesian versus Frequentist Statistics

91

frequentist has since spired into the war of ideas with the Bayesian approach recruiting more

and more adherents from among practitioners and academics. Some Bayesian stalwarts

advocate the approach as the only effective means for making probabilistic inferences and

projections and tackling ranges of risk dilemmas faced by decision makers (Malakoff, 1999

[96]; Jaynes and Bretthorst, 2003 [79]).

The frequentist or classical approach depends on the frequentist interpretation of

probability; that is, the probability of an event is expressed as the ‘limit’ of its frequency of

occurrence as the number of trials tends to infinity. As the definition suggests, repeated and

random trials is the framing context which encumbers the application of the approach to

‘infrequent’ events. The Bayesian approach, au contraire, adopts the subjective definition of

probability; that is, probability represents degree of belief which is a dynamic value subject to

continual changes in the data. The entity’s belief is ‘updated’ in view of the new information

that arrives.

Frequentists recognize probability as a concept applicable to repeatable experiments. Based

on the relative frequency interpretation, a statement like “the probability of observing ‘Head’

in tossing a fair (unbiased) coin is 0.5” implies that if the coin is flipped an unlimited (or a

sufficiently large) number of times, the frequency of ‘Head’ occurrences tends to 0.5. Such

elucidation is often a long way from how probability is perceived by practitioners and even the

experts in academia. The predominant interpretation is that upon tossing the coin, there is a

50% chance that the coin lands ‘Head’. Such perception is not accepted by frequentists since

they view probability as an intrinsic property of the respective event. Frequentists

determinedly avert from making statements about a single trial (i.e., a single toss of the coin,

in this case). The framework does not serve questions like “what is the probability of ‘Head’ in

a given flip”. Nonetheless, it offers the following answer to an assertive seeker: “In a given

trial, the coin will either fall ‘Head’ (with the probability of one) or ‘Tail’ (with the probability

of one) and the probability of observing a ‘Head’ remains constant from a trial to another,

being equal to one-half”.

Page 124: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.2. Bayesian versus Frequentist Statistics

92

The answer supplied by the Bayesian framework to the aforementioned question is more

intuitive and more consistent with scientific reasoning and the widespread cognition of

probability. The Bayesian approach interprets the question of “what is the probability of

‘Head’ in a given flip” as “what is the degree of belief in the coin landing ‘Head’, in a given

trial”. Assuming that the frequency of ‘Head’ occurrences was one-half in a past series of

experiments, the Bayesian framework asserts that “based on the current state of knowledge,

there is a 50% certainty that the coin turns ‘Head’ in the next trial”. Consequently, it emerges

that despite the frequentist persuasion, the probability of an event is not considered as its

intrinsic property, remaining intact through trials, but rather is viewed as a subjective

expression of uncertainty about future outcomes and can be revised in light of the new

information obtained from an extended experiment.

The distinction between the two rival interpretations can further be elaborated upon using

the following example, adopted from Ferson (2004) [55]. Frequentists apprehend probability as

a construct that is only applicable to events occurring in repeatable random processes. The

frequentist approach is, therefore, not inclined to address questions like “does God exist” or,

equivalently, “what is the probability of God’s existence” or “is she guilty of first-degree

murder”. The introduction provided above illuminates the logic behind this limitation: one

cannot design an experiment where the relative frequency of the existence of God is examined.

Nor can one conduct an experiment, consisting of a set of random trials, where she is

‘sometimes’ [55] guilty and ‘sometimes not' [55]. The Bayesian framework, on the other hand,

can easily accommodate such cases since the approach indeed gauges degrees of belief; how

certain one can be that God exists or she is guilty.

Of course, the differences between the two frameworks go beyond philosophical

arguments, and into pragmatic issues, which will be delved into following an introduction to

the Bayesian approach. There exist a plenty of literature offering broad conceptual aid with the

fundamentals of the methodology (e.g., Ferson, 2004 [55]; Rachev et al., 2008 [115]; Damien et

al., 2013 [44]). The reader is encouraged to pursue them for a comprehensive review of the

Page 125: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.3. Basics of Bayesian Inference

93

approach. Furthermore, the author assumes familiarity with the fundamentals of classical

statistics on the part of the reader.

5.3. Basics of Bayesian Inference

Derived based on the Baye’s theorem, the Bayesian approach is a formal representation of

the subjective view of uncertainty and is a principled way for describing the process of

scientific reasoning in conformity with the axioms of probability theory. It offers a profound

structure for inferring from data which is more consistent with the learning mechanism

adopted by the human brain, programmed to learn from experience. The brain has been hard-

wired to judge by drawing from the reservoir of interacting past experiences and updating

them in light of present ones. Such approach forms the core of the Bayesian methodology

where a ‘prior’ knowledge is updated by learning from the current information; the updated

prior knowledge is called the ‘posterior’ knowledge.

The Baye’s rule, which plays the central role in the Bayesian framework, in its most well-

known form is shown below:

𝑓𝑓Ζ(𝜁𝜁|ұ) =𝑓𝑓Ζ(𝜁𝜁) × 𝑓𝑓Ұ(ұ|𝜁𝜁)

𝑓𝑓Ұ(ұ) (5.1)

where, ұ represents a draw from the random variable Ұ, and Ζ denotes the parameter(s) of the

probability density function 𝑓𝑓Ұ(ұ). The principal idea is to derive inferences about the

parameter Ζ assuming that the characteristics of the sampled data are best described by 𝑓𝑓Ұ(ұ).

Note that Ζ itself is treated as a random variable contingent on a series of unknown

parameters, referred to as ‘hyperparameters’ (Rachev et al., 2008 [115]). The prior knowledge

regarding Ζ is summarized into the probability density function 𝑓𝑓Ζ(𝜁𝜁). It is, therefore, called

the ‘prior’ distribution of Ζ. The density function 𝑓𝑓Ζ(𝜁𝜁|ұ) denotes the updated version of the

Page 126: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.3. Basics of Bayesian Inference

94

prior density which is refined using the data (i.e., ұ1, ұ2, ұ3, … , ұ𝑛𝑛) and is termed the

‘posterior’ distribution of Ζ.

The density function 𝑓𝑓Ұ(ұ|𝜁𝜁) indeed represents 𝑓𝑓Ұ(ұ1, ұ2, ұ3, … , ұ𝑛𝑛|𝜁𝜁), where ұi (𝑖𝑖 =

1, … , 𝑛𝑛) designates the 𝑖𝑖th realization of the random variable Ұ. Therefore, the function

𝑓𝑓Ұ(ұ|𝜁𝜁) can be formulated as a function of 𝜁𝜁 since given all the 𝑛𝑛 samples, the only unknown

remains 𝜁𝜁. This new function is termed the ‘likelihood’ function which congregates all the

information held by the data into a single function.

In the Baye’s theorem, the greatest computational intricacy lies in calculating the

denominator in Equation (5.1). It is indeed a normalization factor and can be computed based

on the law of total probability presented below in its continuous form:

𝑓𝑓Ұ(ұ) = � 𝑓𝑓Ұ(ұ|𝜁𝜁) × 𝑓𝑓Ζ(𝜁𝜁)dζ𝜁𝜁

(5.2)

The preceding discussion clarified how the Baye’s rule integrates the prior information with

the accumulated evidence (i.e., data) to assess the overall uncertainty of an event. These two

sources of information are fused together into a posterior distribution and it is the relative

strength of each source that determines the final ‘location’ and ‘dispersion’ of the posterior

distribution. The Strength of a distribution is often gauged by its narrowness (so to speak)

since the focus of a narrow distribution remains on a shorter range of possible values, ruling

out a wider array of data. There is, therefore, a trade-off between the two sources of

knowledge and the posterior distribution inherits its characteristics mainly from the stronger

source (i.e., the most informative source).

As an extreme example, consider a case where the prior belief is represented in the form of

a Dirac delta function. It can mathematically be proven that regardless of the size of the new

Page 127: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.3. Basics of Bayesian Inference

95

evidence accrued, the posterior distribution would remain intact, resembling the shape of the

prior distribution. Such an outcome comes as no surprise since when an analyst expresses the

prior belief as a Dirac delta function, they explicitly communicate a great confidence in the

prior knowledge; hence, data cannot cast any additional light or doubt on such strong

preconception. Perhaps the most ardent debate across the pro-Bayesian and anti-Bayesian

statisticians revolves around prior distributions which are broadly classified as informative and

non-informative priors (O’Hagan, 2008 [102]).

A prior distribution is considered informative if it alters the form of the respective posterior

distribution; that is, in a case where an informative prior is utilized, the inference derived on

the basis of the posterior differs from the one deduced solely based on the data, due to the

influence exercised by the prior knowledge. When considerable prior evidence exists, the

Bayesian approach encourages the analysts to benefit from such prior resources in arriving at

effective solutions; however, in some circumstances, the analyst may not desire to perturb the

posterior distribution with the prior knowledge since (s)he is either skeptical of the prior

information itself or in doubt whether an appropriate prior density has been adopted to present

the prior knowledge (i.e., if the prior knowledge has been translated correctly into a relational

mathematical expression). Drawing upon Laplace’s ‘principle of insufficient reason’ (Ferson,

2004 [55]), such cases can be handled by non-informative priors also referred to as ‘vague’ or

‘diffuse’ priors (Rachev et al., 2008 [115]). They do not influence the corresponding posteriors

and, consequently, the inferences made about the parameter(s), and are genuine indications of

ambiguity and hesitancy about prior beliefs.

Some researchers have put great effort into determining the most suitable non-informative

distributions in various settings (e.g., Berger, 2006 [21]). Although a distribution reflecting an

absolute lack of sensitivity to the prior knowledge seems unachievable, several forms of priors

have been proven to exert least amount of effect on posteriors; uniform distributions are

predominantly adopted to account for prior ignorance. Alternatively, some choose

distributions like normal with considerably large variances (Ferson, 2004 [55]; O’Hagan, 2008

[102]; Rachev et al., 2008 [115]).

Page 128: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.4. Merits of the Bayesian Methodology

96

Worthy of note is that one does not expect the inferences to be sensitive to the form of the

prior distribution if sufficient amount of data exist. In the presence of strong data, the posterior

is primarily determined by the data (i.e., the likelihood function). Each prior is updated in

sequence as a new instance of data arrives and given reasonable initial forms, all differing

priors would eventually tend toward a common posterior form; sequential updating gradually

improves individual priors and moves them from the primary assignment toward the true

posterior.

5.4. Merits of the Bayesian Methodology

As specified earlier, one significant and distinctive aspect of the Bayesian approach is its

capacity to provide a formal and legitimate mechanism to hybridize the prior information with

the sample information upon availability of substantial amount of prior evidence – a feature

which leads to more informative posteriors and enhances the subsequent inferences. This merit

(i.e., use of subjective information) has simultaneously become the focus of the most heated

contention among Bayesians and frequentists. The criticism evoked by the latter party,

however, targets themselves as well since it fails to recognize the role of an analyst’s state of

mind in choosing a specific model or test; frequentists indeed take into account subjective

views, but in an implicit manner, when they devise a model, choose inference rules, speculate

on the value of the parameter(s), etc. Such subjective foundations are considered only natural

by the adherents of the Bayesian approach and they argue that it is much easier to perceive and

scrutinize the impact of subjective beliefs on the final outcomes when the framework

explicitly accounts for them; whereas, such an effect is less clear-cut in the frequentist setting

(O’Hagan, 2008 [102]).

Moving further in the discussion of subjectivity, it is worth stating that consolidating

subjective assertions with the notion of probability does not suggest that the Bayesian

framework is an ensemble of arbitrary ideas like irrational opinions and stiff prejudices.

Subjectivity implies exploiting pieces of information accumulated in the past and Bayesians

Page 129: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.4. Merits of the Bayesian Methodology

97

dismiss any other interpretation of subjectivity emerging from illusory prejudices rather than

facts. Simply put, the Bayesian philosophy is as subjective (or objective) as science is and

aligns more consistently with how scientific theories evolve. This is not to discount the

struggle of science to remain objective but to call attention to how it is yielded. Objectivity is

achieved through developing theories and debating the validity of them. The theories are

refined ceaselessly as new evidence accrues and, finally, converge to an accepted framework

(Howson and Urbach, 1996 [71]; O’Hagan, 2008 [102]). O’Hagan, (2008) [102] comments that

“objectivity in science is really a convergence of subjective opinion, and that agreement may

be only temporary”.

The Bayesian approach is also praised for its intuitive and easily interpretable inferences,

aside from the inherent notion of subjectivity. To clarify the content of this sentence, it is first

required to point out another important distinction between the Bayesian and frequentist

settings which arises from each method’s view of unknown parameters. Although the

Bayesian methodology is often applauded for its capacity to incorporate subjective knowledge,

it was discussed in Section 5.3 that this aspect may be suppressed in some situations by means

of non-informative priors. Even in the case of informative priors, if there are enough data, the

posterior would eventually turn insensitive to the prior. The most distinctive superiority of the

Bayesian approach is indeed its view of the unknown parameters.

The ultimate goal of many data analysis attempts remains deriving statistical inferences, or

‘learning’, about the unknown parameters – the domain where the fundamental deviation

between the two schools of thought occurs. The frequentist setting considers parameters as

fixed numbers where the concept of ‘uncertainty’ is not applicable. On the contrary, Bayesians

treat all the unknown parameters of a model as realizations of random variables. The main

interest of the latter, therefore, lies in eliciting the distributions associated with the respective

parameters. In a nutshell, the Bayesian approach is a fully probabilistic process of learning;

data are used to learn about the parameters and this knowledge is ‘updated’ continually in

view of new evidence.

Page 130: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.4. Merits of the Bayesian Methodology

98

The debate of considering parameters as random variables, rather than fixed numbers,

originates from the differences between two main sources/types of uncertainty: (i) ‘aleatory’

uncertainty (or ‘variability’), (ii) ‘epistemic’ uncertainty (or ‘incertitude’). The former refers to

the uncertainty associated with random variables, in the commonly acknowledged sense.

Sample data, which are generated from a truly repeatable underlying process, inevitably

embrace this kind of uncertainty. Since it arises from the inherent randomness of the system,

increasing the sample size would not reduce the magnitude of the aleatory uncertainty. The

epistemic uncertainty, on the other hand, originates from the modeler’s insufficient or

incomplete knowledge (paucity of knowledge) and can be lessened upon acquisition of more

data. Parameters belong in this latter category; to wit, the modeler remains uncertain regarding

the true value of a parameter and their uncertainty may be reduced and resolved as more

evidence accumulates.

The frequentist setting, which is underlain by the frequency interpretation of probability,

only measures aleatory uncertainties. Implicit in this property of the framework lies

frequentists’ view of the probability distribution of a random variable: it is a resolute construct

and frequentists do not recognize any uncertainty associated with its form; whereas, Bayesians

do not assign any definiteness to the location and dispersion of a probability distribution and

acknowledge that the shape of the distribution can be refined in view of new information. In

short, Bayesians amalgamate parameter uncertainty and data uncertainty (i.e., both types of

uncertainty) and view probability as a comprehensive concept which simultaneously addresses

all sources of uncertainty.

The aforementioned has laid the foundation to explain how the Bayesian setting facilitates

deriving intuitive conclusions which is recognized as its most significant benefit. It is provided

by the capacity to make inferences about the uncertainty of a parameter using the respective

(posterior) density. Inferences are broadly classified as point estimates (such as mean, median,

or mode) and interval estimates (such as credible interval). A considerable uncertainty in a

posterior distribution would, however, render the point estimates uninformative. Thus, the

concept of ‘credible interval’ provides a more suitable context for putting the aptitude of the

Page 131: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.4. Merits of the Bayesian Methodology

99

Bayesian framework into perspective. Credible interval is often considered as the analogue of

‘confidence interval’ in the classical statistics; yet, the two concepts differ critically in

philosophy, with the former identified as a more practical notion. A credible interval, which is

intelligibly obtained from the respective posterior distribution, represents a region whose

probability corresponds to a pre-specified value (e.g., 1 − Ƿ). In mathematical terms, the

interval [𝑙𝑙,𝑢𝑢] stands for the (1 − Ƿ)% credible interval of the parameter Ζ if

ℙ(𝑙𝑙 < Ζ < u|ұ) = �𝑓𝑓Ζ(𝜁𝜁|ұ)d𝜁𝜁𝑢𝑢

𝑂𝑂

= 1 − Ƿ (5.3)

The credible interval thus provides a range of likely values for Ζ and, more importantly, a

forthright response to a question like “In which interval does the parameter reside with the

probability (1 − Ƿ)% (e.g., 95%)”. It is tempting to rush conclusions and proclaim that the

notion of confidence interval addresses the same question; that is to interpret the statement of

“The (1 − Ƿ)% confidence interval of Ζ is [𝑙𝑙,𝑢𝑢�]” as “Ζ falls within the interval of [𝑙𝑙,𝑢𝑢�] with

the probability (1 − Ƿ)%”. However, this insight is incorrect since in classical statistics,

parameters are considered as fixed numbers. Thus, one cannot legitimately make probability

statements about them. The parameter Ζ either belongs (with the probability of one) or does

not belong (with the probability one) in the range of [𝑙𝑙,𝑢𝑢�]. The misconception has its roots in

failing to recognize that confidence interval is a random interval; that is, the lower (𝑙𝑙) and

upper (𝑢𝑢�) bounds are themselves realizations from the random variables 𝐿𝐿� and 𝑈𝑈�,

respectively. The essence of a (1 − Ƿ)% confidence interval and its precise meaning can be

delineated as follows:

If a sufficiently large number of sets of random samples are generated and for each set, the

confidence interval is computed following a given rule of inference, then we are confident that

(1 − Ƿ)% of these random intervals will include the ‘true’ value of the parameter. It is,

Page 132: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.4. Merits of the Bayesian Methodology

100

therefore, the rule that is under evaluation by means of repeated sampling. Such interpretation

is certainly less intuitive and hard to communicate with non-statisticians. It is likewise

cumbersome to derive meaningful conclusions in the risk framework, on the basis of the

preceding correct interpretation. Digging deeper, it becomes evident that confidence interval is

frequently mistaken for credible interval on the part of experts. It is indeed the credible

interval which can reasonably address what they seek to measure about the parameters (Huck,

2009 [73]).

It is readily observable that the hypothesis testing can be conducted more straightforwardly

in light of the concept of credible interval. As a case in point, suppose that the null hypothesis

is testing if the values of a parameter is at least zero. In the Bayesian setting, the probability of

the statement can be calculated based on the respective posterior distribution and the null

hypothesis would be accepted if the obtained probability exceeds a predetermined threshold

(O’Hagan, 2008 [102]). Contrast such intuitive perspective with the frequentist wisdom which

treats parameters as fixed numbers, revoking the validity of any probability statement about

them. The framework, therefore, has to develop a random variable (in the aleatory sense) with

a known density that is linked to the null hypothesis. It is then this random construct, referred

to as ‘test statistic’, which underlies the decision of rejecting or accepting the null hypothesis.

Finding the right test statistic in many settings complicates the problem. Moreover,

frequentists’ hypothesis testing suffers from the same limitation imposed by the confidence

interval: Few really understand it and it does not suit the framework of risk analysis (see, Page

99). When it is stated that the null hypothesis (e.g., Ζ ≥ 0) is rejected at the significance level

of, for example, 5%, a majority would interpret it as ℙ(Ζ ≥ 0|ұ) ≤ 0.05, or equivalently, the

probability that the null hypothesis is ‘true’, given the data, is equal to, or less than, 5%.

Similar to the case of confidence interval, such a perception contradicts the relative frequency

notion of probability since the null hypothesis and parameters, for that matter, are not part of a

repeatable experiment, and hence, probability statements cannot be tied to them (refer to

Section 5.2 for further details). The correct, yet less intuitive and practically transparent,

interpretation would illustrate that if the null hypothesis were ‘true’, the probability of

Page 133: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.5. Merits of the Bayesian Approach in Risk Assessment

101

observing a test statistic whose absolute value is equal to, or greater than, the realized test

statistic must be higher than the pre-set statistical significance (in this case, 5%).

Indirect and distorted responses to inference questions is just one of the shortcomings of the

frequentist school of thought. Other contradictory traits are also associated with it, stemming

from the frequency definition of probability which demands repeatability and is sensitive to

the sampling structure or design of the experiment. The reader is encouraged to consult Berger

et al. (1988) [22] and O’Hagan (2008) [102] in this regard.

5.5. Merits of the Bayesian Approach in Risk Assessment

The contribution of the Bayesian methodology to different fields of practical reasoning

remained burdened by its high level of computational complexity; the process of computing

posteriors, particularly the normalization factor, can be cumbersome and impractical without

computers except for simple problems such as those with ‘conjugate’ priors. Since 1980s,

when powerful computing machines started to gain ground, the Bayesian approach is enjoying

an abrupt upward shift in its applications. It is particularly of interest in both financial and

non-financial risk management. Some statisticians even claim that it is the only probability

paradigm that can properly handle risk assessments. The remainder of this section allocates

more effort to navigate the aforementioned promises of the Bayesian wisdom in the field of

risk analysis.

The context of risk measurement and decision making is often identified by data scarcity –

a domain where the Bayesian approach comes in handy. By virtue of the notion of prior

distribution, Bayesians can make inferences even when empirical data are absent.

Furthermore, sophisticated tasks, like decision making, demand expert knowledge as well as

judgment and intuition to be formulated into objective mathematical terms. In the field of risk

management, where often limited data exist, not only is such expert intervention desirable but

essential. By synthesizing prior beliefs and new evidence, the Bayesian posterior distribution

greatly serves such requisites.

Page 134: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.5. Merits of the Bayesian Approach in Risk Assessment

102

The capacity of the Bayesian approach to measure the epistemic component of uncertainty,

in addition to its aleatory constituent, is another feature that notably contributes to its

reputation as a suitable framework to pursue risk analyses. The following example

demonstrates how the frequentist approach may remain inferior to the Bayesian method in a

setting where it is sought to predict the future return of a portfolio. The future return of a

given portfolio can be considered as a random variable drawn from an underlying population.

Regarding the return as a random variable implies that the solution recognizes the aleatory

uncertainty associated with it. The uncertainty about future data is, however, primarily of the

epistemic type (O’Hagan, 2008 [102]); the parameters of the predictive density, conjectured to

express the likeliness of possible future returns, are themselves uncertain figures and this

aspect is not accounted for by the frequentists (see, Section 5.4, Page 98 for more details).

From the practical point of view, any type of uncertainty would in principle blur the

definiteness of the future and its impact on the final outcome must be jointly scrutinized with

that of any other source. On this account, a risk analyst or a decision maker is, in essence,

seeking a predictive distribution which would hybridize all sources of uncertainty and act as a

basis for deriving informed decisions and generating risk scenarios. The Bayesian framework

is the only setting which caters to such appeal. Bayesian thinkers are consistently concerned

with shaping their beliefs in the form of probability distributions which would in sequence be

blended with other uncertainties, leading to direct probability statements on different scales.

Heretofore, the focal intent of this chapter has been detailing the prominent advantages of

the Bayesian approach. The following discussion highlights and reviews certain features that

accomplished the recent rapid adoption of the methodology by many risk analysts. The list

appears in the study of Ferson (2004) [55]. A more comprehensive list praising the Bayesian

methodology can be located in Berry (1996) [24].

First. As detailed in the previous section, one benefit of the Bayesian framework is its

conformity with the process of inductive reasoning implemented by the human brain.

Inferences derived based on the Bayesian perspective are intuitively reasonable sets of

outcomes which address the original concerns posed by the analysts in a more direct and

Page 135: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.5. Merits of the Bayesian Approach in Risk Assessment

103

concise manner. As discussed in Section 5.2, questions like “How likely is she to have

committed the crime?” cannot be addressed in the frequentist setting, which draws on the

notion of repetitiveness and demands a repeatable random underlying process; one cannot

design an experiment or a sequence of trials where she may be found guilty in some and not

guilty in the rest, and then, equalize the frequency of being judged guilty with the probability

of her committing the crime. Such questions are, nonetheless, welcomed under the Bayesian

philosophy since the setting caters to the interior monologue of the human brain; that is, in

light of the subjective interpretation of probability, the methodology parses the

aforementioned question and reforms it as follows: “what is our degree of belief that she is

guilty?”

Beside a vaster variety of problems that can be accommodated by the Bayesian approach,

its inferences (e.g., credible interval) comply more with the scientific reasoning, compared to

their so-called counterparts in the frequentist setting (in this case, confidence interval). Finally,

in further support of the intuitive nature of the methodology, one can refer to its ability to

absorb all the prior evidence in an explicit manner and revise and refine it as new evidence

arrives. The reader is encouraged to consult Sections 5.3 and 5.4 for more details on these

matters.

Second. The Bayesian philosophy better feeds into the process of decision making. Although

more intricate to develop, Bayesian inferences are formed in a setting where any uncertainty

can be expressed as a probability computed using the posterior density, which is the key tool

for conveying risk information. It is indeed the process of hypothesis testing which facilitates

decision making based on the posterior distributions. In the frequentist setting, decision

making is likewise conducted through hypothesis testing; however, as detailed earlier, the

philosophy of hypothesis testing in the classical setting is less intuitive and more susceptible to

logical errors. Furthermore, it is not admissible to associate probability statements with the

null and alternative hypotheses although they are direct indicatives of the analyst’s concerns

and best frame the examiner’s questions. Consequently, the analyst is only supplied with the

binary answer of ‘accept’ or ‘reject’ and faces a numeric value (i.e., the significance level)

Page 136: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.5. Merits of the Bayesian Approach in Risk Assessment

104

which indeed specifies the reliability of the rule chosen for hypothesis testing, not the

conclusions laid out in the form of hypotheses (see, Section 5.4, Page 100).

Among the shortcomings that accompany hypothesis testing in the frequentist framework,

the following is perhaps the most overlooked one. The classical hypothesis testing incurs two

types of error: (i) Type I error: the probability of rejecting the null hypothesis when it is true,

(ii) Type II error: the probability of failing to reject the null hypothesis while it is false. Type I

error is often considered in the analyses and is perceived as an integral part of the classical

hypothesis testing; whereas, type II error comparatively poses less of a concern. In the context

of risk, however, quantifying the latter error is significant to the same extent as that of the

former. This condition introduces extra complexity with few benefits, as already established:

the computation of type II error requires defining a test statistic (and its corresponding density)

specific to the alternative hypothesis. The process demands choosing an exclusive statement

for the alternative hypothesis and would greatly complicate the implementation of the test.

Since the Bayesian approach tackles the task of decision making on a more fundamental

level, through assigning probability statements to the hypotheses, it is capable of charting a

direct and cognitive path to reach a decision. In addition, rather than yielding the binary

answer of ‘accept’ or ‘reject’, it leads experts to find a range for their degree of belief in the

accuracy of the hypotheses. In the Bayesian point of view, the decision-making route adopted

by frequentists is regarded as a detour; that is, classical statisticians first need to devise a

random process built upon the original decision question. Probabilistic statements can only be

made about this random process, and only in a recursive manner can experts reach inferences

on the original query.

Third. In the risk context, adherents of the Bayesian perspective further proclaim the

superiority of the methodology by drawing on its capacity to explicitly accommodate

subjective views in the process of decision making. Synthesizing the information provided by

the sample and the corresponding prior model into a posterior density, on which subsequent

Page 137: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.5. Merits of the Bayesian Approach in Risk Assessment

105

scrutinies rest, is not only intuitive but also necessary and essential, particularly, when limited

data exist or most of the accessible information is subjective (see, Section 5.3).

The aforementioned features of the Bayesian approach have led to its most significant

application – Bayesian predictive inferences – due to which the methodology has gained

significant ground in different fields of finance, particularly portfolio management and market

risk management (Rachev et al., 2008 [115]). The ability of the Bayesian framework to derive

predictive inferences contributes greatly to its recognition as a tool to improve risk

assessments. Predictive inferences are derived based on the knowledge presented by the

‘predictive distribution’, defined as the distribution of the next datum given a realized

sequence of data and can be estimated using the (joint) posterior distribution of the

parameter(s) of the data generating process. The following equation, adopted from Rachev et

al. (2008) [115], exhibits the mathematical definition of the predictive distribution which serves

as a tool of great flexibility since not only can it be used to derive point predictions but

interval predictions as well:

ᵹҰ+(ұ+|ұ) = � 𝑓𝑓Ұ+(ұ+|𝜁𝜁)𝑓𝑓Ζ(𝜁𝜁|ұ)d𝜁𝜁𝜁𝜁

(5.4)

where, ұ+ stands for the next datum and it is assumed to be a random draw from 𝑓𝑓Ұ+(ұ+|𝜁𝜁).

The function 𝑓𝑓Ζ(𝜁𝜁|ұ) denotes the posterior distribution of the parameter Ζ, computed as

depicted by Equation (5.1). Note that 𝜁𝜁 represents a realization of Ζ which is treated as a

random variable in the Bayesian setting. By virtue of integrating over the values of the

parameter(s), the predictive distribution would solely depend on the past observations (i.e.,

past realizations of the random variable Ұ). In other words, the predictive density absorbs all

the uncertainty associated with the parameter(s) by averaging over their values. Any inference

derived from the predictive distribution would, therefore, reflect the epistemic uncertainty in

Page 138: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.5. Merits of the Bayesian Approach in Risk Assessment

106

addition to the aleatory uncertainty which is accounted for by the sampling distribution. The

classical statistics fails to incorporate the former type of uncertainty. Instead, it conjectures on

the values of the unknown parameters and regards those assumptions as the true values of the

parameters which are then fed into the predictive density.

The increasing prominence of the Bayesian philosophy in the field of risk is indeed

nurtured by its capacity to yield predictive densities which combine all sources of uncertainty.

This characteristic can essentially benefit any analysis involving a series of unknown

parameters that cannot be determined with certainty. One of the fields of risk where the

Bayesian views have penetrated the consciousness of practitioners and academics is the mean-

variance analysis in portfolio selection. In this context, the expected value and variance of the

return of a portfolio are expressed as functions of two sets of parameters: (i) expected returns

of the assets comprising the portfolio (ii) covariances between the returns of the assets in the

portfolio. The classical mean-variance approach would proceed by finding an individual

estimate for each of these parameters based on the available samples and would rest the

subsequent computations on the assumption that these estimates represent the true values of

the parameters. However, since the approach does not recognize the estimation uncertainty or

ambiguity associated with the parameters, it is subject to significant errors in the optimal

weights selected for the portfolio assets. This issue can be alleviated using the approach

described by Equation (6.4); that is, the multivariate predictive distribution of the future

returns of the assets can be obtained by integrating (averaging) over the values of the

aforelisted parameters (i.e., expected returns of the assets and covariances between the returns

of the assets). The predictive density yielded by this process would be independent of the

parameters. The reader is referred to Rachev et al. (2008) [115] for detailed mathematical

descriptions.

Other fields of finance adopt the Bayesian wisdom along similar lines. It has emerged as

the key inference technique in other frameworks of optimal portfolio selection, predicting

returns, gauging market efficiency under different measures, and volatility modeling, just to

name a few. It lies beyond the scope and expertise of this thesis to elaborate on all the listed

Page 139: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 5. Literature Review on Methodology: Bayesian Inference in Risk Assessment

5.5. Merits of the Bayesian Approach in Risk Assessment

107

applications. In the interest of clarity and thoroughness, one additional example will follow

which aligns closely with the objectives pursued by this research.

Conducted by Siu et al. (2004) [134], the study implements the VaR inference procedure in

two cases of linear and non-linear portfolios and demonstrates how the Bayesian approach

improves the inferences. The methodology involves choosing a model which intends to

describe the stochastic process of the portfolio value and treats its underlying parameters as

random variables. The posterior joint distribution of these parameters is then computed by

hybridizing the objective knowledge inferred from the historical data and the subjective

knowledge (i.e., a prior density) provided by the authors. The estimated posterior joint

distribution is subsequently utilized to find the predictive distribution of the future portfolio

value which in turn serves as the basis of VaR calculation. The VaR figure can be updated in

sequence as more data accrue. The outcomes of the research indicate the superiority of the

Bayesian VaR model to its classical counterparts and regards the Bayesian VaR model as a

‘good alternative’ to the classical approaches.

As a closing remark for this chapter, it is possible to observe that the features of the

Bayesian framework offer a unique setting for dealing with the limits of the risk analysis of

IPOs. Essentially, there exist limited data, yet often strong beliefs regarding performative and

ostensive aspects (Feldman and Pentland, 2003 [54]) of individual IPOs. Such prior views are

typically formed by scrutinizing the counterpart aspects of the comparables of the respective

IPO. The Bayesian methodology corresponds such a setting in many respects and would serve

as a key tool for deriving Bayesian VaR inferences about IPOs. To the best of the author's

knowledge, to this day, this research avenue has remained unexplored. Data paucity has

slackened risk estimations particularly through mathematically sensible approaches and has

rendered some of the existing attempts ad hoc or inadequate due to their case-specific or

extremely simplifying assumptions (e.g., Abdou and Dicle, 2007 [1]; Arnold et al., 2010 [10]).

This study is the first of its kind whose focal intent is to unravel the IPO risk analysis from the

Bayesian perspective. Sections 6.3 and 6.4 detail the steps involved in developing this phase of

the methodology.

Page 140: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

108

Chapter 6. METHODOLOGY

6.1. Review of Objectives

The main objective of this thesis is to quantify the short-term risk of investing in IPOs.

Here, ‘risk’ is defined as a value-at-risk measure which is computed based on the distribution

of the IPOI’s stock price 90 trading days subsequent to the day of issuing (PDF90IPOI). The

comprehensive literature review in Section 3.2 established that using comparables would be a

prudent approach to quantify the short-term risk of investing in IPOs. Based on the research

presented, the comparables approach is predominantly used to set target prices and provides

estimates with roughly identical accuracy to other mathematically intensive valuation

approaches. The prevalence and strength of the method in projecting future prices indicates

that it can, likewise, be a supreme approach to finding the distribution of prices at a given time

in the post-IPO market.

The methodology, which fulfills the main objective, pursues the following specific goals:

(i) Phase I: The IPOI’s comparables are identified and the IPOI’s similarity to each of

them is quantified through developing a ‘distance function’ (see, Section 6.2).

(ii) Phase II: PDF90IPOI is estimated in a recursive Bayesian framework which utilizes the

outcomes of Phase I (see, Section 6.3).

(iii) The ‘distance function’ is calibrated by backtesting, using a recursive-cyclic

algorithm. It iteratively runs Phase I and Phase II across different IPOIs and draws

upon the Bayesian framework to find the optimum model parameter whose

likelihood is maximum given the realized market prices (see, Section 6.4).

6.2. Phase I: Comparable Selection

In this research, the comparables method forms the building block of the short-term risk

estimation. Therefore, a more accurate screening of comparables would eventually translate

Page 141: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

109

into improved risk estimates. This research seeks a process which facilitates a multi-

dimensional comparison of firms in an objective manner. Section 3.3 presents a detailed

review of the existing methodologies for identifying comparables and lists any shortcoming

that may make the respective technique incompatible with the specific objectives of this

research. To put into brief terms, the industry-based approaches can be considered the most

prevalently applied methods of comparable selection; however, they indeed approximate the

performance of the firm of interest with the overall characteristics of the respective industry,

disregarding firm-specific details such as operating and financial features. Moreover, most of

these techniques do not have the capacity to simultaneously compare multiple aspects of the

operations of the firm of interest with the corresponding mean industry characteristics.

Further review of the literature returned two studies which base the selection of

comparables on multiple idiosyncratic accounting and financial factors: (i) Bhojraj and Lee

(2002) [25], (ii) Simak (2000) [133]. Bhojraj and Lee’s (2002) [25] methodology is regression-

based, and consequently, encompasses the shortcomings associated with regression such as ex-

ante definition of the form of the function. Furthermore, the methodology is designed such

that it requires market information which does not exist in the case of IPOs. Section 3.3

outlines the technique and exhaustively discusses further limitations associated with it (see,

Pages 40-44).

Simak’s (2000) [133] study adopts the DEA framework to perform a multi-dimensional

comparison of characteristics across different firms. The methodology and its merits and

demerits are discussed at great length in Section 3.3 (see, Pages 44-47). The comparable

selection algorithm proposed in this thesis is motivated by the principles of Simak’s (2000) [133] work. As discussed earlier, the DEA approach is a non-parametric approach which

endogenously establishes a relationship between the inputs and outputs, and thus, prevails over

the limitations imposed by the regression analysis. The DEA approach also accounts for the

economies of scale and would choose comparables such that they have the same economies of

scale as the firm of interest in addition to resembling it in terms of financial performance.

Page 142: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

110

Table 6-1. Input-Output Data Used by the CCR Model in an Input-Oriented Analysis

DMU Input 1 Input 2 Output

A 4 3 1

B 7 3 1

C 8 1 1

D 4 2 1

E 2 4 1

F 10 1 1

The comparables are ideally expected to resemble all aspects of the firm of interest;

however, in practice, such comparables do not exist. This problem can be mitigated by

reducing the comparison criteria to limited pre-determined factors; yet, the complication

continues to hold since it is hard to achieve a comparable firm which matches the firm of

interest in all such factors. Thus, some dimensions must be assigned higher priority, and

hence greater weights, in the comparison. The challenge lies in identifying these dimensions

and the respective weights. Perhaps, the most significant benefit of a DEA-based comparison

technique comes from the capacity of DEA to endogenously determine such weights.

As indicated in Section 3.3, a DEA-based comparison method selects comparables from a

set of firms which share identical ‘efficient peers’. This approach can be justified as follows:

The weights received by inputs and outputs of an inefficient DMU are similar to the weights

assigned to the inputs and outputs of the respective efficient peers. Hence, DMUs which are

associated with the same ‘reference set’ would dispense similar levels of significance to the

respective inputs and outputs. This point can be further illustrated using the ‘primal’ form of a

simple input-oriented CCR (or CRS) model. The data are collected from Example 2.2 of

Cooper et al. (2007) [41] and are presented in Table 6-1.

Page 143: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

111

The following six equations present the main constraints of the ‘primal’ input-oriented

CCR model:

A: 𝑊𝑊𝑂𝑂 ≤ 4𝑊𝑊𝐼𝐼1 + 3𝑊𝑊𝐼𝐼2 (6.1)

B: 𝑊𝑊𝑂𝑂 ≤ 7𝑊𝑊𝐼𝐼1 + 3𝑊𝑊𝐼𝐼2 (6.2)

C: 𝑊𝑊𝑂𝑂 ≤ 8𝑊𝑊𝐼𝐼1 + 1𝑊𝑊𝐼𝐼2 (6.3)

D: 𝑊𝑊𝑂𝑂 ≤ 4𝑊𝑊𝐼𝐼1 + 2𝑊𝑊𝐼𝐼2 (6.4)

E: 𝑊𝑊𝑂𝑂 ≤ 2𝑊𝑊𝐼𝐼1 + 4𝑊𝑊𝐼𝐼2 (6.5)

F: 𝑊𝑊𝑂𝑂 ≤ 10𝑊𝑊𝐼𝐼1 + 1𝑊𝑊𝐼𝐼2 (6.6)

Here, the parameters 𝑊𝑊𝐼𝐼1, 𝑊𝑊𝐼𝐼2 and 𝑊𝑊𝑂𝑂 denote the weights of Input 1, Input 2 and the Output,

respectively. The constraints define the shared feasible region as the half-space lying above the

efficient hyperplanes, C, D and E (see, Figure 6-1). Figure 6-2 complements Figure 6-1 by

adding the hyperplane of one of the inefficient DMUs (i.e., DMU A). Note that the hyperplane

A bears the closest proximity to the efficient hyperplanes at the intersection of hyperplanes D

and E and this demonstration visualizes why DEA chooses DMUs D and E as the ‘efficient

peers’ of DMU A. Figure 6-2 also describes why DEA would choose similar final weights for

DMUs A, D, and E.

Prior to presenting the DEA-based model developed in this thesis for choosing

comparables, Section 6.2.1 engages in describing the pool of candidates for acting as the

comparables of the IPOI. Section 6.2.2, subsequently, establishes the list of inputs and outputs.

The list of inputs and outputs are scrutinized in advance of outlining the model since its

structural form would be impacted by the characteristics of these inputs or outputs such as

positivity or negativity or the degree of discretionarity associated with each.

Page 144: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

112

Figure 6-1. This figure demonstrates the efficient hyperplanes determined by the constraints presented in

Equations (6.1) to (6.6). The shared feasible region lies above the efficient hyperplanes.

Page 145: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

113

Figure 6-2. This figure complements Figure 6-1 by adding the hyperplane of DMU A which is represented by

Equation (6.1). DMU A is an inefficient DMU, and thus, its hyperplane lies below the hyperplanes of the

efficient DMUs C, D, and E. The set of ‘efficient peers’ of DMU A comprises DMUs D and E. Note that the

intersection of DMUs D and E is where the hyperplane A has the closest proximity to the efficient frontier. DEA,

therefore, determines similar final weights for DMUs D, E, and A.

Page 146: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

114

6.2.1. Pool of Candidates

In this research, the pool of past IPOs is screened in order to find suitable comparables for

the IPO of interest (IPOI); that is, this study searches for the firms which were comparable to

the IPOI at their issue time. Choosing past IPOs as comparables offers a significant advantage

over selecting concurrent public firms to act as comparables. It enables the methodology to

control for IPO-specific conditions which impact the IPO offer price and short-term post-IPO

market prices. Note that the risk analysis proposed in this thesis draws on the post-IPO market

histories of the comparables in order to estimate the distribution of the IPOI’s stock price 90

trading days subsequent to the issuing day (see, Section 6.3). Past IPOs would provide a more

trustworthy basis for approximating such a distribution since IPO-specific characteristics are

reflected in their post-IPO market histories. Some of these characterises are listed below:

(i) Underwriters positively affect the success of IPOs. They gauge market sentiment

through road shows and underwriter-specific networks and can exercise

considerable latitude to shape the final offer price based on the investors’ perception

of the respective stock. Furthermore, underwriters may positively impact the

market’s interpretation of financial statements and other disclosures.

(ii) Earnings management is another prevalent phenomenon in IPOs. Issuers take

advantage of income increasing accounting choices in order to represent themselves

in the best light possible.

(iii) Issuers often enter into a price stabilization agreement with the underwriters; that is,

during a ‘relatively brief time limit’ (Geddes, 2003) [57], the underwriter may engage

in share repurchase, at the IPO offer price, in order to balance the excessive selling

pressure in the immediate post-IPO market (see, Section 2.5.2).

The impact of all the aforementioned practices and alike would be taken into account in the

risk analysis if the IPOI is compared to other IPOs instead of contemporary public firms.

In addition to controlling for structural characteristics of IPOs, limiting the pool of

candidates to past IPOs would guarantee access to sufficient historical market data for the

Page 147: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

115

comparable IPOs. Such data indeed represent the post-IPO fate of them and can be used to

estimate the distribution of the stock price for the respective IPOI, at the time point of 90

trading days after the issuing day (see, Section 6.3).

To further support the notion, this research submits that an IPOI and its prospective

comparable would be matched more closely in terms of growth potential and risk if the set of

comparables to IPOI comprises past IPOs, as opposed to concurrent public firms. An IPOI

may be in accordance with a public firm with respect to accounting and financial

fundamentals. Yet, different perceptions of future performance may be associated with each of

them, particularly if the public firm is considered to be at its equilibrium stage. Therefore, in

light of the differences in investors’ perspective of the growth potential and risk of the two

firms, it would be more prudent to compare the IPOI to other IPOs. Such approach can be

supported by the work of Teoh et al. (1998) [135] and Kim and Ritter (1999) [87]. In different

settings, they reported improved valuation outcomes when the IPO under scrutiny is matched

with other recent IPOs or past IPOs from the same industry.

In conclusion to this section, it is worth stressing that the data of the candidate IPOs would

belong to different points in time. To further elaborate on the pool of candidates and the data

set-up, a snapshot of the data is presented in Figure 6-3. In this case, the IPOI took place in

1997. The DEA model would select the comparables from the pool depicted below which

comprises the IPOs preceding the IPOI. The inputs and outputs are, therefore, adjusted for

inflation choosing the issue year of the IPOI as the base year. Furthermore, as explained in the

following section, GDP and the sector-specific index would control for economic productivity

and the performance of the respective sector that can impact other DEA inputs and outputs

such as net income.

6.2.2. Variable Selection

The comprehensive literature review presented in Section 3.4 (see, Pages 47-52) identifies

key value indicators in the market. The multi-dimensional comparable selection can be

founded on these factors. The survey demonstrates the significant impact of accounting

Page 148: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

116

variables on contemporary market prices and how recent accounting information can shape

investors’ expectations of future. It is perceived that there is a basis for the bids in the primary

(pre-IPO) and secondary (post-IPO) markets and investor pricing is significantly correlated by

the immediately preceding financial performance of the respective firm. Factors deemed as

key value drivers can be summarized as shown below:

(i) Revenue

(ii) Gross profit

(iii) Selling, general and administrative expenses

(iv) Net income

(v) Shareholder’s equity

(vi) Operating cash flow

Figure 6-3. A Snapshot of the Pool of Candidates for a Given IPOI in the ‘Health Care’ Sector.

Page 149: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

117

The choice of DEA inputs and outputs in this research draws upon the above-indicated

factors. The input-output classification of these elements is performed based on the following

principle: A factor is considered an ‘input’ (‘output’) if the efficient peers of a hypothetical

inefficient firm are expected to operate under reduced (increased) levels of the respective

factor. As a case in point, net income fits into the output category since an efficient firm is

expected to earn more income compared to its inefficient peers.

As discussed thoroughly in Chapter 4, DEA is a non-parametric technique which does not

require a priori knowledge of the functional form of the production process. However, inputs

and outputs should be selected appropriately so as to express the true production process. To

meet this criterion and based on the pre-IPO data availability, the inputs and outputs are

configured as follows (see, Figure 6-4):

Inputs

- Capital Expenditure (CAPX)

- Change in Working Capital Other than Cash (Change in WCOC)

- Debt in Current and Long-Term Liabilities (DCL)

- Gross Domestic Product

- S&P Sector Index

Outputs

- Operating Income before Depreciation (EBITDA)

- Net Income (NI)

- Operating Cash Flow (OANCF)

- Change in Cash and Cash Equivalents (CHECH)

- Shareholders’ Equity (SE)

Page 150: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

118

Note that for each IPO, the point factors pertain to the date of the last audited annual

statements released before the offering. Any period accounting factor, or any accounting

variable representing a change, is extracted from the two consecutive audited annual financial

Figure 6-4. General Layout of the DEA Model

Page 151: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

119

statements which immediately preceded the issue date. The granular discussion of data is

provided in Chapter 7. The following points are worth remarking regarding the inputs and

outputs:

Note 1. Annual Capital Expenditure (CAPX):

This input accounts for investment in fixed assets (i.e., property, plant and equipment).

Note 2. Change in Working Capital Other than Cash (ΔWCOC):

In addition to investments in long-term assets, represented by CAPX, a firm likewise

invests in net current assets or ‘net working capital’ (Ross et al., 2010 [122]). The

amount of such investment is represented by ΔWCOC. Equation (6.7), shown below,

presents the components which form ΔWCOC.

A positive value for ΔWCOC would imply that the firm has experienced a cash

outflow for the respective period. This cash outflow can be considered an investment

in the operations. Note that ΔWCOC does not include the changes in the ‘cash’

account which also belongs to the class of ‘current assets’. Due to the significance of

cash, most analysts tend to address this component independently (Ross et al., 2010

[122]). Considering that the increase in cash is the ultimate incentive of any business,

the change in ‘cash and cash equivalents’ (CHECH) is treated as a DEA output in this

study. CHECH represents the net cash flow from all operating, investing, and

financing activities in the respective period. It is important to note that a positive value

of CHECH would imply a cash inflow.

Change in Working Capital Other than Cash (6.7)

= Accounts Receivable: (Decrease) Increase

+ Inventory: (Decrease) Increase

+ Accounts Payable and Accrued Liabilities: (Increase) Decrease

Page 152: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

120

+ Income Taxes – Accrued: (Increase) Decrease

+ Assets and Liabilities: Other (Net Change)

Note 3. Debt (DCL) and Stockholders’ Equity (SE):

Financial leverage can raise the return to common shareholders, and hence, it is often

desirable to include some debt in the capital structure provided that the yielded return

on assets exceeds the after-tax cost of debt. Elevated debt levels, however, would

increase the bankruptcy risk as well as the credit risk, burdening further access to the

capital market. One measure that puts the level of debt into perspective is the debt-to-

equity ratio. It can be used to examine a firm’s long-term solvency risk and assess how

risky its capital structure is. The two constituents of this ratio should, therefore, be

factored in for identifying comparables for an IPOI. To further elaborate on these two

factors, it must be stated that in this research, DCL aggregates long-term debt and its

short-term component which is due in one-year (i.e., long-term debt due in one year).

Shareholders’ equity represents the total claim on the assets of a firm and consists of

the two accounts of preferred equity and common equity.

Note 4. Gross Domestic Product (GDP) and the S&P 500 Sector-Specific Index:

As noted earlier, the firms participating in the DEA analysis do not belong to the same

point in time. These two factors, therefore, proxy the state of the economy and the

mode of the respective sector at the issuing times of individual firms.

In the language of DEA, GDP and the sector index can be considered ‘non-

discretionary’ factors since they do impact the performance of different DMUs; yet,

they cannot be varied at the discretion of the analyst or the management team. The

reader is referred to Section 4.4 for a thorough discussion on this subject. Drawing on

the principle of input-output classification, introduced above, GDP and the sector

index are treated as inputs. This set up would imply that the efficient peers of a given

inefficient firm are expected to operate in a similar or inferior environment. Reversing

Page 153: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

121

the logic would suggest that a group of firms functioning in a more favorable

environment, and hence, yielding superior outputs would not form a suitable set of

efficient peers. The reason is that due to adverse environmental conditions, the

corresponding inefficient firms may not be able to achieve the improved targets which

are set by the efficient peers.

Note 5. S&P 500 Sector-Specific Index:

The S&P 500 sector-specific indices are computed using the large-cap companies in

the respective sectors. In this study, each sector comprises companies with different

sizes of market cap. Therefore, the ‘S&P 1500 composite’ sector-specific index would

be more representative of the state of the respective sector since it is derived based on a

combination of small-, mid- and large-cap companies. However, the scrutiny of S&P

1500 composite indices in all S&P sectors indicates that the data on these indices do

not cover the entire study period (1990-2012) and are supplied only after 1994 in the

COMPUSTAT database (see, Section 7.2). Therefore, the analysis is conducted using

the S&P 500 sector-specific index.

Note 6. Operating Income Before Depreciation (EBITDA):

Commonly referred to as EBITDA, this output denotes ‘net sales’; that is, ‘sales’ net of

‘cost of goods sold’ and ‘selling, general and administrative expenses’.

As a closing remark for this section, it should be reiterated that the DEA model in Figure

6-4 can be viewed as a black box which utilizes DCL, CAPX, and ΔWCOC – investments in

long-term and net current assets – to yield net income, operating cash flow, and the other

elements shown in Figure 6-4.

6.2.3. The DEA Model

At the onset of this section, it is worth restating the aim of this DEA model. It pursues the

objective of identifying the firms which were comparable to the IPOI at their issuing times.

Page 154: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

122

Each DMU is represented by one point in the 10-dimensional space of inputs and outputs. The

DEA model would map these points into a space with coordinates of 𝜆𝜆 and 𝜓𝜓. The parameter

𝜓𝜓 represents a new efficiency measure. It is determined as the value of the objective function

in the new DEA model (introduced below). It, however, continues to act as a measure of the

distance from the efficient frontier. Using two DEA parameters, 𝜆𝜆 and 𝜓𝜓, a new ‘similarity’

metric would then be developed to gauge the degree of similarity of these comparables to the

IPOI.

The scrutiny of DEA inputs and outputs in Section 6.2.2 establishes that the problem

encompasses non-discretionary factors as well as negative numbers. The DEA model is, thus,

required to address these two characteristics of the data. The following two subsections

(Sections 6.2.3.1 and 6.2.3.2) are focused on these aspects of the DEA analysis. In addition, in

this study, the efficiency score approximates the distance from the ‘fully efficient’ frontier. It

should, therefore, incorporate all sources of inefficiency. To meet this prerequisite, this

research will choose a slacks-based model. The radial models would not make suitable

candidates to serve this goal since in this study, the efficiency score of ‘1’ should indicate that

the firm under assessment is positioned on the ‘strong defining or fully efficient hyperplanes’

(Seidl and Sommersguter-Reichmann, 2011 [129]); whereas, in the case of radial models, the

efficiency score of ‘1’ is the ‘necessary’, but not the ‘sufficient’, condition to situate a firm on

the fully efficient hyperplanes. Under these models, a firm associated with the efficiency score

of ‘1’ may be located on the weakly efficient hyperplanes due the unaccounted for slacks.

6.2.3.1. Negative Data

In this thesis, some DEA inputs and outputs comprise a mix of negative and positive

numbers. Thus, the DEA model is expected to have the capacity to cope with this

characteristic of the data. The DEA framework was originally developed in the context of non-

negative inputs and outputs. Section 4.3 (see, Pages 64-79) comprehensively reviews the

approaches suggested for extending this framework to negative numbers. These methods can

be broadly classified as follows:

Page 155: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

123

(i) data translation

(ii) treatment of negative outputs (inputs) as positive inputs (outputs)

(iii) individual models such as ‘semi-oriented radial measure’ (SORM) proposed by

Emrouznejad et al. (2010) [49] or ‘variant of the radial measure’ (VRM) suggested by

Cheng et al. (2013) [38].

(iv) range directional models

Section 4.3 details these approaches and presents the respective shortcomings. Here, the

shortcomings associated with the second approach are revisited. Due to its simple theory, this

approach has gained a wide application; however, it is associated with significant limitations.

The SORM model was derived based on similar principles; that is, the model disaggregates the

vector of each input and output into two sub-vectors of positive and negative numbers. Thus,

in addition to the model-specific limitations, noted in Section 4.3.3, the SORM model would

suffer from similar defects.

Method (ii), indicated above, is mathematically impaired due to the following results:

(i) The most apparent shortcoming is the artificial increase in the number of inputs or

outputs which can adversely impact the optimization problem. This limitation

would be more pronounced in the case of sparse DMUs engaging in the analysis.

(ii) The most fundamental shortcoming associated with this approach can be explained

by referring to the basic goal pursued by DEA: It is employed to evaluate the

relative performance of different firms or decision making units. The performance

of each DMU can be represented by an efficiency score bounded between ‘0’ and

‘1’, with ‘1’ indicating ‘best performers’ which form a piece-wise linear frontier

and ‘envelope’ the inefficient DMUs. The best performers not only serve as

references to identify sources of inefficiency but also as benchmarks for the

inefficient DMUs to strive for. Dividing an individual input or output into two sub-

vectors would lead to the targets and efficiency scores which are fundamentally

different from what could have been achieved if the input or output vector were kept

intact, integrating both negative and positive numbers.

Page 156: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

124

This point can best be substantiated by the following example. Consider ‘net

income’ as a DEA output. This example seeks maximizing the net income, and it is,

thus, treated as an output. Suppose a percentage of the DMUs under scrutiny have

reported negative incomes for the period. Following the above-indicated steps

would lead to two sub-vectors. The sub-vector containing positive net incomes

would be considered as an output; whereas, the sub-vector including the absolute

values of the negative net incomes would be treated as an input since the problem

seeks minimizing the absolute values of these negative numbers. The best target that

the above-indicated DEA model could possibly set for this latter sub-vector would

be ‘zero’ net income. The reason is that the positive net incomes are located on the

output side, and from the pure mathematical point of view, bear zero relevance to

those on the input side. The optimization problem, therefore, does not consider the

DMUs with positive net incomes as potential targets for the DMUs reporting

negative incomes.

Using the same logic, it can be inferred that the efficiency scores and other

optimization results would, likewise, be fundamentally different for the DMUs with

positive net incomes. In analyzing these DMUs, the optimization problem assumes

that the minimum possible net income is ‘zero’. Based on the points highlighted

above, it can be readily asserted that allocating negative and positive numbers to

separate sub-vectors fundamentally changes the nature of the problem and puts

forward an essentially different affair.

(iii) To further comprehend the complications imposed by splitting an input or output

based on the sign, one can again refer to the original theory of DEA. Known as one

of the most significant functions of DEA, it unfolded the possibility of

simultaneously addressing multiple dimensions of a DMU’s activity through its

capacity to accommodate multiple inputs and outputs. In practice, it is perhaps

hardly feasible for a DMU to simultaneously minimize all its inputs and/or

maximize all its outputs. The DEA offers guidance as to which inputs and outputs

should receive prime attention. In the example provided above, if the vector of net

Page 157: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

125

income remains intact, a DMU’s negative income would be collectively analyzed

with the rest of its inputs and outputs. The DEA may deem other inputs or outputs

more effective on the firm’s operation, and hence, its results may advise the firm to

focus on those aspects to improve its performance.

(iv) The final shortcoming regarding this approach concerns ‘zero’ place holders. Refer

back to the example of ‘net income’ mentioned above. The presence of zeros which

simply act as the place holders of negative incomes would skew the optimization

outcomes; the DEA would tend to choose the DMUs associated with these zeros as

efficient units, particularly in an input-oriented model, since they have indeed

achieved the minimum value possible.

As concluded in the literature review of Section 4.3, the ‘range directional’ models – RDM

and MSBM suggested by Silva Portela et al. (2004) [132] and Sharp et al. (2007) [131],

respectively – offer the most comprehensive framework to cope with negative data,

particularly, in cases comprising a mix of positive and negative numbers. The range

directional models hold the characteristic of translation invariance while preserving the unit-

invariance feature as well as the true production process. In addition, the efficiency scores

yielded by these methods can be readily interpreted similar to the efficiency scores of

conventional DEA models such as BCC, CCR and SBM. Thus, the efficiency scores resulted

from the range directional models would suit the function considered for the efficiency score

in this thesis; that is, the efficiency score of a firm acts as an approximation for its distance

from the efficient frontier. It should, however, be noted that the RDM model is a radial model.

As detailed above, the efficiency score provided by such models may not incorporate all

sources of inefficiency. It may, therefore, represent the distance from a ‘weakly efficient’

hyperplane; whereas, this study aims at gauging the distance of any firm from the ‘strongly

efficient’ hyperplanes. Considering the limitations posed by radial models, this research would

draw on the slacks-based model of MSBM (Sharp et al., 2007 [131]) to estimate the similarity

metric. The model is represented below by Equations (6.8) to (6.12):

Page 158: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

126

Model 1: Modified Slacks-Based Measure [131]

min 𝜌𝜌 =𝜆𝜆, 𝑠𝑠−,𝑠𝑠+

�1 − 1𝑚𝑚∑ 𝑠𝑠𝑖𝑖−

𝑅𝑅𝑥𝑥𝑖𝑖𝑜𝑜𝑚𝑚𝑖𝑖=1 �

�1 + 1𝑠𝑠 ∑

𝑠𝑠𝑟𝑟+𝑅𝑅𝑦𝑦𝑟𝑟𝑜𝑜

𝑠𝑠𝑟𝑟=1 �

(6.8)

Eq. 6 [131]

Subject to 𝑠𝑠− + 𝑋𝑋𝜆𝜆 = 𝑥𝑥𝑜𝑜 (6.9)

𝑌𝑌𝜆𝜆 − 𝑠𝑠+ = 𝑦𝑦𝑜𝑜 (6.10)

𝑒𝑒𝜆𝜆 = 1 (6.11)

𝜆𝜆 ≥ 0, 𝑠𝑠− ≥ 0, 𝑠𝑠+ ≥ 0 (6.12)

The notation of this model is consistent with the one presented in Chapter 4. There exist 𝑚𝑚

inputs, 𝑠𝑠 outputs and 𝐽𝐽 firms. The parameters 𝑅𝑅𝑥𝑥𝑖𝑖𝑜𝑜 and 𝑅𝑅𝑦𝑦𝑟𝑟𝑜𝑜 denote ‘ranges of possible

improvement’ for input 𝑖𝑖 and output 𝑟𝑟 of DMUo, respectively and can be expressed as follows:

𝑅𝑅𝑥𝑥𝑖𝑖𝑜𝑜 = 𝑥𝑥𝑖𝑖𝑜𝑜 − min𝐽𝐽�𝑥𝑥𝑖𝑖𝑟𝑟�

(6.13)

𝑅𝑅𝑦𝑦𝑟𝑟𝑜𝑜 = max𝐽𝐽

�𝑦𝑦𝑟𝑟𝑟𝑟� − 𝑦𝑦𝑟𝑟𝑜𝑜 (6.14)

where, 𝑟𝑟 = 1, … , 𝑠𝑠

𝑖𝑖 = 1, … ,𝑚𝑚

𝑗𝑗 = 1, … , 𝐽𝐽

Page 159: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

127

Here, 𝑅𝑅𝑥𝑥𝑖𝑖𝑜𝑜 and 𝑅𝑅𝑦𝑦𝑟𝑟𝑜𝑜 are indeed equivalents of 𝑔𝑔𝑥𝑥𝑖𝑖𝑜𝑜 and 𝑔𝑔𝑦𝑦𝑟𝑟𝑜𝑜, presented in Equations (4.68)

and (4.69). The reader is referred to Section 4.3.5 (see, Pages 74-79) for further details on the

mechanism of the range directional models.

6.2.3.2. Non-Discretionary Factors

The problem at hand includes non-discretionary (ND) inputs: GDP and sector index. As

stated earlier, the comparable firms are expected to resemble the IPOI in terms of discretionary

factors such as period net income and net working capital (see, Section 6.2.2). These

comparable firms are likewise expected to operate in a similar environment. To be more

particular, consider a firm with a similar or higher net income which is compared to the IPOI.

The figure of net income encompasses the impact of the market. Thus, if the firm is running

during a ‘hot’ period of a sector, its net income is indeed not comparable to the net income of

the IPOI operating in a ‘cold’ period.

Section 4.4 comprehensively reviews the DEA studies which focused on the ND factors.

The objective functions suggested by these studies bear no direct relevance to the ND

variables. These factors impact the objective function only indirectly by means of their

presence as the constraints of the optimization program. Such approach is supported by the

viewpoint that the ND factors are exogenously fixed and cannot be enhanced at the discretion

of the management team. Thus, such factors should not be incorporated into the efficiency

score which provides a consolidated insight into the aspects which can be improved. The DEA

models addressing the ND factors can be classified as below (see, Section 4.4 for further

details).

Non-Radial Models

(i) The slacks associated with the ND factors are set to be zero which would a priori

obligate the respective dimensions to be on the efficient frontier.

(ii) The constraints of zero ND slacks are relaxed; yet, the ND slacks are not allowed in

the objective function.

Page 160: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

128

The first group ensures that the firm under scrutiny matches its efficient target with

respect to the exogenously fixed ND factors. In theory, this class may appear to

offer the supreme solution to the problem of ND variables. Further scrutiny of this

class, however, revealed that the optimizations would often be infeasible due to the

strict constraints. This point is further discussed below. It was also observed that in

most of the cases where the optimizations were executed successfully, these models

tend to set the firm under investigation as its own target. Such result is often

expected with stringent constraints.

Under the second class, the ND slacks are prone to be set such that significant

differences exist between the environments of the firm under scrutiny and its peers.

Radial Models:

This setting accommodates the characteristics of the ND factors by excluding the

radial efficiency parameter from the ND constraints. Thus, the ND variables cannot

be improved radially. If the analysis involves the second phase which is slack

maximization (see, Section 4.2), this group would suffer from the same

shortcomings associated with the non-radial models.

This study searches for a methodology which chooses the efficient target such that it

coincides with the respective firm of interest in terms of the ND factors. The methodology

should, however, resolve the infeasibility issue caused by too many stringent constraints,

forcing the ND slacks to be zero. To the best of the author’s knowledge, no DEA study has

accomplished the aforementioned objectives. The subsequent discussion outlines the

methodology proposed by this study to tackle this problem. It should be noted that this

discussion assumes ‘variable returns to scale’ (see, Section 4.1) since the analysis includes

firms of different sizes and scales.

The DEA analysis can be viewed as carrying out the following two tasks simultaneously:

(i) outlining the efficient frontier, and hence, the efficient firms, (ii) identifying the efficient

peers and computing the efficient targets for respective inefficient firms. To execute these two

Page 161: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

129

tasks, DEA would run 𝐽𝐽 optimizations given that the analysis includes 𝐽𝐽 DMUs. If the efficient

frontier (i.e., the convex hull of the existing DMUs) were provided in advance, the analysis

would then transform into a new problem: a mapping problem. This problem would address

how to map the inefficient DMUs onto the convex hull enveloping them and how to determine

the appropriate efficient targets for the respective inefficient DMUs. The constraints in the

DEA model indeed set the mapping rules. To provide examples, consider Equations (6.9) and

(6.10). These two equations indicate that the efficient target of DMUo is the one which utilizes

less inputs and produces more outputs than DMUo. These criteria would be met by more than

one point and it is the objective function that specifies which target should be selected. The

optimization program presented by Equations (6.8) to (6.12) tends to choose the farthest target

which satisfies all the constraints.

The previous paragraph clarifies that one can distinguish between the determination of the

efficient frontier and conducting the efficiency analysis, both of which are carried out

concurrently by DEA. To further clarify the point of the previous paragraph, it is important to

note that the efficient frontier (i.e., the convex hull) can be obtained solely based on the

geometrical properties of the data, with no need for 𝐽𝐽 optimizations. Based on this

interpretation, it can be concluded that if the structure of the convex hull is known, then by

means of an optimization problem, one can identify the desired efficient target for the

inefficient firm of interest. Figure 6-5 visualizes this point using a two-input and one-output

case. For the sake of simplicity, it is assumed that the DMUs generate identical outputs. The

vertical axis represents a non-discretionary input and the horizontal axis describes a

discretionary input. Figure 6-5 demonstrates that DMU E can be improved in different

directions and each direction would result in a different efficient target (e.g., E1′ ). It is the

constraints which determine the route of enhancement. If the ND slacks are set to zero, route 4

would be selected which leads to the efficient target of E4′ .

Page 162: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

130

In this thesis, a DEA model is used to set the rules for choosing the routes and efficient

targets (see, Equations (6.15) to (6.21) below). It is assumed that the efficient frontier is

known, and hence, each DEA analysis would only incorporate the DMU of interest (i.e., E)

and the efficient DMUs (i.e., A, B, C, and D). The constraints of this new DEA model can be

set so that it leads to efficient targets with desired characteristics. As an example, the ND slack

of DMU E can be set to zero in advance. In this approach, however, the optimization would

not be deteriorated by an excessive number of constraints. The reason is that the efficient

frontier is known a priori and any hyperplane represented by the constraints would eventually

intersect the frontier.

Figure 6-5. A Demonstration of Several Potential Improvement Directions for DMU E

Page 163: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

131

The methodology presented above assumes that the efficient frontier is known a priori. In

order to identify the efficient DMUs forming the efficient frontier, this thesis uses the MSBM

model (Sharp et al., 2007 [131]) presented above (Equations (6.8) to (6.12)). At this stage, the

analysis does not distinguish between discretionary and non-discretionary factors since this

stage pursues the sole purpose of determining the frontier. Figure 6-5 further clarifies this

point. In this simple plot, where the efficient frontier is readily observable, the discretionary

and non-discretionary factors are treated identically.

Once the efficient DMUs are detected by Model 1, the model presented in the following

page (i.e., Model 2) is used for each inefficient DMU of interest (shown by the subscript ‘𝑜𝑜’

below). The parameters 𝜋𝜋 and 𝜈𝜈 in Model 2 denote the numbers of discretionary and non-

discretionary inputs, respectively. The numbers of discretionary and non-discretionary outputs

are correspondingly displayed by 𝜉𝜉 and 𝜔𝜔. The number of DMUs used in each optimization is

reduced to 𝜅𝜅. This implies that the number of efficient DMUs identified by Model 1 is ‘𝜅𝜅 −

1’. The rest of the parameters preserve their meanings. The symbol ‘ ̂ ’ emphasizes that the

corresponding vectors have been trimmed down since only the efficient DMUs and one

inefficient DMU participate in each optimization.

Note. Model 2 is prone to be infeasible if the route of improvement dictated by the constraints

only intersects the weakly efficient hyperplanes of the efficient frontier. If for a given

inefficient DMU of interest, the optimization does not converge to a solution, the equality

signs in Equations (6.17) and (6.19) are relaxed to the corresponding non-strict inequality

signs. In this research, this change is consistent with the underlying principle of the DEA

framework which aims at setting achievable targets for an inefficient DMU since it implies

that the efficient peers operated in an environment inferior to that of the inefficient DMU.

Therefore, they can be accepted as potential targets for the inefficient DMU.

Page 164: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

132

Model 2

min𝜓𝜓 =𝜆𝜆 � , �̂�𝑠−,�̂�𝑠+

�1 − 1𝜋𝜋∑

�̂�𝑠𝑖𝑖−

𝑅𝑅�𝑥𝑥𝑖𝑖𝑜𝑜𝜋𝜋𝑖𝑖=1 �

�1 + 1𝜉𝜉 ∑

�̂�𝑠𝑟𝑟+𝑅𝑅�𝑦𝑦𝑟𝑟𝑜𝑜

𝜉𝜉𝑟𝑟=1 �

(6.15)

Subject to 𝑥𝑥�𝑖𝑖𝑜𝑜 = �𝑥𝑥�𝑖𝑖𝑖𝑖�̂�𝜆𝑖𝑖

𝜅𝜅

𝑖𝑖=1

+ �̂�𝑠𝑖𝑖− 𝑖𝑖 ∈ D

(6.16) 𝑖𝑖 = 1, … ,𝜋𝜋

𝑥𝑥�𝑂𝑂𝑜𝑜 = �𝑥𝑥�𝑂𝑂𝑖𝑖�̂�𝜆𝑖𝑖

𝜅𝜅

𝑖𝑖=1

𝑙𝑙 ∈ ND (6.17)

See, ‘Note’ on Page 131 𝑙𝑙 = 1, … , 𝜈𝜈

𝑦𝑦�𝑟𝑟𝑜𝑜 = �𝑦𝑦�𝑟𝑟𝑖𝑖�̂�𝜆𝑖𝑖

𝜅𝜅

𝑖𝑖=1

− �̂�𝑠𝑟𝑟+ 𝑟𝑟 ∈ D

(6.18) 𝑟𝑟 = 1, … , 𝜉𝜉

𝑦𝑦�𝑢𝑢𝑜𝑜 = �𝑦𝑦�𝑢𝑢𝑖𝑖�̂�𝜆𝑖𝑖

𝜅𝜅

𝑖𝑖=1

𝑢𝑢 ∈ ND (6.19)

See, ‘Note’ on Page 131 𝑢𝑢 = 1, … ,𝜔𝜔

�𝜆𝜆𝑖𝑖

𝜅𝜅

𝑖𝑖=1

= 1 (6.20)

𝜆𝜆𝑖𝑖 ≥ 0, 𝑗𝑗 = 1, … , 𝜅𝜅

�̂�𝑠𝑖𝑖− ≥ 0, 𝑖𝑖 ∈ D and 𝑖𝑖 = 1, … ,𝜋𝜋 (6.21)

�̂�𝑠𝑟𝑟+ ≥ 0, 𝑟𝑟 ∈ D and 𝑟𝑟 = 1, … , 𝜉𝜉

Page 165: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

133

6.2.4. Efficient IPOI Treatment

Based on the original principle suggested by Simak (2000) [133], each firm under

investigation would be compared to the following set of firms: It consists of other inefficient

firms which share identical efficient peers with the firm under investigation. The set also

includes the efficient peers themselves. As immediately becomes apparent, one of the

limitations of Simak’s (2000) [133] methodology is that it is not applicable to the cases where

the firm of interest is efficient itself (see, Section 3.3).

This problem is addressed in this thesis using an approach motivated by the slacks-based

measure of ‘super-efficiency’ in DEA (Tone, 2002 [137]; Cooper et al., 2007 [41]). Since DEA

assigns the efficiency score of ‘1’ to all efficient DMUs, the ‘super-efficiency’ method was

first proposed to rank the efficient DMUs (Andersen and Petersen, 2003 [8]). In order to gauge

the super-efficiency of an efficient DMU of interest (DMUo), this method essentially removes

DMUo from the set of DMUs and searches for a reference point P which would then perform

as the basis for measuring the super-efficiency score of DMUo. The point P is less efficient

than DMUo and lies in the new production possibility set; that is, the inputs (outputs)

associated with P are more (less) than the inputs (outputs) of DMUo. It is the objective

function and the constraints which would specify where P is desired to be located in the new

production possibility set.

Model 3, introduced in the next two pages (i.e., Equations (6.22) to (6.32)), solves a non-

linear optimization to find a point similar to P which would act as an analogue of the IPOI.

This point, referred to as U, is defined to be the closest point to IPOI which fulfills the

following conditions: (i) it shares identical non-discretionary inputs and outputs with the IPOI,

(ii) it lies on the new efficient frontier obtained after omitting the IPOI . The efficient IPOI

would then be approximated by point U and the set of comparables of the IPOI would consist

of (i) the new efficient DMUs whose weighted sum builds the point U, (ii) the inefficient

DMUs associated with these new efficient DMUs. Based on the discussion in Section 6.2.3.2,

it can be readily verified that the efficient DMUs identified by Model 3 are also efficient under

Models 1 and 2 if the IPOI is eliminated.

Page 166: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

134

Model 3

min�̅�𝑥, 𝑦𝑦�

1 + 1𝜋𝜋∑

�̅�𝑥𝑖𝑖 − �̿�𝑥𝑖𝑖max𝑁𝑁�𝑥𝑥𝑖𝑖𝑛𝑛� − min

𝑁𝑁�𝑥𝑥𝑖𝑖𝑛𝑛�

𝜋𝜋𝑖𝑖=1

1 − 1𝜉𝜉 ∑

𝑦𝑦�𝑟𝑟 − 𝑦𝑦�𝑟𝑟min𝑁𝑁�𝑦𝑦𝑟𝑟𝑛𝑛� − max

𝑁𝑁�𝑦𝑦𝑟𝑟𝑛𝑛�

𝜉𝜉𝑟𝑟=1

(6.22)

Subject to −�̅�𝑥𝑖𝑖 + � 𝑥𝑥𝑖𝑖𝑖𝑖𝜆𝜆𝑖𝑖

𝜅𝜅−1

𝑖𝑖=1𝑖𝑖≠IPOI

≤ 0 𝑖𝑖 ∈ D

(6.23) 𝑖𝑖 = 1, … ,𝜋𝜋

�̅�𝑥𝑂𝑂 = � 𝑥𝑥𝑂𝑂𝑖𝑖𝜆𝜆𝑖𝑖

𝜅𝜅−1

𝑖𝑖=1𝑖𝑖≠IPOI

𝑙𝑙 ∈ ND (6.24)

See, ‘Note’ on Page

136 𝑙𝑙 = 1, … , 𝜈𝜈

𝑦𝑦�𝑟𝑟 − � 𝑦𝑦𝑟𝑟𝑖𝑖𝜆𝜆𝑖𝑖

𝜅𝜅−1

𝑖𝑖=1𝑖𝑖≠IPOI

≤ 0 𝑟𝑟 ∈ D

(6.25) 𝑟𝑟 = 1, … , 𝜉𝜉

𝑦𝑦�𝑢𝑢 = � 𝑦𝑦𝑢𝑢𝑖𝑖𝜆𝜆𝑖𝑖

𝜅𝜅−1

𝑖𝑖=1𝑖𝑖≠IPOI

𝑢𝑢 ∈ ND (6.26)

See, ‘Note’ on Page

136 𝑢𝑢 = 1, … ,𝜔𝜔

� 𝜆𝜆𝑖𝑖

𝜅𝜅−1

𝑖𝑖=1𝑖𝑖≠IPOI

= 1 (6.27)

�̅�𝑥𝑂𝑂 = �̿�𝑥𝑂𝑂 𝑙𝑙 ∈ ND (6.28)

See, ‘Note’ on Page

136 𝑙𝑙 = 1, … , 𝜈𝜈

Page 167: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

135

𝑦𝑦�𝑢𝑢 = 𝑦𝑦�𝑢𝑢 𝑢𝑢 ∈ ND (6.29)

See, ‘Note’ on Page

136 𝑢𝑢 = 1, … ,𝜔𝜔

�̿�𝑥𝑖𝑖 ≤ �̅�𝑥𝑖𝑖 ≤ max𝑁𝑁�𝑥𝑥𝑖𝑖𝑛𝑛�

𝑖𝑖 ∈ D

(6.30) 𝑖𝑖 = 1, … ,𝜋𝜋

𝑛𝑛 = 1, … ,𝑁𝑁

min𝑁𝑁

{𝑦𝑦𝑟𝑟𝑛𝑛} ≤ 𝑦𝑦�𝑟𝑟 ≤ 𝑦𝑦�𝑟𝑟

𝑟𝑟 ∈ D

(6.31) 𝑟𝑟 = 1, … , 𝜉𝜉

𝑛𝑛 = 1, … ,𝑁𝑁

𝜆𝜆𝑖𝑖 ≥ 0 𝑗𝑗 = 1, … , (𝜅𝜅 − 1)

𝑗𝑗 ≠ IPOI (6.32)

In Model 3, the point U is represented by discretionary and non-discretionary inputs and

outputs of �̅�𝑥𝑖𝑖, �̅�𝑥𝑂𝑂, 𝑦𝑦�𝑟𝑟, 𝑦𝑦�𝑢𝑢, respectively. The discretionary and non-discretionary inputs and

outputs of the IPOI are correspondingly denoted by �̿�𝑥𝑖𝑖, �̿�𝑥𝑂𝑂, 𝑦𝑦�𝑟𝑟, 𝑦𝑦�𝑢𝑢. In the list of DMUs, the

IPOI is identified by the index of ‘IPOI’. Consistent with Model 2, there exist 𝜋𝜋 discretionary

inputs, 𝑣𝑣 non-discretionary inputs, 𝜉𝜉 discretionary outputs, and 𝜔𝜔 non-discretionary outputs.

The function max𝑁𝑁

(or min𝑁𝑁

) finds the maximum (or minimum) value of the respective input or

output across all the DMUs which were initially present in the analysis; that is, 𝑁𝑁 represents

the total number of DMUs in the original set which was supplied to Model 1. It is indeed equal

to 𝐽𝐽 in Model 1 and has been replaced by a new symbol in Model 3 to make a clear distinction

between 𝑗𝑗 = 1, … , (𝜅𝜅 − 1) and 𝑛𝑛 = 1, … ,𝑁𝑁. As spelled out above, (𝜅𝜅 − 1) designates the

number of efficient DMUs identified by Model 1 and 𝑁𝑁 counts all the DMUs fed into Model 1.

Page 168: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

136

Note. The optimization presented as Model 3 would be infeasible if the efficient IPOI resides

on the intersection of weakly and strongly efficient hyperplanes. In such a case, the equality

signs in Equations (6.24), (6.26), (6.28), and (6.29) are relaxed to the corresponding non-strict

inequality signs. The set of comparables of the efficient IPOI is then assumed to consist of the

new efficient DMUs corresponding to the point U and the inefficient DMUs associated with

these new efficient DMUs.

Beyond serving the specific features of the problem at hand, Model 3 expands the ‘super-

efficiency’ literature in DEA through satisfying all the following conditions simultaneously:

First. It is unit-invariant.

Second. It is translation-invariant and capable of handling mixes of negative and positive

values while keeping the vector containing the mix intact; that is, the proposed model does

not require segmenting a given vector of input or output into two sub-vectors of negative

and non-negative elements.

Third. It accommodates non-discretionary inputs and outputs.

Fourth. In compliance with the notion of super-efficiency, the objective function is

designed to yield a super-efficiency score of greater than unity. Furthermore, as indicated

above, in the context of super-efficiency, it is often desired to measure the super-

efficiency of an efficient DMU on the basis of a reference point which bears the closest

proximity to it (e.g., Tone, 2002 [137]). The objective function in Model 3 meets this

criterion and leads the optimization program to identify the closest reference point for the

respective efficient DMU, while preserving the properties of unit-invariance and

translation invariance.

To the best of the author’s knowledge, a model which fulfills all the aforementioned

requirements is unprecedented in the super-efficiency literature of DEA. The most intricate

challenge lies in devising the objective function. This research has developed and examined

various alternatives, none of which conforms to all the specifications listed above. Particularly,

the fourth criterion, which seeks the closest reference point, must be approached with care. As

Page 169: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

137

a case in point, the limitation of one candidate solution is outlined below. It caters to the

characteristics of unit-invariance and translation-invariance yet fails to properly address the

fourth condition.

The next equation presents the objective function suggested by Tone (2002) [137]. The

notation has been modified for the sake of consistency with Model 3.

min 𝛿𝛿 =

1𝜋𝜋∑

�̅�𝑥𝑖𝑖�̿�𝑥𝑖𝑖

𝜋𝜋𝑖𝑖=1

1𝜉𝜉 ∑

𝑦𝑦�𝑟𝑟𝑦𝑦�𝑟𝑟

𝜉𝜉𝑟𝑟=1

(6.33) [137]

As can be observed, Tone’s (2002) [137] objective function is unit-invariant; however, it is not

translation invariant and cannot be employed in the case of negative numbers. Thus, motivated

by the principles of range directional models, Equation (6.33) is modified into the following

form:

min 𝛿𝛿 =

1𝜋𝜋∑

𝑅𝑅�̅�𝑥𝑖𝑖𝑅𝑅�̿�𝑥𝑖𝑖

𝜋𝜋𝑖𝑖=1

1𝜉𝜉 ∑

𝑅𝑅𝑦𝑦�𝑟𝑟𝑅𝑅𝑦𝑦�𝑟𝑟

𝜉𝜉𝑟𝑟=1

(6.34)

In Equation (6.34), the ‘ranges of possible improvement’ for the inputs and outputs of a given

DMU can be computed using the two equations presented below:

𝑅𝑅𝑥𝑥𝑖𝑖 = 𝑥𝑥𝑖𝑖 − min𝐽𝐽�𝑥𝑥𝑖𝑖𝑟𝑟�

(6.35)

𝑅𝑅𝑦𝑦𝑟𝑟 = max𝐽𝐽

�𝑦𝑦𝑟𝑟𝑟𝑟� − 𝑦𝑦𝑟𝑟 (6.36)

Page 170: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

138

where, 𝑟𝑟 = 1, … , 𝜉𝜉

𝑖𝑖 = 1, … ,𝜋𝜋

𝑗𝑗 = 1, … , 𝐽𝐽

Equation (6.34) is both translation-invariant and unit-invariant. Yet, it violates the fourth

criterion posed above. As discussed earlier, the objective function is used to pick the ‘best’

case in the domain determined by the constraints. In minimizing Equation (6.34), the

optimization would tend to choose the closest inputs but the farthest outputs to the efficient

DMU of interest (or IPOI); whereas, the goal is to find the closest point to the efficient DMU

in terms of both inputs and outputs. Moreover, in Equation (6.34), 𝑅𝑅�̿�𝑥𝑖𝑖 or 𝑅𝑅𝑦𝑦�𝑟𝑟 may happen to

be zero which would render the optimization infeasible.

6.2.5. Algorithm of Phase I

The focus of all the previous sections was on developing the DEA models which suit the

specific objectives of this thesis as well as the characteristics of the inputs and outputs. This

section outlines how these models are implemented to compute the values of 𝜓𝜓 and 𝜆𝜆 for the

IPOI and its comparables.

The sector under scrutiny is assumed to include 𝑛𝑛 IPOs which took place at different

points in time. The first IPOI is then selected from this set such that the number of past IPOs is

sufficient to reliably conduct the DEA optimizations. The ID of 𝑏𝑏 is assigned to this IPOI.

Once this step is terminated, and the values of 𝜓𝜓 and 𝜆𝜆 are obtained for the IPOI and its

comparables, the analysis selects the IPO succeeding the first IPOI as the new IPOI. The same

steps are repeated to compute the values of 𝜓𝜓 and 𝜆𝜆 for the new IPOI and its comparables. The

remaining IPOs are selected sequentially in an identical manner for 𝜓𝜓 and 𝜆𝜆 calculations. The

outcomes generated in this step will later be required to calibrate the similarity metric (see,

Section 6.4). The process is visualized by the flowchart presented in Figure 6-6.

Page 171: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

139

The DEA analysis is sensitive to the presence of outliers. They affect the shape of the

efficient frontier and, consequently, the results. In this thesis, the “Jackstrap” approach

proposed by De Sousa and Stošić (2005) [45] is adopted to detect outliers. This bootstrapping

methodology draws on a measure referred to as ‘leverage’ to identify outliers. In essence,

‘leverage’ gauges the collective change in the efficiency scores of the retained DMUs when

one DMU is removed from the analysis. The reader is referred to the original paper for the

specifics. The “Jackstrap” method can be completely automated, and Figure 6-7 and Figure

6-8 demonstrate how the approach is implemented in this thesis.

It should be noted that the analysis in this research is designed such that if the IPOI is

identified as an outlier, the algorithm exits the process of outlier removal. The analysis is then

iterated without outlier detection on account of the fact that if the IPOI is an outlier itself,

other outliers are potentially more relevant to the IPOI than non-outliers.

The steps illustrated in Figure 6-6, Figure 6-7, and Figure 6-8 are implemented in the

MATLAB programming software. The written ‘functions’ can be categorized into three

separate libraries. The first library performs data mining and all the data processing operations.

The second library implements the DEA models. The third library performs the analysis of

determining the values of 𝜓𝜓 and 𝜆𝜆 in the format presented in Figure 6-6 to Figure 6-8. A total

of approximately 7000 lines of script were written to implement the ensuing flowcharts.

Once the values of 𝜓𝜓 and 𝜆𝜆 are determined for the IPOI and its comparables, the distance

between the IPOI and each of its comparables is calculated in the new space of 𝜓𝜓 and 𝜆𝜆’s,

using the following equation:

𝐷𝐷𝑐𝑐𝑐𝑐2 = �𝜆𝜆𝑐𝑐 − 𝜆𝜆𝑐𝑐�

2+ 𝛼𝛼(𝜓𝜓𝑐𝑐 − 𝜓𝜓𝑐𝑐)2

𝑐𝑐 = 1, … , 𝜂𝜂 (6.37)

Page 172: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

140

where, 𝐷𝐷𝑐𝑐𝑐𝑐 refers to the distance between the IPOI under scrutiny, identified by the ID of 𝑏𝑏,

and the comparable firm 𝑐𝑐. The number of comparables of the IPOI is denoted by 𝜂𝜂. The

parameter 𝛼𝛼 is a constant determined by calibrating the model (see, Section 6.4). Each term in

Equation (6.37) is an equivalent of the ‘Euclidean distance’ in a transformed space. The

distance 𝐷𝐷 is indeed a measure of the similarity between the IPOI and its individual

comparables. The most similar firm (or the closest comparable) would, therefore, be the one

which has the minimum distance from the IPOI in the transformed space.

Page 173: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

141

Figure 6-6. This flowchart presents the sequence of steps involved in Phase I of the analysis. The MATLAB

programming software is used to implement all the steps. The written ‘functions’ are classified in three different

libraries: (i) Data Mining and Data Processing, (ii) DEA Models, (iii) Analysis. In total, the libraries comprise

approximately 7000 lines of script.

Page 174: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

142

Figure 6-6. Continued.

Page 175: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

143

Figure 6-7. The outlier detection algorithm of De Sousa and Stošić (2005) [45] is implemented as

demonstrated by this flowchart.

Page 176: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

144

Figure 6-7. Continued.

Page 177: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.2. Phase I: Comparable Selection

145

Figure 6-8. This flowchart outlines the process of computing ‘threshold’, indicated in Figure 6-7.

Page 178: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.3. Phase II: Assessment of Short-Term Risk

146

6.3. Phase II: Assessment of Short-Term Risk

The ultimate objective of this thesis is to quantify the short-term risk of investing in IPOs.

Here, ‘risk’ is defined as a ‘value at risk’ (VAR) measure determined based on PDF90IPOI. This

abbreviation stands for the distribution of the IPOI’s stock price at the time point of 90 trading

days subsequent to the issue day. This section outlines the steps involved in estimating

PDF90IPOI.

6.3.1. Stock Pricing Model

The model of geometric Brownian motion (GBM) is selected to estimate PDF90IPOI. There

exist other mathematically intricate models that can be used to model stock prices. Abidin and

Jaffar (2012) [2] provide a comprehensive list of such models. Two examples from their list

would be ‘clustering-genetic fuzzy system’ (CGFS) (Hadavandi et al., 2010 [63]) and ‘Markov–

Fourier grey model’ (MFGM) (Hsu et al., 2009 [72]). The study indicates that such models are

only suitable for predicting long-term prices or the price of the ‘next day’.

The necessity of complex models in projecting long-term trends is not exclusive to stock

prices. Another example would be interest rate estimation. This framework is also

theoretically rigorous since interest rate-based derivatives, such as bonds and interest rate

swaps, are often long-term. Thus, the respective models are required to incorporate factors

such as long-term average and rate of mean reversion, all of which contribute to the

complexity of the framework.

GBM is deemed a reasonable model for finding the short-term evolution of stock prices.

The most well-known and prevalently-used option pricing formula – the Black-Scholes-

Merton formula (Black and Scholes, 1973 [27]; Merton, 1973 [98]) - draws upon GBM to model

stock prices. Considering that options are short-term investments, it can be inferred that GBM

is a suitable candidate to predict short-term stock prices. Perhaps the most important motive

for choosing GBM is its prevalent use among investors by virtue of its simplicity (Abidin and

Jaffar (2012) [2]). The outcomes of this model can potentially impact the stock trends since the

Page 179: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.3. Phase II: Assessment of Short-Term Risk

147

investors tend to base their analyses of future prices on GBM. This practice can consequently

influence their trading strategies as well as the types and amounts of the orders placed by

them.

To further justify the use of GBM in the context of this research, it should be noted that it is

advantageous in the following two respects:

(i) The assessment of risk draws on a short-term post-IPO history, and it is discussed

above that GBM is befitting in such cases.

(ii) The 90-day time window used by the methodology starts from the first trading day

which implies that no history exists prior to the start of this window. Thus, factors

such as long-term average or rate of mean reversion are not applicable in this case.

Considering that such factors would pose no concern in this problem, simpler

models can be adopted to describe short-term trends, and GBM is selected to serve

this purpose.

The stochastic process of price (𝑉𝑉) is said to follow GBM if it satisfies the following

stochastic differential equation:

d𝑉𝑉𝑡𝑡 = 𝜇𝜇𝑉𝑉𝑡𝑡d𝑡𝑡 + 𝜎𝜎𝑉𝑉𝑡𝑡d𝑍𝑍𝑡𝑡 (6.38)

In Equation (6.38), the parameter 𝜇𝜇 is a constant representing the expected rate of return for

the stock. The expected drift rate in 𝑉𝑉 can then be expressed as 𝜇𝜇𝑉𝑉. The volatility of the stock

price is denoted by the constant 𝜎𝜎. Note that the parameters 𝜇𝜇 and 𝜎𝜎 are draws from the

random variables Μ and Σ, respectively. The variable 𝑍𝑍 follows a Wiener process, and the

change in 𝑍𝑍 during an infinitesimally small time period (d𝑡𝑡) can be shown by the following

equation:

Page 180: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.3. Phase II: Assessment of Short-Term Risk

148

d𝑍𝑍𝑡𝑡 = 𝜖𝜖√d𝑡𝑡 (6.39)

where, the parameter 𝜖𝜖 represents a random variable which has a standard normal distribution

N(0,1).

6.3.2. Distribution of Stock Price 90 Days after the Issuing Day

At the onset of this section it was indicated that this research pursues the specific goal of

finding the distribution of the IPOI’s price at the time point of 90 trading days after the issuing

day. This price, henceforth, is referred to as 𝑉𝑉90. In order to estimate the distribution of 𝑉𝑉90,

one of the required pieces of information is the starting point of the stochastic process of price;

that is, the closing price of the first trading day (𝑉𝑉1). Prior to delving into the details of finding

the distribution of 𝑉𝑉90, the subsequent three paragraphs explain the conjecture on the possible

distribution of 𝑉𝑉1.

As will be discussed in Chapter 7, the data of the ‘Health Care’ sector (Sector 35 under

GICS) are used to examine the proposed methodology in this thesis. For all the 327 IPOs,

which span the time period of 1990-2012, the histogram of the following ratio is plotted in

Figure 6-9: closing price of the first day to IPO offer price. The histogram closely resembles a

log-normal distribution.

Note that the mode of approximately ‘1’ in Figure 6-9 could have been caused by

underwriters’ price stabilizing activities. If the IPOI is a ‘broke issue’ and the underwriters

have been granted the ‘overallotment’ option, they can limit the stock supply in the market by

buying back some of the ‘overalloted’ shares. This practice would prevent or delay the fall of

prices below the offer price (see, Section 2.5.2) and would lead to a first-day closing price in

the vicinity of the offer price. It can, therefore, be concluded that the closing price of the first

day would not often reflect the true degree of overpricing. If the stock price is overpriced with

respect to its short-term performance, this property is expected to be captured by a negative

drift in GBM.

Page 181: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.3. Phase II: Assessment of Short-Term Risk

149

Based on the observation depicted in Figure 6-9, it is assumed that the ratio of closing price of

the first trading day (𝑉𝑉1) to offer price (𝑉𝑉offer) is a random variable which follows a log-normal

distribution with the corresponding normal distribution of N(𝜉𝜉,𝜔𝜔2). Consequently, the

distribution of 𝑉𝑉1 for each IPO can be assumed to be log-normal since the offer price is known

a priori for the corresponding IPO. Note that the parameters 𝜉𝜉 and 𝜔𝜔 are themselves assumed

to be draws (realizations) from the random variables of Ξ and Ω, respectively. The parameter

𝑣𝑣1 denotes a realization from the log-normal distribution of the random variable 𝑉𝑉1.

As established, the price is assumed to follow a GBM. The only unknown parameters of the

GBM are 𝜇𝜇, 𝜎𝜎 and 𝑣𝑣1. If these parameters are known, the conditional PDF of 𝑉𝑉90 given 𝜇𝜇, 𝜎𝜎

Figure 6-9. The figure presents the histogram of the ratio of ‘closing price of the first day to offer price’

for 327 IPOs in the ‘health care’ sector. The IPOs span the time period of 1990-2012.

Page 182: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.3. Phase II: Assessment of Short-Term Risk

150

and 𝑣𝑣1 (i.e., ℒ𝑉𝑉90(𝑣𝑣90|Μ ≈ 𝜇𝜇, Σ ≈ 𝜎𝜎,𝑉𝑉1 ≈ 𝑣𝑣1)) can be obtained as explained below. The

approximate equality Μ ≈ 𝜇𝜇 can be more precisely expressed as 𝜇𝜇 < Μ ≤ 𝜇𝜇 + d𝜇𝜇. The

approximate equalities of Σ ≈ 𝜎𝜎 and 𝑉𝑉1 ≈ 𝑣𝑣1 can be similarly interpreted.

For arbitrary values of 𝜇𝜇, 𝜎𝜎 and 𝑣𝑣1, where 𝑣𝑣1 denotes the starting point of the stochastic

process at the time 𝑡𝑡 = 1, the analytic solution of the stochastic differential equation of GBM

is presented below:

𝑉𝑉𝑡𝑡𝑣𝑣1

= exp��𝜇𝜇 −𝜎𝜎2

2� 𝑡𝑡 + 𝜎𝜎𝑍𝑍𝑡𝑡� (6.40)

→ 𝑉𝑉𝑡𝑡 = exp �ln 𝑣𝑣1 + �𝜇𝜇 −𝜎𝜎2

2� 𝑡𝑡 + 𝜎𝜎𝑍𝑍𝑡𝑡�

Equation (6.40) implies that (ln𝑉𝑉𝑡𝑡) is a draw from the normal distribution N �ln𝑣𝑣1 +

�𝜇𝜇 − 𝜎𝜎2

2� 𝑡𝑡,𝜎𝜎2𝑡𝑡�. The variable 𝑉𝑉𝑡𝑡 is, thus, log-normally distributed. The log-normal

distribution of 𝑉𝑉𝑡𝑡 can be expressed as follows:

ℒ𝑉𝑉𝑡𝑡(𝑣𝑣𝑡𝑡|Μ ≈ 𝜇𝜇, Σ ≈ 𝜎𝜎,𝑉𝑉1 ≈ 𝑣𝑣1, 𝑡𝑡) (6.41)

=1

√2𝜋𝜋 𝜎𝜎√𝑡𝑡𝑣𝑣𝑡𝑡exp

⎜⎛− �ln 𝑣𝑣𝑡𝑡𝑣𝑣1

− �𝜇𝜇 − 𝜎𝜎22 � 𝑡𝑡�

2

2𝜎𝜎2𝑡𝑡

⎟⎞

Given the parameters 𝜇𝜇, 𝜎𝜎, and 𝑣𝑣1, the distribution of 𝑉𝑉90 can be estimated using Equation

(6.41). However, these three parameters are not deterministic and represent realizations from

Page 183: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.3. Phase II: Assessment of Short-Term Risk

151

the random variables Μ, Σ, and 𝑉𝑉1. Particularly, 𝑉𝑉1 is a draw from a distribution whose

parameters are themselves random. As discussed above, the dynamics of the ratio of first

closing price to offer price is best described by a log-normal distribution whose two unknown

parameters (𝜉𝜉 and 𝜔𝜔) denote realizations from the random variables Ξ and Ω, respectively.

Considering that Ξ and Ω could potentially be correlated with Μ and Σ, their distributions must

be considered jointly in order to find the distribution of 𝑉𝑉90. The function 𝑓𝑓Μ,Σ,Ξ,Ω(𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔)

designates this joint distribution and is employed to derive the distribution of 𝑉𝑉90 (i.e.,

PDF90IPOI) as demonstrated by the equation established below:

𝑓𝑓𝑉𝑉90(𝑣𝑣90) (6.42)

= � � � � � � ℒ𝑉𝑉90(𝑣𝑣90|Μ ≈ 𝜇𝜇, Σ ≈ 𝜎𝜎,𝑉𝑉1 ≈ 𝑣𝑣1) 𝑓𝑓V1(𝑣𝑣1|𝜉𝜉,𝜔𝜔)d𝑣𝑣1

+∞

𝑣𝑣1=−∞

�+∞

𝜇𝜇=−∞

+∞

𝜎𝜎=−∞

+∞

𝜉𝜉=−∞

+∞

𝜔𝜔=−∞

× 𝑓𝑓Μ,Σ,Ξ,Ω(𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔) d𝜇𝜇d𝜎𝜎d𝜉𝜉d𝜔𝜔

where, 𝑓𝑓𝑉𝑉90(𝑣𝑣90) symbolizes PDF90IPOI. Note that ℒ𝑉𝑉90 denotes the conditional distribution of

the random variable 𝑉𝑉90, given the event of (Μ ≈ 𝜇𝜇, Σ ≈ 𝜎𝜎,𝑉𝑉1 ≈ 𝑣𝑣1); whereas, 𝑓𝑓𝑉𝑉90(𝑣𝑣90)

represents the unconditional distribution of 𝑉𝑉90. It is computed by integrating (averaging) over

all values of the underlying parameters and, hence, bears no dependency on any of them.

Through the integration process, Equation (6.42) indeed accounts for the ‘epistemic’

uncertainty associated with 𝑉𝑉90 (see, Section 5.4, Page 98). The only untackled component of

Equation (6.42) remains 𝑓𝑓Μ,Σ,Ξ,Ω(𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔) and the next section (Section 6.3.2.1) is focused on

outlining the approach suggested for its estimation.

6.3.2.1. Estimating the Joint Distribution of the Parameters

Prior to delineating the method used to estimate 𝑓𝑓Μ,Σ,Ξ,Ω(𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔), it should be noted that

the IPOs which issued prior to the 1st IPOI are, henceforth, referred to as ‘pioneers’. Pioneers

Page 184: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.3. Phase II: Assessment of Short-Term Risk

152

can be considered as the starting ‘pool of candidates’ (see, Section 6.2.1); no IPO from this list

is scrutinized as an IPOI. For further clarification on the terminology, consider Figure 6-10.

This figure depicts the ‘pool of candidates’ for each IPOI using the data of the ‘health care’

sector. To ensure the reliability of the outcomes of Phase I, which is developed in the DEA

framework, the first 51 IPOs are reserved as the starting pool of candidates (i.e., the pioneer

group).

Estimating 𝑓𝑓Μ,Σ,Ξ,Ω(𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔) involves two main steps: (i) estimating a prior distribution for

𝑓𝑓Μ,Σ,Ξ,Ω(𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔), (ii) estimating a posterior distribution for 𝑓𝑓Μ,Σ,Ξ,Ω(𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔) through

updating the prior distribution using the Bayesian framework. The two steps are detailed in the

next subsections (i.e., Sections 6.3.2.1.1 and 6.3.2.1.2).

6.3.2.1.1. Estimating the Posterior Distributions

In order to estimate 𝑓𝑓Μ,Σ,Ξ,Ω(𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔), first it is assumed that there exists a prior

distribution for 𝑓𝑓Μ,Σ,Ξ,Ω(𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔) which is known. It is explained later how this prior

distribution is estimated for individual IPOs (see, Section 6.3.2.1.2). The Bayesian framework

is subsequently employed to update the prior distribution of each IPO. The updating process is

carried out using the 90-day realized trajectory of stock prices for the respective IPO. The

starting point of each trajectory is the closing price of the first trading day (𝑣𝑣1), and the

frequency of data is daily. Below, it is demonstrated mathematically how to update the prior

distribution:

𝑓𝑓Μ,Σ,Ξ,Ωpost𝑔𝑔 �𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔|𝑢𝑢, ln

𝑣𝑣2𝑣𝑣1

, ln𝑣𝑣3𝑣𝑣2

, … , ln𝑣𝑣𝑛𝑛−2𝑣𝑣𝑛𝑛−3

, ln𝑣𝑣𝑛𝑛−1𝑣𝑣𝑛𝑛−2

, ln𝑣𝑣𝑛𝑛𝑣𝑣𝑛𝑛−1

� (6.43)

=P �𝑢𝑢, ln𝑣𝑣2𝑣𝑣1

, ln𝑣𝑣3𝑣𝑣2, … , ln𝑣𝑣𝑛𝑛−2𝑣𝑣𝑛𝑛−3

, ln𝑣𝑣𝑛𝑛−1𝑣𝑣𝑛𝑛−2, ln 𝑣𝑣𝑛𝑛

𝑣𝑣𝑛𝑛−1�𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔�𝑓𝑓Μ,Σ,Ξ,Ω

prior𝑔𝑔 (𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔)

∫ ∫ ∫ ∫ P �𝑢𝑢, ln𝑣𝑣2𝑣𝑣1, … , ln 𝑣𝑣𝑛𝑛−1𝑣𝑣𝑛𝑛−2

, ln 𝑣𝑣𝑛𝑛𝑣𝑣𝑛𝑛−1

��̅�𝜇,𝜎𝜎�, 𝜉𝜉̅,𝜔𝜔��+∞𝜇𝜇�=−∞

+∞𝜎𝜎�=−∞

+∞𝜉𝜉�=−∞

+∞𝜔𝜔�=−∞ 𝑓𝑓Μ,Σ,Ξ,Ω

Prior𝑔𝑔��̅�𝜇,𝜎𝜎�, 𝜉𝜉̅,𝜔𝜔��d�̅�𝜇d𝜎𝜎�d𝜉𝜉̅d𝜔𝜔�

Page 185: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.3. Phase II: Assessment of Short-Term Risk

153

1 1 1 1

1 1

2 2 2 2 2 2

3 3 3 3 3 3

⋮ ⋮ ⋮ ⋮ ⋮ ⋮

50 50 50 50 50 50

51 51 51 51 51 51

52

(IPOI1) 52 52 52 52 52

53

(IPOI2) 53 53 53 53

54

(IPOI3) 54 54 54

55

(IPOI4) 55 55

⋱ 56

56

57 57

⋮ ⋮

323 323

324 324

325 325

326

(IPOI275) 326

327 (IPOI276)

Pool

of C

andi

date

s for

IPO

I 1

Figure 6-10. Using the data of the ‘health care’ sector (Sector 35 under GICS), this figure provides an

overview of the IPOIs and their ‘pools of candidates’. Each column presents the IPOs used to carry out

Phase I and Phase II for the respective IPOI. The reader is referred to Section 6.2.1 for further

information on the pool of candidates. The first 51 IPOs are referred to as ‘pioneers’.

Pool

of C

andi

date

s for

IPO

I 2

Pool

of C

andi

date

s for

IPO

I 3

Pool

of C

andi

date

s for

IPO

I 4

Page 186: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.3. Phase II: Assessment of Short-Term Risk

154

In Equation (6.43), the variable 𝑢𝑢 stands for the ratio of closing price of the first trading day

(𝑣𝑣1) to offer price (𝑣𝑣offer). The subscript 𝑔𝑔 identifies the IPO under scrutiny and indicates that

the respective variable/distribution is ascribed to the 𝑔𝑔th IPO. The realized trajectory of the 𝑔𝑔th

IPO is a vector of size 𝑛𝑛. The function 𝑓𝑓Μ,Σ,Ξ,Ωprior𝑔𝑔(𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔) represents the prior form of

𝑓𝑓Μ,Σ,Ξ,Ω(𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔) for the 𝑔𝑔th IPO, and Equation (6.43) solves for the respective posterior

distribution which is denoted by 𝑓𝑓Μ,Σ,Ξ,Ωpost𝑔𝑔 �𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔|𝑢𝑢, ln 𝑣𝑣2

𝑣𝑣1, … , ln 𝑣𝑣𝑛𝑛

𝑣𝑣𝑛𝑛−1�.

Since 𝑓𝑓Μ,Σ,Ξ,Ωprior𝑔𝑔(𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔) is assumed to be known at this stage, the only unknown in

Equation (6.43) is the probability P �𝑢𝑢, ln 𝑣𝑣2𝑣𝑣1

, ln 𝑣𝑣3𝑣𝑣2

, … , ln 𝑣𝑣𝑛𝑛−2𝑣𝑣𝑛𝑛−3

, ln 𝑣𝑣𝑛𝑛−1𝑣𝑣𝑛𝑛−2

, ln 𝑣𝑣𝑛𝑛𝑣𝑣𝑛𝑛−1

�𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔� which

can be calculated as detailed next.

For a given combination of 𝜇𝜇, 𝜎𝜎 and two consecutive observed prices (i.e., 𝑣𝑣𝑖𝑖 and 𝑣𝑣𝑖𝑖+1,

where 𝑖𝑖 = 1, … , (𝑛𝑛 − 1)), the following equation can be inferred from the analytic solution of

GBM, shown by Equation (6.40):

𝜖𝜖𝑖𝑖 =ln𝑣𝑣𝑖𝑖+1𝑣𝑣𝑖𝑖

− �𝜇𝜇 − 𝜎𝜎22 � Δ𝑡𝑡

𝜎𝜎√Δ𝑡𝑡 (6.44)

𝑖𝑖 = 1, … , (𝑛𝑛 − 1)

Equation (6.44) indicates that if 𝑣𝑣𝑖𝑖, 𝑣𝑣𝑖𝑖+1, 𝜇𝜇 and 𝜎𝜎 are given, the value of 𝜖𝜖𝑖𝑖 can be determined

recursively by substituting 𝑣𝑣𝑖𝑖, 𝑣𝑣𝑖𝑖+1, 𝜇𝜇 and 𝜎𝜎 back into the analytic solution of GBM. The

variable 𝜖𝜖𝑖𝑖 indeed represents the value of the 𝑖𝑖th random draw from the standard normal

distribution which, in conjunction with 𝑣𝑣𝑖𝑖, 𝜇𝜇 and 𝜎𝜎, led to the observed price of 𝑣𝑣𝑖𝑖+1.

Page 187: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.3. Phase II: Assessment of Short-Term Risk

155

Based on Equation (6.44), it can be concluded that the probability

P �𝑢𝑢, ln 𝑣𝑣2𝑣𝑣1

, ln 𝑣𝑣3𝑣𝑣2

, … , ln 𝑣𝑣𝑛𝑛−2𝑣𝑣𝑛𝑛−3

, ln 𝑣𝑣𝑛𝑛−1𝑣𝑣𝑛𝑛−2

, ln 𝑣𝑣𝑛𝑛𝑣𝑣𝑛𝑛−1

�𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔�, which appears in Equation (6.43), can

be computed as follows:

P �𝑢𝑢, ln𝑣𝑣2𝑣𝑣1

, ln𝑣𝑣3𝑣𝑣2

, … , ln𝑣𝑣𝑛𝑛−2𝑣𝑣𝑛𝑛−3

, ln𝑣𝑣𝑛𝑛−1𝑣𝑣𝑛𝑛−2

, ln𝑣𝑣𝑛𝑛𝑣𝑣𝑛𝑛−1

�𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔� (6.45)

= P(𝑢𝑢, 𝜖𝜖1, 𝜖𝜖2, … , 𝜖𝜖𝑛𝑛−2, 𝜖𝜖𝑛𝑛−1|𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔)

= P(𝑢𝑢|𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔) × P(𝜖𝜖1|𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔) × … × P(𝜖𝜖𝑛𝑛−1|𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔)

= 𝑓𝑓𝑈𝑈logN(𝑢𝑢; 𝜉𝜉,𝜔𝜔)d𝑢𝑢 × 𝑓𝑓ΕN(𝜖𝜖1; 0,1)d𝜖𝜖 × … × 𝑓𝑓ΕN(𝜖𝜖𝑛𝑛−1; 0,1)d𝜖𝜖

= 1

𝜔𝜔𝑢𝑢√2𝜋𝜋exp�−

(ln𝑢𝑢 − 𝜉𝜉)2

2𝜔𝜔2 �d𝑢𝑢 ×1

(2𝜋𝜋)(𝑛𝑛−1)/2 exp�− 12�𝜖𝜖𝑖𝑖2𝑛𝑛−1

𝑖𝑖=1

�d𝜖𝜖𝑛𝑛−1

where, the probability P(𝑢𝑢, 𝜖𝜖1, 𝜖𝜖2, 𝜖𝜖3, … , 𝜖𝜖𝑛𝑛−2, 𝜖𝜖𝑛𝑛−1|𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔) is calculated as the product of

the probabilities of the realized points since they are independent from each other. The

computation of each term in Equation (6.45) is straightforward: Each element of the (𝑛𝑛 − 1)-

tuple of (𝜖𝜖1, 𝜖𝜖2, 𝜖𝜖3, … , 𝜖𝜖𝑛𝑛−2, 𝜖𝜖𝑛𝑛−1) is a sample from the standard normal distribution whose

PDF is denoted by 𝑓𝑓ΕN(𝜖𝜖; 0,1). With regard to 𝑢𝑢, as explained earlier, it is derived from a log-

normal distribution, represented by 𝑓𝑓𝑈𝑈logN(𝑢𝑢; 𝜉𝜉,𝜔𝜔), in Equation (6.45).

6.3.2.1.2. Estimating the Prior Distributions

In the case of the ‘pioneers’ (i.e., the first 51 IPOs in Figure 6-10), the prior distribution

required to solve Equation (6.43) is assumed to be ‘uniform’ – drawing on the notion of ‘non-

informative’ priors detailed in Section 5.3. Note that for each IPO, this distribution is

subsequently updated based on the respective IPO’s realized market data. Hence, the final

outcomes are not significantly sensitive to the primary form conjectured for the distribution. In

Page 188: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.3. Phase II: Assessment of Short-Term Risk

156

order to estimate the prior distribution for the IPOs which occur subsequent to the ‘pioneers’,

the following method is used:

Starting from the 1st IPOI, its prior distribution is approximated by the posterior distribution

of its closest comparable. The closest comparable is defined to be a preceding IPO which has

the minimum distance from the respective IPOI in the 𝜆𝜆 and 𝜓𝜓 space (see, Section 6.2.5). As

presented by Equation (6.37) (repeated below), the distance between the IPOI under scrutiny,

identified by the ID of 𝑏𝑏, and its comparable 𝑐𝑐 can be quantified as follows:

𝐷𝐷𝑐𝑐𝑐𝑐2 = �𝜆𝜆𝑐𝑐 − 𝜆𝜆𝑐𝑐�2

+ 𝛼𝛼(𝜓𝜓𝑐𝑐 − 𝜓𝜓𝑐𝑐)2 (6.37)

Example: Consider the 1st IPOI (i.e., the 52nd IPO). The set of comparables to this IPOI

consists of the following IPOs: [8, 16, 20, 46]. The 20th IPO exhibits the minimum distance

from the 1st IPOI; that is, 𝐷𝐷20,52 is the minimum of all 𝐷𝐷𝑐𝑐,52’s. Thus, the prior distribution of

the 1st IPOI (i.e., 𝑓𝑓Μ,Σ,Ξ,Ωprior52(𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔)) is assumed to be equal to the posterior distribution of the

20th IPO (i.e., 𝑓𝑓Μ,Σ,Ξ,Ωpost20 �𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔|𝑢𝑢20, ln 𝑣𝑣20,2

𝑣𝑣20,1, ln 𝑣𝑣20,3

𝑣𝑣20,2, … , ln 𝑣𝑣20,𝑛𝑛−2

𝑣𝑣20,𝑛𝑛−3, ln 𝑣𝑣20,𝑛𝑛−1

𝑣𝑣20,𝑛𝑛−2, ln 𝑣𝑣20,𝑛𝑛

𝑣𝑣20,𝑛𝑛−1�).

Once the 1st IPOI (i.e., the 52nd IPO) is realized, its prior distribution is updated based on

the respective stock prices observed in the market. The estimated posterior distribution of the

52nd IPO is then recorded to be used when it is selected as the closest comparable to a

subsequent IPOI. In that event, the posterior distribution of the 52nd IPO would act as the prior

distribution for the corresponding IPOI. The approach outlined above for finding the prior and

posterior distributions of the 1st IPOI can similarly be applied to any IPOI which issues after

the 1st IPOI.

Page 189: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.4. Calibrating the Distance Equation

157

6.4. Calibrating the Distance Equation

In order to estimate the distance between a given IPOI and its comparables, the following

equation was suggested in Phase I:

𝐷𝐷𝑐𝑐𝑐𝑐2 = �𝜆𝜆𝑐𝑐 − 𝜆𝜆𝑐𝑐�

2+ 𝛼𝛼(𝜓𝜓𝑐𝑐 − 𝜓𝜓𝑐𝑐)2

𝑐𝑐 = 1, … , 𝜂𝜂 (6.37)

where, the subscript 𝑐𝑐 identifies the respective comparable and 𝜂𝜂 stands for the total number of

comparables for the corresponding IPOI which is assigned the ID of 𝑏𝑏. This section is focused

on outlining the approach proposed to estimate the optimum value of 𝛼𝛼. The calibration

process seeks to find a model (i.e., 𝛼𝛼) under which the likelihood of the observed prices (i.e.,

𝑉𝑉90’s) is maximum.

First, 𝛼𝛼 is assigned a value. Then, the proposed methodology in Phase II is used to obtain

the distribution of each IPOI’s stock price at the time point of 90 days after the issue day (i.e.,

𝑓𝑓𝑉𝑉90(𝑣𝑣90)). Subsequently, the probability of each IPOI’s realized price at day 90 is computed

using its 𝑓𝑓𝑉𝑉90(𝑣𝑣90). Considering the ‘health care’ sector as an example (see, Figure 6-10), the

above steps imply that there exist a total of 276 distributions (i.e., 𝑓𝑓𝑉𝑉90(𝑣𝑣90)’s) under a given

𝛼𝛼. Consequently, there exist 276 numbers, each specifying the probability of the realized 𝑣𝑣90

of the respective IPOI using its specific 𝑓𝑓𝑉𝑉90(𝑣𝑣90).

Second, the likelihood of the realized prices under the assumed 𝛼𝛼 is computed as shown

below:

P�𝛼𝛼�𝑉𝑉90IPOI1 ≈ 𝑣𝑣90

IPOI1 ,𝑉𝑉90IPOI2 ≈ 𝑣𝑣90

IPOI2 , … ,𝑉𝑉90IPOIG ≈ 𝑣𝑣90

IPOIG� (6.46)

=P�𝑉𝑉90

IPOI1 ≈ 𝑣𝑣90IPOI1 ,𝑉𝑉90

IPOI2 ≈ 𝑣𝑣90IPOI2 , … ,𝑉𝑉90

IPOIG ≈ 𝑣𝑣90IPOIG�𝛼𝛼�P(𝛼𝛼)

P�𝑉𝑉90IPOI1 ≈ 𝑣𝑣90

IPOI1 ,𝑉𝑉90IPOI2 ≈ 𝑣𝑣90

IPOI2 , … ,𝑉𝑉90IPOIG ≈ 𝑣𝑣90

IPOIG�

Page 190: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.4. Calibrating the Distance Equation

158

In this equation, G refers to the total number of IPOIs. It is equal to 276 in the case of the

‘health care’ sector. The approximate equality of 𝑉𝑉90IPOI1 ≈ 𝑣𝑣90

IPOI1 abbreviates the following

inequality: 𝑣𝑣90IPOI1 < 𝑉𝑉90

IPOI1 ≤ 𝑣𝑣90IPOI1 + d𝑣𝑣90.

The steps indicated above are iterated using different values of 𝛼𝛼. It is assumed that all the

values of 𝛼𝛼 are equally likely and the variable 𝛼𝛼 follows a uniform distribution which implies

that P(𝛼𝛼) is a constant. Therefore, maximizing Equation (6.46) translates into maximizing the

probability P�𝑉𝑉90IPOI1 ≈ 𝑣𝑣90

IPOI1 ,𝑉𝑉90IPOI2 ≈ 𝑣𝑣90

IPOI2 , … ,𝑉𝑉90IPOIG ≈ 𝑣𝑣90

IPOIG|𝛼𝛼� which expresses the

likelihood of 𝛼𝛼. Therefore, the optimum 𝛼𝛼 for the distance function (i.e., Equation (6.37)) is

the 𝛼𝛼 whose likelihood is maximum, given the observed values of 𝑉𝑉90 across all the IPOIs.

Given 𝛼𝛼, the following equation always holds for the probability

P�𝑉𝑉90IPOI1 ≈ 𝑣𝑣90

IPOI1 ,𝑉𝑉90IPOI2 ≈ 𝑣𝑣90

IPOI2 , … ,𝑉𝑉90IPOIG ≈ 𝑣𝑣90

IPOIG|𝛼𝛼�:

P�𝑉𝑉90IPOI1 ≈ 𝑣𝑣90

IPOI1 ,𝑉𝑉90IPOI2 ≈ 𝑣𝑣90

IPOI2 , … ,𝑉𝑉90IPOIG ≈ 𝑣𝑣90

IPOIG|𝛼𝛼� (6.47)

= P�𝑉𝑉90IPOIG ≈ 𝑣𝑣90

IPOIG�𝑉𝑉90IPOI1 ≈ 𝑣𝑣90

IPOI1 , … ,𝑉𝑉90IPOIG−1 ≈ 𝑣𝑣90

IPOIG−1 ,𝛼𝛼�

× P�𝑉𝑉90IPOIG−1 ≈ 𝑣𝑣90

IPOIG−1�𝑉𝑉90IPOI1 ≈ 𝑣𝑣90

IPOI1 , … ,𝑉𝑉90IPOIG−2 ≈ 𝑣𝑣90

IPOIG−2 ,𝛼𝛼�

× P�𝑉𝑉90IPOIG−2 ≈ 𝑣𝑣90

IPOIG−2�𝑉𝑉90IPOI1 ≈ 𝑣𝑣90

IPOI1 , … ,𝑉𝑉90IPOIG−3 ≈ 𝑣𝑣90

IPOIG−3 ,𝛼𝛼�

× P�𝑉𝑉90IPOI3 ≈ 𝑣𝑣90

IPOI3�𝑉𝑉90IPOI1 ≈ 𝑣𝑣90

IPOI1 ,𝑉𝑉90IPOI2 ≈ 𝑣𝑣90

IPOI2 ,𝛼𝛼�

× P�𝑉𝑉90IPOI2 ≈ 𝑣𝑣90

IPOI2�𝑉𝑉90IPOI1 ≈ 𝑣𝑣90

IPOI1 ,𝛼𝛼�

× P�𝑉𝑉90IPOI1 ≈ 𝑣𝑣90

IPOI1�𝛼𝛼)

Page 191: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.4. Calibrating the Distance Equation

159

It has now become clear why the probability of the realized 𝑣𝑣90 for each IPOI was

computed above; each probability term on the RHS of Equation (6.47) indeed represents the

probability of the realized 𝑣𝑣90 for the respective IPOI. The following example further clarifies

the concept. The probability of the realized 𝑣𝑣90 for the 2nd IPOI, calculated based on

𝑓𝑓𝑉𝑉90IPOI2(𝑣𝑣90), is indeed equal to P�𝑉𝑉90

IPOI2 ≈ 𝑣𝑣90IPOI2�𝑉𝑉90

IPOI1 ≈ 𝑣𝑣90IPOI1 ,𝛼𝛼�. The reason is that when

the 2nd IPOI is scrutinized, the first IPOI has already taken place and 𝑣𝑣90IPOI1 has been observed.

As a conclusion to this section, the flowchart in Figure 6-11 provides an overview of the

calibration process.

Page 192: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 6. Methodology 6.4. Calibrating the Distance Equation

160

Figure 6-11. This flowchart provides an overview of the process of calibrating the ‘distance function’ and

estimating the optimum value for 𝛼𝛼 (see, Equation (6.37)).

Page 193: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

161

Chapter 7. DATA

7.1. Fundamental Financial Data

As indicated in Chapter 6, the pre-IPO fundamental financial data are required to conduct

the analysis. For a given firm, at least one pre-IPO audited annual data point is required for

each input and output. After scrutinizing the capabilities of various search engines and the

availability of pre-IPO data in each, the following sources were selected to collect data:

(i) Bloomberg Terminal

(ii) Wharton Research Data Services (WRDS)

(iii) Securities Data Company PlatinumTM (SDC PlatinumTM)

The Bloomberg database offers a comprehensive custom search for the IPOs in a specific

industry and/or during a particular time period. In Bloomberg, however, limited pre-IPO

financial data are available for the list of IPOs returned by the search. Bloomberg has started

the scrutiny of prospectuses since 2011. The lists of U.S. IPOs were downloaded in 10

categories each representing one Bloomberg industry. These lists required pruning from

spinoffs/carve-outs, shelf registrations and non-original IPOs. The ‘new issues’ database of

SDC PlatinumTM was used for this purpose. The 9-digit CUSIPs of the Bloomberg lists were

fed into the SDC PlatinumTM and the SDC matches were retrieved. Spinoffs/carve-outs, shelf

registrations and non-original IPOs were then excluded from the lists of returned matches. In

addition, the SDC PlatinumTM was used to limit the lists to the U.S. IPOs whose country of

domicile is also U.S.

SDC PlatinumTM consists of different databases which provide detailed information on new

issues, mergers and acquisitions, private equity, and more. The data on IPOs and secondary

offerings of common stocks are limited to the cases with the ‘book-building’ mechanism

which is dominant in the U.S. (see, Section 2.4 for the details of the mechanism). SDC is

Page 194: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 7. Data 7.1. Fundamental Financial Data

162

updated daily. Some elements may include data from 1970 to present; however, for many

elements, the data are reliably available for the last ten to fifteen years.

Using the sources such as SEC filings and prospectuses, the ‘new issues’ database of SDC

provides the following variables and many more: offer price, aftermarket prices, some of the

elements from the financial statements, numbers of primary and secondary shares offered. The

database also includes information on the underwriters’ and issuers’ profiles such as industry

and location.

In order to retrieve the respective pre-IPO audited annual data, the CUSIP lists returned

from the SDC PlatinumTM were subsequently supplied into the COMPUSTAT database of

Wharton Research Data Services (WRDS). As indicated earlier, the sector-specific indices are

required for the risk analysis in this thesis. Up to this stage, the Bloomberg sector

classification was preserved, and the CUSIPs were clustered in 10 Bloomberg sectors. Due to

limited sector-specific index data on Bloomberg sectors, the macro list of all the CUSIPs was

re-classified into 10 S&P sectors, using the Global Industry Classification Standard (GICS)

codes.

Compared to other databases such as FACTSET, S&P Capital IQ, and Thomson Reuters

Financial Database, the COMPUSTAT database of WRDS is more comprehensive in terms of

inclusion of pre-IPO data; however, there still exists a considerable number of firms with no

pre-IPO information. These firms were identified and eliminated.

Table 7-1 presents the number of firms retained under each S&P sector once all the

modifications and configurations were implemented. The numbers in the second column of

Table 7-1 represent the case where the computer program examines the availability of only

one pre-IPO audited annual datum for each input and output. It should, however, be noted that

in case of balance sheet items, two pre-IPO data points are required in order to calculate the

‘period’ changes for the respective variables. This extra criterion further trims the number of

firms. The fourth column in Table 7-1 contains these numbers.

Page 195: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 7. Data 7.1. Fundamental Financial Data

163

Table 7-1. The table presents the number of remaining IPOs under each S&P sector after completing the process

of data mining. The third and fifth columns display the time periods spanned by the respective IPOs.

GICCD (S&P Sector)

No. of IPOs

(1 pre-IPO

data point)

Temporal

Range

(1 pre-IPO

data point)

No. of IPOs

(2 pre-IPO

data points)

Temporal

Range

(2 pre-IPO

data points)

55 (Utilities) 7 1990-1998 1 1993

50 (Telecommunication

Services) 61 1990-2010 35 1993-2010

15 (Materials) 80 1988-2012 36 1991-2012

10 (Energy) 88 1988-2011 52 1993-2011

30 (Consumer Staples) 96 1988-2013 42 1991-2013

40 (Financials) 187 1989-2012 111 1992-2012

20 (Industrials) 363 1988-1992 162 1990-2012

25 (Consumer

Discretionary) 560 1988-2013 208 1991-2012

35 (Health Care) 604 1988-2012 327 1990-2012

45 (Information

Technology) 1017 1987-2013 501 1990-2013

Page 196: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 7. Data 7.2. Sector Index and GDP

164

7.2. Sector Index and GDP

Four sector-specific indices were retrieved for each S&P sector: (i) S&P 1500 Composite,

(ii) S&P 400, (iii) S&P 500, (iv) S&P 600. The index values were provided at a minimum

frequency of monthly in the COMPUSTAT database of WRDS. The geometric mean was used

in order to convert the monthly periodicity to annual frequency.

Among the four indices of each sector, S&P 1500 Composite would serve as the most

representative index since each sector comprises firms of diverse market capitalizations.

However, as demonstrated in Table 7-2, the time periods spanned by S&P 1500 Composite do

not fully cover the periods shown in Table 7-1. The widest ranges of data are available for the

S&P 500 index. Therefore, the analysis is based on this class of indices.

There exists an extensive list of firms from different sectors and of different types that can

be utilized to examine the proposed models. IPOs can be categorized into two broad classes:

IPOs in mature industries and IPOs in growing industries. The characteristics of these two

classes should be considered in selecting an appropriate industry/sector for the analysis. The

IPOs in mature industries suggest a robust setting to test the models; the performance of the

proposed methodology can be examined in an environment with less information asymmetry

and more established firms going public. Sector 35 (health care) is the most populated sector

which holds the aforementioned characteristics. Hence, this thesis selects the ‘health care’

sector for examining the methodology.

The annual ‘real’ GDP data were obtained from U.S. Department of Commerce: Bureau of

Economic Analysis [53]. The GDP implicit price deflator was used to adjust the data for

inflation. In each case, the base year was set to be the issue year of the respective IPOI (see,

Section 6.2.1). Considering that this thesis includes the fundamental financial data of a diverse

set of firms, the GDP deflator was deemed to be more reliable to carry out inflation

adjustments as opposed to other common measures such as consumer price index (CPI). The

GDP deflator is a metric obtained based on GDP which, itself, is representative of all the final

goods and services produced within the geographical boundaries of an economy. CPI, on the

Page 197: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 7. Data 7.2. Sector Index and GDP

165

other hand, is derived based on a basket of household goods and services. It would, therefore,

provide a less comprehensive insight into changes in the level of prices.

Table 7-2. Each cell displays the time period spanned by the respective sector and index in COMPUSTAT.

S&P Sector S&P 1500 S&P 400 S&P 500 S&P 600

55 (Utilities) 1994-2013 1995-2013 1983-2013 1994-2013

50

(Telecommunication

Services)

1994-2013 1994-2013 1991-2013 1994-2013

15 (Materials) 1994-2013 1994-2013 1991-2013 1994-2013

10 (Energy) 1994-2013 1994-2013 1986-2013 1994-2013

30 (Consumer

Staples) 1994-2013 1994-2013 1991-2013 1994-2013

40 (Financials) 1994-2013 1994-2013 1983-2013 1994-2013

20 (Industrials) 1994-2013 1994-2013 1984-2013 1994-2013

25 (Consumer

Discretionary) 1994-2013 1994-2013 1991-2013 1994-2013

35 (Health Care) 1994-2013 1994-2013 1987-2013 1994-2013

45 (Information

Technology) 1994-2013 1994-2013 1986-2013 1994-2013

Page 198: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

166

Chapter 8. RESULTS

8.1. Results of Phase I: Comparable Selection

As established in Chapter 7, the proposed methodology is examined using the IPOs in the

‘health care’ sector. The process of data mining results in 327 U.S. IPOs spanning the time

period of 1990-2012. The first IPO selected for the analysis is the 52nd IPO on the list; that is,

the ID of the IPOI is 52 or 𝑏𝑏 = 52 (see, Section 6.2.5 and Figure 6-6). The ‘pool of candidates’

for the first IPOI, therefore, consists of 51 past IPOs. Once the algorithm identifies and records

the comparables for the first IPOI with the ID of 𝑏𝑏 = 52, the 53rd IPO on the list would be

selected as the next IPOI and the process is iterated (refer to Chapter 6 for further details).

As detailed in Section 6.2.5, the MATLAB programming software is used to implement

Models 1, 2, and 3 as well as all the steps demonstrated by the flowcharts in Figure 6-6 to

Figure 6-8. In sum, approximately 7000 lines of script are written to execute Phase I. The

reader is referred to Chapter 6 for the specifics.

The proposed methodology for Phase I is run in three different settings:

RUN 1. Run 1 does not execute the two steps of outlier detection and efficient IPOI

treatment (see, Sections 6.2.4 and 6.2.5). Consequently, if an IPOI is identified as

efficient, it is excluded from the macro list of all the 327 IPOs. The reason can be

articulated as follows: No comparables can be identified for an efficient IPOI which

does not undergo the process of ‘efficient IPOI treatment’. Consequently, no prior

and posterior joint distributions of the parameters Μ, Σ, Ξ, and Ω (i.e.

𝑓𝑓Μ,Σ,Ξ,Ω(𝜇𝜇,σ, ξ,ω)) can be estimated for the respective efficient IPOI. Thus, should

it be selected as the closest comparable of a succeeding IPOI, it does not hold the

required information to engage in estimating the same distribution for the new IPOI.

RUN 2. The efficient IPOIs are treated; however, the analysis does not execute the

outlier detection step.

Page 199: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.1. Results of Phase I: Comparable Selection

167

RUN 3. The efficient IPOIs are treated and the outliers are detected using the

“Jackstrap” approach proposed by De Sousa and Stošić (2005) [45] (see, Sections

6.2.4 and 6.2.5). The number of cycles of bootstrapping is 100. Side analyses

indicated that in the setting of this thesis, the value of ‘leverage’ for a DMU does

not change significantly by increasing the number of bootstraps beyond 100. It

should be noted that the “Jackstrap” approach is computationally expensive and

time consuming. To provide a perspective on the run time, consider the 52nd IPO

scrutinized using 100 cycles where each random sample (without replacement)

contains 35 DMUs; that is, 𝑞𝑞 = 100, 𝑚𝑚 = 35, 𝑏𝑏 = 52 in Figure 6-6 to Figure 6-8. In

this case, the minimum number of optimizations executed for a single IPOI would

be 100×35×34. If the total number of IPOIs (i.e., 276) and the growing pool of

candidates are taken into account, it can be inferred that the process is considerably

time consuming.

It should be noted that the contribution of phase I to the risk analysis of IPOs can only be

tested once Phase II is completed. In this section, however, the outcomes of Phase I can be

examined in terms of the robustness of the methodology in choosing the comparables. Figure

8-1 visualizes the number of comparables associated with each IPOI under RUN 1, RUN 2

and RUN 3. The efficient IPOI treatment and the outlier detection tend to smooth out the

erratic changes in the number of comparables per IPOI. Particularly pronounced is the impact

of outlier detection which is executed in RUN 3. The increased number of comparables per

IPOI in RUN 1 can be attributed to the following cause:

In DEA, once an efficient DMU is excluded, based on its spatial position with respect to the

remaining efficient DMUs, the production possibility set may shrink. The efficient frontier

would consequently be reconfigured and new efficient peers and efficient hyperplanes would

be assigned to the inefficient DMUs which were previously associated with the excluded

efficient DMU. The reshaping may impact the classification of the inefficient DMUs with

identical efficient peers. To further clarify this point, consider Figure 8-2.

Page 200: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.1. Results of Phase I: Comparable Selection

168

52 100 150 200 250 300 3270

10

20

30

RUN 1

ID of IPOI

No.

of C

ompa

rabl

es

52 100 150 200 250 300 3270

5

10

15

20

RUN 2

ID of IPOI

No.

of C

ompa

rabl

es

52 100 150 200 250 300 3270

5

10

15

20

ID of IPOI

RUN 3

No.

of C

ompa

rabl

es

Figure 8-1. The bar plots visualize the number of comparables associated with the respective IPOIs whose IDs are presented on the horizontal axis.

Page 201: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.1. Results of Phase I: Comparable Selection

169

In Figure 8-2, when the efficient DMU G is excluded, the production possibility set shrinks.

Moreover, the inefficient DMUs which were previously associated with either of the two

hyperplanes GJL and DGJ, are now jointly ‘enveloped’ by the new and larger hyperplane DJL.

This mathematical property gives rise to the difference between the first and second panels in

Figure 8-1; under RUN 1, once an IPOI is identified as an efficient unit, it is excluded from

the ‘pool of candidates’ of any subsequent IPOI. The eliminated efficient IPO might continue

to serve as an efficient unit if it were preserved in the pool of candidates. Its removal can,

therefore, impact the comparables associated with the subsequent IPOIs.

The third panel in Figure 8-1 reflects the impact of detecting outliers in addition to treating

efficient IPOIs. As indicated in Section 6.2.5, outliers significantly impact the form and level

of the frontier. The outliers cap the production possibility set at a higher level. Consequently,

they mask out the impact of other DMUs which could potentially set closer targets for the

respective inefficient IPOI. The production possibility set is segmented more coarsely in the

presence of outliers and their removal can lead to a more detailed clustering of DMUs.

Considering that the comparables for any inefficient IPOI are set to be the firms which share

the same efficient hyperplane with the IPOI, a more detailed clustering would lead to a finer

selection of comparables. This impact is reflected in the smoother bar plot of RUN 3.

Worthy of note is the smaller variance and steady trend depicted by RUN 3 which is more

coherent with practical rationality. Compare the lower panel in Figure 8-1 to the upper two

ones where RUN 1 and RUN 2 exhibit rather increasing trends, implying that a more recent

IPOI is more likely to be linked with a larger group of comparables. The growth of number of

comparables with size of the pool of candidates, nevertheless, seems incompatible with

practical wisdom and intuition since given sufficient data for each IPOI, one does not expect a

latter IPOI to be associated with, on average, more comparables in contrast with a former IPOI

which took place a few years earlier.

Page 202: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.1. Results of Phase I: Comparable Selection

170

Figure 8-2. This figure depicts the efficient frontier for the data presented in Table 1.5 of Cooper et al. (2007) [41].

Note that the output ‘Inpatients’ has been eliminated which facilitates visualizing the frontier in 3 dimensions. It is

illustrated how the shape of the frontier changes by the exclusion of the efficient DMU G. The production

possibility set, which is capped by the efficient frontier, spans a smaller space subsequent to the removal of the

efficient DMU G. The segmentation of the frontier changes as well; the number of efficient hyperplanes decreases

in this example.

Page 203: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.1. Results of Phase I: Comparable Selection

171

To further examine the robustness of the methodology in choosing comparables, the

following test is carried out: Once the set of comparables for each IPOI is identified, the

algorithm continues to identify and record the comparables of the respective IPO in the

subsequent analyses where it participates only as a member of the pool of candidates of

another IPOI. To elaborate more, consider IPOI with the ID of 𝑏𝑏 = 52. The 52nd IPO is present

in the pool of candidates of the 53rd IPOI, 54th IPOI and other succeeding IPOIs. The

comparables of the 52nd IPO can be identified following the same principles applied in the

case of IPOIs. In the next step, this test computes the average size of all the sets of

comparables identified for the 52nd IPO. The analysis, subsequently, finds the union of all

these sets and calculates the ratio of the average size to the size of the union set. This ratio will

be referred to as the ‘average-union’ ratio, hereafter. The same steps are iterated for each of

the remaining IPOIs. Figure 8-3 exhibits the ratios for individual IPOs in RUN 3.

A value closer to unity for the average-union ratio indicates less variability in the size of the

set of comparables selected for the respective IPO. The ‘average-union’ ratio would be equal

to unity if the set of comparables remains intact for an IPO through all the executions. Ratios

greater than 70% were observed for 72.24%, 65.72% and 64.86% of the IPOIs in RUN 1,

RUN 2 and RUN 3, respectively.

The goal pursued by the previous analysis was to investigate the changes in the size of the

set of comparables for individual IPOs as the number of IPOs participating in the analysis

grows. It is likewise interesting to inspect the changes in the composition of these sets of

comparables. Figure 8-4 visualizes the change in the structure of the comparables of the 52nd

IPO. The horizontal axis represents the number of IPOs that form the pool of candidates for

the respective analysis of comparables. The vertical axis exhibits the union of all the sets of

comparables identified for the 52nd IPO through all the cycles of executions shown on the

horizontal axis. The red color indicates which comparables are selected for the 52nd IPO in the

corresponding round of execution. To elaborate more, consider the number 52 on the

horizontal axis. It indicates that 52 IPOs were included in the analysis. It additionally implies

that the 52nd IPO engages as an IPOI in this cycle. The number 53 indicates that one new IPO

Page 204: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.1. Results of Phase I: Comparable Selection

172

(i.e., the 53rd IPO) is added to the previous list. In this new set of size 53, the comparables of

the 52nd IPO are selected following the same method used to identify its comparables when it

was an IPOI. The rest of Figure 8-4 was completed by iterating the same approach.

Figure 8-4 demonstrates that the composition of the set of comparables for the 52nd IPO is

only affected by the addition of new IPOs. The new IPOs change the production possibility set

such that some of the former comparables are replaced with the new IPOs (not the previously

existing ones). Further scrutiny of the data reveals that compared to the retained former

comparables, the excluded former comparables are often positioned farther from the IPO

under investigation (i.e., the 52nd IPO in Figure 8-4) in the 𝜆𝜆 space. Therefore, the new pieces

of information tend to remove these comparables from the set of comparables of the IPO under

Figure 8-3. This figure demonstrates the ‘average-union’ ratio for individual IPOs.

Page 205: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.1. Results of Phase I: Comparable Selection

173

investigation. As a case in point, consider the 20th IPO. It almost persistently continues to

perform as a comparable for the 52nd IPO. The examination of 𝜆𝜆 results indicates that for the

first 96 cases (i.e., where the number of participating IPOs varies between 52 and 147), the

20th IPO bears the minimum distance from the 52nd IPO in the 𝜆𝜆 space. In the subsequent cases

where the 20th IPO is present, it is observed to be the second closely positioned IPO to the 52nd

IPO in the 𝜆𝜆 space, dominated by the 148th IPO.

In Figure 8-4, the set of comparables for the 52nd IPO remains stable during the first 15

cycles of execution. The addition of the 67th IPO changes the structure of the comparables set

from [8, 16, 20, 46] to [8, 20, 46, 67]. In other words, on its entry, the 67th IPO replaces the

16th IPO in the set of comparables. The 67th IPO is overridden by the 88th IPO once it is

supplied. The 88th IPO, in turn, is overruled by the 106th IPO upon its addition to the list. Note

that the 8th, 20th, and 46th IPOs continue to serve as comparables during these sequences of

execution until the 8th IPO is dominated by the 148th IPO upon its entry, and the 46th IPO

concedes its position to the 170th IPO once it engages as one of the candidates. The 20th IPO,

however, preserves its status as a comparable. The rest of Figure 8-4 can be interpreted in an

alike fashion. Similar graphs were observed for other randomly selected IPOs (i.e., the 86th,

109th and 201st IPOs) which attest to the robustness of the proposed methodology in choosing

comparables.

Page 206: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.1. Results of Phase I: Comparable Selection

174

Figure 8-4. This figure provides an overview of the changes in the composition of the comparables for the 52nd IPO. The horizontal axis

represents the number of candidates (i.e., the IPOs participating in the respective analysis of comparables). The vertical axis exhibits the

union of all the sets of comparables obtained for the 52nd IPO using the candidate sets of different sizes. Under a given set of candidates,

the selected comparables are colored in red.

Page 207: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

175

8.2. Results of Phase II: Assessment of Short-Term Risk

8.2.1. Review of the Methodology of Phase II

The previous section examined the robustness of the methodology of Phase I (i.e.,

comparable selection). The focus of this section shifts to scrutinizing Phase II which utilizes

the outcomes of Phase I to quantify the short-term risk of investing in an IPO. It has now been

established that Phase I pursues the goal of identifying a group of comparables for every IPOI

on the list of 276 IPOIs. Phase I was originally inspired by the need for a scientific basis which

would compensate for the lack of sufficient history required to carry out the risk analysis of

individual IPOIs. It allows for finding a given IPOI’s closest comparable and utilizing its

relevant historical data in deriving inferences about the respective IPOI. Evidently, the risk of

each IPOI on this list is analyzed on the basis of the preceding IPOs since if one stands at the

issue time of a given IPOI, none of the succeeding IPOs has yet taken place. Figure 6-10

provides a visual aid for demonstrating the process and is frequently cited in this Section. It is,

therefore, duplicated here for convenience. Each column in this figure depicts a separate round

of execution which analyzes the risk of the IPOI highlighted in orange, drawing upon all the

IPOs listed in the corresponding upper cells. Further details on this figure can be found in

Section 6.3.2.1.

Section 8.2 aims at presenting the outcomes of Phase II. The methodology is detailed at

length in Sections 6.2 and 6.3. The reader is encouraged to review these sections since they

form the foundation of the discussions that follow and contain comprehensive expositions of

the risk assessment approach proposed in this research. As a glance at Chapter 6 would

indicate, the procedure is sophisticated and not definable by simple formulae. Therefore, in the

interest of space and not exhausting the patience of an attentive reader, the author must be

excused for not restating all the steps and for restricting the content of this section to a brief

review that follows.

Page 208: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

176

1 1 1 1

1 1

2 2 2 2 2 2

3 3 3 3 3 3

⋮ ⋮ ⋮ ⋮ ⋮ ⋮

50 50 50 50 50 50

51 51 51 51 51 51

52

(IPOI1) 52 52 52 52 52

53

(IPOI2) 53 53 53 53

54

(IPOI3) 54 54 54

55

(IPOI4) 55 55

⋱ 56 56

57 57

⋮ ⋮

323 323

324 324

325 325

326

(IPOI275) 326

327 (IPOI276)

Figure 6-10. Using the data of the ‘health care’ sector (Sector 35 under GICS), this figure provides an

overview of the IPOIs and their ‘pools of candidates’. Each column presents the IPOs used to carry out

Phase I and Phase II for the respective IPOI. The reader is referred to Section 6.2.1 for further information

on the pool of candidates. The first 51 IPOs are referred to as ‘pioneers’.

Pool

of C

andi

date

s for

IPO

I 3

Pool

of C

andi

date

s for

IPO

I 4

Pool

of C

andi

date

s for

IPO

I 2

Pool

of C

andi

date

s for

IPO

I 1

Page 209: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

177

In concise terms, this research aims at estimating the unconditional probability density

function of price at the time horizon of 90 days (𝑉𝑉90), for each of the 276 IPOIs depicted on

Figure 6-10. This density is referred to as PDF90IPOI and would subsequently act as the basis of

VaR computations for the corresponding IPOI. Approximately 4000 lines of script have been

written to execute Phase II.

The random behaviour of the stock price (𝑉𝑉𝑡𝑡) has been assumed to be described by a GBM.

Thus, if all the parameters of the GBM (i.e., 𝜇𝜇, 𝜎𝜎, and 𝑣𝑣1) were given, PDF90IPOI would be a log-

normal distribution, computed using Equation (6.41). Nonetheless, these parameters are not

deterministic but indeed realizations from the random variables Μ, Σ, and 𝑉𝑉1. Particularly, 𝑉𝑉1

has been assumed to come from a log-normal distribution whose underlying parameters (i.e.,

hyperparameters) are themselves random. Since the offer price is known a priori for a given

IPO, the ratio of first closing price to offer price would likewise follow a log-normal

distribution whose hyperparameters are designated with 𝜉𝜉 and 𝜔𝜔, representing the mean and

standard deviation of the associated normal distribution, respectively. Considering that 𝜉𝜉 and

𝜔𝜔 could potentially be correlated with 𝜇𝜇 and 𝜎𝜎, their distributions must be considered jointly in

order to find PDF90IPOI. This joint probability density function is denoted by 𝑓𝑓Μ,Σ,Ξ,Ω(𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔),

and if it is supplied for an IPOI, its predictive density PDF90IPOI can be computed using

Equation (6.42).

The Bayesian perspective offers a unique framework to tackle this problem (see, Chapter 5,

particularly, Section 5.5). For each IPOI, there exist limited data, yet often strong prior beliefs

regarding its performative and ostensive aspects (Feldman and Pentland, 2003 [54]) – a setting

that can best be served by the Bayesian approach. Such prior views typically emerge by

examining the performances of the respective comparables, and inspired by this insight, this

thesis approximates the prior joint density of a given IPOI (e.g., IPOI100 in Figure 6-10) by the

posterior joint density of its closest comparable. The estimated prior joint density was the only

missing piece in the puzzle of PDF90IPOI100, and when it is supplied, one can proceed with

deriving VaR inferences. Once IPOI100 itself hits the market and accumulates sufficient

Page 210: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

178

history, its prior joint density is updated by Equation (6.43), using its 90-day price trajectory.

The obtained posterior joint density is then recorded to serve as the prior joint density of an

ensuing IPOI whose closest comparable is selected to be IPOI100 (see, Section 6.3.2.1).

The next challenge of the research appears in a case where the closest comparable of an

IPOI (e.g., IPOI1) resides in the category of ‘pioneers’, highlighted in green in Figure 6-10.

None of the pioneers is scrutinized under Phase I (see, Section 6.2.1). Consequently, the

selected pioneer would not be associated with any ‘closest comparable’ whose posterior joint

density could play the role of the prior joint density for the selected pioneer. Evidently, when

there is no prior joint density associated with the selected pioneer, no posterior density can be

found, to be adopted as the prior density of IPOI1. To address this challenge, this research

draws on the notion of ‘non-informative’ priors (see, Section 5.3, Page 95) and attributes a

‘uniform’ density to all the pioneers to serve as their prior joint density. This uniform density

is individually updated for each pioneer, using the respective 90-day price trajectory, and the

yielded posteriors would be reserved for cases like IPOI1. The next section expounds how the

range of each dimension of the uniform joint density (i.e., Μ, Σ, Ξ, and Ω) is set.

8.2.2. Uniform Prior Joint Density

Aside from possible ranges for the random variables of Μ, Σ, Ξ, and Ω, little is known about

them. The Bayesian analysis would, therefore, bound itself to exercise no preference over any

specific combination of the values, residing within these intervals, which may be realized.

This ignorance can be reflected by defining the prior joint density as a uniform density (see,

Section 5.3) over a four-dimensional volume spanned by the ranges of Μ, Σ, Ξ, and Ω; that is,

RΜ, RΣ, RΞ, RΩ, respectively. The prior joint density in the case of pioneers can then be

expressed mathematically as shown below:

𝑓𝑓Μ,Σ,Ξ,Ωprior (𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔) =

I(𝜇𝜇 ∈ RΜ)RΜ

×I(𝜎𝜎 ∈ RΣ)

RΣ×

I(𝜉𝜉 ∈ RΞ)RΞ

×I(𝜔𝜔 ∈ RΩ)

RΩ (8.1)

Page 211: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

179

In Equation (8.1), I(. ) stands for an indicator function which is assigned the value of 1

when its condition is true and equals 0 if its argument returns false. The next two subsections

(i.e., Sections 8.2.2.1 and 8.2.2.2) are devoted to explaining how this research estimates RΜ,

RΣ, RΞ and RΩ.

8.2.2.1. Estimating 𝐑𝐑𝚳𝚳 and 𝐑𝐑𝚺𝚺

The ranges of expected daily rate of return (i.e., RM) and daily volatility (i.e., RΣ) are

estimated according to the steps described below:

Step 1. For each pioneer, the prices are inflation-adjusted over the respective 90-day

period, choosing day 90 as the base day. Inflation rate is an economic variable

that is perceived to be impacting stock returns. Since the posterior joint density

of a pioneer – which eventually emerges based on these returns – can potentially

be adopted by a succeeding IPOI as its prior joint density, correcting for

inflation would facilitate such cross-time transmissions. Drawing on the same

logic, prior to any analysis, the realized 90-day sequence of each IPOI is

likewise adjusted for inflation on the basis of the corresponding 𝑣𝑣90. The

inflation adjustments are conducted using the GDP deflator data.

Step 2. For each pioneer, the random behaviour of price is modeled by a GBM whose

analytic solution is presented in Chapter 6 (see, Equation (6.40)). In individual

cases, the parameters 𝜇𝜇 and 𝜎𝜎 are calibrated relying on the corresponding 90-day

price trajectory and using the method of maximum likelihood estimation (MLE)

whose key idea is to find a pair (�̂�𝜇,𝜎𝜎�) that would produce the best fit to the

observed historical returns. The details of the calibration process can be found

below in Section 8.2.2.1.1.

Step 3. For each pioneer, the 95% confidence intervals of the estimates of 𝜇𝜇 and 𝜎𝜎 are

computed, which reflect the degree of certainty in the rule yielding the

confidence intervals (see, Section 8.2.2.1.1).

Page 212: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

180

Step 4. Executing Steps 1 to 3 for all the 51 pioneers would produce 51 pairs of (�̂�𝜇𝑖𝑖,𝜎𝜎�𝑖𝑖),

51 confidence intervals for 𝜇𝜇 (i.e., �̂�𝜇𝐿𝐿𝑖𝑖 ≤ 𝜇𝜇𝑖𝑖 ≤ �̂�𝜇𝑈𝑈𝑟𝑟), and 51 confidence intervals

for 𝜎𝜎 (i.e., 𝜎𝜎�𝐿𝐿𝑖𝑖 ≤ 𝜎𝜎𝑖𝑖 ≤ 𝜎𝜎�𝑈𝑈𝑟𝑟), where 𝑗𝑗 = 1,…, 51 designates the respective pioneer

and �̂�𝜇𝐿𝐿𝑖𝑖 and �̂�𝜇𝑈𝑈𝑟𝑟 denote the lower and upper bounds of the confidence interval

for 𝜇𝜇𝑖𝑖. Similarly, 𝜎𝜎�𝐿𝐿𝑖𝑖 and 𝜎𝜎�𝑈𝑈𝑟𝑟 represent the boundaries of the confidence interval

for 𝜎𝜎𝑖𝑖.

Step 5. In order to ensure that RΜ and RΣ are sufficiently broad ranges which

encompass all the typical values that can be attributed to expected daily rate of

return and daily volatility, the union of all the corresponding sets of confidence

intervals are computed to serve as RΜ and RΣ; that is, RΜ encloses the union of

all the 51 intervals of [�̂�𝜇𝐿𝐿𝑖𝑖 , �̂�𝜇𝑈𝑈𝑟𝑟], and RΣ is set to be the collection of all the

values residing in the 51 intervals of [𝜎𝜎�𝐿𝐿𝑖𝑖,𝜎𝜎�𝑈𝑈𝑟𝑟]. The process is visualized in

Figure 8-5.

Page 213: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

181

Figure 8-5. This figure visualizes the process of determining RΜ and RΣ.

Page 214: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

182

Key Consideration 1.

Since calibration in each case is conducted based on a 90-day price trajectory, the

confidence intervals may be wide to reflect the uncertainty around the respective 𝜇𝜇 and 𝜎𝜎

estimations. Particularly pronounced is the impact of sample size on the width of the

confidence interval of 𝜇𝜇 (see, Equation (8.7)); that is to say, generally, only would a long

history yield a reliable estimate of 𝜇𝜇, and indeed, a wider, rather than tighter, confidence

interval is logically more coherent with the interpretation of the expected rate of return; if one

could estimate 𝜇𝜇 with a high level of precision and accuracy, then, one could confidently

speculate on the direction and magnitude of the drift which is, by nature, an idealistic scenario.

Nonetheless, this research is not concerned with achieving a dead-on (so to speak) single

estimate for 𝜇𝜇 or 𝜎𝜎, in the case of every pioneer, but it seeks a sufficiently wide, yet focused,

global interval that encompasses all the typical values of 𝜇𝜇 or 𝜎𝜎, for all pioneers. In other

words, this research endeavours to obtain the broadest narrow interval such that any value that

resides beyond its bounds can safely be claimed to be associated with a negligible probability.

This goal would adequately be met even if the respective 95% confidence interval of each

pioneer were used as its specific RΜ or RΣ, which would eventually lead to pioneer-specific

uniform prior joint densities. Yet, this study lies on the conservative side by adopting the union

of all such intervals to serve as the global estimate of RΜ or RΣ.

It is worthy of remarking that the foremost prior joint density is chosen to be uniform to

reflect prior ignorance. For each pioneer, this prior joint density is subsequently updated by the

respective ‘likelihoods’, leading to a more concentrated posterior density compared to the

preliminary prior. A side analysis indicates that in the case of each pioneer, after sequentially

updating the uniform density for (90-1) times, the mass of the probability density retreats from

the boundaries toward the center, supporting the initial claim that beyond RΜ and RΣ, the

probabilities are negligible.

In further support of the proposed path for estimating RΜ and RΣ, it must be stated that the

foremost prior joint density utilized here does not express a complete ignorance; this research

Page 215: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

183

has adopted a comprehensive approach to ensure that the narrowest conservative intervals are

approximated for RΜ and RΣ. The Bayesian framework indeed allows for less complex approaches,

without trading much accuracy for simplicity. One extension of such views to this research would

imply setting RΜ and RΣ such that they simply represent dominant beliefs regarding the inflation-

adjusted 𝜇𝜇 and 𝜎𝜎 in the ‘health care’ sector. Below, one such example is presented, adopted from

Rachev et al. (2008) [115]. The approach appears inferior to the one suggested by this thesis, yet is

prevalently used.

Suppose that the random behaviour of the monthly return of an asset is best described by the

normal N(𝑟𝑟, 𝑠𝑠2) and the returns are identically and independently distributed. Further, assume that

the variance 𝑠𝑠2 is given; whereas, the mean 𝑟𝑟 is not deterministic and itself represents a realization

from a probability distribution whose prior form is conjectured to be N(𝑎𝑎,𝑑𝑑2). The problem,

therefore, boils down to finding 𝑎𝑎 and 𝑑𝑑 which would subsequently determine the dynamics of 𝑟𝑟.

In this case, the authors (Rachev et al., 2008) [115] estimated the two parameters 𝑎𝑎 and 𝑑𝑑 by solely

relying on their perspective of the market. As an illustration, if the median of average monthly

return is viewed to be typically 1%, it would imply that 𝑎𝑎 is equal to 1%. Furthermore, the

parameter 𝑑𝑑 can be specified based on a belief such as ‘the probability of observing an average

monthly return less than 0.5% is not more than 25%’. Such speculation can be translated into the

statement that the first quartile equals 0.5% upon which 𝑑𝑑 can be calculated.

Stated in a concise manner, this research does not pursue to find a single strictly correct estimate

for 𝜇𝜇 or 𝜎𝜎, nor is it concerned with deriving the long-term approximation of 𝜇𝜇 or 𝜎𝜎. On the

contrary, it aims at acquiring two intervals which would reflect the prior evidence on the ranges of

typical values for short-term expected daily rate of return and daily volatility. In such a context,

one expects to achieve reliable outcomes even if these intervals are simply set based on arbitrary

facts and anecdotes or dominant views in the industry – an anticipation that emerges from the

capacity of the Bayesian methodology to sequentially update and refine the prior knowledge and

filter out irrelevant prior information. Yet, in a more prudent approach, this thesis utilizes the 95%

confidence intervals and takes the conservatism one step forward by determining RΜ and RΣ as the

union of all the respective confidence intervals.

Page 216: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

184

8.2.2.1.1. Maximum Likelihood Estimation Method

For the sake of thoroughness, the method of maximum likelihood estimation employed to

approximate 𝜇𝜇 and 𝜎𝜎 is outlined below. The reader is encouraged to consult texts of statistics

for a detailed review. Assume that the 90-day price trajectory of IPO𝑖𝑖 is represented as

follows: (𝑣𝑣𝑖𝑖,1,𝑣𝑣𝑖𝑖,2, 𝑣𝑣𝑖𝑖,3, … , 𝑣𝑣𝑖𝑖,90). This 90-tuple is indeed a realization from the random

sequence of (𝑉𝑉𝑖𝑖,1,𝑉𝑉𝑖𝑖,2,𝑉𝑉𝑖𝑖,3, … ,𝑉𝑉𝑖𝑖,90). The maximum likelihood estimates of 𝜇𝜇 and 𝜎𝜎 are defined

as the values which would maximize the probability of observing the vicinity of the array

(𝑣𝑣𝑖𝑖,1, 𝑣𝑣𝑖𝑖,2, 𝑣𝑣𝑖𝑖,3, … , 𝑣𝑣𝑖𝑖,90).

As can be deduced from the analytic solution of the stochastic differential equation of GBM

(see, Equation (6.40)), the logarithmic period returns (i.e., ln 𝑉𝑉𝑟𝑟,2

𝑉𝑉𝑟𝑟,1, ln 𝑉𝑉𝑟𝑟,3

𝑉𝑉𝑟𝑟,2, … , ln 𝑉𝑉𝑟𝑟,90

𝑉𝑉𝑟𝑟,89) are

identically and independently distributed, and their dynamics can be described by the

normal N(𝑟𝑟, 𝑠𝑠2) = N��𝜇𝜇 − 𝜎𝜎2

2� Δ𝑡𝑡,𝜎𝜎2Δ𝑡𝑡�. By virtue of the i.i.d. assumption for the sequence

(ln 𝑉𝑉𝑟𝑟,2

𝑉𝑉𝑟𝑟,1, ln 𝑉𝑉𝑟𝑟,3

𝑉𝑉𝑟𝑟,2, … , ln 𝑉𝑉𝑟𝑟,90

𝑉𝑉𝑟𝑟,89), the joint distribution of the logarithmic returns for IPO𝑖𝑖 can be

expressed by the following equation:

𝑓𝑓𝑋𝑋𝑟𝑟,1,𝑋𝑋𝑟𝑟,2,…,𝑋𝑋𝑟𝑟,89�𝑥𝑥𝑖𝑖,1, 𝑥𝑥𝑖𝑖,2, … , 𝑥𝑥𝑖𝑖,89�𝑟𝑟, 𝑠𝑠� = �𝑓𝑓𝑋𝑋𝑟𝑟,𝑖𝑖�𝑥𝑥𝑖𝑖,𝑖𝑖�𝑟𝑟, 𝑠𝑠�89

𝑖𝑖=1

(8.2)

where, the random variable 𝑋𝑋 stands for logarithmic return, and 𝑥𝑥 denotes a realization from

𝑋𝑋.

When an observed sequence (𝑥𝑥𝑖𝑖,1, 𝑥𝑥𝑖𝑖,2, … , 𝑥𝑥𝑖𝑖,89) is substituted in the joint density function

of IPO𝑖𝑖, shown by Equation (8.2), the only unknowns would remain 𝑟𝑟 and 𝑠𝑠 (consequently, 𝜇𝜇

and 𝜎𝜎). The resultant function is referred to as the likelihood function and can be maximized in

Page 217: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

185

𝑟𝑟 and 𝑠𝑠 to obtain the maximum likelihood estimates, denoted by �̂�𝑟 and �̂�𝑠, respectively. Note

that the product shown in the preceding equation can become considerably small. Thus, to

avoid numerical problems, one needs to maximize the log-likelihood function. In the case of

GBM, the normality assumption yields a deterministic form for 𝑓𝑓𝑋𝑋1,𝑋𝑋2,…,𝑋𝑋89(𝑥𝑥1, 𝑥𝑥2, … , 𝑥𝑥89|𝑟𝑟, 𝑠𝑠)

and leads to the following familiar expressions for �̂�𝑟 and �̂�𝑠:

�̂�𝑟 =1𝑛𝑛� 𝑥𝑥𝑖𝑖

𝑛𝑛=89

𝑖𝑖=1

(8.3)

�̂�𝑠2 =1

𝑛𝑛 − 1� (𝑥𝑥𝑖𝑖 − �̂�𝑟)2𝑛𝑛=89

𝑖𝑖=1

(8.4)

Subsequently, 𝜇𝜇 and 𝜎𝜎 can be estimated as shown below:

�̂�𝑟 = ��̂�𝜇 −𝜎𝜎�2

2�Δ𝑡𝑡 (8.5)

�̂�𝑠2 = 𝜎𝜎�2Δ𝑡𝑡 (8.6)

Suppose 𝑄𝑄𝑛𝑛 represents the average of i.i.d. normal random variables; i.e. 𝑄𝑄𝑛𝑛 = 𝑋𝑋1+⋯+𝑋𝑋𝑛𝑛𝑛𝑛

.

Since the average itself is normally distributed �𝑄𝑄𝑛𝑛~N �𝑟𝑟, 𝑠𝑠2

𝑛𝑛��, the (1 − Ƿ) × 100 %

confidence interval of 𝑟𝑟 can be achieved using the following equation:

Page 218: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

186

𝑞𝑞𝑛𝑛 − CDFSN,Ƿ/2−1 �̂�𝑠

√𝑛𝑛≤ 𝑟𝑟 ≤ 𝑞𝑞𝑛𝑛 + CDFSN,Ƿ/2

−1 �̂�𝑠√𝑛𝑛

(8.7)

where, 𝑞𝑞𝑛𝑛 represents an observation from the random variable 𝑄𝑄𝑛𝑛 and can indeed be

approximated by the maximum likelihood estimate of 𝑟𝑟. The variable CDFSN,Ƿ/2−1 denotes the

absolute value of the inverse cumulative density function (CDF−1) of the standard normal

distribution (SN), at the cumulative probability of Ƿ/2. It is equal to 1.96 if Ƿ = 0.05.

The confidence interval of 𝑠𝑠2 can be estimated as shown below:

(𝑛𝑛 − 1)�̂�𝑠2

CDF𝜒𝜒𝑛𝑛−12 ,(1−Ƿ/2) −1 ≤ 𝑠𝑠2 ≤

(𝑛𝑛 − 1)�̂�𝑠2

CDF𝜒𝜒𝑛𝑛−12 ,Ƿ/2 −1 (8.8)

In this equation, CDF𝜒𝜒𝑛𝑛−12 ,Ƿ/2 −1 and CDF𝜒𝜒𝑛𝑛−12 ,(1−Ƿ/2)

−1 respectively represent the lower and upper

Ƿ/2 × 100 % quantiles of the chi-square distribution with the degree of freedom of (𝑛𝑛 − 1).

Equation (8.8) is built upon the principles that the square of a standard normal random

variable follows a chi-square distribution and the sum of 𝑛𝑛 independent chi-square random

variables continues to be a chi-square random variable with the degree of freedom of (𝑛𝑛 − 1).

8.2.2.2. Estimating 𝐑𝐑𝚵𝚵 and 𝐑𝐑𝛀𝛀.

It was established in Section 6.3.2 that the inflation-adjusted ratio of ‘closing price of the

first day’ to ‘IPO offer price’ can be assumed to follow a log-normal distribution. This

conclusion was reached through scrutinizing the ratios of all the 327 IPOs from the ‘health

care’ sector. Drawing on this knowledge, RΞ and RΩ can be determined by fitting a lognormal

distribution to the ratios of the pioneers and computing the 95% confidence intervals of ω and

Page 219: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

187

ξ ; that is, RΩ = 𝜔𝜔�𝐿𝐿 ≤ ω ≤ 𝜔𝜔�𝑈𝑈 and RΞ = 𝜉𝜉𝐿𝐿 ≤ ξ ≤ 𝜉𝜉𝑈𝑈, where, 𝜔𝜔�𝐿𝐿 and 𝜔𝜔�𝑈𝑈 denote the lower

and upper bounds of the confidence interval of 𝜔𝜔, respectively. Likewise, 𝜉𝜉𝐿𝐿 and 𝜉𝜉𝑈𝑈 mark the

lower and upper boundaries of the confidence interval of 𝜉𝜉.

Finally, the approaches outlined in this section yield the following estimations for RΜ, RΣ,

RΞ, and RΩ:

RΜ: [−0.0196, 0.0211]

RΣ: [0.0175, 0.0848]

RΞ: [0.0419, 0.1310]

RΩ: [0.1325, 0.1969]

As detailed above, the objective was to find the narrowest broad intervals beyond which reside

the least probable values; that is, any combination of (𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔) that falls outside the domain

bounded by RΜ, RΣ, RΞ, and RΩ can reliably be assigned the probability of zero. It is

worthwhile to reiterate the insight that for a given pioneer, the uniform joint density defined

by RΜ, RΣ, RΞ, and RΩ (see, Equation (8.1)) is subsequently updated in light of the

corresponding observed data. Likewise, it gets more refined and becomes more concentrated

as it further propagates in time, across the succeeding IPOIs. The new evidence from each

relevant future IPOI can be regarded as a numerical sieve which revises the existent

probabilities and further pulls the mass of the joint density from the boundaries of RΜ, RΣ, RΞ,

and RΩ, and toward the center.

Finally, as a guide for the practical aspects of the procedure, it should be indicated that each

of the intervals shown above is divided into 100 equal segments. Consequently, 108,

computationally burdensome, unique combinations of (𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔) are achieved which can be

utilized to execute Equations (6.42) and (6.43) for individual IPOs. Below, two more pieces of

Page 220: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

188

information are provided on the nuts and bolts of the process of estimating PDF90IPOI for each

IPOI. These ranges are deemed suitable and adequately conservative to cater to potential

extreme scenarios and address the tails of the distribution PDF90IPOI:

0.1 ≤ 𝑉𝑉1/𝑉𝑉offer ≤ 10

0.0001 ≤ 𝑉𝑉90/𝑉𝑉offer ≤ 10

The resolution of 𝑉𝑉1/𝑉𝑉offer is set as 1/100 of its range, and the interval of 𝑉𝑉90/𝑉𝑉offer is

segmented into 600 parts which are of equal length on the logarithmic scale.

Page 221: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

189

Key Consideration 2.

It can be inferred from Section 8.2.2 that ‘empirical’ expected rate of return (i.e., 𝜇𝜇) will

be employed in estimating PDF90IPOI, and hence, in deriving VaR inferences. As a side note,

it is worthwhile to give insight as to why VaR estimations are always developed in the

‘real-world’, as opposed to the ‘risk-neutral’, mode. To delineate, one must first pinpoint

the underlying principles of the risk-neutral approach which, in essence, emerged as a

means for pricing. It is indeed a forward-looking concept drawing on the notion of ‘no-

arbitrage’ to determine a price which ensures that no ex-ante arbitrage profit can be locked

in by any of the investing parties.

As a simple case in point, consider an investment in a European call option in a setting

with no transaction costs and dividends. The risk-neutral concept relies upon the notion of

creating a ‘riskless’ portfolio which would earn a return equal to the risk-free interest rest

in the absence of arbitrage opportunities. In the case of pricing a call option on a stock,

such a riskless position can be established by holding a short position on the option and a

long position in Δ shares of the same stock. The number of shares, Δ, is set such that there

is no uncertainty regarding the value of this portfolio at the end of an infinitesimal period;

hence, its return should correspond to the risk-free rate. In this setting, one can solve for the

only remaining unknown; that is, the price of the option at the beginning of the period.

Delta neutrality or hedge of the position is only temporary and can be preserved by

‘rebalancing’ the portfolio in sufficient frequency through dynamically adjusting the

number of the stocks (i.e., Δ).

The risk-neutral concept indeed offers a framework where pricing can be carried out

with the assumption that the investors are risk-neutral and as long as the riskless position is

maintained, the ‘empirical’ expected rate of return, or the individuals’ risk preferences, are

irrelevant. From a practical perspective, pricing in a setting where both parties are risk-

Page 222: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

190

neutral is only intuitive since the outcome would reconcile the incentives of both agents

and remain compatible with the incentive structure of the game of market exchange.

In contrary to the objective pursued by the risk-neutral wisdom – which is setting up

instantaneously hedged, and hence riskless, portfolios for the sole sake of pricing – VaR

focuses on estimating the risk in an event where a position remains unhedged. It is indeed

the very characteristic of risk-neutrality that renders VaR computations irrelevant in the

risk-neutral world. Ultimately, it is a no-arbitrage world, where investors are concerned

with constant hedging and realizing risk-free returns. They are, therefore, certain about

their risk and do not seek, nor do they need, a measure like VaR to gauge the riskiness of

their positions.

Outside the risk-neutral context, however, lies the ‘real-world’, where not all the

positions are hedged and investors trade increased risks for increased rewards. It is a world

where an investor may wish to hold (lock) a risky position, with no or partial protection, in

an attempt to potentially achieve a higher profit. Within such a setting, individual risk

preferences, and empirical expected rate of return (i.e., 𝜇𝜇), play a central role and investors

can benefit from a concept like VaR to gauge their exposure. In a nutshell, it can be stated

that regardless of the pricing techniques utilized to arrive at fair price estimates, the gain or

loss of a given portfolio, at a predetermined time horizon, is ultimately a function of the

empirical expected rate of return. Consequently, it can be concluded that the respective

gain-loss distribution as well as VaR inferences are contingent on ‘real’, not ‘risk-neutral’,

probabilities. The reader is encouraged to consult the following literature for more

explanatory comments on why a risk-neutral mode would not offer a suitable framework

for risk management: Bliss and Panigirtzoglou (2004) [28], Vincent-Humphreys and Noss

(2012) [138], Hull (2012) [75], Giordano and Siciliano (2013) [60].

Page 223: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

191

8.2.3. Calibrating the Distance Function

Previous discussions have made it plain that one requisite for estimating a given IPOI’s

predictive density (i.e., PDF90IPOI) is its closest comparable’s posterior joint density of the

parameters. Since each IPOI is associated with more than one comparable, a metric is required

to measure the similarity between the IPOI and each of its comparables. Equation (6.37),

suggested in Section 6.3, serves this goal and assists with pinpointing the best match for the

given IPOI. Below, it is repeated for the ease of reference:

𝐷𝐷𝑐𝑐𝑐𝑐2 = �𝜆𝜆𝑐𝑐 − 𝜆𝜆𝑐𝑐�

2+ 𝛼𝛼(𝜓𝜓𝑐𝑐 − 𝜓𝜓𝑐𝑐)2

𝑐𝑐 = 1, … , 𝜂𝜂 (6.37)

Recall that in this equation, 𝜂𝜂 represents the total number of comparables associated with an

IPOI, marked by the ID of 𝑏𝑏, and the subscript 𝑐𝑐 designates the ID of the comparable under

scrutiny. The closest comparable is selected as the one which leads to the minimum 𝐷𝐷𝑐𝑐𝑐𝑐.

Phase I computes the values of 𝜆𝜆 and 𝜓𝜓 for a given IPOI and its comparables. Thus, the

only unknown of Equation (6.37) remains 𝛼𝛼 which is estimated using a maximum likelihood

approach (see, Section 6.4). The algorithm of the method is visualized by the flowchart

presented in Figure 6-11. Concisely, the calibration process seeks an estimate of 𝛼𝛼 which

would maximize the probability of the realized sequence of 𝑉𝑉90’s: P�𝑉𝑉90IPOI1 ≈ 𝑣𝑣90

IPOI1 ,𝑉𝑉90IPOI2 ≈

𝑣𝑣90IPOI2 , … ,𝑉𝑉90

IPOI276 ≈ 𝑣𝑣90IPOI276|𝛼𝛼�. To further make sense of the calibration process, consider

Figure 8-6. This figure exhibits the PDF90IPOI of the last IPOI (i.e., the 276th IPOI) which is

generated by temporarily assigning the value of 1.5 to 𝛼𝛼; that is, the closest comparable of this

IPOI is identified based on the assumption that 𝛼𝛼 = 1.5 in Equation (6.37), and subsequently,

the corresponding PDF90IPOI is estimated by executing Phase II. It is noteworthy that the area

under the curve of any estimated PDF90IPOI should be computed to examine whether it equals

Page 224: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

192

unity; otherwise, the estimated PDF90IPOI must be normalized. In this research, every estimated

PDF90IPOI met the condition of unit area under the curve, and hence, no normalization was

carried out.

Figure 8-6. The graph visualizes the estimated PDF90IPOI for the 276th IPOI (i.e., 327th IPO). The dashed line

marks the observed price on day 90 (𝑣𝑣90realized) normalized by the respective inflation-adjusted offer price

(𝑣𝑣offer). Its position is an indicative of the probability that would be associated with the event 𝑣𝑣90realized/

𝑣𝑣offer ≈ 1.015 if the proposed methodology of this thesis is employed and if 𝛼𝛼 is set equal to 1.5. A similar

process is undertaken for each IPOI in order to achieve the value of the yielded PDF90IPOI at the respective

𝑣𝑣90realized/𝑣𝑣offer. These values are recorded to be subsequently utilized in determining the likelihood of 𝛼𝛼 =

1.5 following the steps described within the text.

Page 225: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

193

Next, in a backtesting approach, the probability of observing the vicinity of the realized V90

(i.e., 𝑣𝑣90realized < 𝑉𝑉90 ≤ 𝑣𝑣90realized + d𝑣𝑣) is recorded; that is, PDF90IPOI × d𝑣𝑣. To avoid numerical

complications, the value of the probability density function (PDF90IPOI) at 𝑣𝑣90realized is stored, as

opposed to a small probability pertaining to the event 𝑣𝑣90realized < 𝑉𝑉90 ≤ 𝑣𝑣90realized + d𝑣𝑣. Since

d𝑣𝑣 is defined as an infinitesimally small change in price which remains consistent across

different assumptions of 𝛼𝛼, it would be cancelled out if it were maintained in the process.

The aforementioned steps are iterated for individual IPOIs, and in an alike fashion, for each

IPOI, the value of its PDF90IPOI at the respective realized 𝑉𝑉90 is retrieved. The iteration process

returns an array of length 276 for a given 𝛼𝛼, which comprises the PDF values at the

corresponding 𝑣𝑣90realized’s. The final step involves calculating the logarithm of the product of

all the elements in this array. As is readily perceived, the logarithm function is used to avoid

obtaining an excessively small decimal which is inevitable due to the dimension of the

problem; the estimated product indeed represents the value of a joint pdf – which spans the

ranges of 276 variables in a (276+1)-dimensional space – at a given point whose position is

uniquely determined by 276 coordinates. The reader is encouraged to review Equations (6.46)

and (6.47) for the mathematical expression of the approach.

With each update on the value of 𝛼𝛼, a new likelihood would be yielded by the above-

indicated process. In other words, the probability of observing the sample sequence of

276 𝑣𝑣90realized’s varies under different assumptions of 𝛼𝛼. Figure 8-7 represents the outcomes of

the calibration process. The likelihood function reaches its maximum at 𝛼𝛼 ≅ 1.5, and hence,

the maximum likelihood estimate of 𝛼𝛼 is equal to 1.5.

The preceding discussions provided insight that Phase II is a computationally intensive

process. After exhaustively optimizing the algorithm to attain a satisfactory speed, it takes

approximately 20 hours to obtain each point depicted on Figure 8-7, using a machine with 16.0

GB of RAM and the processor of Intel(R) Core(TM) i7-3770 CPU @ 3.4 GHz. In the interest

of efficient execution of the calibration process, it was first implemented using the following 9

values of 𝛼𝛼: [0, 0.5, 1, 1.5, 2, 2.5, 3, 4, 5]. Then, in an ‘adaptive search’ manner, the resolution

Page 226: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

194

was increased within the ranges deemed the most susceptible to change in order to further

explore the impact of changing 𝛼𝛼.

Note that starting at the vicinity of 𝛼𝛼 = 3, the log-likelihood function stabilizes which can

be explained as follows: For a given IPOI, beyond a certain threshold, the 𝛼𝛼 term (i.e., the

second term) in Equation (6.37) would dominate, effectively identifying the closest

comparable as the one whose efficiency score (𝜓𝜓) falls the closest to the given IPOI’s 𝜓𝜓.

Figure 8-7. This graph exhibits the log-likelihood function of 𝛼𝛼. Each depicted point is obtained by executing

the following two main steps: Step 1 assigns a value to 𝛼𝛼 and executes phase I and phase II for each of the

276 IPOIs which results in 276 individual PDF90IPOI’s under the postulated 𝛼𝛼. Step 2 computes the logarithm

of the product of 276 elements each representing the value of the PDF90IPOI of a given IPOI at the respective

realized 𝑉𝑉90. The likelihood function hits its maximum at 𝛼𝛼 ≅ 1.5.

Page 227: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

195

Therefore, beyond that threshold, the closest comparable for the specified IPOI would remain

intact. Now, suppose that LIST𝛼𝛼 is a vector of length 276 with each cell populated by the ID of

the closest comparable of the corresponding IPOI. It can be deduced from Figure 8-7, once 𝛼𝛼

hits the value of approximately 3, the 𝛼𝛼 term dominates the similarity metric in all the 276

IPOI cases and LIST𝛼𝛼 remains unchanged from then onward. If the array of closest

comparables (i.e., LIST𝛼𝛼) stays fixed, so does the list of posterior joint densities of the

parameters, required to run Phase II. Recall that in estimating the PDF90IPOI for a given IPOI, its

closest comparable’s posterior joint density is adopted as the prior joint density of the IPOI.

Hence, if the posterior density does not vary, the ensuing PDF90IPOI would likewise remain

invariant, leading to a stable log-likelihood quantity.

To further scrutinize Figure 8-7, it is worth directing attention to the small variations

between the values depicted on the vertical axis. Since these numbers represent logarithmic

likelihoods, their differences only appear insignificant; yet, the quotients derived based on

them are significant. To elaborate more, consider the differences between the values of the

log-likelihood function at 𝛼𝛼 = 0, 𝛼𝛼 = 1.5, and 𝛼𝛼 = 2.5. Refer to the corresponding likelihoods

as 𝐴𝐴, 𝐵𝐵, and 𝐶𝐶, respectively. The following subtractions then result from Figure 8-7:

log𝐵𝐵 − log𝐴𝐴 = 0.4122 →𝐵𝐵𝐴𝐴

= 100.4122 = 2.5832

log𝐵𝐵 − log𝐶𝐶 = 0.8923 →𝐵𝐵𝐶𝐶

= 100.8923 = 7.8037

The former subtraction indicates that the maximum likelihood is 2.6 times greater than the

likelihood at 𝛼𝛼 = 0, where the second term in Equation (6.37) indeed does not exist.

According to the latter subtraction, the maximum likelihood is approximately 8 times the

minimum likelihood. Both examples reflect that the presence of the second term (i.e., the 𝜓𝜓

Page 228: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

196

term), in addition to the 𝜆𝜆 term, affords a higher accuracy and precision for identifying closest

comparables, estimating PDF90IPOI’s, and consequently, quantifying risk.

The descending trend of the log-likelihood function after 𝛼𝛼 ≅ 1.5 is another observation

worth accentuating. In the multi-dimensional space of operating and financial characteristics,

each firm can be represented by one point. Phase I maps such a point onto a new space where

it is represented with the new coordinates of 𝜆𝜆s and 𝜓𝜓. The similarity analysis is conducted in

this new space. As 𝛼𝛼 grows, it causes more emphasis to be placed on the 𝜓𝜓 term of the

similarity metric, diminishing the impact of the 𝜆𝜆 term which incorporates other dimensions of

the comparison. The process of comparable selection is, therefore, adversely affected by the

suppression of the role of 𝜆𝜆, leading to less accurate estimated PDF90IPOI’s and, consequently,

lower log-likelihoods.

To complete the discussion on Figure 8-7, the reader is provided a window into its

microstructure and the implication of the maximum likelihood estimate of 𝛼𝛼 in terms of the

accuracy of individual PDF90IPOI’s. Figure 8-8 serves as an example of the typical theme that

one encounters when plotting the PDF90IPOI’s produced under the maximum likelihood estimate

of 𝛼𝛼 (i.e., 𝛼𝛼 = 1.5) and the ones yielded using other assumptions of 𝛼𝛼 (e.g., 𝛼𝛼 = 0). As

depicted, a narrower PDF90IPOI is accomplished when 𝛼𝛼 = 1.5, which is more desirable since it

indicates less variance and more certainty about the value of 𝑉𝑉90.

A critical eye may point out that in Figure 8-8, the realized 𝑉𝑉90 (𝑣𝑣90realized) lands closer to

the mode of PDF90276thIPOI generated with 𝛼𝛼 = 1.5 and, hence, argue that the red curve is a

superior forecast since 𝑣𝑣90realized comes closer to its most likely value. This logic is

incompatible with the philosophy underlying the notion of likelihood function in statistics: it

does not matter how closely the realized quantity hits the mode. What matters is the

probability assigned to (or predicted for) the realized value. Drawing on this principle, in a

case such as Figure 8-9, the red curve would maintain its superiority although its mode is more

distant from the realized 𝑉𝑉90.

Page 229: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

197

As a closing remark to this section, it is worthwhile to investigate the trend of the maximum

likelihood estimate of 𝛼𝛼 as the number of IPOIs participating in the calibration process

changes. Using all the 276 IPOIs, the above-indicated calibration intended to find a suitable

approximation for 𝛼𝛼 which can subsequently be presented as a property of the distance

function, eliminating the burden of the extra step of calibration in measuring the risk of other

IPOs. However, one may choose to calibrate the distance function each time it is engaged in

estimating the predictive density of a given IPOI. To elaborate, suppose that one aims at

quantifying the risk of the 100th IPO (i.e., the 49th IPOI). Drawing upon the outcomes shown

above, the analysis can be carried out by assuming 𝛼𝛼 is equal to 1.5 a priori.

Figure 8-8. The graph visualizes the two PDF90IPOI’s estimated for the 276th (last) IPOI under the two

assumptions of 𝛼𝛼 = 0 and 𝛼𝛼 = 1.5. A narrower PDF is achieved with the maximum likelihood estimate of 𝛼𝛼

(𝛼𝛼 = 1.5), which is more desirable since it indicates less variance and more certainty about the value of 𝑉𝑉90.

Page 230: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

198

Alternatively, one can find the maximum likelihood estimate of 𝛼𝛼 for this specific problem.

The latter approach would require repeating all the aforementioned calibration steps except

that under each assumption of 𝛼𝛼, there exist only 48 PDF90IPOI’s (and the respective 48

𝑣𝑣90realized’s), as opposed to 276 PDF90IPOI’s (and the corresponding 𝑣𝑣90realized’s).

Figure 8-9. Similar to Figure 8-8, this plot visualizes the two estimated PDF90IPOI’s for the 17th IPOI when 𝛼𝛼 =

0 and 𝛼𝛼 = 1.5. The graph serves as an example of the case where the maximum likelihood estimate of the

parameter (𝛼𝛼 = 1.5) has produced a narrower PDF; yet, the mode of the PDF is comparatively remote from

the realized price, marked by the dashed line. Such an observation should not be viewed as an element of

concern about the accuracy of the PDF since the concept of maximum likelihood rests the basis of accuracy

comparison upon examining the assigned probability to the realized event, and not upon the deviation of the

realized event from the most likely value (mode) of the PDF.

Page 231: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

199

Figure 8-10 summarizes the calibration results produced under varying sets of engaged

IPOs (i.e., PDF90IPOI’s). Sensibly, in the case of the first IPOI (i.e., the 52nd IPO), the distance

function cannot be calibrated since there exists no preceding IPO with a predictive density. It

should be noted that the set of 𝛼𝛼 values, to act as the potential candidates for the maximum

likelihood estimate of 𝛼𝛼, has remained intact which implies that the vector of 𝛼𝛼 in Figure 8-10

is identical to the one used for Figure 8-7. This restriction arises from the computational

tedium of estimating PDF90IPOI’s under a given 𝛼𝛼. Since the PDF90IPOI’s have been computed and

stored for all the elements of the macro set of 276 IPOIs, under individual 𝛼𝛼 values displayed

in Figure 8-7, one can benefit from them to straightforwardly run the calibration process using

subsets of the macro set; otherwise, for any new candidate of 𝛼𝛼, one ought to first estimate the

respective PDF90IPOI’s which considerably increases the execution time.

Figure 8-10 demonstrates more fluctuations for the early subsets, where the calibrations

were conducted on the basis of smaller groups of IPOs, and as the number of participating

IPOs increases, the value of the maximum likelihood estimate of 𝛼𝛼 approaches 1.5. The rather

sharp convergence can be attributed to the limited number of the 𝛼𝛼 candidates available for the

analysis. Note that the two 𝛼𝛼 candidates at the vicinity of 𝛼𝛼 = 1.5 are 𝛼𝛼 = 1.25 and 𝛼𝛼 = 1.7,

which have evidently been dominated by 𝛼𝛼 = 1.5. Perhaps a smoother convergence behaviour

would be recorded if a higher resolution of 𝛼𝛼 were computationally affordable.

To support the stability of the convergence behaviour depicted in Figure 8-10, this thesis

draws upon a well-known theorem in estimation theory: Cramer-Rao Lower Bound (CRLB)

(Panik, 2005 [106]; Hardle and Simar, 2012 [67]). It can be used to compute the lower bound for

the variance of the posited maximum likelihood ‘estimator’ of 𝛼𝛼0F

1. A desirable property of any

maximum likelihood estimator is that in limit (i.e., for asymptotically large samples), it would

achieve the lower bound given by the Cramer-Rao theorem (Kay, 1993 [83]). In this theorem,

the variance is expressed as the inverse of the expected value of the Hessian matrix of the

likelihood function with respect to the corresponding parameter(s) (in this case, 𝛼𝛼). 1 The author is grateful to Dr. Yuri Lawryshyn and his PhD student, Ali Bashiri, for bringing this theorem to

her notice.

Page 232: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

200

A bootstrapping process was conducted to find the expected value of the Hessian matrix. In

each bootstrapping cycle, the requisite log-likelihood arrays for individual 𝛼𝛼’s were generated

by resampling with replacement from the corresponding original log-likelihood sets. The

obtained variances, under 102, 103, and 106 cycles of bootstrapping, can sequentially be listed

as follows: 0.0153, 0.0147, and 0.0150, which respectively translate into the standard

deviations 0.1238, 0.1211, and 0.1224. Drawing upon the depicted consistency between the

results, the standard deviation can be assumed to be approximately equal to 0.12. The reported

standard deviation specifies the long-run deviation of the proposed maximum likelihood

estimate of 𝛼𝛼 from its true value. In other words, for an asymptotically large sample of IPOs,

the expected margin of error on the maximum likelihood estimate of 𝛼𝛼 is given by CRLB.

Noticeable in Figure 8-10 is a small cluster of size 9 where the individual maximum

likelihood estimates of 𝛼𝛼 are reported to be 0.5. Within this cluster, individual calibrations are

carried out using the following subsets, respectively: 52nd IPO to 141st IPO, 52nd IPO to 142nd

IPO, …, 52nd IPO to 149th IPO. A more detailed investigation of the corresponding log-

likelihood functions reveals that 0.5 indeed represents the ‘argument’ of a local maximum,

which either wins the position of the global maximum with a slight margin over 1.5 (see,

Figure A-7), or holds this position jointly with 1.5 (see, Figure A-8). Thus, although it first

appeared to be an anomaly, the maximum likelihood estimate of 0.5 is indeed consistent with

the rest of the results exhibited by Figure 8-10. Appendix A further elucidates this point and

summarizes the outcomes of this side analysis. Based on the scrutiny of the underlying log-

likelihood functions, it is concluded that the local maximum at 0.5 is ‘unstable’ and will be

dominated by the one at 1.5 once the subset size rises beyond a threshold (in this problem,

148).

In substance, Figure 8-10 depicts that in the majority of the cases, the maximum likelihood

estimate of 𝛼𝛼 falls within the range 1.25 ≤ 𝛼𝛼 ≤ 1.9, and as the pool of engaged IPOs grows in

size, 𝛼𝛼 = 1.5 is persistently selected as the maximum likelihood estimate of 𝛼𝛼. It is, therefore,

concluded that 𝛼𝛼 = 1.5 presents a reasonable approximation for the coefficient of the 𝜓𝜓 term

in the distance function and can be utilized to skip the extra step of calibration.

Page 233: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

201

As a final consideration on the topic of calibration, it is worth noting that one must

distinguish between the ‘relative’ and ‘standalone’ accuracy of individual sets of PDF90IPOI’s,

estimated using different assumptions of 𝛼𝛼. A set of PDF90IPOI’s, produced under a given value

of 𝛼𝛼, may collectively be more accurate compared to another set of the same size, generated

using a different assumption of 𝛼𝛼; however, regardless of the relative superiority or inferiority

in accuracy, both sets may be inaccurate (or accurate). The next section (Section 8.2.4)

provides a comprehensive discussion on how one can examine the standalone accuracy of a set

of PDF90IPOI’s and reports the outcomes of the tests for the set of 276 PDF90IPOI’s obtained under

the maximum likelihood estimate of 𝛼𝛼 (i.e., 𝛼𝛼 = 1.5). Appendix A presents the outputs of

these tests for the cases where 𝛼𝛼 assumes a value from the following set: [0, 0.5, 1.25, 1.9, 3,

4, 5]. In concise terms, the results indicate that if one selects an 𝛼𝛼 value remote from 1.5,

namely 0, 0.5, 1.9, or ≥ 3, and subsequently estimates the PDF90IPOI’s of all the 276 IPOIs, the

resultant PDF90IPOI’s would fail to pass the accuracy tests. The implication of the outcomes of

the tests is two-fold: They provide evidence on the degree of the standalone accuracy of the

individual sets of PDF90IPOI’s, produced under different assumptions of 𝛼𝛼. In addition, they

signify the importance of the role of 𝛼𝛼 in the proposed risk methodology, through

demonstrating how changes in the value of 𝛼𝛼 can influence the standalone accuracy of the

corresponding sets of estimated PDF90IPOI’s.

Page 234: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

202

Figure 8-10. This graph exhibits the outcomes of the calibration process for varying sets of engaged IPOs.

As displayed on the horizontal axis, the set producing the first point on the plot consists of the 52nd IPO,

only. The second point on the figure is generated using the set comprising the 52nd and 53rd IPOs. In a like

fashion, the remaining points are obtained by sequentially increasing the size of the set of participating IPOs

by one. For each set, the maximum likelihood estimate of 𝛼𝛼 is found through repeating the maximum

likelihood calibration process described in the text. As a case in point, consider the label 52-100, on the

horizontal axis; it implies that the corresponding calibration process was carried out on the basis of the

PDF90IPOI’s estimated for the 52nd IPO to the 100th IPO, or equivalently, the 1st IPOI to the 49th IPOI, under the

individual 𝛼𝛼 values depicted on the vertical axis. For this particular set of length 49, the likelihood function

reaches its maximum at the value of 𝛼𝛼 = 1.25; that is, the maximum likelihood estimate of 𝛼𝛼 is equal to

1.25. This figure demonstrates that despite early fluctuations in the value of the maximum likelihood

estimate of 𝛼𝛼, the trend remains steady at 𝛼𝛼 = 1.5 once the number of participating IPOs rises beyond a

certain threshold (i.e., 148).

Page 235: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

203

8.2.4. Scrutiny of Estimated Probability Density Functions

Scrutinizing the accuracy of the estimated PDF90IPOI’s is undoubtedly the most important

piece of the puzzle that this research endeavours to solve. The calibration process pursues the

goal of finding a parameter estimate (i.e., 𝛼𝛼 = 1.5) under which the probability of the realized

prices on day 90 is maximum. Hence, its outcome can merely be interpreted in terms of the

‘relative’ accuracy of the PDFs; that is, the PDFs produced using the assumption of 𝛼𝛼 = 1.5

are typically more accurate compared to the ones yielded under other values of 𝛼𝛼. However,

such statement does not shed any light on the ‘standalone’ accuracy of the PDFs generated

under the assumption of 𝛼𝛼 = 1.5. This section aims at tackling this aspect.

Backtesting forms the basis of any attempt to examine the proposed methodology since past

events are valuable sources of knowledge and the success of the methodology in achieving its

objectives can be judged by analyzing its performance under historical scenarios. The topic of

scrutiny is approached from three different angles which are discussed successively in separate

sections; i.e., Sections 8.2.4.1, 8.2.4.2, and 8.2.4.3.

8.2.4.1. Test of Uniform Cumulative Probabilities

The proposed methodology yields a specific PDF90IPOI for each IPOI, and only is a single

observation available for each PDF90IPOI; that is, the respective realized price on day 90

(𝑣𝑣90realized). Suppose for a moment that instead of 276 individual PDF90IPOI’s and 276 𝑣𝑣90realized’s,

there were a single estimated PDF, referred to as PDF, and 276 samples deemed to be

random draws from PDF. One straightforward approach to verify the correctness of this

claim would be gauging the deviation of the empirical PDF, built upon the 276 samples, from

the hypothesized PDF. If the deviation is proven to be insignificant, then the original claim

cannot be rejected.

This research, however, faces a different setting: here, one attempts to examine the

accuracy of a methodology (a black box, so to speak) that produces different PDFs, for each of

Page 236: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

204

which only a single value would subsequently be realized. The test of uniform cumulative

probabilities is an approach that can be employed in such a case in order to examine the

validity of the methodology, and consequently its outcomes. This approach draws on a concept

referred to as ‘probability-integral transform’ which is proven below for the specific problem

at hand. For an extended elaboration, an interested reader may refer to statistics texts such as

Gibbons and Chakraborti (2005) [58]. By means of this approach, one can restructure the

problem such that it resembles the simple above-indicted example where a single PDF (PDF)

was involved.

For each IPOI, its 𝑣𝑣90realized is associated with a cumulative probability value (CPV) which

is computed based on the corresponding estimated PDF90IPOI. To clarify further, CPV can be

expressed as follows:

CPV𝑖𝑖 = P�𝑉𝑉90 ≤ 𝑣𝑣𝑖𝑖,90

realized�

𝑗𝑗 = 1, … ,276 (8.9)

where, 𝑗𝑗 designates the ID of the IPOI under consideration and ranges from 1 to 276. The

variable 𝑣𝑣𝑖𝑖,90realized denotes the realized price of the 𝑗𝑗th IPOI on day 90.

The primary step of the test of uniform cumulative probabilities involves computing CPVs

for all the 276 IPOIs. The test indicates that if the methodology has produced

accurate PDF90IPOI’s in all the 276 cases, the empirical PDF of the computed 276 CPVs must

resemble the uniform PDF spanning the interval of [0,1]. To phrase more precisely, the

methodology would pass the accuracy test if it is proven that the observed set of 276 CPVs has

been drawn from a reference population with uniform density between 0 and 1. The following

proposition elucidates the origin of this test:

Page 237: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

205

Proposition 1.

CPV is a random variable which is uniformly distributed over the interval of [0,1].

Proof.

The following lemma must first be established in order to accomplish the proof of

Proposition 1.

Lemma 1.

Suppose 𝑋𝑋 and 𝑌𝑌 are random variables whose PDFs are denoted by 𝑓𝑓𝑋𝑋(𝑥𝑥) and

𝑓𝑓𝑌𝑌(𝑦𝑦), respectively. Regard 𝑌𝑌 as a function of �𝑌𝑌 = ℊ(𝑋𝑋)�. The PDF of 𝑌𝑌 can,

therefore, be expressed as below:

𝑓𝑓𝑌𝑌(𝑦𝑦) =𝑓𝑓𝑋𝑋�ℊ−1(𝑦𝑦)�ℊ′�ℊ−1(𝑦𝑦)�

(8.10)

Proof.

𝑦𝑦 = ℊ(𝑥𝑥)

→ � 𝑥𝑥 = ℊ−1(𝑦𝑦)

d𝑦𝑦 = ℊ′(𝑥𝑥)d𝑥𝑥𝑓𝑓𝑌𝑌(𝑦𝑦)d𝑦𝑦 = 𝑓𝑓𝑋𝑋(𝑥𝑥)d𝑥𝑥

(1)(2)(3)

(2)�� d𝑥𝑥 =

d𝑦𝑦ℊ′(𝑥𝑥)

(1)�� d𝑥𝑥 =

d𝑦𝑦ℊ′�ℊ−1(𝑦𝑦)�

Page 238: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

206

(1),(2)→(3)�⎯⎯⎯⎯⎯⎯� 𝑓𝑓𝑌𝑌(𝑦𝑦)d𝑦𝑦 =

𝑓𝑓𝑋𝑋�ℊ−1(𝑦𝑦)�d𝑦𝑦ℊ′�ℊ−1(𝑦𝑦)�

In order to prove proposition 1, once again assume that 𝑋𝑋 is a random variable with a

PDF represented by 𝑓𝑓𝑋𝑋(𝑥𝑥). Suppose 𝑥𝑥� is a sample drawn from 𝑓𝑓𝑋𝑋(𝑥𝑥). Note that prior to

sampling, 𝑥𝑥� was itself a random variable which is equal to 𝑋𝑋 in distribution. Denote

this random variable as 𝑋𝑋�. It can readily be perceived that the cumulative probability

value of 𝑋𝑋� is likewise a random variable. Designate this latter random variable by 𝑌𝑌

which can mathematically be expressed as 𝑌𝑌 = Ƒ𝑋𝑋(𝑋𝑋�), where Ƒ denotes the continuous

cumulative density function (CDF) of 𝑋𝑋. The PDF of 𝑌𝑌 can then be found using

Lemma 1 (see, Equation (8.10)):

𝑓𝑓𝑌𝑌(𝑦𝑦) = 𝑓𝑓𝑋𝑋�Ƒ𝑋𝑋−1(𝑦𝑦)�Ƒ𝑋𝑋 ′ �Ƒ𝑋𝑋−1(𝑦𝑦)�

(8.11)

= 𝑓𝑓𝑋𝑋(𝑥𝑥�) Ƒ𝑋𝑋 ′ (𝑥𝑥�)

= 𝑓𝑓𝑋𝑋(𝑥𝑥�) 𝑓𝑓𝑋𝑋(𝑥𝑥�)

= 1

According to Equation (8.11), 𝑓𝑓𝑌𝑌(𝑦𝑦) is equal to unity which implies that it is a

uniform over a range of width 1. Considering that 𝑌𝑌 represents a cumulative

probability value, it must be non-negative and less than, or equal to, unity. Thus, the

uniform density, 𝑓𝑓𝑌𝑌(𝑦𝑦), spans the interval of [0,1].

Page 239: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

207

In the context of this thesis, CPV, defined by Equation (8.9), acts as the

counterpart of 𝑌𝑌. Likewise, 𝑣𝑣90realized can be regarded as the analogue of 𝑥𝑥�. Thus, prior

to its realization, it is considered a random variable which can be denoted by 𝑉𝑉�90. In

the light of this analogy, it is clear that akin to the cumulative probability of 𝑋𝑋� which

is a random variable with the uniform PDF on [0,1], the cumulative probability value

of 𝑉𝑉�90 (i.e., CPV) is uniformly distributed over the unit interval of [0,1] .

Proposition 1 proved that CPV is a uniform random variable on [0,1]. Thus, each CPV𝑖𝑖 ,

computed based on Equation (8.9), is indeed a sample from the uniform PDF on [0,1]. The

problem has now boiled down to a familiar one: There exists a theoretical PDF, which is

the uniform on [0,1], and a sample set of length 276, which encompasses the observed

CPVs. If the proposed methodology is estimating the PDF90IPOI’s accurately, the empirical

PDF constructed on the basis of the observed CPVs must resemble the uniform PDF on

[0,1]. The next subsection focuses on the scrutiny of this aspect.

8.2.4.1.1. Non-Parametric Statistical Tests to Examine the Uniformity Assumption

Figure 8-11 presents the histogram of the 276 CPVs computed using Equation (8.9). The

range of each bin can be regarded as one ‘event’. According to Proposition 1, in theory, these

events must exhibit equal frequency; that is, the sample data must be distributed evenly

between the bins. A visual inspection indicates that there exists a sufficient resemblance

between the shape of this histogram and a theoretical discrete uniform distribution over the

range of [0,1]. However, one may turn to statistical tests as firm bases for deriving substantive

and solid conclusions about the uniformity of the CPVs, and consequently, the accuracy of the

proposed methodology for estimating PDF90IPOI. The following hypothesis testing is, therefore,

conducted:

H0: The observed set of 276 CPVs has been drawn from a reference population with

uniform density between 0 and 1.

Page 240: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

208

H1: The observed set does not come from the reference density indicated under H0.

Two non-parametric tests are adopted to execute the hypothesis testing:

(i) Pearson’s chi-square test to assess ‘goodness of fit’

(ii) One-sample Kolmogorov-Smirnov test (K-S test)

Pearson’s chi-square test to assess ‘goodness of fit’:

In the case of the former test, Pearson’s chi-square test, the null hypothesis can more

precisely be stated as follows: the observed sequence of 276 CPVs has been drawn from a

population where all mutually exclusive events occur with equal frequency. As a case in point,

consider the histogram depicted by Figure 8-11. The range of each bin on the x-axis marks an

event which is mutually exclusive with the others, and the frequency of each event is defined

as the portion of the total observations it holds. The histogram in Figure 8-11 consists of 20

events. As an example, 0.05 < CPV≤ 0.1 outlines one event whose observed and theoretical

frequencies can respectively be calculated as 14/276 = 0.0507 and 1/20 = 0.05.

At par with the majority of statistical tests, Pearson’s chi-square test defines a test statistic

which performs as a measure of similarity between the observed sample set and the reference

or theoretical set. The test statistic resembles ‘sum of squared errors’ and is obtained by first

computing individual differences between the observed frequency of each event and its

expected frequency in the reference or theoretical uniform histogram. For the example

indicated above, the difference would be (0.0507 − 0.05). The squared differences are then

summed and normalized with the theoretical frequency in order to achieve the test statistic.

Page 241: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

209

It has been proven that under the null hypothesis, the test statistic follows a chi-square

distribution whose degree of freedom is equal to the number of events (bins) reduced by one

since there exists one constraint; that is, the sum of frequencies equals unity. Once the test

statistic is computed, one must refer to the chi-square distribution with the respective degree of

freedom and fetch the probability of observing a value equal to, or greater than, the test

statistic. This probability is referred to as p-value. A p-value greater than a pre-specified

significance level, Ƿ, would indicate that there is not a significant evidence against the null

hypothesis. Hence, H0 cannot be rejected at the significance level of Ƿ. For practical purposes,

Ƿ is often set to be 0.05. Further details regarding Pearson’s chi-square test can be accessed in

Figure 8-11. The figure presents the histogram of the CPVs computed for the 276 IPOIs (i.e., 52nd IPO to

327th IPO), using Equation (8.9).

Page 242: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

210

the vast literature addressing this topic. The author has benefited from Panneerselvam (2004)

[107] and Altiok and Melamed (2007) [6].

One-sample Kolmogorov-Smirnov test (K-S test):

Evidently, Pearson’s chi-square test is sensitive to the classification of the events, or

simply, the number of bins. Consequently, it would generate the most reliable outcomes when

data are abundant. The K-S test offers an alternative approach which resolves this issue. It

does not rely on a histogram (or PDF); rather, it compares the empirical CDF with the

reference CDF, and hence, is capable of coping with smaller samples.

In its ‘one-sample’ case, the K-S test is well-known for its capacity to examine whether a

sample comes from a normal distribution; yet, the theory is applicable to any reference

distribution, including uniform distributions. The null hypothesis of the one-sample K-S test

can more precisely be phrased as follows: the empirical CDF of the sample ‘agrees’ with the

CDF of the reference density. The null hypothesis is tested against the alternative hypothesis

according to which, the empirical CDF of the sample does not agree with the reference CDF.

The K-S test statistic is defined as the maximum absolute distance between the reference

CDF and the empirical CDF. The test statistic indeed acts as a measure to gauge the goodness

of fit, and as inferred from the definition, smaller values are more desirable. If ň denotes the

total number of observations and the test statistic is referred to as Šň, under the null hypothesis,

√ňŠň converges in distribution to the ‘Kolmogorov’ distribution. An ‘insignificant’ p-value

would cast doubt on the validity of the null hypothesis, and if the pre-defined significance

level (e.g., Ƿ = 0.05) exceeds the p-value, the null hypothesis would be rejected. The reader is

encouraged to consult texts of statistics for further information on the K-S test (e.g.,

Panneerselvam, 2004 [107]; Altiok and Melamed, 2007 [6]).

Page 243: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

211

Test Results:

The following p-values were obtained for the realized set of 276 CPVs:

Pearson’s chi-square test: p-value = 0.9710

K-S test: p-value = 0.8893

Both p-values are satisfactorily large and well beyond the significance levels set in practice.

The null hypothesis is, therefore, accepted from which it follows that the proposed

methodology (i.e., Phase I and Phase II) has accomplished accurate and reliable PDFs.

8.2.4.2. Test of VaR Break Frequency

Due to its comprehensiveness, the test of uniform CPVs alone would suffice to establish the

accuracy of the estimated PDFs, and hence, the subsequent inferences; however, putting that

aside, the methodology can further be scrutinized by examining the realized frequency of

‘VaR breaks’. Below, it is expounded why this test offers a suitable means to evaluate the

suggested methodology.

The ultimate goal of this research is to offer a methodology for quantifying the short-term

risk of IPOs. This objective is accomplished by devising a VaR approach which computes a

90-day VaR at the confidence level of 100 × (1 − ῤ)%. (e.g., 95% if ῤ = 0.05). Recall from

Section 5.1 that a 90-day VaR represents a threshold price (or gain-loss) which can be

interpreted as follows: with the probability of (1 − ῤ) (e.g., 0.95), the price of day 90 will not

descend below the level designated by VaR. The process of estimating PDF90IPOI’s, in its

entirety, can be regarded as a black box. For each IPOI, this black box produces a PDF which

would then act as the basis of the respective VaR analysis. VaR is essentially a quantile and its

accuracy is contingent upon, and attributed to, the accuracy of the underlying PDF (in this

case, PDF90IPOI). Therefore, in examining a VaR system, it is indeed the underlying PDF which

is assessed.

Page 244: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

212

Examining the frequency of ‘VaR breaks’ is a prevalently-used ‘backtesting’ approach to

evaluate various VaR models (e.g., Prem et al., 2010 [113]; Pinedo and Walter, 2013 [112] ). If for

a given case, the estimated VaR threshold exceeds the realized price (or gain-loss), a VaR

break has taken place. To warrant reliance, the ‘realized’ frequency of VaR breaks under a

given VaR approach is expected to be at par with the ‘nominal’ frequency of VaR breaks (i.e.,

ῤ).

The essence of the notion of ‘VaR break frequency’ rests on the frequency interpretation of

probability. Suppose that there exists an accurate VaR system which produces 95% VaRs on a

daily ‘time horizon’. If this system is used every day to obtain a VaR estimate for the next day,

the long-run frequency of VaR breaks would converge to 5%. In the context of this research,

the time horizon of the VaR system is 90 days and it has generated 276 VaR estimates. Table

8-1 reports the realized VaR break frequencies under the nominal frequencies of ῤ = 5%,

ῤ = 10%, ῤ = 15%, and ῤ = 20%. As presented, the computed VaR break frequencies

closely resemble their nominal counterparts which attests to the validity of the proposed

methodology (black box) of this thesis. Note that due to the relatively small size of the sample

(i.e., 276 VaR estimates), one would not expect the realized frequencies to be exactly at par

with the nominal frequencies since the latter indeed denote long-run frequencies.

8.2.4.3. Test of Impact of Comparable Selection and Bayesian Updating

on Estimated PDFs

The two tests presented in the previous sections demonstrated the accuracy of the proposed

methodology of this thesis (i.e., Phase I and Phase II). There still remains one more avenue

worth exploring; that is, the impact of incorporating the two steps of comparable selection and

Bayesian updating on the quality of the estimated PDF90IPOI’s. These two steps have

undoubtedly enhanced the mathematical rigor of achieving the PDF90IPOI for a given IPOI. The

question is what their contribution has been to the accuracy of the final outcomes. This section

aims at addressing this aspect and inspects the effectiveness of these two steps in delivering

the mission of this research.

Page 245: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

213

For the sake of fluency, it is advantageous to first review the logic underlying the

comparable selection and Bayesian updating steps. This thesis pursued a methodology to

estimate the probability density function of a short-term price or return for a given IPOI. This

task was, however, encumbered with lack of sufficient historical evidence and knowledge – an

inherent characteristic of most IPOs. To cope with this challenge, the closest comparable of

each IPOI was appointed to compensate for the requisite, yet non-existent, data. Though, the

impact of the closest comparable was kept to a minimum through only adopting its joint

density of parameters. Recall that this joint density was indeed in its posterior form; the

Bayesian approach assisted with this step and served to refine and revise the prior joint density

of the closest comparable before it was dispensed to the given IPOI.

Table 8-1. For each nominal frequency (ῤ) on the leftmost column, the rightmost counterpart presents the

realized frequency which is computed as the quotient of the middle column over the total number of the

IPOIs (i.e., 276).

Nominal

Frequency of

VaR Breaks (ῤ)

VaR Breaks

(Number

out of 276)

Realized

Frequency of

VaR Breaks

0.05 16 0.0580

0.1 30 0.1087

0.15 41 0.1486

0.2 55 0.1993

Page 246: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

214

In order to demonstrate the significance of the steps of comparable selection and Bayesian

updating, it is next investigated how the methodology would perform if they were excluded. It

would shrink into the following steps:

First.

For a given IPO (i.e., IPO𝑖𝑖, where 𝑘𝑘 = 52 , …, 327), there exist (𝑘𝑘 − 1) preceding realized

IPOs; i.e., when analyzing IPO𝑖𝑖, none of the succeeding IPOs has yet taken place. This set

of length (𝑘𝑘 − 1) is utilized to obtain the respective uniform joint density of the

parameters (i.e., 𝑓𝑓Μ,Σ,Ξ,Ω(𝜇𝜇,𝜎𝜎, 𝜉𝜉,𝜔𝜔)), following the same steps used to estimate the uniform

prior joint density for the ‘pioneers’ in Section 8.2.2. The only minor modification is that

here, the uniform joint density is obtained using all of the (𝑘𝑘 − 1) former IPOs, as

opposed to using the pioneers only. The process is sophisticated and hence the brevity of a

short review may cause some confusion. The reader is, therefore, encouraged to consult

Section 8.2.2 for the specifics.

Second.

The first step yields individual uniform joint densities for the IPOs on the list. The

respective uniform joint density of each IPO serves as the joint density required in

Equation (6.42) to compute PDF90IPOI. Note that for a given IPO, its uniform joint density

was obtained using all of its preceding IPOs, with no reference to the closest comparable.

Further note that the respective uniform joint density was directly plugged into Equation

(6.42) without undergoing any Bayesian updating process.

The new methodology implicitly assumes that there exists a theoretical global joint density of

the parameters and attempts to approximate it by an empirical global joint density. For each

IPOI, this empirical joint density is developed using the ‘sample events’ dispensed from all of

the preceding realized IPOs; in the case of each former IPO, the combination of the confidence

intervals of its parameters represents one ‘sample event’ drawn from the theoretical global

density. Such logic is frequently deployed in empirical studies. Ruud (1993) [125] and Asquith

Page 247: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

215

et al. (1998) [12] can be listed as examples in the context of IPOs. In contrast to the new

methodology, the methodology proposed in Phase I and Phase II associates specific joint

densities with individual IPOIs and assumes that the joint density of a given IPOI is unique

and can only be approximated by the posterior joint density of its closest comparable.

The outcomes of the new methodology are evaluated based on the same tests used to

examine the results of the former methodology:

(i) test of uniform cumulative probability values (Section 8.2.4.1)

(ii) test of VaR break frequency (Section 8.2.4.2)

For the sake of clarity, henceforth, the new methodology, which excludes the two steps of

comparable selection and Bayesian updating, is referred to as ‘simplified’, and the former

methodology, which preserves the two steps, is called ‘comprehensive’.

Test of Uniform Cumulative Probability Values

Figure 8-12 is the counterpart of Figure 8-11, plotted on the basis of the outcomes of the

‘simplified’ methodology. The histogram does not evoke the shape of a uniform distribution;

yet, the nonparametric tests of K-S and Pearson’s chi-square were employed to test the null

hypothesis that ‘the observed set of 276 CPVs has been drawn from a reference population

with uniform density between 0 and 1’. The tests would offer some perspective on the

probability of the observed sample set coming from the uniform distribution of [0,1] (see,

Section 8.2.4.1.1). The following p-values were yielded by the tests:

Pearson’s chi-square test: p-value = 1.86 × 10−6

K-S test: p-value = 0.00024

which provide no significant evidence to accept the null hypothesis. It can, therefore, be

inferred that the ‘simplified’ methodology fails to estimate PDF90IPOI’s accurately which, in

consequence, would adversely impact the accuracy of the VaR outcomes, as shown next.

Page 248: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

216

Test of VaR Break Frequency

Table 8-2 presents the results of the analysis of VaR break frequency. The realized frequencies

fall inferior to the nominal frequencies, to a great extent. This observation implies that the

VaR inferences, developed on the basis of the ‘simplified’ methodology, overestimate the risk

at all the four confidence levels of 95%, 90%, 85%, 80%.

The outcomes of the test of VaR break frequency can further be supported by a closer

inspection of Figure 8-12. It can be deduced from this figure that the PDF90IPOI’s estimated

Figure 8-12. This figure is produced following the same steps used for Figure 8-11. The only difference

remains the underlying methodology which generates the CPVs. Figure 8-11 depicts the histogram of the

CPVs computed using the ‘comprehensive’ methodology; whereas, here, the graph is formed on the basis of

the CPVs resulted from the ‘simplified’ methodology.

Page 249: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

217

using the ‘simplified’ methodology associate comparatively greater probabilities with the

rare events on the left tail of the distributions (e.g., 0.1 < 𝑉𝑉90𝑣𝑣offer

≤ 0.1 + d𝑣𝑣90𝑣𝑣offer

). Since VaR, at

any confidence level, is in essence a quantile, a ‘heavier’ left tail would yield relatively low

(i.e., pessimistic) VaR thresholds which are hard to break. The comparatively lower numbers

of VaR breaks in Table 8-2 attest to this feature.

The accuracy of the outcomes of the ‘comprehensive’ methodology was previously

established (see, Sections 8.2.4.1 and 8.2.4.2). The analysis conducted in this section

demonstrates that not only does the ‘comprehensive’ methodology yield reliable outcomes, but

also all the steps it engages to obtain such outputs are essential and effective. Excluding the

two steps of comparable selection and Bayesian updating would lead to a VaR model which

Table 8-2. This table can be regarded as the counterpart of Table 8-1. The latter reports the realized VaR

break frequencies under the ‘comprehensive’ methodology; whereas, this table exhibits the outcomes of the

test of VaR break frequency for the ‘simplified’ methodology.

Nominal

Frequency of

VaR Breaks (ῤ)

VaR Breaks

(Number

out of 276)

Realized

Frequency of VaR

Breaks under the

‘Simplified’

Methodology

0.05 3 0.0109

0.1 9 0.0326

0.15 14 0.0507

0.2 23 0.0833

Page 250: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 8. Results 8.2. Results of Phase II: Assessment of Short-Term Risk

218

overestimates risk. Pessimism bias in risk estimations would deter the participation of

investors, even the ones with a moderate or high tolerance for risk, or may impose excessive

hedging requirements, leading to waste of resources.

Page 251: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

219

Chapter 9. CONCLUSIONS AND FUTURE WORK

9.1. In Conclusion: IPO Risk Analysis Unravelled Using

the Bayesian Perspective

As the closing remarks of this thesis, it is worth reiterating that the focal intent of this

research was to design and backtest a methodology for quantifying the short-term risk of IPOs.

Devising a risk methodology for IPOs – which would yield a mathematically literate

interpretation of risk – was often burdened by the scarcity of data. This thesis commenced

with the objective of solving this problem. It was founded on a foresight that one potential

solution lies in developing a methodology which would bridge the two familiar notions

indicated below.

First Notion. Comparable-based methods are prevalently utilized in various analysis

domains. It was established in Chapter 3 why founding the risk analysis on comparables

offers a promising approach to quantify the market price-based risk of IPOs. Therein,

evidence was provided for the predominance of the comparable-based approach among

analysts and its potentially significant role on market prices. It was, therefore, concluded

that a comparable-based approach can act as the building block of the risk assessment of

IPOs.

Second Notion. The Bayesian framework offers a unique setting for risk assessments in

various fields including finance. It is particularly advantageous when there exist limited

data, yet, strong prior belief – the two characteristics inherent in most IPOs. The capacity

of the Bayesian approach to measure the ‘epistemic’ component of uncertainty, in addition

to its ‘aleatory’ constituent, is another feature that notably contributes to its reputation as a

suitable framework to pursue risk analyses (see, Section 5.5).

In an environment of data paucity, the first notion would assist with identifying the most

suitable source of prior knowledge (i.e., the closest comparable). The second notion offers a

Page 252: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.1. In Conclusion: IPO Risk Analysis Unravelled Using the Bayesian Perspective

220

setting where the prior belief can be formulated and refined to be utilized in the risk

assessment of a given IPO. The following remarks further elaborate on the function of the

second notion in fulfilling the objectives sought by this thesis.

In this study, the stochastic process of price is modeled by a GBM with indeterministic

parameters. Despite the frequentist school of thought, the Bayesian framework views

parameters as random variables whose uncertainty is of the ‘epistemic’ type. Such setting

would allow one to structure the joint density required to describe the probabilities of all

possible combinations of the model parameters. Subsequently, one can proceed with

estimating the predicative density of price (i.e., PDF90IPOI) by averaging over all combinations

of the parameters with individual weights set as the probabilities of observing the vicinity of

the respective combinations (see, Section 5.5). The predictive density yielded by this approach

has indeed hybridized parameter uncertainty and data uncertainty (i.e., both types of

uncertainty) which leads to inferences like VaR that can be viewed as comprehensive

concepts, simultaneously addressing all sources of uncertainty. Thus, not only does this study

offer a novel systematic VaR methodology to quantify risk of IPOs, but by virtue of the latter

feature, it further serves to enhance the accuracy of the proposed VaR approach; that is, in

estimating the uncertainty of price at a specified short-term horizon, it recognizes that the

definiteness of price is blurred by the epistemic uncertainty, in addition to the aleatory

uncertainty, and hence, proceeds with unraveling the IPO risk analysis using the Bayesian

VaR wisdom.

The risk methodology of this thesis comprises two phases (i.e., Phase I and Phase II) and

addresses the aforementioned two notions as described below.

The first notion is handled in Phase I (see, Section 6.2) which yields two significant

contributions: The primary contribution is a DEA-based multi-dimensional ‘similarity’

metric, developed for selecting the closest comparable (i.e., the most suitable source of

prior knowledge). In working out the details of the similarity metric and in searching for a

DEA super-efficiency model which would satisfy the specific features and constraints of

Page 253: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.1. In Conclusion: IPO Risk Analysis Unravelled Using the Bayesian Perspective

221

the problem under consideration, an opportunity for a new model was recognized. This new

DEA model of super-efficiency forms the secondary contribution of Phase I. It is

worthwhile to note that Phase I has been developed within the framework of DEA;

however, despite its conventional usage as a tool of efficiency analysis, DEA here serves as

a mapping means. In the multi-dimensional space of operating and financial characteristics,

each firm can be represented by one point. DEA maps this point into a new space whose

coordinates are DEA parameters. The mapping facilitates multi-dimensional comparisons

between firms which then lead to the selection of closest comparables.

The second notion is addressed in Phase II (see, Section 6.3) which focuses on the chief

contribution of this thesis; that is, measuring the short-term risk of IPOs in a Bayesian

setting.

Subsequent to designing the methodology, the validation studies were conducted. The

suggested methodology was implemented and examined by means of approximately 11000

lines of program code in MATLAB (a trademark of MathWorks®). The process of data

mining returned 327 U.S. IPOs (1990-2012) in the ‘health-care’ sector which were engaged

towards this goal (see, Chapter 7). Below, the major pieces of the scrutiny work and the

respective outcomes are concisely reviewed (see, Chapter 8 for more details).

Piece 1: Phase I

Phase I is implemented in three different settings, referred to as RUN 1, RUN 2,

and RUN 3. The latter – where both steps of efficient IPO treatment (see, Section

6.2.4) and outlier detection (see, Section 6.2.5) are executed – offers a more

theoretically coherent structure and consequently leads to outcomes more consistent

with practical experience. It was, therefore, chosen for the rest of the analyses.

In Section 8.1, the robustness of the comparable selection method is demonstrated

through three different qualitative evaluations: (i) inspection of the number of

comparables associated with individual IPOIs, (ii) scrutiny of ‘average-union’

Page 254: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.1. In Conclusion: IPO Risk Analysis Unravelled Using the Bayesian Perspective

222

ratios, (iii) systematic examination of the changes in the composition of the set of

comparables for a given IPOI, induced by sequentially enhancing the ‘pool of

candidates’ of the given IPOI through adding a new IPO to its pool in each

sequence. It was established that the composition of the set of comparables for a

given IPOI tends to remain intact as the respective pool of candidates grows in size.

The composition may only change with the entry of a new IPO which, upon its

arrival, would replace an existent comparable. This observation attests to the power

of Phase I in finely selecting the comparables which play a key role in Phase II.

Piece 2: Phase II and Model Calibration

Phase II aims at estimating the PDF of each IPOI’s stock price at day 90

(i.e., PDF90IPOI). This PDF subsequently acts as the basis of 90-day VaR estimations.

One significant piece of information required to carry out Phase II is supplied by

Phase I; that is, the ID of the closest comparable to a given IPOI. This knowledge is

then utilized in a Bayesian setting to accomplish the goal pursued by Phase II.

Finding the closest comparable of each IPOI mandates defining a similarity metric

which would act as a measure for proximity and compute the distance between the

IPOI and each of its comparables. The distance function developed in Phase I serves

this aim, yielding a smaller distance for a closer comparable. To reduce uncertainty,

the parameter of the distance function is calibrated using a maximum likelihood

approach. Section 8.2.3 describes how the calibration is executed and presents the

outcomes. The procedure returns the value of 1.5 as the maximum likelihood

estimate of 𝛼𝛼. Recall that in the distance function, 𝛼𝛼 represents the coefficient of the

second term which gauges the difference between the efficiency scores. Hence, a

value of 𝛼𝛼 greater than zero, and unity for that matter, affirms the significance of

modelling the compromise between 𝜓𝜓 and 𝜆𝜆 and validates the notion of

incorporating the efficiency score as one dimension of comparison, postulated

originally in Sections 3.3 and 6.2.

Page 255: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.1. In Conclusion: IPO Risk Analysis Unravelled Using the Bayesian Perspective

223

Piece 3: Scrutiny of Estimated 𝐏𝐏𝐏𝐏𝐏𝐏𝟗𝟗𝟗𝟗𝐈𝐈𝐏𝐏𝐈𝐈𝐈𝐈’s

This piece is undoubtedly the most significant and critical of all the pieces since it

examines the ‘standalone’ accuracy of the estimated PDF90IPOI’s. Any test selected to

accomplish this task would simultaneously authenticate the accuracy of VaR

estimates since VaR is essentially a quantile and its credibility stems from the

accuracy of its underlying PDF. Section 8.2.4.1 proposes a backtesting technique

which scrutinizes the correctness of the estimated PDF90IPOI’s through inspecting the

PDF of the cumulative probability values (CPVs) associated with the realized 𝑉𝑉90’s.

For a given IPOI, the CPV of the realized 𝑉𝑉90 is computed based on the respective

estimated PDF90IPOI. It is proven that the theoretical PDF of the CPVs is uniform over

the interval of [0,1] (see, Proposition 1, Section 8.2.4.1). From Proposition 1, it

follows that if the proposed methodology of this thesis is projecting reliable

PDF90IPOI’s, one must fail to reject the null hypothesis ‘the observed set of 276 CPVs

has been drawn from a reference population with uniform density between 0 and 1’

against the alternative ‘the observed set does not come from the reference density’.

The following two statistical tests were employed to perform the hypothesis testing:

(i) Pearson’s chi-square test to assess ‘goodness of fit’, (ii) One-sample

Kolmogorov-Smirnov test (K-S test). The tests yielded the p-values of 0.9710 and

0.8893, respectively. Both p-values are sufficiently large to warrant the acceptance

of the null hypothesis and, hence, ensure that the proposed methodology conforms

to the prescribed standards.

Computing the frequencies of ‘VaR break’ was another approach undertaken to

further scrutinize the credibility of the estimated PDF90IPOI’s. The frequency

interpretation of probability can be employed to illuminate the rationale behind this

technique. Suppose that the entire process of quantifying risk can be regarded as a

‘system’ which estimates a 90-day VaR for a given IPOI. For an accurate system –

if it is utilized in 𝑛𝑛 cases of IPOIs to estimate the corresponding 90-day VaRs at the

Page 256: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.1. In Conclusion: IPO Risk Analysis Unravelled Using the Bayesian Perspective

224

confidence level of (1 − ῤ) – the frequency of VaR breaks must converge to ῤ as 𝑛𝑛

tends to infinity. The following VaR break frequencies were respectively reported

for the 90-day VaRs with the confidence levels of 95%, 90%, 85%, 80%: 0.0580,

0.1087, 0.1486, 0.1993. Due to the limited sample size, one does not expect the

realized frequencies to be exactly at par with the nominal frequencies (i.e., ῤ); yet,

they are sufficiently (and satisfactorily) close to validate the VaR approach.

In closing the preceding recapitulating discussion, it is noteworthy to point out that any

model is essentially just an abstraction of the reality it intends to reproduce. Therefore,

regardless of its complexity, a model could only be useful if it reflects the reality. By means of

the validation process, one can ensure that the model provides a sensible approximation of the

observed phenomena and conforms to the rules of the underlying system as well as the

objectives, principles and preferences of the original problem. These criteria have been

achieved by the risk methodology of this thesis. Here, the main objective was the

mathematical modelling of the short-term risk of IPOs and the validity and accuracy of the

methodology was established through backtesting analyses. One day, one might design a new

model to solve the same problem – a frequent scenario in the history of science. The point

indicated above then implies that this model would have to undergo the same backtesting and

scrutiny process as the proposed methodology of this research did. The model would only be

proven accurate and reliable if it meets the pre-set criteria and demonstrates satisfactory test

outcomes.

Page 257: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.2. Future Research Directions and Application Prospects

225

9.2. Future Research Directions and Application

Prospects

The focus of this section can best be summarized as follows: The first section (i.e., Section

9.2.1) presents further applications of this research. The discussion then proceeds to briefly

outline potential future research directions (see, Section 9.2.1.3).

9.2.1. Extension of Applications

The primary motivation and application of the proposed methodology in this thesis is to

measure the short-term risk of a portfolio consisting of a single IPO. The applications are,

however, broad. Particularly pronounced is the capacity of the methodology to facilitate and

improve the risk analysis of portfolios comprising other assets, in addition to IPO stocks.

Furthermore, the methodology can be employed towards developing hedging strategies for

portfolios containing IPO shares. As clarified in the following sections, these two applications

form the basis for developing an approach which decomposes the VaR computed for an IPO

into its systematic and unsystematic components and estimates the contribution of each source

to the risk of the IPO. With inevitable brevity, the following three sections (i.e., Sections

9.2.1.1 to 9.2.1.3) are devoted to indicating the main points of the aforelisted applications.

9.2.1.1. Risk Measurement of Portfolios Containing IPOs

For various reasons, skilled investors may be persuaded to hold a diversified portfolio of

stocks of IPO and public companies. One such incentive could be diversifying away the risk of

IPOs. Regardless of the underlying motivation for building such portfolios, the investors

would seek to properly measure the corresponding risk. This task is, nevertheless, hard to

accomplish due to lack of sufficient market history which impedes developing a perspective

about the correlation between the returns of IPO and non-IPO stocks. The methodology

offered in this thesis provides a path towards mitigating this problem, and the remainder of

Page 258: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.2. Future Research Directions and Application Prospects

226

this section attempts to map out a more explicit formulation of this path using a simple

example:

Consider a portfolio of two assets: stocks of company IP and stocks of company PU.

Company PU is a mature public company with sufficient trading data, and company IP is on

the verge of going public. For the sake of simplicity, assume that the weights of the assets are

pre-set. The approach presented below aims at estimating the short-term risk of this portfolio.

In accordance with the methodology proposed in this thesis, the short-term risk is defined as a

90-day VaR computed based on the probability density function of the value of the portfolio

90 days after the issuing day of company IP. This probability density function is, hereafter,

referred to as PDF90Portfolio.

One prerequisite piece of information for estimating PDF90Portfolio is the correlation between

the (logarithmic) returns of company IP and company PU over the 90-day period commencing

on the first day of trade for company IP. The correlation can be approximated using the closest

comparable of company IP as detailed next:

Suppose Phase I has identified company CO as the closest comparable to company IP.

Recall that company CO is selected as the closest comparable in terms of its pre-IPO

characteristics, and at present, it is a public company with an acceptable length of market

history. The short-term correlation between the returns of company IP and company PU can be

approximated as the correlation between the returns of company CO, realized over the

respective 90-day post-IPO period and the returns of company PU, observed during the same

90-day window. Once the correlation between the returns of company IP and company PU is

estimated, one can proceed with developing an approach for estimating PDF90Portfolio. Drawing

upon the methodology proposed in this thesis, one such approach is presented below.

Page 259: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.2. Future Research Directions and Application Prospects

227

The value of the portfolio at time 𝑡𝑡, denoted by 𝑉𝑉𝑡𝑡P, can mathematically be demonstrated as

follows:

𝑉𝑉𝑡𝑡P = 𝑤𝑤PU𝑉𝑉𝑡𝑡PU + 𝑤𝑤IP𝑉𝑉𝑡𝑡IP (9.1)

where, the superscripts PU and IP designate company PU and company IP, respectively. The

variable 𝑤𝑤 stands for the weight of an asset in the portfolio, and 𝑉𝑉𝑡𝑡 represents value at time 𝑡𝑡.

Based on Equation (9.1), an infinitesimally small change in the value of the portfolio,

d𝑉𝑉𝑡𝑡P, over an infinitesimal time interval, d𝑡𝑡, can be expressed as shown below:

d𝑉𝑉𝑡𝑡P = 𝑤𝑤PUd𝑉𝑉𝑡𝑡PU + 𝑤𝑤IPd𝑉𝑉𝑡𝑡IP (9.2)

Recall that in Section 6.3, the stochastic process of price was modelled as a GBM.

Drawing upon the same assumption, d𝑉𝑉𝑡𝑡PU and d𝑉𝑉𝑡𝑡IP can be described as follows:

d𝑉𝑉𝑡𝑡PU = 𝜇𝜇PU𝑉𝑉𝑡𝑡PUd𝑡𝑡 + 𝜎𝜎PU𝑉𝑉𝑡𝑡PUd𝑍𝑍1,𝑡𝑡 (9.3)

d𝑉𝑉𝑡𝑡IP = 𝜇𝜇IP𝑉𝑉𝑡𝑡IPd𝑡𝑡 + 𝜎𝜎IP𝑉𝑉𝑡𝑡IPd𝑍𝑍2,𝑡𝑡 (9.4)

where, the parameters 𝜇𝜇 and 𝜎𝜎 are constants representing the expected rate of return and

volatility, respectively, and the variable 𝑍𝑍 follows a Wiener process. Substituting Equations

(9.3) and (9.4) into Equation (9.2) and rearranging yield the following equation:

Page 260: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.2. Future Research Directions and Application Prospects

228

d𝑉𝑉𝑡𝑡P = 𝑤𝑤PU𝑉𝑉𝑡𝑡PU�𝜇𝜇PUd𝑡𝑡 + 𝜎𝜎PUd𝑍𝑍1,𝑡𝑡� (9.5)

+ 𝑤𝑤IP𝑉𝑉𝑡𝑡IP�𝜇𝜇IPd𝑡𝑡 + 𝜎𝜎IPd𝑍𝑍2,𝑡𝑡�

Equation (9.5) can be written in terms of two uncorrelated Wiener processes. To this end,

d𝑍𝑍2,𝑡𝑡 is replaced by the following two terms:

d𝑍𝑍2,𝑡𝑡 = 𝜌𝜌d𝑍𝑍1,𝑡𝑡 + �1 − 𝜌𝜌2d𝑍𝑍3,𝑡𝑡 (9.6)

where, 𝑍𝑍3,𝑡𝑡 denotes a new Wiener process that is uncorrelated with 𝑍𝑍1,𝑡𝑡, and 𝜌𝜌 denotes the

correlation coefficient between the logarithmic returns of company PU and company IP, over

the time interval discussed above. By substituting Equation (9.6) into Equation (9.5), the

stochastic differential equation of the value of the portfolio, 𝑉𝑉𝑡𝑡P, takes the final form shown

below:

d𝑉𝑉𝑡𝑡P = �𝑤𝑤PU𝑉𝑉𝑡𝑡PU𝜇𝜇PU + 𝑤𝑤IP𝑉𝑉𝑡𝑡IP𝜇𝜇IP�d𝑡𝑡 (9.7)

+ �𝑤𝑤PU𝑉𝑉𝑡𝑡PU𝜎𝜎PU + 𝑤𝑤IP𝑉𝑉𝑡𝑡IP𝜎𝜎IP𝜌𝜌�d𝑍𝑍1,𝑡𝑡

+ 𝑤𝑤IP𝑉𝑉𝑡𝑡IP𝜎𝜎IP�1 − 𝜌𝜌2d𝑍𝑍3,𝑡𝑡

Subsequently, one can solve for the conditional PDF of the portfolio value on day 90,

given a specific combination of the parameters. The unconditional PDF of the portfolio value

on day 90 (i.e., PDF90Portfolio) can then be estimated in the same fashion as PDF90IPOI, using

Equation (6.42), if the following two considerations are taken into account:

Page 261: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.2. Future Research Directions and Application Prospects

229

Consideration 1. Company PU introduces two additional parameters into Equation (6.42).

Company PU represents a mature public company, and its sufficient trading history

provides some confidence regarding estimates of the respective parameters (i.e., 𝜇𝜇PU and

𝜎𝜎PU). It is, therefore, assumed that the parameters are not random and can be treated as

fixed values. This assumption reduces the number of parameters required to estimate

PDF90Portfolio and simplifies the solution, to a great extent. In the case of company IP (i.e.,

the IPO), the same approach as the one proposed in Chapter 6 is employed; that is, the

parameters 𝜇𝜇IP and 𝜎𝜎IP are assumed to be draws from the random variables ΜIPand ΣIP,

respectively. In addition, to better reflect the specific characteristics of the first day of

trade, the random variable of the ratio of first closing price to offer price is assumed to

follow a log-normal distribution whose underlying parameters (i.e., hyperparameters) are

random themselves. The hyperparameters are denoted by 𝜉𝜉IP and 𝜔𝜔IP which represent

realizations from the random variables ΞIP and ΩIP, correspondingly. Since ΞIP and ΩIP

could potentially be correlated with ΜIPand ΣIP, their distributions must be considered

jointly in order to find PDF90Portfolio. This joint distribution is designated with the function

𝑓𝑓ΜIP,ΣIP,ΞIP,ΩIP(𝜇𝜇IP,𝜎𝜎IP, 𝜉𝜉IP,𝜔𝜔IP).

Consideration 2. A closed-form expression for the conditional PDF of the portfolio value on

day 90 is absent.

Equation (6.42) was developed for a portfolio of a single asset (i.e., a given IPOI) whose

stochastic process of price movements was assumed to follow a GBM. Therefore, for a

given combination of the parameters, the conditional PDF of price on day 90 could be

expressed in closed form; it was a log-normal distribution, denoted by

ℒ𝑉𝑉90(𝑣𝑣90|Μ ≈ 𝜇𝜇, Σ ≈ 𝜎𝜎,𝑉𝑉1 ≈ 𝑣𝑣1) which can more precisely be written as

ℒ𝑉𝑉90IP(𝑣𝑣90IP�ΜIP ≈ 𝜇𝜇IP, ΣIP ≈ 𝜎𝜎IP,𝑉𝑉1IP ≈ 𝑣𝑣1IP). On the contrary, in the case of a portfolio

comprising two GBM assets, it seems hard to find a closed-form expression for the

conditional PDF of the portfolio value on day 90. It is left to be explored if the stochastic

Page 262: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.2. Future Research Directions and Application Prospects

230

differential equation of the portfolio value, presented in Equation (9.7), can be solved

analytically. At this stage, due to the absence of such a solution, for each combination of

the parameters, a Monte Carlo simulation can be performed to estimate the conditional

PDF of the portfolio value on day 90.

With these two considerations, Equation (6.42) is modified into the following form:

𝑓𝑓𝑉𝑉90P (𝑣𝑣90P ) (9.8)

= � � � � � 𝒢𝒢𝑉𝑉90P (𝑣𝑣90P �ΜIP ≈ 𝜇𝜇IP, ΣIP ≈ 𝜎𝜎IP,𝑉𝑉1IP ≈ 𝑣𝑣1IP, 𝜇𝜇PU,𝜎𝜎PU) +∞

𝑣𝑣1IP=−∞

+∞

𝜇𝜇IP=−∞

+∞

𝜎𝜎IP=−∞

+∞

𝜉𝜉IP=−∞

+∞

𝜔𝜔IP=−∞

× 𝑓𝑓V1IP(𝑣𝑣1IP|𝜉𝜉IP,𝜔𝜔IP) d𝑣𝑣1IP × 𝑓𝑓ΜIP,ΣIP,ΞIP,ΩIP(𝜇𝜇IP,𝜎𝜎IP, 𝜉𝜉IP,𝜔𝜔IP) d𝜇𝜇IPd𝜎𝜎IPd𝜉𝜉IPd𝜔𝜔IP

where, 𝑉𝑉90P denotes the value of the portfolio on day 90. The function 𝑓𝑓𝑉𝑉90P (𝑣𝑣90P ) represents

PDF90Portfolio which indeed stands for the unconditional PDF of 𝑉𝑉90P . The conditional PDF of

𝑉𝑉90P is presented as the function 𝒢𝒢𝑉𝑉90P (𝑣𝑣90P �ΜIP ≈ 𝜇𝜇IP, ΣIP ≈ 𝜎𝜎IP,𝑉𝑉1IP ≈ 𝑣𝑣1IP, 𝜇𝜇PU,𝜎𝜎PU) . The rest

of the notation remains consistent with the one presented earlier.

9.2.1.2. Hedging the IPO Investment Risk

The unconditional PDF (i.e., PDF90Portfolio) estimated above, can be utilized to develop

hedging strategies. In the previous section, it was assumed that the proportion of each asset in

the portfolio is predetermined, and the analysis aimed at quantifying the short-term risk of the

portfolio under the pre-set weights. One can, however, formulate an optimization problem to

find the optimum set of weights which minimizes the 90-day VaR of the portfolio, at a given

confidence level. To elaborate, for each combination of the weights, the steps indicated above

would be repeated to estimate PDF90Portfolio and the respective VaR. The set of weights under

Page 263: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.2. Future Research Directions and Application Prospects

231

which the portfolio exhibits the minimum VaR, or conversely the minimum risk, is selected as

the optimum set. One can partially hedge against the risk of the IPO by creating a portfolio on

the basis of the optimum weights. The optimization problem can be augmented with an

additional constraint to reflect the expected portfolio return desired by the investor. This

approach would offer a means to achieve the right balance between risk and reward.

9.2.1.3. IPO VaR Decomposition

One can benefit from the techniques discussed in the previous two sections (i.e., Sections

9.2.1.1 and 9.2.1.2) to decompose the computed VaR for a given IPOI into partial VaRs. A

case in point would be estimating the relative contributions of the systematic and unsystematic

components to the total VaR of IPOI. One possible approach to accomplish this goal is

described below:

Suppose that a 90-day VaR is computed for a given IPOI which was referred to as company

IP in the previous sections. Further suppose that a portfolio of assets IP and PU is created

using the optimum weights which minimize the portfolio VaR. Recall that the abbreviation PU

stands for ‘public’ and represents a mature public company. The difference between the VaR

of IPOI and the VaR of the portfolio can be considered as a proxy for the unsystematic

component of the risk of IPOI since it is the risk that vanishes upon diversification. It must be

noted that the suggested method serves as a tentative solution. In practice, often a relatively

large portfolio is required to completely eliminate the unsystematic risk.

9.2.2. Future Research: Incorporating Information from Other

Comparables

Phase I (see, Section 6.2) identifies a set of comparables for a given IPOI. The closest

comparable is selected as the one which exhibits the minimum distance from IPOI, where

‘distance’ is calculated using the similarity metric presented in Equation (6.37). Recall that the

proposed methodology only appoints the closest comparable’s posterior joint PDF of

Page 264: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.2. Future Research Directions and Application Prospects

232

parameters to act as the (prior) joint PDF of parameters required to estimate PDF90IPOI for the

respective IPOI. Stated otherwise, the information held by the rest of the comparables is

dismissed. A future research may focus on modifying Phase I such that it utilizes the

supplementary information offered by the remaining comparables. The scope of Phase I may

be extended in one of the following two directions:

Direction 1. Virtual Firm

The outcomes of Phase I can be used to form a virtual firm which would be positioned

closer to the given IPOI than any of the respective comparables. One can then proceed with

the virtual firm, as opposed to the closest comparable, to conduct Phase II. The following

details ought to be worked out before the method can be adopted in the analysis of the

short-term risk of IPOs:

The concept of virtual firm can be illustrated using a small-scale example. Figure 9-1.a

visualizes a simple one-input and one-output DEA example with a variable returns to scale

(VRS) frontier of efficiency. Suppose that firm G represents IPOI. Its set of comparables

would then consist of firms A, B, E, and F, encircled in grey (see, Section 6.2 for more

details). The virtual firm would be located in the convex hull of the set of comparables and

is defined as the convex combination of A, B, E, and F which falls the closest to firm G.

Firm K in Figure 9-1.b designates the virtual firm in this case. To attain a suitable

compromise, an additional optimization problem must be solved which would search for

the optimum vector of weights to be engaged towards building the virtual firm.

The virtual firm, however, adds to the intricacy of the analysis. If the GBM assumption is

maintained, the price process of the virtual firm can be modelled as the convex combination

of the GBM processes of the comparables. Equation (9.5) merges two GBM processes but

can be extended to address more. As discussed there, the result would not be a GBM

anymore. It can readily be deduced that the process of estimating the requisite, yet non-

existent, joint PDF of the parameters for a given IPOI becomes greatly complicated due to

the following reasons: (i) the price models for the virtual firm and for the IPOI are

Page 265: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.2. Future Research Directions and Application Prospects

233

(a)

(b)

Figure 9-1. The figure illustrates a one-input and one-output DEA example with a variable returns to

scale frontier of efficiency. Firm G represents the firm of interest whose comparables are encircled

in grey in the upper graph. The lower graph depicts the convex hull of the set of comparables as well

as the ‘virtual’ firm which resides on the boundary of the convex hull and is designated by Firm K.

Page 266: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.2. Future Research Directions and Application Prospects

234

incompatible, and so are the corresponding underlying parameters, (ii) a closed-form

solution for the stochastic differential equation of the virtual firm seems hard to achieve

which would increase the complexity of the Bayesian updating process, described in

Section 6.3.2.1. Both reasons call for extra refining steps in order to prepare the prior joint

PDF of the parameters of the virtual firm to be dispensed to the given IPOI.

Besides, the notion of virtual firm would convolute the problem in another way: it

introduces extra parameters pertaining to the other comparables. The increase in the number

of parameters would significantly magnify the computational burden.

As a side note, it must be stated that the total number of outstanding shares varies across

different firms and it impacts the per-share prices. Hence, one may use ‘market

capitalization’, as opposed to ‘price’, when combining the GBM processes. The achieved

model would then describe the evolution of the market capitalization of the virtual firm.

Direction 2. Compound Posterior Joint Density

Direction 2 can be considered as an alternative to Direction 1, presented above. Likewise,

it offers an approach which would utilize the information available from all the

comparables to enhance the proposed methodology in Chapter 6. Therein, only the closest

comparable’s posterior joint PDF of parameters was adopted as a proxy for the (prior) joint

PDF required to estimate PDF90IPOI. One can, however, incorporate the information held by

the remaining comparables into the analysis through building a compound posterior joint

PDF of parameters which can subsequently replace the posterior joint PDF of the closest

comparable in the methodology presented in Chapter 6. In order to form a compound

posterior joint PDF, first, the posterior joint PDFs of the remaining comparables must be

computed in the same fashion as that for the closest comparable. The compound posterior

joint PDF is then formed as a convex combination of the posterior joint PDFs of all the

comparables.

Page 267: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.2. Future Research Directions and Application Prospects

235

The challenge of the approach suggested here lies in determining the coefficients of

combination. The posterior joint PDFs ought to be combined on the basis of the probability

of observing each which must bear some relevance to the distance between the

corresponding comparable and IPOI. To this end, one can define a function which would

relate the distance between a given IPOI and each comparable (calculated using Equation

(6.37)) to the probability of observing the posterior joint PDF of the respective comparable.

Inspired by the Boltzmann distribution, one possible solution is demonstrated below:

𝑃𝑃𝑐𝑐𝑐𝑐 =e−𝛽𝛽𝐷𝐷𝑐𝑐𝑏𝑏

2

∑ 𝑒𝑒−𝛽𝛽𝐷𝐷𝑐𝑐𝑏𝑏2𝜂𝜂𝑐𝑐=1

(9.9)

where, 𝑃𝑃 represents the probability of the event where the prior joint PDF of IPOI is equal

to the posterior joint PDF of the respective comparable. In line with Equation (6.37), IPOI

is identified by the subscript 𝑏𝑏 and the comparable is marked by the subscript 𝑐𝑐, where

𝑐𝑐 = 1, … , 𝜂𝜂 and the variable 𝜂𝜂 indicates the total number of comparables associated with

IPOI. The distance between IPOI (i.e., 𝑏𝑏) and the corresponding comparable (i.e., 𝑐𝑐) is

denoted by 𝐷𝐷𝑐𝑐𝑐𝑐. The parameter 𝛽𝛽 is an unknown constant that can be estimated in a

calibration process. Note that according to Equation (9.9), as the distance between a given

IPOI and its comparable approaches infinity, the probability of the respective two joint

PDFs resembling each other tends to zero. Furthermore, the closer a comparable is

positioned to IPOI, the higher is the probability that its posterior joint PDF is identical to

the prior joint PDF of IPOI. Both inferences indicate that Equation (9.9) conforms to the

concept of ‘distance’ defined in this thesis. It, however, serves as a provisional solution,

pending further scrutiny and evaluation.

It is worthy of pointing out that the discussions presented under both ‘Direction 1’ and

‘Direction 2’ assume that the method of selecting the set of comparables for a given IPOI

Page 268: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Chapter 9. Conclusions and Future Work 9.2. Future Research Directions and Application Prospects

236

remains intact and as described in Chapter 6. In future, one may try developing a different

method for this purpose; that is, investigating the possibility of incorporating new DEA

models into the process. Cases in point would be the recently suggested DEA models by

Aparicio, 2014 [9], which pursue the objective of finding the closest efficient target for an

inefficient DMU, or the BAM model proposed by Cooper et al., 2011 [40] (see, Section 4.3.5).

Once customized to meet the specific requirements of this research, such models could

potentially significantly improve the results and hence, the author puts forth the idea of

incorporating them in Phase I as a promising research avenue to be explored.

In conclusion, it must be reiterated that the latter section of this chapter closes the document

by first presenting more applications which further elaborate on the significance of this

research. Subsequently, the focus shifts to proposing two approaches that can be employed

towards benefiting from the extra knowledge provided by the comparables other than the

closest comparable. The discussion also outlines the methodological challenges concerning the

development and application of these methods. It bears repeating that this research has

succeeded in achieving its intent and has produced accurate and satisfactory outcomes in the

absence of the suggested extensions. These additional steps would considerably increase the

mathematical and computational intricacy of the process, where the accuracy gain is uncertain.

Thus, prior to incorporating them into the existing methodology, one must consider the trade-

off between increased complexity and superior outcomes.

Page 269: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

237

REFERENCES

[1] Abdou K. and Dicle M. F., "Do Risk Factors Matter in the IPO Valuation?",

Journal of Financial Regulation and Compliance, vol. 15, no. 1, pp. 63- 89,

2007.

[2] Abidin S. N. Z. and Jaffar M. M., "A Review on Geometric Brownian

Motion in Forecasting the Share Prices in Bursa Malaysia", World Applied

Sciences Journal, vol. 17, pp. 87-93, 2012.

[3] Aggarwal R. , "Stabilization Activities by Underwriters after Initial Public

Offerings", The Journal of Finance, vol. 55, no. 3, pp. 1075-1103, 2000.

[4] Alford A. W., "The Effect of the Set of Comparable Firms on the Accuracy

of the Price-Earnings Valuation Method", Journal of Accounting Research, vol.

30, no. 1, pp. 94-108, 1992.

[5] Ali A. I. and Seiford L. M., "Translation Invariance in Data Envelopment

Analysis", Operations Research Letters, vol. 9, no. 6, pp. 403-405 , 1990.

[6] Altiok T. and Melamed B. , Simulation Modeling and Analysis with

ARENA.: Academic Press, Elsevier Inc., 2007.

[7] Anadol B. , "Valuing Private Companies: A Dea Approach", University of

Toronto, MASc Dissertation 2000.

[8] Andersen P. and Petersen N. C., "A Procedure for Ranking Efficient Units

in Data Envelopment Analysis", Management Science, vol. 39, no. 10, pp.

1261-1264 , 1993.

[9] Aparicio J. , Borras F. , Ortiz L. , and Pastor J. T., "Benchmarking in

Healthcare: An Approach Based on Closest Targets", in Managing Service

Page 270: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

238

Productivity: Using Frontier Efficiency Methodologies and Multicriteria

Decision Making for Improving Service Performance, Emrouznejad A. and

Cabanda E. , Eds.: Springer, 2014, pp. 67-92.

[10] Arnold T. , Fishe R. P.H., and North D. , "The Effects of Ambiguous

Information on Initial and Subsequent IPO Returns", Financial Management,

vol. 39, no. 4, pp. 1497-1519, 2010.

[11] Arthurs J. D., Hoskisson R. E., Busenitz L. W., and Johnson R. A.,

"Managerial Agents Watching Other Agents: Multiple Agency Conflicts

Regarding Underpricing in IPO Firms", Academy of Management Journal, vol.

51, no. 2, pp. 277-294, 2008.

[12] Asquith D. , Jones J. D., and Kieschnick R. , "Evidence on Price

Stabilization and Underpricing in Early IPO Returns", The Journal of Finance,

vol. 53, no. 5, pp. 1759-1773, 1998.

[13] Asquith P. , Mikhail M. B., and Andrea S. A., "Information Content of

Equity Analyst Reports", Journal of Financial Economics, vol. 75, no. 2, pp.

245–282, 2005.

[14] Ayayi A. , "The 180-Day Lock-Up Period and Insiders’ Equity Selling",

Journal of Economics and Finance , vol. 29, no. 1, pp. 32-45 , 2005.

[15] Banker R. D., Charnes A. , and Cooper W. W., "Some Models for

Estimating Technical and Scale Inefficiencies in Data Envelopment Analysis",

Management Science, vol. 30, no. 9, pp. 1078-1092, 1984.

[16] Banker R. D. and Morey R. C., "Efficiency Analysis for Exogenously Fixed

Inputs and Outputs", Operations Research, vol. 34, no. 4, pp. 513-521, 1986.

[17] Baron D. P., "A Model of the Demand for Investment Banking Advising

Page 271: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

239

and Distribution Services for New Issues", The Journal of Finance, vol. 37, no.

4, pp. 955-976, 1982.

[18] Baron D. P. and Holmström B. , "The Investment Banking Contract for New

Issues Under Asymmetric Information: Delegation and the Incentive Problem",

The Journal of Finance, vol. 35, no. 5, pp. 1115-1138, 1980.

[19] Bartov E. , Mohanram P. , and Seethamraju C. , "Valuation of Internet

Stocks: An IPO Perspective", Journal of Accounting Research, vol. 40 , no. 2,

pp. 321-346, 2002.

[20] Benveniste L. M. and Spindt P. A., "How Investment Bankers Determine

the Offer Price and Allocation of New Issues", Journal of Financial

Economics, vol. 24, no. 2, pp. 343-361, 1989.

[21] Berger J. , "The Case for Objective Bayesian Analysis", Bayesian Analysis,

vol. 1, no. 3, pp. 385-402, 2006.

[22] Berger J. O. et al., "The Likelihood Principle", Lecture Notes-Monograph

Series, vol. 6, 1988.

[23] Berk J. B., Green R. C., and Naik V. , "Valuation and Return Dynamics of

New Ventures", The Review of Financial Studies, vol. 17, no. 1, pp. 1-35,

2004.

[24] Berry D. A., Statistics: A Bayesian Perspective.: Duxbury Press, 1996.

[25] Bhojraj S. and Lee C. M. C., "Who Is My Peer? A Valuation-Based

Approach to the Selection of Comparable Firms", Journal of Accounting

Research, vol. 40, no. 2, pp. 407-439, 2002.

[26] Biais B. and Faugeron-Crouzet A. M., "IPO Auctions: English, Dutch,...

French, and Internet", Journal of Financial Intermediation, vol. 11, no. 1, pp.

Page 272: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

240

9–36, 2002.

[27] Black F. and Scholes M. , "The Pricing of Options and Corporate

Liabilities", Journal of Political Economy , vol. 81, no. 3, pp. 637-654, 1973.

[28] Bliss R. R. and Panigirtzoglou N. , "Option-Implied Risk Aversion

Estimates", The Journal of Finance, vol. 59, no. 1, pp. 407-446, 2004.

[29] Boatsman J. R. and Baskin E. F., "Asset Valuation with Incomplete

Markets", The Accounting Review, vol. 56, no. 1, pp. 38-53, 1981.

[30] Bogetoft P. and Otto, L., Benchmarking with DEA, SFA, and R.: Springer

Science+Business Media, 2011.

[31] Brau J. C. and Fawcett S. E., "Initial Public Offerings: An Analysis of

Theory and Practice", The Journal of Finance, vol. 61, no. 1, pp. 399-436,

2006.

[32] Brav A. and Lehavy R. , "An Empirical Analysis of Analysts' Target Prices:

Short-term Informativeness and Long-term Dynamics", The Journal of

Finance, vol. 58, no. 5, pp. 1933-1968, 2003.

[33] Certo S. T., Holcomb T. R., and Holmes Jr. R. M., "IPO Research in

Management and Entrepreneurship: Moving the Agenda Forward", Journal of

Management, vol. 35, no. 6, pp. 1340–1378, 2009.

[34] Chambers R. G., Chung Y. , and Färe R. , "Benefit and Distance Functions",

Journal of Economic Theory, vol. 70, no. 2, pp. 407-419, 1996.

[35] Chambers R. G., Chung Y. , and Färe R. , "Profit, Directional Distance

Functions, and Nerlovian Efficiency", Journal of Optimization Theory and

Applications, vol. 98, no. 2, pp. 351-364, 1998.

Page 273: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

241

[36] Charnes A. , Cooper W. W., and Rhodes E. , "Measuring the Efficiency of

Decision Making Units", European Journal of Operational Research, vol. 2,

no. 6, pp. 429-444, 1978.

[37] Charnes A. , Cooper W. W., Rousseau J. , and Semple J. , "Data

Envelopment Analysis and Axiomatic Notions of Efficiency and Reference

Sets", Center for Cybernetic Studies, University of Texas, CCS Research

Report 558, 1987.

[38] Cheng G. , Zervopoulos P. , and Qian Z. , "A Variant of Radial Measure

Capable of Dealing with Negative Inputs and Outputs in Data Envelopment

Analysis", European Journal of Operational Research, vol. 225, no. 1, pp.

100-105, 2013.

[39] Chung Y. H., Färe R. , and Grosskopf S. , "Productivity and Undesirable

Outputs: A Directional Distance Function Approach", Journal of

Environmental Management, vol. 51, no. 3, pp. 229-240, 1997.

[40] Cooper W. W., Pastor J. T., Borras F. , Aparicio J. , and Pastor D. , "BAM:

A Bounded Adjusted Measure of Efficiency for Use with Bounded Additive

Models", Journal of Productivity Analysis, vol. 35, no. 2, pp. 85-94, 2011.

[41] Cooper W. W., Seiford L. M., and Tone K. , Data Envelopment Analysis: A

Comprehensive Text with Models, Applications, References and DEA-Solver

Software.: Springer Science+Business Media, 2007.

[42] Cornelli F. and Goldreich D. , "Bookbuilding and Strategic Allocation", The

Journal of Finance, vol. 56, no. 6, pp. 2337–2369, 2001.

[43] Cornelli F. and Goldreich D. , "Bookbuilding: How Informative Is the Order

Book?", The Journal of Finance, vol. 58, no. 4, pp. 1415–1443, 2003.

Page 274: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

242

[44] Damien P. , Dellaportas P. , Polson N. G., and Stephens D. M., Bayesian

Theory and Applications.: Oxford University Press, 2013.

[45] De Sousa M. and Stošić B. , "Technical Efficiency of the Brazilian

Municipalities: Correcting Nonparametric Frontier Measurements for

Outliers", Journal of Productivity Analysis, vol. 24, no. 2, pp. 157-181, 2005.

[46] DeAngelo L. E., "Equity Valuation and Corporate Control", The Accounting

Review, vol. 65, no. 1, pp. 93-112, 1990.

[47] Dowd K. , Beyond Value at Risk: The New Science of Risk Management.:

John Wiley & Sons Ltd., 1999.

[48] Draho J. , The IPO Decision: Why and How Companies Go Public.: Edward

Elgar Publishing Limited, 2004.

[49] Emrouznejad A. , Anouze A. L., and Thanassoulis E. , "A Semi-Oriented

Radial Measure for Measuring the Efficiency of Decision Making Units with

Negative Data, Using DEA", European Journal of Operational Research, vol.

200, no. 1, pp. 297-304, 2010.

[50] Fama E. F. and French K. R., "Industry Costs of Equity", Journal of

Financial Economics, vol. 43, no. 2, pp. 153-193, 1997.

[51] Farrell M. J., "The Measurement of Productive Efficiency", Journal of the

Royal Statistical Society. Series A (General) , vol. 120, no. 3, pp. 253-290,

1957.

[52] Farzipoor Saen R. , "Developing a Nondiscretionary Model of Slacks-Based

Measure in Data Envelopment Analysis", Applied Mathematics and

Computation, vol. 169, no. 2, pp. 1440–1447, 2005.

[53] (2013) Federal Reserve Bank of St. Louis. [Online].

Page 275: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

243

http://research.stlouisfed.org/fred2/series/GDPC1/downloaddata?cid=106

[54] Feldman M. S. and Pentland B. T., "Reconceptualizing Organizational

Routines as a Source of Flexibility and Change", Administrative Science

Quarterly, vol. 48, no. 1, pp. 94-118, 2003.

[55] Ferson S. , "Bayesian Methods in Risk Assessment", Dominique

GUYONNET, Head of Waste & Storage Unit, Working Paper 2004.

[56] Garfinkel J. A., "IPO Underpricing, Insider Selling and Subsequent Equity

Offerings: Is Underpricing a Signal of Quality?", Financial Management, vol.

22, no. 1, pp. 74-83, 1993.

[57] Geddes R. , IPOs and Equity Offerings. Oxford: Butterworth-Heinemann,

2003.

[58] Gibbons J. D. and Chakraborti S. , Nonparametric Statistical Inference.:

Taylor & Francis e-Library, 2005.

[59] Gilson S. C., Hotchkiss E. S., and Ruback R. S., "Valuation of Bankrupt

Firms", Review of Financial Studies, vol. 13, no. 1, pp. 43-74, 2000.

[60] Giordano L. and Siciliano G. , "Real-World and Risk-Neutral Probabilities

in the Regulation on the Transparency of Structured Products", Commissione

Nazionale per le Società e la Borsa (CONSOB), Working Paper 2013.

[61] Golany B. and Roll Y. , "Some Extensions of Techniques to Handle Non-

Discretionary Factors in Data Envelopment Analysis", The Journal of

Productivity Analysis, vol. 4, no. 4, pp. 419-432, 1993.

[62] Group RiskMetrics , "Risk-Metrics Technical Document", J.P. Morgan,

1986.

Page 276: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

244

[63] Hadavandi E. , Shavandi H. , and Ghanbari A. , "Integration of Genetic

Fuzzy Systems and Artificial Neural Networks for Stock Price Forecasting",

Knowledge-Based Systems, vol. 23, no. 8, pp. 800-808, 2010.

[64] Hall J. , "The Impact of Growth, Volatility and Competitive Advantage on

the Value of Equity Investments and Their Embedded Options", University of

Queensland, PhD Thesis 2005.

[65] Hand J. R.M., "Profits, Losses and the Non-Linear Pricing of Internet

Stocks", Working Paper 2000. [Online]. SSRN: http://ssrn.com/abstract=204875 or

http://dx.doi.org/10.2139/ssrn.204875

[66] Hand J. R.M., "The Market Valuation of Biotechnology Firms and

Biotechnology R&D", Working Paper 2001. [Online]. SSRN:

http://ssrn.com/abstract=294471 or http://dx.doi.org/10.2139/ssrn.294471

[67] Hardle W. K. and Simar L. , Applied Multivariate Statistical Analysis, 3rd

ed.: Springer, 2012.

[68] Holton G. A., "Closed Form Value at Risk", Contingency Analysis, pp. 1-10,

1996.

[69] Holton G. , "Subjective Value at Risk", Risks and Rewards Newsletter, no.

31, pp. 14-16, 1998.

[70] Houston J. , James C. , and Karceski J. , "What a Difference a Month

Makes: Stock Analyst Valuations Following Initial Public Offerings", Journal

of Financial and Quantitative Analysis, vol. 41, no. 1, pp. 111-137, 2006.

[71] Howson C. and Urbach P. , Scientific Reasoning: The Bayesian Approach,

2nd ed.: Open Court Publishing Company, 1996.

[72] Hsu Y. T., Liu M. C., Yeh J. , and Hung H. F., "Forecasting the Turning

Page 277: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

245

Time of Stock Market Based on Markov–Fourier Grey Model", Expert Systems

With Applications, vol. 36, no. 4, pp. 8597-8603, 2009.

[73] Huck S. W., Statistical Misconceptions.: Taylor & Francis Group, LLC,

2009.

[74] Hughes P. J. and Thakor A. V., "Litigation Risk, Intermediation, and the

Underpricing of Initial Public Offerings", Review of Financial Studies, vol. 5,

no. 4, pp. 709-742, 1992.

[75] Hull J. C., Risk Management and Financial Institutions, 3rd ed.: John Wiley

& Sons, Inc., 2012.

[76] Ibbotson R. G., "Price Performance of Common Stock New Issues", Journal

of Financial Economics, vol. 2, no. 3, pp. 235-272, 1975.

[77] Ibbotson R. G. and Jaffe J. F., "'Hot Issue' Markets", The Journal of

Finance, vol. 30, no. 4, pp. 1027-1042, 1975.

[78] Ibbotson R. G. and Ritter J. R., "Initial Public Offerings", in North-Holland

Handbooks of Operations Research and Management Science, Jarrow R. A.,

Maksimovic V. A., and Ziemba W. T., Eds., 1995, pp. 993-1016.

[79] Jaynes E. T. and Bretthorst G. L., Probability Theory: The Logic of

Science.: Cambridge University Press, 2003.

[80] Jorion P. , Value at Risk: The New Benchmark for Managing Financial Risk,

3rd ed.: McGraw-Hill Companies, Inc., 2006.

[81] Jorion P. and Talmor C. , "Value Relevance of Financial and Non Financial

Information in Emerging Industries: The Changing Role of Web Traffic Data",

Working Paper 2001.

Page 278: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

246

[82] Kaplan S. N. and Ruback R. S., "The Valuation of Cash Flow Forecasts: An

Empirical Analysis", The Journal of Finance, vol. 50, no. 4, pp. 1059-1093,

1995.

[83] Kay S. , Fundamentals of Statistical Signal Processing, Volume I:

Estimation Theory.: Prentice Hall, 1993, vol. 1.

[84] Kerstens K. and De Woestyne V. , "Negative Data in DEA: A Simple

Proportional Distance Function Approach", The Journal of the Operational

Research Society, vol. 62, no. 7, pp. 1413-1419, 2011.

[85] Khindanova I. N. and Rachev S. T., "Value at risk: Recent advances", in

Handbook on Analytic-Computational Methods in Applied Mathematics.:

University of California, Santa Barbara, Department of Economics, 2000.

[86] Khindanova I. N., Rachev S. , and Schwartz E. , "Stable Modeling of Value

at Risk", Mathematical and Computer Modelling, vol. 34, no. 9-11, pp. 1223-

1259, 2001.

[87] Kim M. and Ritter J. R., "Valuing IPOs", Journal of Financial Economics,

vol. 53, no. 3, pp. 409-437, 1999.

[88] Koopmans T. C., "Analysis of Production as an Efficient Combination of

Activities", in Activity Analysis of Production and Allocation., 1951, pp. 33-97.

[Online]. http://www.policonomics.com/wp-content/uploads/Analysis-of-Production-

as-an-Efficient-Combination-of-Activities.pdf

[89] Lee C. , "Underwriter Reputation and the Decision to Go Public", Journal of

Finance and Accountancy, vol. 6, pp. 108-128, 2011.

[90] Ljungqvista A. P. and Wilhelm Jr. W. J., "IPO Allocations: Discriminatory

or Discretionary?", Journal of Financial Economics, vol. 65, no. 2, pp. 167–

Page 279: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

247

201, 2002.

[91] Ljungqvist A. P., Jenkinson T. , and Wilhelm Jr. W. J., "Global Integration

in Primary Equity Markets: The Role of U.S. Banks and U.S. Investors", The

Review of Financial Studies, vol. 16, no. 1, pp. 63-99 , 2003.

[92] Ljungqvist A. , Nanda V , and Singh R. , "Hot Markets, Investor Sentiment,

and IPO Pricing", The Journal of Business, vol. 79, no. 4, pp. 1667-1702, 2006.

[93] Loughran T. and Ritter J. R., "The New Issues Puzzle", The Journal of

Finance, vol. 50, no. 1, pp. 23-51, 1995.

[94] Lovell C.A. K., "Measuring the Macroeconomic Performance of the

Taiwanese Economy", International Journal of Production Economics, vol. 39,

no. 1-2, pp. 165–178, 1995.

[95] Lovell C.A. K. and Pastor J. T., "Units Invariant and Translation Invariant

DEA Models", Operations Research Letters, vol. 18, no. 3, pp. 147–151, 1995.

[96] Malakoff D. , "Bayes Offers a 'New' Way to Make Sense of Numbers",

Science, vol. 286, no. 5444, pp. 1460-1464 , 1999.

[97] Maya C. , "In Search of the True Value of a Start-Up Firm: Creative

Destruction and Real Options Approach", in Proceedings of the 8th Annual

International Conference on Real Options, 2004.

[98] Merton R. C., "Theory of Rational Option Pricing", The Bell Journal of

Economics and Management Science , vol. 4, no. 1, pp. 141-183, 1973.

[99] Mikkelson W. H., Partch, M. M. and Shah K. , "Ownership and Operating

Performance of Companies That Go Public", Financial Economics, vol. 44, no.

3, pp. 281-307, 1997.

Page 280: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

248

[100] Muñiza M. , Paradi J. , Ruggiero J. , and Yang Z. , "Evaluating Alternative

DEA Models Used to Control for Non-Discretionary Inputs", Computers &

Operations Research, vol. 33, no. 5, pp. 1173–1183, 2006.

[101] Myers S. C., "Finance Theory and Financial Strategy", Strategic

Management, vol. 14, no. 1, pp. 126-137, 1984.

[102] O'Hagan A. , "The Bayesian Approach to Statistics", in Handbook of

Probability: Theory and Applications.: Sage Publications, Inc., 2008.

[103] Ohlson J. A., "Eamings, Book Values, and Dividends in Equity Valuation",

Contemporary Accounting Research, vol. 11, no. 2, pp. 661–687, 1995.

[104] Pagano M. , Panetta F. , and Zingales L. , "Why Do Companies Go Public?

An Empirical Analysis", The Journal of Finance, vol. 53, no. 1, pp. 27-64,

1998.

[105] Palepu K. G. and Healy P. M., Business Analysis and Valuation: Using

Financial Statements, 4th ed.: Thomson South-Western, 2008.

[106] Panik M. J., Advanced Statistics from an Elementary Point of View.:

Elsevier Academic Press, 2005.

[107] Panneerselvam R. , Research Methodology.: Prentice-Hall, 2004.

[108] Pastor J. T., "Translation Invariance in Data Envelopment Analysis: A

Generalization", Annals of Operations Research, vol. 66, pp. 93-102, 1996.

[109] Pastor J. T. and Aparicio J. , "Translation Invariance in Data Envelopment

Analysis", in Data Envelopment Analysis: A Handbook of Models and

Methods, Zhu J. , Ed.: Springer Verlag, 2015.

[110] Pastor J. T. and Ruiz J. L., "Variables With Negative Values In DEA", in

Page 281: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

249

Modeling Data Irregularities and Structural Complexities in Data

Envelopment Analysis, Zhu J. and Cook W. D., Eds.: Springer, 2007, ch. 4, pp.

63-84.

[111] Pastor J. T., Ruiz J. L., and Sirvent I. , "An Enhanced DEA Russell Graph

Efficiency Measure", European Journal of Operational Research, vol. 115, no.

3, pp. 596-607, 1999.

[112] Pinedo M. and Walter I. , Eds., Global Asset Management: Strategies, Risks,

Processes, and Technologies.: SimCorp StrategyLab, 2013.

[113] Prem K. P. et al., "Risk Measures Constituting a Risk Metrics which

Enables Improved Decision Making: Value-at-Risk", Journal of Loss

Prevention in the Process Industries, vol. 23, no. 2, pp. 211-219, 2010.

[114] Purnanandam A. K. and Swaminathan B. , "Are IPOs Really Underpriced?",

The Review of Financial Studies, vol. 17, no. 3, pp. 811-848, 2004.

[115] Rachev S. T., Hsu S. J. J., Bagasheva B. S., and Fabozzi F. J., Bayesian

Methods in Finance.: John Wiley & Sons, Inc., 2008.

[116] Ray S. C., "Data Envelopment Analysis, Nondiscretionary Inputs and

Efficiency: An Alternative Interpretation", Socio-Economic Planning Sciences,

vol. 22, no. 4, pp. 167-176, 1988.

[117] (2013) Renaissance Capital: The IPO Expert. [Online]. http://www.renaissancecapital.com/RenCap/AboutUs/WhyIPOs.aspx

[118] Ritter J. R., "The 'Hot Issue' Market of 1980", The Journal of Business, vol.

57, no. 2, pp. 215-240, 1984.

[119] Ritter J. R., "The Long-Run Performance of Initial Public Offerings", The

Journal of Finance, vol. 46, no. 1, pp. 3-27, 1991.

Page 282: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

250

[120] Roosenboom P. , "How Do Underwriters Value Initial Public Offerings? An

Empirical Analysis of the French IPO Market", Contemporary Accounting

Research, vol. 24, no. 4, pp. 1217–1243, 2007.

[121] Roosenboom P. , "Valuing and Pricing IPOs", Journal of Banking and

Finance, vol. 36, no. 6, pp. 1653-1664, 2012.

[122] Ross S. A., Westerfield R. W., Jordan B. D., and Roberts G. S.,

Fundamentals of Corporate Finance.: McGraw-Hill Ryerson, 2010.

[123] Ruggiero J. , "Non-Discretionary Inputs in Data Envelopment Analysis",

European Journal of Operational Research, vol. 111, no. 3, pp. 461–469,

1998.

[124] Ruggiero J. , "On the Measurement of Technical Efficiency in the Public

Sector", European Journal of Operational Research, vol. 90, no. 3, pp. 553–

565, 1996.

[125] Ruud J. S., "Underwriter Price Support and the IPO Underpricing Puzzle",

Journal of Financial Economics, vol. 34, no. 2, pp. 135-151, 1993.

[126] Scheel H. , "Undesirable Outputs in Efficiency Valuations", European

Journal of Operational Research, vol. 132, no. 2, pp. 400-410, 2001.

[127] Schwartz E. S. and Moon M. , "Rational Pricing of Internet Companies",

Financial Analysts Journal, vol. 56, no. 3, pp. 62-75, 2000.

[128] Schwartz E. S. and Moon M. , "Rational Pricing of Internet Companies

Revisited", The Financial Review, vol. 36, no. 4, pp. 7-26, 2001.

[129] Seidl I. and Sommersguter-Reichmann M. , "Visualizing Production

Surfaces in 3D Diagrams", Advances in Operations Research, vol. 2011, pp. 1-

17, 2011.

Page 283: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

251

[130] Seiford L. M. and Zhu J. , "Modeling Undesirable Factors in Efficiency

Evaluation", European Journal of Operational Research, vol. 142, no. 1, pp.

16–20, 2002.

[131] Sharp J. A., Meng W. , and Liu W. , "A Modified Slacks-Based Measure

Model for Data Envelopment Analysis with 'Natural' Negative Outputs and

Inputs", The Journal of the Operational Research Society, vol. 58, no. 12, pp.

1672-1677, 2007.

[132] Silva Portela M. C. A. , Thanassoulis E. , and Sim G. , "Negative Data in

DEA: A Directional Distance Approach Applied to Bank Branches", The

Journal of the Operational Research Society, vol. 55, no. 10, pp. 1111-1121,

2004.

[133] Simak P. C., "Inverse and Negative DEA and their Application to Credit

Risk Evaluation", University of Toronto, PhD Dissertation 2000.

[134] Siu T. K., Tong H. , and Yang H. , "On Bayesian Value at Risk: From

Linear to Non-Linear Portfolios", Asia-Pacific Financial Markets, vol. 11, pp.

161-184, 2004.

[135] Teoh S. H., Wong T. J., and Rao G. R., "Are Accruals during Initial Public

Offerings Opportunistic?", Review of Accounting Studies, vol. 3, no. 1-2, pp.

175–208, 1998.

[136] Tone K. , "A Slacks-Based Measure of Efficiency in Data Envelopment

Analysis", European Journal of Operational Research, vol. 130, no. 3, pp.

498-509, 2001.

[137] Tone K. , "A Slacks-Based Measure of Super-Efficiency in Data

Envelopment Analysis", European Journal of Operational Research, vol. 143,

no. 1, pp. 32-41, 2002.

Page 284: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

252

[138] Vincent-Humphreys R. and Noss J. , "Estimating Probability Distributions

of Future Asset Prices: Empirical Transformations from Option-Implied Risk-

Neutral to Real-World Density Functions", Working Paper No. 455 2012.

[139] Wu L. C., Ong C. S., and Hsu Y.W. , "Knowledge-Based Organization

Evaluation", Decision Support Systems, vol. 45, no. 3, pp. 541-549, 2008.

[140] Yang Z. and Paradi J. C., "Cross Firm Bank Branch Benchmarking Using

'Handicapped' Data Envelopment Analysis to Adjust for Corporate Strategic

Effects", in Proceedings of the 39th Hawaii International Conference on

System Sciences, 2006. [Online]. http://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=01579362

[141] Zhu J. , "Multi-Factor Performance Measure Model with an Application to

Fortune 500 Companies", European Journal of Operational Research, vol.

123, no. 1, pp. 105-124, 2000.

Page 285: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

253

Appendix A. COMPLEMENTARY DETAILS ON THE

CALIBRATION OF THE DISTANCE FUNCTION

A.1 Introduction

Section 8.2.3 detailed the calibration process developed to find the maximum likelihood

estimate of the parameter 𝛼𝛼 which resides in the distance function (see, Equation (6.37)). This

Appendix summarizes the outcomes of the side analyses performed to further elucidate the

mechanism underlying Figure 8-7 and Figure 8-10. First, Section A.2 reports the results of the

two tests of ‘uniform cumulative probabilities’ (see, Section 8.2.4.1) and ‘VaR break

frequency’ (see, Section 8.2.4.2), conducted on the basis of all the 276 IPOIs, for the cases

where 𝛼𝛼 is set to values other than 1.5. Second, Section A.3 focuses on a small cluster depicted

in Figure 8-10, which appears to be an anomaly. For any subset that falls under this cluster, the

corresponding calibration outputs 0.5 as the maximum likelihood estimate of 𝛼𝛼. Section A.3

provides some visual aid to demonstrate the features of the log-likelihood functions associated

with this cluster.

A.2 Supplementary Discussion on Section 8.2.4: Scrutiny

of Estimated 𝐏𝐏𝐏𝐏𝐏𝐏𝟗𝟗𝟗𝟗𝐈𝐈𝐏𝐏𝐈𝐈𝐈𝐈’s under other Values of 𝜶𝜶

The calibration process, developed in Section 8.2.3, reported 1.5 as the maximum

likelihood estimate of 𝛼𝛼 when all the 276 IPOIs participate in the analysis. Further analysis

indicated that 1.5 can be presented as a property of the distance function and may be utilized

to skip the computationally intensive step of calibration; that is, the risk assessment of a given

IPO can be performed by assuming 𝛼𝛼 = 1.5, a priori.

Subsequently, Section 8.2.4 evaluated the standalone accuracy of the set of 276 PDF90IPOI’s

produced under 𝛼𝛼 = 1.5. In this section, the same process is iterated using the following values

Page 286: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Appendix A. Complementary Details on the Calibration of the Distance Function

A.2 Scrutiny of Estimated Probability Density Functions: Other Values of 𝛼𝛼

254

of 𝛼𝛼: [0, 0.5, 1.25, 1.9, ≥ 3]. It has already been established in Section 8.2.3 that the PDF90IPOI’s

generated using 𝛼𝛼 = 1.5 are typically more accurate compared to the ones obtained using the

above-indicated 𝛼𝛼 values. Here, however, the goal is to investigate whether the individual sets

of PDF90IPOI’s associated with these 𝛼𝛼 values would pass the tests of ‘uniform cumulative

probabilities’ and ‘VaR break frequency’.

Figure A-1 to Figure A-5, which are presented at the end of this subsection, visualize the

results of the test of uniform cumulative probability values (CPVs) for the individual 𝛼𝛼 values.

The corresponding outcomes of the K-S test are displayed in Table A-1. Recall that for a given

𝛼𝛼, the K-S test is conducted under the null hypothesis that ‘the observed set of 276 CPVs has

been drawn from a reference population with uniform density between 0 and 1’ (see, Section

8.2.4.1.1).

Table A-1. The table reports the individual K-S p-values computed under different assumptions of 𝛼𝛼. For a

given 𝛼𝛼, the null hypothesis can be stated as ‘the observed set of 276 CPVs has been drawn from a reference

population with uniform density between 0 and 1’. Results indicate that in the cases of 0, 0.5, 1.9, and ≥ 3,

there is no strong evidence in favour of the null hypothesis, which suggests that the respective observed

sample set would be considered a rare event if one assumed that the null hypothesis is valid. It can, therefore,

be concluded that the individual sets of PDF90IPOI’s produced using these 𝛼𝛼 values have failed to fulfil the

objectives of the test of uniform CPVs, which implies that no convincing evidence exists to support the

standalone accuracy of each of these sets.

𝜶𝜶 0 0.5 1.25 1.5 1.9 ≥ 3

p-Value of the K-S Test 0.0195 0.0269 0.3048 0.8893 0.0082 0.0018

Page 287: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Appendix A. Complementary Details on the Calibration of the Distance Function

A.2 Scrutiny of Estimated Probability Density Functions: Other Values of 𝛼𝛼

255

Through the visual inspection of the shapes of the histograms and drawing upon the p-

values reported in Table A-1, it is concluded that there is no significant evidence to accept the

null hypothesis in the cases of the individual sets of PDF90IPOI’s obtained using the following

values of 𝛼𝛼: [0, 0.5, 1.9, ≥ 3]. The null hypothesis cannot, however, be rejected for the set of

PDF90IPOI’s generated with the assumption of 𝛼𝛼 = 1.25. The values of 𝛼𝛼 can be sorted in the

ascending order of the respective p-values: [≥ 3, 1.9, 0, 0.5, 1.25]; only the last member of the

set (i.e., 𝛼𝛼 = 1.25) exhibits a p-value which provides sufficient confidence in the accuracy of

the corresponding set of PDF90IPOI’s. Yet, the p-value reported in the case of 𝛼𝛼 = 1.25 (i.e.,

0.3048) still falls inferior to the p-value obtained for 𝛼𝛼 = 1.5 (i.e., 0.8893).

The results of the test of VaR break frequency are reported in Table A-2. As presented, at

all the four confidence levels of 95%, 90%, 85%, 80%, the risk is ‘overestimated’ in the cases

Table A-2. Each cell presents the realized VaR break frequency under the corresponding 𝛼𝛼 and nominal

frequency of VaR breaks. The exhibited frequencies can be interpreted in a similar fashion to those displayed

in Table 8-1 (see, Section 8.2.4.2).

Realized

Frequency of

VaR Breaks

𝜶𝜶

0 0.5 1.25 1.5 1.9 ≥ 3

Nom

inal

Fre

quen

cy o

f

VaR

Bre

aks

0.05 0.0290 0.0326 0.0399 0.0580 0.0761 0.0978

0.1 0.0580 0.0616 0.0906 0.1087 0.1377 0.1594

0.15 0.0978 0.1014 0.1341 0.1486 0.2138 0.2391

0.2 0.1341 0.1377 0.1848 0.1993 0.2609 0.2826

Page 288: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Appendix A. Complementary Details on the Calibration of the Distance Function

A.2 Scrutiny of Estimated Probability Density Functions: Other Values of 𝛼𝛼

256

of 𝛼𝛼 = 0 and 0.05, and to a lesser degree, under 𝛼𝛼 = 1.25. For the higher values of 𝛼𝛼 (i.e., 1.9

and ≥ 3), the realized VaR break frequencies are the indicative of risk ‘underestimation’.

The implications of the outcomes of both tests can be recapitulated as follows: First, using

the set of all IPOIs, the highest accuracy was achieved under the 𝛼𝛼 estimate equal to 1.5. As

the value of 𝛼𝛼 deviates from 1.5, so does the standalone accuracy of the corresponding set of

estimated PDF90IPOI’s. In Section 8.2.3, some possible underlying mechanisms leading to such

results are discussed at length. Second, it can straightforwardly be verified that 𝛼𝛼 has the

capacity to influence the standalone accuracy of estimated PDF90IPOI’s, which attests to the

efficacy of the second term of the distance function (i.e., the 𝜓𝜓 dimension) in selecting the

closest comparable.

Page 289: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Appendix A. Complementary Details on the Calibration of the Distance Function

A.2 Scrutiny of Estimated Probability Density Functions: Other Values of 𝛼𝛼

257

Figure A-1. The histogram is built using 276 CPVs which are computed on the basis of the respective set of 276

PDF90IPOI’s, estimated under the assumption of 𝛼𝛼 = 0.

Figure A-2. The histogram is built using 276 CPVs which are computed on the basis of the respective set of 276

PDF90IPOI’s, estimated under the assumption of 𝛼𝛼 = 0.5.

Page 290: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Appendix A. Complementary Details on the Calibration of the Distance Function

A.2 Scrutiny of Estimated Probability Density Functions: Other Values of 𝛼𝛼

258

Figure A-3. The histogram is built using 276 CPVs which are computed on the basis of the respective set of 276

PDF90IPOI’s, estimated under the assumption of 𝛼𝛼 = 1.25.

Figure A-4. The histogram is built using 276 CPVs which are computed on the basis of the respective set of 276

PDF90IPOI’s, estimated under the assumption of 𝛼𝛼 = 1.9.

Page 291: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Appendix A. Complementary Details on the Calibration of the Distance Function

A.2 Scrutiny of Estimated Probability Density Functions: Other Values of 𝛼𝛼

259

Figure A-5. The histogram is built using 276 CPVs which are computed on the basis of the respective set of 276

PDF90IPOI’s, estimated under the assumption of 𝛼𝛼 ≥ 1.9. Recall from Section 8.2.3 that starting at the vicinity of

𝛼𝛼 = 3, the log-likelihood function stabilizes which indicates that the second term (i.e., the 𝛼𝛼 term) has dominated

the first term of the similarity metric in all the 276 IPOI cases. Therefore, the individual sets of PDF90IPOI’s

estimated from this point onward remain unchanged, which consequently translates into invariant sets of CPVs,

and hence, identical CPV histograms.

Page 292: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Appendix A. Complementary Details on the Calibration of the Distance Function

A.3 Behaviour of Log-Likelihood Functions of 𝛼𝛼

260

A.3 Behaviour of Log-Likelihood Functions of 𝜶𝜶

As indicated earlier, the focal intent of this section is to investigate the structure and

characteristics of the log-likelihood functions underlying a cluster of 9 consecutive points

depicted in Figure 8-10. The common feature of the points within this cluster is their

comparatively low maximum likelihood estimate of 𝛼𝛼 (i.e., 𝛼𝛼 = 0.5), which called for further

scrutiny; the first subset of the IPOs which exhibits this feature comprises the 52nd to 141st

IPOs. The behaviour of the log-likelihood functions associated with the points within this

cluster can be categorized into three different types:

First. Figure A-6 visualizes an example of the first type of behaviour. It depicts the log-

likelihood function obtained on the basis of the subset consisting of the 52nd to 141st IPOs.

Similar shapes were observed for the succeeding two subsets (i.e., 52nd IPO to 142nd IPO

and 52nd IPO to 143rd IPO). As illustrated in Figure A-6, 𝛼𝛼 = 0.5 represents the ‘argument’

(arg) of a local maximum which has won the position of global arg maximum from 1.8,

with only a marginal difference. Note that in Figure 8-10, the maximum likelihood estimate

of 𝛼𝛼 is reported as 1.8 immediately before plunging to the level of 0.5. Although no local

maximum occurs at 1.5, its likelihood is situated very close to the local maximums; in

Figure A-6, the difference between the log-likelihood values at 𝛼𝛼 = 0.5 and 𝛼𝛼 = 1.5 is

equal to 0.03, which translates into 1.07 on the linear scale.

Second. The sets of 52nd IPO to 144th IPO and 52nd IPO to 145th IPO fall under this class.

The log-likelihood function of the former set is presented in Figure A-7. Similar to the

‘first’ type, 𝛼𝛼 = 0.5 marks both the local and global maximums. Here, however, the second

local maximum occurs at 𝛼𝛼 = 1.5 whose likelihood value continues to remain close to that

of the global maximum.

Third. The rest of the subsets within this cluster belong to the third category. As a case in

point, consider Figure A-8 which depicts the log-likelihood function of the subset

Page 293: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Appendix A. Complementary Details on the Calibration of the Distance Function

A.3 Behaviour of Log-Likelihood Functions of 𝛼𝛼

261

comprising the 52nd to 146th IPOs. As demonstrated, 𝛼𝛼 = 0.5 and 𝛼𝛼 = 1.5 share the position

of maximum likelihood estimate of 𝛼𝛼.

To further elaborate on the microstructure of Figure 8-10, refer to Figure A-9. This figure

presents the log-likelihood function of the subset consisting of the 52nd to 150th IPOs. For this

subset, the maximum likelihood estimate of 𝛼𝛼 has reverted to 1.8. Although 𝛼𝛼 = 0.5 may still

be considered the argument of a local maximum, it is associated with a significantly lower log-

likelihood value compared to the new global maximum. Likewise, the gap between the

likelihood values at 𝛼𝛼 = 0.5 and 𝛼𝛼 = 1.5 has widened, and here, 1.5 is showing a noticeably

higher likelihood value. Note that the deviation of the likelihood of 𝛼𝛼 = 1.5 from the global

maximum continues to remain small.

In short, the core of all the aforementioned details can be stated as follows: When 0.5 is

selected as the maximum likelihood estimate of 𝛼𝛼, there exists another local maximum which

either rests at the same level as the likelihood of 𝛼𝛼 = 0.5, or is surpassed by it with a slight

margin. Therefore, although it first appeared to be an anomaly, the maximum likelihood

estimate of 0.5 is indeed consistent with the rest of the results exhibited by Figure 8-10.

Considering these observations in conjunction with the trend of the maximum likelihood

estimate of 𝛼𝛼, demonstrated in Figure 8-10, it can be concluded that the maximum at 𝛼𝛼 = 0.5

is unstable and will be dominated by the one at 1.5 once the subset size rises beyond a certain

threshold.

Page 294: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Appendix A. Complementary Details on the Calibration of the Distance Function

A.3 Behaviour of Log-Likelihood Functions of 𝛼𝛼

262

Figure A-6. The graph demonstrates the log-likelihood function of 𝛼𝛼 computed on the basis of the subset

consisting of the 52nd to 141st IPOs. This figure serves as an example of the ‘first type’ of behaviour described on

Page 260. Similar behaviours have been recorded for the two succeeding subsets (i.e., 52 to 142 and 52 to 143).

Figure A-7. The graph demonstrates the log-likelihood function of 𝛼𝛼 computed on the basis of the subset

consisting of the 52nd to 144th IPOs. This figure serves as an example of the ‘second type’ of behaviour described

on Page 260. A similar behaviour has been observed for the set of the 52nd to 145th IPOs.

Page 295: A HYBRID BAYESIAN AND DATA-ENVELOPMENT-ANALYSIS … · quantified as a value-at-risk (VaR) inferred from the probability density function of price on day 90 (i.e., PDF. 90IPOI). This

Appendix A. Complementary Details on the Calibration of the Distance Function

A.3 Behaviour of Log-Likelihood Functions of 𝛼𝛼

263

Figure A-8. The graph demonstrates the log-likelihood function of 𝛼𝛼 computed on the basis of the subset

consisting of the 52nd to 146th IPOs. This figure serves as an example of the ‘third type’ of behaviour described on

Page 260. Similar behaviours have been recorded for the three succeeding subsets (i.e., 52 to 147, 52 to 148, and

52 to 149).

Figure A-9. The graph exhibits the log-likelihood function of 𝛼𝛼 computed on the basis of the subset consisting of

the 52nd to 150th IPOs. More details regarding this plot are provided on Page 261.