Call Letter for Reliable Techno Designs

Embed Size (px)

Citation preview

  • 8/14/2019 Call Letter for Reliable Techno Designs

    1/31

    Hi,

    ----------------------------------------------------------------------------------------------------------------------

    To,

    Reliable Techno-Designs Pvt. Ltd.

    BG/SEI 11/03, MIDC Bhosari,

    Near BSNL Office, Pune - 411026Landline:-020-30684519, Mob:- 9922912177

    Sub: Interview of the short listed candidate for your requirement ofAsst.Manager/ Deputy Manager Press Tool Designg

    Dear Tushar Sir,

    We are pleased to send the candidate bearing this letter,

    Mr Mangesh C. Girgaonkar

    who is short-listed by us for the interview. His CV is attached herewith for yourready reference.

    We await your valued feedback at your earliest convenience.

    With Best Regards,

    Mrs.Swet

    ha.F.Hottinavar

    HR (Sr.Recruiter)

    For Mandar Sants PlacementServices

    Random variable

    MANDAR SANTSPLACEMENT SERVICES.

    VIJIGEESHA, 17, PRASHANT NAGAR,Behind KAKA HALWAI, L.B.S. Road,

    Pune - 411030CONTACT:- 020-24332433, 9922435001

  • 8/14/2019 Call Letter for Reliable Techno Designs

    2/31

    Broadly, there are two types of random variable discrete and continuous. Discreterandom variables take on one of a set of specific values, each with some probabilitygreater than zero. Continuous random variables can be realized with any of a range ofvalues (e.g., a real number between zero and one), and so there are several ranges (e.g. 0to one half) that have a probability greater than zero of occurring.

    A random variable has either an associated probability distribution (discrete randomvariable) or probability density function (continuous random variable).

    Random Variable

    The outcome of an experiment need not be a number, for example, the outcome

    when a coin is tossed can be 'heads' or 'tails'. However, we often want to

    represent outcomes as numbers. A random variable is a function that associates

    a unique numerical value with every outcome of an experiment. The value of the

    random variable will vary from trial to trial as the experiment is repeated.

    There are two types of random variable - discrete and continuous. A random

    variable has either an associated probability distribution (discrete random

    variable) or probability density function (continuous random variable).

    Examples--A coin is tossed ten times. The random variable X is the number of tails thatare noted. X can only take the values 0, 1, ..., 10, so X is a discrete random variable.

    1. A light bulb is burned until it burns out. The random variable Y is its

    lifetime in hours. Y can take any positive real value, so Y is a continuous

    random variable.

    Expected Value

    The expected value (or population mean) of a random variable indicates its

    average or central value. It is a useful summary value (a number) of the

    variable's distribution.

    Stating the expected value gives a general impression of the behaviour of somerandom variable without giving full details of its probability distribution (if it is

    discrete) or its probability density function (if it is continuous).

    Two random variables with the same expected value can have very different

    distributions. There are other useful descriptive measures which affect the shape

    of the distribution, for example variance.

    http://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#discvar#discvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#contvar#contvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#variance#variancehttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#discvar#discvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#contvar#contvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#variance#variance
  • 8/14/2019 Call Letter for Reliable Techno Designs

    3/31

    The expected value of a random variable X is symbolised by E(X) or .

    If X is a discrete random variable with possible values x1, x2, x3, ..., xn, and p(xi)

    denotes P(X = xi), then the expected value of X is defined by:

    where the elements are summed over all values of the random variable X.

    If X is a continuous random variable with probability density function f(x), then the

    expected value of X is defined by:

    Example

    Discrete case : When a die is thrown, each of the possible faces 1, 2, 3, 4, 5, 6

    (the xi's) has a probability of 1/6 (the p(xi)'s) of showing. The expected value of

    the face showing is therefore:

    = E(X) = (1 x 1/6) + (2 x 1/6) + (3 x 1/6) + (4 x 1/6) + (5 x 1/6) + (6 x 1/6)

    = 3.5

    Notice that, in this case, E(X) is 3.5, which is not a possible value of X.

    See also sample mean.

    Variance

    The (population) variance of a random variable is a non-negative number which

    gives an idea of how widely spread the values of the random variable are likely to

    be; the larger the variance, the more scattered the observations on average.

    Stating the variance gives an impression of how closely concentrated round the

    expected value the distribution is; it is a measure of the 'spread' of a distribution

    about its average value.

    Variance is symbolised by V(X) or Var(X) or

    The variance of the random variable X is defined to be:

    where E(X) is the expected value of the random variable X.

    http://www.stats.gla.ac.uk/steps/glossary/presenting_data.html#sampmeanhttp://www.stats.gla.ac.uk/steps/glossary/presenting_data.html#sampmeanhttp://www.stats.gla.ac.uk/steps/glossary/presenting_data.html#sampmean
  • 8/14/2019 Call Letter for Reliable Techno Designs

    4/31

    Notes

    a. the larger the variance, the further that individual values of the random

    variable (observations) tend to be from the mean, on average;

    b. the smaller the variance, the closer that individual values of the randomvariable (observations) tend to be to the mean, on average;

    c. taking the square root of the variance gives the standard deviation, i.e.:

    d. the variance and standard deviation of a random variable are always non-

    negative.

    See also sample variance.

    Probability Distribution

    The probability distribution of a discrete random variable is a list of probabilities

    associated with each of its possible values. It is also sometimes called the

    probability function or the probability mass function.

    More formally, the probability distribution of a discrete random variable X is a

    function which gives the probability p(xi) that the random variable equals xi, for

    each value xi:

    p(xi) = P(X=xi)

    It satisfies the following conditions:

    a.

    b.

    Cumulative Distribution Function

    http://www.stats.gla.ac.uk/steps/glossary/presenting_data.html#sampvarhttp://www.stats.gla.ac.uk/steps/glossary/presenting_data.html#sampvar
  • 8/14/2019 Call Letter for Reliable Techno Designs

    5/31

    All random variables (discrete and continuous) have a cumulative distribution

    function. It is a function giving the probability that the random variable X is less

    than or equal to x, for every value x.

    Formally, the cumulative distribution function F(x) is defined to be:

    for

    For a discrete random variable, the cumulative distribution function is found by

    summing up the probabilities as in the example below.

    For a continuous random variable, the cumulative distribution function is the

    integral of its probability density function.

    Example

    Discrete case : Suppose a random variable X has the following probability

    distribution p(xi):

    xi 0 1 2 3 4 5

    p(xi) 1/32 5/32 10/32 10/32 5/32 1/32

    This is actually a binomial distribution: Bi(5, 0.5) or B(5, 0.5). The cumulative

    distribution function F(x) is then:

    xi 0 1 2 3 4 5

    F(xi) 1/32 6/32 16/32 26/32 31/32 32/32

    F(x) does not change at intermediate values. For example:

    F(1.3) = F(1) = 6/32

    F(2.86) = F(2) = 16/32

    Probability Density Function

    The probability density function of a continuous random variable is a function

    which can be integrated to obtain the probability that the random variable takes a

    value in a given interval.

  • 8/14/2019 Call Letter for Reliable Techno Designs

    6/31

    More formally, the probability density function, f(x), of a continuous random

    variable X is the derivative of the cumulative distribution function F(x):

    Since it follows that:

    If f(x) is a probability density function then it must obey two conditions:

    a. that the total probability for all possible values of the continuous random

    variable X is 1:

    b. that the probability density function can never be negative: f(x) > 0 for all x.

    Discrete Random Variable

    A discrete random variable is one which may take on only a countable number of

    distinct values such as 0, 1, 2, 3, 4, ... Discrete random variables are usually (but

    not necessarily) counts. If a random variable can take only a finite number of

    distinct values, then it must be discrete. Examples of discrete random variables

    include the number of children in a family, the Friday night attendance at a

    cinema, the number of patients in a doctor's surgery, the number of defective

    light bulbs in a box of ten.

    Compare continuous random variable.

    Continuous Random Variable

    A continuous random variable is one which takes an infinite number of possiblevalues. Continuous random variables are usually measurements. Examples

    include height, weight, the amount of sugar in an orange, the time required to run

    a mile.

    http://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#contvar#contvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#contvar#contvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#contvar#contvar
  • 8/14/2019 Call Letter for Reliable Techno Designs

    7/31

    Compare discrete random variable.

    Independent Random Variables

    Two random variables X and Y say, are said to be independent if and only if thevalue of X has no influence on the value of Y and vice versa.

    The cumulative distribution functions of two independent random variables X and

    Y are related by

    F(x,y) = G(x).H(y)

    where

    G(x) and H(y) are the marginal distribution functions of X and Y for all

    pairs (x,y).

    Knowledge of the value of X does not effect the probability distribution of Y and

    vice versa. Thus there is no relationship between the values of independent

    random variables.

    For continuous independent random variables, their probability density functions

    are related by

    f(x,y) = g(x).h(y)

    where

    g(x) and h(y) are the marginal density functions of the random variables X

    and Y respectively, for all pairs (x,y).

    For discrete independent random variables, their probabilities are related by

    P(X = xi ; Y = yj) = P(X = xi).P(Y=yj)

    for each pair (xi,yj).

    Probability-Probability (P-P) Plot

    A probability-probability (P-P) plot is used to see if a given set of data follows

    some specified distribution. It should be approximately linear if the specifieddistribution is the correct model.

    The probability-probability (P-P) plot is constructed using the theoretical

    cumulative distribution function, F(x), of the specified model. The values in the

    http://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#discvar#discvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#cdf#cdfhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#probdistn#probdistnhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#contvar#contvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#discvar#discvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#cdf#cdfhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#discvar#discvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#cdf#cdfhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#probdistn#probdistnhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#contvar#contvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#discvar#discvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#cdf#cdf
  • 8/14/2019 Call Letter for Reliable Techno Designs

    8/31

    sample of data, in order from smallest to largest, are denoted x (1), x(2), ..., x(n). For

    i = 1, 2, ....., n, F(x(i)) is plotted against (i-0.5)/n.

    Compare quantile-quantile (Q-Q) plot.

    Quantile-Quantile (QQ) Plot

    A quantile-quantile (Q-Q) plot is used to see if a given set of data follows some

    specified distribution. It should be approximately linear if the specified distribution

    is the correct model.

    The quantile-quantile (Q-Q) plot is constructed using the theoretical cumulative

    distribution function, F(x), of the specified model. The values in the sample of

    data, in order from smallest to largest, are denoted x(1), x(2), ..., x(n). For i = 1, 2,

    ....., n, x(i) is plotted against F-1((i-0.5)/n).

    Compare probability-probability (P-P) plot.

    Normal Distribution

    Normal distributions model (some) continuous random variables. Strictly, a

    Normal random variable should be capable of assuming any value on the real

    line, though this requirement is often waived in practice. For example, height at a

    given age for a given gender in a given racial group is adequately described by a

    Normal random variable even though heights must be positive.

    A continuous random variable X, taking all real values in the range is said

    to follow a Normal distribution with parameters and if it has probability

    density function

    We write

    This probability density function (p.d.f.) is a symmetrical, bell-shaped curve,

    centred at its expected value . The variance is .

    http://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#qqplot#qqplothttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#cdf#cdfhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#cdf#cdfhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#ppplot#ppplothttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#contvar#contvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#pdf#pdfhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#qqplot#qqplothttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#cdf#cdfhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#cdf#cdfhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#ppplot#ppplothttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#contvar#contvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#pdf#pdf
  • 8/14/2019 Call Letter for Reliable Techno Designs

    9/31

    Many distributions arising in practice can be approximated by a Normal

    distribution. Other random variables may be transformed to normality.

    The simplest case of the normal distribution, known as the Standard Normal

    Distribution, has expected value zero and variance one. This is written as N(0,1).

    Examples

    Poisson Distribution

    Poisson distributions model (some) discrete random variables. Typically, a Poisson

    random variable is a count of the number of events that occur in a certain time interval orspatial area. For example, the number of cars passing a fixed point in a 5 minute interval,or the number of calls received by a switchboard during a given period of time.

    A discrete random variable X is said to follow a Poisson distribution with parameter m,written X ~ Po(m), if it has probability distribution

    where

    http://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#expval#expvalhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#variance#variancehttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#discvar#discvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#expval#expvalhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#variance#variancehttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#discvar#discvar
  • 8/14/2019 Call Letter for Reliable Techno Designs

    10/31

    x = 0, 1, 2, ..., n

    m > 0.

    The following requirements must be met:

    a. the length of the observation period is fixed in advance;

    b. the events occur at a constant average rate;

    c. the number of events occurring in disjoint intervals are statistically

    independent.

    The Poisson distribution has expected value E(X) = m and variance V(X) = m; i.e.

    E(X) = V(X) = m.

    The Poisson distribution can sometimes be used to approximate the Binomial

    distribution with parameters n and p. When the number of observations n is

    large, and the success probability p is small, the Bi(n,p) distribution approaches

    the Poisson distribution with the parameter given by m = np. This is useful since

    the computations involved in calculating binomial probabilities are greatly

    reduced.

    Examples

    Binomial Distribution

    Binomial distributions model (some) discrete random variables.

    http://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#indepevents#indepeventshttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#binodistn#binodistnhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#binodistn#binodistnhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#discvar#discvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#indepevents#indepeventshttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#binodistn#binodistnhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#binodistn#binodistnhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#discvar#discvar
  • 8/14/2019 Call Letter for Reliable Techno Designs

    11/31

    Typically, a binomial random variable is the number of successes in a series of

    trials, for example, the number of 'heads' occurring when a coin is tossed 50

    times.

    A discrete random variable X is said to follow a Binomial distribution withparameters n and p, written X ~ Bi(n,p) or X ~ B(n,p), if it has probability

    distribution

    where

    x = 0, 1, 2, ......., n

    n = 1, 2, 3, .......

    p = success probability; 0 < p < 1

    The trials must meet the following requirements:

    a. the total number of trials is fixed in advance;

    b. there are just two outcomes of each trial; success and failure;

    c. the outcomes of all the trials are statistically independent;

    d. all the trials have the same probability of success.

    The Binomial distribution has expected value E(X) = np andvariance V(X) =

    np(1-p).

    Examples

    http://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#outcome#outcomehttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#indepevents#indepeventshttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#expval#expvalhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#variance#variancehttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#variance#variancehttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#outcome#outcomehttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#indepevents#indepeventshttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#expval#expvalhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#variance#variance
  • 8/14/2019 Call Letter for Reliable Techno Designs

    12/31

    Geometric Distribution

    Geometric distributions model (some) discrete random variables. Typically, a

    Geometric random variable is the number of trials required to obtain the first

    failure, for example, the number of tosses of a coin untill the first 'tail' is obtained,

    or a process where components from a production line are tested, in turn, until

    the first defective item is found.

    A discrete random variable X is said to follow a Geometric distribution with

    parameter p, written X ~ Ge(p), if it has probability distribution

    P(X=x) = px-1

    (1-p)x

    where

    x = 1, 2, 3, ...

    p = success probability; 0 < p < 1

    The trials must meet the following requirements:

    a. the total number of trials is potentially infinite;

    b. there are just two outcomes of each trial; success and failure;

    c. the outcomes of all the trials are statistically independent;

    d. all the trials have the same probability of success.

    The Geometric distribution has expected value E(X)= 1/(1-p) and variance

    V(X)=p/{(1-p)2}.

    The Geometric distribution is related to the Binomial distribution in that both are

    based on independent trials in which the probability of success is constant and

    equal to p. However, a Geometric random variable is the number of trials until the

    first failure, whereas a Binomial random variable is the number of successes in n

    trials.

    http://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#discvar#discvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#outcome#outcomehttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#indepevents#indepeventshttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#expval#expvalhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#variance#variancehttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#binodistn#binodistnhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#discvar#discvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#outcome#outcomehttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#indepevents#indepeventshttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#expval#expvalhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#variance#variancehttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#binodistn#binodistn
  • 8/14/2019 Call Letter for Reliable Techno Designs

    13/31

    Examples

    Uniform Distribution

    Uniform distributions model (some) continuous random variables and (some)

    discrete random variables. The values of a uniform random variable are uniformly

    distributed over an interval. For example, if buses arrive at a given bus stop every

    15 minutes, and you arrive at the bus stop at a random time, the time you wait for

    the next bus to arrive could be described by a uniform distribution over the

    interval from 0 to 15.

    A discrete random variable X is said to follow a Uniform distribution with

    parameters a and b, written X ~ Un(a,b), if it has probability distribution

    P(X=x) = 1/(b-a)

    where

    x = 1, 2, 3, ......., n.

    A discrete uniform distribution has equal probability at each of its n values.

    A continuous random variable X is said to follow a Uniform distribution with

    parameters a and b, written X ~ Un(a,b), if its probability density function is

    constant within a finite interval [a,b], and zero outside this interval (with a less

    than or equal to b).

    The Uniform distribution has expected value E(X)=(a+b)/2 and variance {(b-

    a)2}/12.

    http://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#contvar#contvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#discvar#discvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#contvar#contvarhttp://www.stats.gla.ac.uk/steps/glossary/probability_distributions.html#discvar#discvar
  • 8/14/2019 Call Letter for Reliable Techno Designs

    14/31

    Example

    Central Limit Theorem

    The Central Limit Theorem states that whenever a random sample of size n is

    taken from anydistribution with mean and variance , then the sample mean

    will be approximatelynormally distributed with mean and variance /n. The

    larger the value of the sample size n, the better the approximation to the normal.

    This is very useful when it comes to inference. For example, it allows us (if the

    sample size is fairly large) to use hypothesis tests which assume normality even

    if our data appear non-normal. This is because the tests use the sample mean

    , which the Central Limit Theorem tells us will be approximately normally

    distributed.

    Measures of central tendancy: sample mean sample median modeDefinition: if our sample isx1, x2, . . . , xn then

    x= sample mean =x1 +x2 + .+xn/n

    Mean, Mode, Median, and Standard

    Deviation

  • 8/14/2019 Call Letter for Reliable Techno Designs

    15/31

  • 8/14/2019 Call Letter for Reliable Techno Designs

    16/31

    measure is the median. The median is the middle score. If we have an even number ofevents we take the average of the two middles. The median is better for describing thetypical value. It is often used for income and home prices.

    Example

    Suppose you randomly selected 10 house prices in the South Lake Tahoe area. Your areinterested in the typical house price. In $100,000 the prices were

    2.7, 2.9, 3.1, 3.4, 3.7, 4.1, 4.3, 4.7, 4.7, 40.8

    If we computed the mean, we would say that the average house price is 710,000.Although this number is true, it does not reflect the price for available housing in SouthLake Tahoe. A closer look at the data shows that the house valued at 40.8 x $100,000 =$4.08 million skews the data. Instead, we use the median. Since there is an even numberof outcomes, we take the average of the middle two

    3.7 + 4.1= 3.9

    2

    The median house price is $390,000. This better reflects what house shoppers shouldexpect to spend.

    There is an alternative value that also is resistant to outliers. This is called thetrimmed

    mean which is the mean after getting rid of the outliers or5% on the top and 5% on thebottom. We can also use the trimmed mean if we are concerned with outliers skewing thedata, however the median is used more often since more people understand it.

    Example:

    At a ski rental shop data was collected on the number of rentals on each of tenconsecutive Saturdays:

    44, 50, 38, 96, 42, 47, 40, 39, 46, 50.

    To find the sample mean, add them and divide by 10:

    44 + 50 + 38 + 96 + 42 + 47 + 40 + 39 + 46 + 50= 49.2

    10

  • 8/14/2019 Call Letter for Reliable Techno Designs

    17/31

    Notice that the mean value is not a value of the sample.

    To find the median, first sort the data:

    38, 39, 40, 42, 44, 46, 47, 50, 50, 96

    Notice that there are two middle numbers 44 and 46. To find the median we take theaverage of the two.

    44 + 46Median = = 45

    2

    Notice also that the mean is larger than all but three of the data points. The mean isinfluenced by outliers while the median is robust.

    Variance, Standard Deviation and Coefficient of Variation

    The mean, mode, median, and trimmed mean do a nice job in telling where the center ofthe data set is, but often we are interested in more. For example, a pharmaceuticalengineer develops a new drug that regulates iron in the blood. Suppose she finds out thatthe average sugar content after taking the medication is the optimal level. This does notmean that the drug is effective. There is a possibility that half of the patients havedangerously low sugar content while the other half have dangerously high content.Instead of the drug being an effective regulator, it is a deadly poison. What thepharmacist needs is a measure of how far the data is spread apart. This is what thevariance and standard deviation do. First we show the formulas for these measurements.Then we will go through the steps on how to use the formulas.

    We define the variance to be

    and thestandard deviation to be

    Variance and Standard Deviation: Step by Step

  • 8/14/2019 Call Letter for Reliable Techno Designs

    18/31

    1. Calculate the mean, x.2. Write a table that subtracts the mean from each observed value.3. Square each of the differences.4. Add this column.5. Divide by n -1 where n is the number of items in the sample This is the

    variance.6. To get thestandard deviationwe take the square root of the variance.

    Example

    The owner of the Ches Tahoe restaurant is interested in how much people spend at therestaurant. He examines 10 randomly selected receipts for parties of four and writesdown the following data.

    44, 50, 38, 96, 42, 47, 40, 39, 46, 50

    He calculated the mean by adding and dividing by 10 to get

    x = 49.2

    Below is the table for getting the standard deviation:

    x x - 49.2 (x - 49.2 )2

    44 -5.2 27.04

    50 0.8 0.64

    38 11.2 125.44

    96 46.8 2190.24

    42 -7.2 51.84

    47 -2.2 4.84

    40 -9.2 84.64

    39 -10.2 104.04

    46 -3.2 10.24

    50 0.8 0.64

    Total 2600.4

  • 8/14/2019 Call Letter for Reliable Techno Designs

    19/31

    Now

    2600.4= 288.7

    10 - 1

    Hence the variance is 289 and the standard deviation is the square root of 289 = 17.

    What this means is that most of the patrons probably spend between $32.20 and $66.20.

    The sample standard deviation will be denoted by s and thepopulation standard deviation

    will be denoted by the Greek letter.

    The sample variance will be denoted by s2 and the population variance will be denoted by

    2.

    The variance and standard deviation describe how spread out the data is. If the data alllies close to the mean, then the standard deviation will be small, while if the data isspread out over a large range of values, s will be large. Having outliers will increase thestandard deviation.

    One of the flaws involved with the standard deviation, is that it depends on the units thatare used. One way of handling this difficulty, is called the coefficient of variation whichis the standard deviation divided by the mean times 100%

    CV = 100%

    In the above example, it is

    17100% = 34.6%

    49.2

    This tells us that the standard deviation of the restaurant bills is 34.6% of the mean.

  • 8/14/2019 Call Letter for Reliable Techno Designs

    20/31

    arithmetic mean

    Definition

    Simple average,equalto the sum of all values divided by the number of values.

    The geometric mean, inmathematics, is a type ofmeanoraverage, which indicates thecentral tendency or typical value of a set of numbers. It is similar to the arithmetic mean,

    which is what most people think of with the word"average," except that instead of adding the set of numbersand then dividing the sum by the count of numbers in the

    set, n, the numbers are multiplied and then the nth root of the resultingproduct is taken.

    For instance, the geometric mean of two numbers, say 2 and 8, is just the square root (i.e.,the second root) of their product, 16, which is 4. As another example, the geometric meanof1, , and is the cube root (i.e., the third root) of their product (0.125), which is .

    teps involved in constructing a frequency

    polygon

    The most common method of graphingresearch data is to construct a frequencypolygon. The first step in constructing afrequency polygon is to list all scores and totabulate how many subjects received eachscore. Steps involved in constructing afrequency polygon are: 1) list all scores andtabulate how many subjects received each score,

    2) place all the scores on a horizontal axis, atequal intervals from lowest score to the highest,3) place the frequencies of scores at equalintervals on the vertical axis, starting with zero,4) for each score, find the point where the scoreintersects with its frequency of occurrence andmake a dot, 5) connect all the dots with straightlines.

    In fact, along with the development ofcomputer technology, there are several graphicdesigning software packages that can generalize

    the frequency polygon easily, such asSpreadsheet in ClarisWorks and Excel inMicrosoft office.Three measures of central tendency

    Measures of central tendency give theresearcher a convenient way of describing a setof data with a single number. Three mostfrequently encountered indices of central

    Geometric mean

    http://www.investorwords.com/347/average.htmlhttp://www.investorwords.com/347/average.htmlhttp://www.businessdictionary.com/definition/equal.htmlhttp://www.businessdictionary.com/definition/equal.htmlhttp://www.investorwords.com/5209/value.htmlhttp://en.wikipedia.org/wiki/Mathematicshttp://en.wikipedia.org/wiki/Mathematicshttp://en.wikipedia.org/wiki/Mathematicshttp://en.wikipedia.org/wiki/Meanhttp://en.wikipedia.org/wiki/Meanhttp://en.wikipedia.org/wiki/Averagehttp://en.wikipedia.org/wiki/Arithmetic_meanhttp://en.wikipedia.org/wiki/Radical_(mathematics)http://en.wikipedia.org/wiki/Radical_(mathematics)http://en.wikipedia.org/wiki/Radical_(mathematics)http://en.wikipedia.org/wiki/Product_(mathematics)http://www.investorwords.com/347/average.htmlhttp://www.businessdictionary.com/definition/equal.htmlhttp://www.investorwords.com/5209/value.htmlhttp://en.wikipedia.org/wiki/Mathematicshttp://en.wikipedia.org/wiki/Meanhttp://en.wikipedia.org/wiki/Averagehttp://en.wikipedia.org/wiki/Arithmetic_meanhttp://en.wikipedia.org/wiki/Radical_(mathematics)http://en.wikipedia.org/wiki/Product_(mathematics)
  • 8/14/2019 Call Letter for Reliable Techno Designs

    21/31

    tendency are the mode, the median, and themean. The mode is appropriate for themeasurement of nominal data, the median for

    the ordinal data, and the mean for the interval

    or ratio data.

    Quartile deviation: In "research talk", thequartile deviation is one-half of the differencebetween the upper quartile and the lowerquartile in a distribution. In English, the upperquartile is the 75th percentile; it means there are75% scores below than that point. Bysubtracting the lower quartile from the upperquartile and then dividing the result by two, weget a measure of variability. If the quartiledeviation is small the scores are close together,whereas if the quartile deviation is large the

    scores are more spread out. The quartiledeviation is a more stable measure of variabilitythan the range and is appropriate whenever themedian is appropriate. Standard Deviation: The standarddeviation is the square root of the variance,which is based on the distance of each scorefrom the mean. It is appropriate when the datarepresent an interval or ratio scale. It is themost stable measure of variability and takes intoaccount each and every score. Measuringstandard deviation is to find out how far eachscore is from the mean, that is, subtracting themean from each score. Steps for calculating thestandard deviation are: 1) Find out N, thenumber of subjects, 2) Calculate the sum of thescores, 3) square each score, 3) Add all thesquares, to get the sum of squares of the scores,4) Square the sum of the scores and divide bythe number of scores (we have a measure ofvariability called variance), 5) Subtract thevariance from the sum of the squares of scoresto get the sum of the squares (SS), 6) divide theSS by N-1. A small standard deviationindicates the scores are close together and alarge standard deviation indicates that the scoresare more spread out.

  • 8/14/2019 Call Letter for Reliable Techno Designs

    22/31

  • 8/14/2019 Call Letter for Reliable Techno Designs

    23/31

    0 - 9 1

    10 - 19 2

    20 - 29 3

    30 - 39 4

    40 - 49 5

    50 - 59 4

    60 - 69 3

    70 - 79 2

    80 - 89 2

    90 - 99 1

    To construct the histogram, groups are plotted on thex axis and their frequencies on theyaxis. The following is a histogram of the data in the above frequency table.

    Histogram

    Information Conveyed by Histograms

    Histograms are useful data summaries that convey the following information:

    The general shape of the frequency distribution (normal, chi-square, etc.)

    Symmetry of the distribution and whether it is skewed

    Modality - unimodal, bimodal, or multimodal

    The histogram of the frequency distribution can be converted to a probability distributionby dividing the tally in each group by the total number of data points to give the relativefrequency.

    The shape of the distribution conveys important information such as the probabilitydistribution of the data. In cases in which the distribution is known, a histogram that does

  • 8/14/2019 Call Letter for Reliable Techno Designs

    24/31

    not fit the distribution may provide clues about a process and measurement problem. Forexample, a histogram that shows a higher than normal frequency in bins near one end andthen a sharp drop-off may indicate that the observer is "helping" the results by classifyingextreme data in the less extreme group.

    Yourcontinued donations keep Wikipedia running!

    From Wikipedia, the free encyclopedia

    Jump to: navigation, search

    Probability theory is the branch ofmathematics concerned with analysis ofrandomphenomena.[1] The central objects of probability theory arerandom variables,stochasticprocesses, andevents: mathematical abstractions ofnon-deterministic events or measuredquantities that may either be single occurrences or evolve over time in an apparentlyrandom fashion. Although an individual coin toss or the roll of adieis a random event, ifrepeated many times the sequence of random events will exhibit certain statistical

    patterns, which can be studied and predicted. Two representative mathematical resultsdescribing such patterns are the law of large numbers and the central limit theorem.

    As a mathematical foundation forstatistics, probability theory is essential to many humanactivities that involve quantitative analysis of large sets of data. Methods of probabilitytheory also apply to description of complex systems given only partial knowledge of theirstate, as in statistical mechanics. A great discovery of twentieth centuryphysics was theprobabilistic nature of physical phenomena at atomic scales, described inquantummechanics.

    Probability theory

    http://wikimediafoundation.org/wiki/Fundraising?source=enwiki_00http://en.wikipedia.org/wiki/Probability_theory#column-onehttp://en.wikipedia.org/wiki/Probability_theory#searchInputhttp://en.wikipedia.org/wiki/Mathematicshttp://en.wikipedia.org/wiki/Mathematicshttp://en.wikipedia.org/wiki/Statistical_randomnesshttp://en.wikipedia.org/wiki/Probability_theory#cite_note-0http://en.wikipedia.org/wiki/Random_variablehttp://en.wikipedia.org/wiki/Random_variablehttp://en.wikipedia.org/wiki/Stochastic_processhttp://en.wikipedia.org/wiki/Stochastic_processhttp://en.wikipedia.org/wiki/Stochastic_processhttp://en.wikipedia.org/wiki/Event_(probability_theory)http://en.wikipedia.org/wiki/Event_(probability_theory)http://en.wikipedia.org/wiki/Determinismhttp://en.wikipedia.org/wiki/Dicehttp://en.wikipedia.org/wiki/Dicehttp://en.wikipedia.org/wiki/Dicehttp://en.wikipedia.org/wiki/Law_of_large_numbershttp://en.wikipedia.org/wiki/Central_limit_theoremhttp://en.wikipedia.org/wiki/Central_limit_theoremhttp://en.wikipedia.org/wiki/Statisticshttp://en.wikipedia.org/wiki/Statistical_mechanicshttp://en.wikipedia.org/wiki/Statistical_mechanicshttp://en.wikipedia.org/wiki/Physicshttp://en.wikipedia.org/wiki/Quantum_mechanicshttp://en.wikipedia.org/wiki/Quantum_mechanicshttp://en.wikipedia.org/wiki/Quantum_mechanicshttp://wikimediafoundation.org/wiki/Fundraising?source=enwiki_00http://en.wikipedia.org/wiki/Probability_theory#column-onehttp://en.wikipedia.org/wiki/Probability_theory#searchInputhttp://en.wikipedia.org/wiki/Mathematicshttp://en.wikipedia.org/wiki/Statistical_randomnesshttp://en.wikipedia.org/wiki/Probability_theory#cite_note-0http://en.wikipedia.org/wiki/Random_variablehttp://en.wikipedia.org/wiki/Stochastic_processhttp://en.wikipedia.org/wiki/Stochastic_processhttp://en.wikipedia.org/wiki/Event_(probability_theory)http://en.wikipedia.org/wiki/Determinismhttp://en.wikipedia.org/wiki/Dicehttp://en.wikipedia.org/wiki/Law_of_large_numbershttp://en.wikipedia.org/wiki/Central_limit_theoremhttp://en.wikipedia.org/wiki/Statisticshttp://en.wikipedia.org/wiki/Statistical_mechanicshttp://en.wikipedia.org/wiki/Physicshttp://en.wikipedia.org/wiki/Quantum_mechanicshttp://en.wikipedia.org/wiki/Quantum_mechanics
  • 8/14/2019 Call Letter for Reliable Techno Designs

    25/31

    Contents

    [hide]

    1 History

    2 Treatment

    o

    2.1 Discrete probability distributionso 2.2 Continuous probability distributions

    o 2.3 Measure-theoretic probability theory

    3 Probability distributions

    4 Convergence of random variables

    5 Law of large numbers

    6 Central limit theorem

    7 See also

    8 References

    9 Bibliography

    [edit] History

    The mathematical theory ofprobability has its roots in attempts to analyse games ofchanceby Gerolamo Cardano in the sixteenth century, and by Pierre de Fermat andBlaise Pascal in the seventeenth century (for example the "problem of points").

    Initially, probability theory mainly considered discrete events, and its methods weremainlycombinatorial. Eventually, analytical considerations compelled the incorporationofcontinuous variables into the theory. This culminated in modern probability theory,the foundations of which were laid by Andrey Nikolaevich Kolmogorov. Kolmogorovcombined the notion ofsample space, introduced byRichard von Mises, andmeasure

    theory and presented his axiom system for probability theory in 1933. Fairly quickly thisbecame the undisputed axiomatic basisfor modern probability theory.[2]

    [edit] Treatment

    Most introductions to probability theory treat discrete probability distributions andcontinuous probability distributions separately. The more mathematically advancedmeasure theory based treatment of probability covers both the discrete, the continuous,any mix of these two and more.

    [edit] Discrete probability distributions

    Main article:Discrete probability distribution

    Discrete probability theory deals with events that occur in countable sample spaces.

    Examples: Throwing dice, experiments with decks of cards, and random walk.

    http://toggletoc%28%29/http://en.wikipedia.org/wiki/Probability_theory#Historyhttp://en.wikipedia.org/wiki/Probability_theory#Treatmenthttp://en.wikipedia.org/wiki/Probability_theory#Discrete_probability_distributionshttp://en.wikipedia.org/wiki/Probability_theory#Continuous_probability_distributionshttp://en.wikipedia.org/wiki/Probability_theory#Measure-theoretic_probability_theoryhttp://en.wikipedia.org/wiki/Probability_theory#Probability_distributionshttp://en.wikipedia.org/wiki/Probability_theory#Convergence_of_random_variableshttp://en.wikipedia.org/wiki/Probability_theory#Law_of_large_numbershttp://en.wikipedia.org/wiki/Probability_theory#Central_limit_theoremhttp://en.wikipedia.org/wiki/Probability_theory#See_alsohttp://en.wikipedia.org/wiki/Probability_theory#Referenceshttp://en.wikipedia.org/wiki/Probability_theory#Bibliographyhttp://en.wikipedia.org/w/index.php?title=Probability_theory&action=edit&section=1http://en.wikipedia.org/wiki/Probabilityhttp://en.wikipedia.org/wiki/Game_of_chancehttp://en.wikipedia.org/wiki/Game_of_chancehttp://en.wikipedia.org/wiki/Game_of_chancehttp://en.wikipedia.org/wiki/Gerolamo_Cardanohttp://en.wikipedia.org/wiki/Pierre_de_Fermathttp://en.wikipedia.org/wiki/Blaise_Pascalhttp://en.wikipedia.org/wiki/Problem_of_pointshttp://en.wikipedia.org/wiki/Combinatoricshttp://en.wikipedia.org/wiki/Combinatoricshttp://en.wikipedia.org/wiki/Mathematical_analysishttp://en.wikipedia.org/wiki/Andrey_Nikolaevich_Kolmogorovhttp://en.wikipedia.org/wiki/Sample_spacehttp://en.wikipedia.org/wiki/Sample_spacehttp://en.wikipedia.org/wiki/Richard_von_Miseshttp://en.wikipedia.org/wiki/Richard_von_Miseshttp://en.wikipedia.org/wiki/Richard_von_Miseshttp://en.wikipedia.org/wiki/Measure_theoryhttp://en.wikipedia.org/wiki/Measure_theoryhttp://en.wikipedia.org/wiki/Measure_theoryhttp://en.wikipedia.org/wiki/Kolmogorov_axiomshttp://en.wikipedia.org/wiki/Axiom_systemhttp://en.wikipedia.org/wiki/Axiom_systemhttp://en.wikipedia.org/wiki/Probability_theory#cite_note-1http://en.wikipedia.org/wiki/Probability_theory#cite_note-1http://en.wikipedia.org/w/index.php?title=Probability_theory&action=edit&section=2http://en.wikipedia.org/w/index.php?title=Probability_theory&action=edit&section=3http://en.wikipedia.org/wiki/Discrete_probability_distributionhttp://en.wikipedia.org/wiki/Countablehttp://en.wikipedia.org/wiki/Dicehttp://en.wikipedia.org/wiki/Deck_of_cardshttp://en.wikipedia.org/wiki/Random_walkhttp://toggletoc%28%29/http://en.wikipedia.org/wiki/Probability_theory#Historyhttp://en.wikipedia.org/wiki/Probability_theory#Treatmenthttp://en.wikipedia.org/wiki/Probability_theory#Discrete_probability_distributionshttp://en.wikipedia.org/wiki/Probability_theory#Continuous_probability_distributionshttp://en.wikipedia.org/wiki/Probability_theory#Measure-theoretic_probability_theoryhttp://en.wikipedia.org/wiki/Probability_theory#Probability_distributionshttp://en.wikipedia.org/wiki/Probability_theory#Convergence_of_random_variableshttp://en.wikipedia.org/wiki/Probability_theory#Law_of_large_numbershttp://en.wikipedia.org/wiki/Probability_theory#Central_limit_theoremhttp://en.wikipedia.org/wiki/Probability_theory#See_alsohttp://en.wikipedia.org/wiki/Probability_theory#Referenceshttp://en.wikipedia.org/wiki/Probability_theory#Bibliographyhttp://en.wikipedia.org/w/index.php?title=Probability_theory&action=edit&section=1http://en.wikipedia.org/wiki/Probabilityhttp://en.wikipedia.org/wiki/Game_of_chancehttp://en.wikipedia.org/wiki/Game_of_chancehttp://en.wikipedia.org/wiki/Gerolamo_Cardanohttp://en.wikipedia.org/wiki/Pierre_de_Fermathttp://en.wikipedia.org/wiki/Blaise_Pascalhttp://en.wikipedia.org/wiki/Problem_of_pointshttp://en.wikipedia.org/wiki/Combinatoricshttp://en.wikipedia.org/wiki/Mathematical_analysishttp://en.wikipedia.org/wiki/Andrey_Nikolaevich_Kolmogorovhttp://en.wikipedia.org/wiki/Sample_spacehttp://en.wikipedia.org/wiki/Richard_von_Miseshttp://en.wikipedia.org/wiki/Measure_theoryhttp://en.wikipedia.org/wiki/Measure_theoryhttp://en.wikipedia.org/wiki/Kolmogorov_axiomshttp://en.wikipedia.org/wiki/Axiom_systemhttp://en.wikipedia.org/wiki/Probability_theory#cite_note-1http://en.wikipedia.org/w/index.php?title=Probability_theory&action=edit&section=2http://en.wikipedia.org/w/index.php?title=Probability_theory&action=edit&section=3http://en.wikipedia.org/wiki/Discrete_probability_distributionhttp://en.wikipedia.org/wiki/Countablehttp://en.wikipedia.org/wiki/Dicehttp://en.wikipedia.org/wiki/Deck_of_cardshttp://en.wikipedia.org/wiki/Random_walk
  • 8/14/2019 Call Letter for Reliable Techno Designs

    26/31

    Classical definition: Initially the probability of an event to occur was defined as numberof cases favorable for the event, over the number of total outcomes possible in anequiprobable sample space.

    For example, if the event is "occurrence of an even number when a die is rolled", the

    probability is given by , since 3 faces out of the 6 have even numbers and eachface has the same probability of appearing.

    Modern definition: The modern definition starts with aset called the sample space,which relates to the set of allpossible outcomes in classical sense, denoted by

    . It is then assumed that for each element , an intrinsic

    "probability" value is attached, which satisfies the following properties:

    1.

    2.

    That is, the probability functionf(x) lies between zero and one for every value ofx in thesample space , and the sum off(x) over all valuesx in the sample space is exactlyequal to 1. An event is defined as any subset of the sample space . The probabilityof the event defined as

    So, the probability of the entire sample space is 1, and the probability of the null event is0.

    The function mapping a point in the sample space to the "probability" value iscalled a probability mass function abbreviated as pmf. The modern definition does nottry to answer how probability mass functions are obtained; instead it builds a theory thatassumes their existence.

    [edit] Continuous probability distributions

    Main article: Continuous probability distribution

    Continuous probability theory deals with events that occur in a continuous samplespace.

    Classical definition: The classical definition breaks down when confronted with thecontinuous case. See Bertrand's paradox.

    Modern definition: If the sample space is the real numbers ( ), then a function calledthe cumulative distribution function (orcdf) is assumed to exist, which gives

    http://en.wikipedia.org/wiki/Dicehttp://en.wikipedia.org/wiki/Sethttp://en.wikipedia.org/wiki/Sethttp://en.wikipedia.org/wiki/Sample_spacehttp://en.wikipedia.org/wiki/Element_(mathematics)http://en.wikipedia.org/wiki/Event_(probability_theory)http://en.wikipedia.org/wiki/Subsethttp://en.wikipedia.org/wiki/Probability_mass_functionhttp://en.wikipedia.org/w/index.php?title=Probability_theory&action=edit&section=4http://en.wikipedia.org/wiki/Continuous_probability_distributionhttp://en.wikipedia.org/wiki/Bertrand's_paradox_(probability)http://en.wikipedia.org/wiki/Real_numbershttp://en.wikipedia.org/wiki/Cumulative_distribution_functionhttp://en.wikipedia.org/wiki/Dicehttp://en.wikipedia.org/wiki/Sethttp://en.wikipedia.org/wiki/Sample_spacehttp://en.wikipedia.org/wiki/Element_(mathematics)http://en.wikipedia.org/wiki/Event_(probability_theory)http://en.wikipedia.org/wiki/Subsethttp://en.wikipedia.org/wiki/Probability_mass_functionhttp://en.wikipedia.org/w/index.php?title=Probability_theory&action=edit&section=4http://en.wikipedia.org/wiki/Continuous_probability_distributionhttp://en.wikipedia.org/wiki/Bertrand's_paradox_(probability)http://en.wikipedia.org/wiki/Real_numbershttp://en.wikipedia.org/wiki/Cumulative_distribution_function
  • 8/14/2019 Call Letter for Reliable Techno Designs

    27/31

    for arandom variableX. That is,F(x) returns the probability thatXwill be less than or equal tox.

    The cdf must satisfy the following properties.

    1. is amonotonically non-decreasing,right-continuous function;

    2.

    3.

    If is differentiable, then the random variableXis said to have a probability density

    function orpdfor simply density

    For a set , the probability of the random variableXbeing in is defined as

    In case the probability density function exists, this can be written as

    Whereas thepdfexists only for continuous random variables, the cdfexists for all randomvariables (including discrete random variables) that take values on

    These concepts can be generalized formultidimensional cases on and othercontinuous sample spaces.

    [edit] Measure-theoretic probability theory

    The raison d'tre of the measure-theoretic treatment of probability is that it unifies thediscrete and the continuous, and makes the difference a question of which measure isused. Furthermore, it covers distributions that are neither discrete nor continuous normixtures of the two.

    An example of such distributions could be a mix of discrete and continuous distributions,for example, a random variable which is 0 with probability 1/2, and takes a value fromrandom normal distribution with probability 1/2. It can still be studied to some extent by

    considering it to have a pdf of , where [x] is the Kronecker deltafunction.

    http://en.wikipedia.org/wiki/Random_variablehttp://en.wikipedia.org/wiki/Random_variablehttp://en.wikipedia.org/wiki/Monotonic_functionhttp://en.wikipedia.org/wiki/Monotonic_functionhttp://en.wikipedia.org/wiki/Monotonic_functionhttp://en.wikipedia.org/wiki/Right-continuoushttp://en.wikipedia.org/wiki/Differentiablehttp://en.wikipedia.org/wiki/Probability_density_functionhttp://en.wikipedia.org/wiki/Probability_density_functionhttp://en.wikipedia.org/wiki/Dimensionhttp://en.wikipedia.org/w/index.php?title=Probability_theory&action=edit&section=5http://en.wikipedia.org/wiki/Kronecker_deltahttp://en.wikipedia.org/wiki/Random_variablehttp://en.wikipedia.org/wiki/Monotonic_functionhttp://en.wikipedia.org/wiki/Right-continuoushttp://en.wikipedia.org/wiki/Differentiablehttp://en.wikipedia.org/wiki/Probability_density_functionhttp://en.wikipedia.org/wiki/Probability_density_functionhttp://en.wikipedia.org/wiki/Dimensionhttp://en.wikipedia.org/w/index.php?title=Probability_theory&action=edit&section=5http://en.wikipedia.org/wiki/Kronecker_delta
  • 8/14/2019 Call Letter for Reliable Techno Designs

    28/31

    Other distributions may not even be a mix, for example, theCantor distribution has nopositive probability for any single point, neither does it have a density. The modernapproach to probability theory solves these problems using measure theory to define theprobability space:

    Given any set , (also called sample space) and a-algebra on it, a measurePiscalled a probability measure if

    1. is non-negative;

    2.

    If is a Borel -algebra then there is a unique probability measure on for any cdf, andvice versa. The measure corresponding to a cdf is said to be induced by the cdf. Thismeasure coincides with the pmf for discrete variables, and pdf for continuous variables,making the measure-theoretic approach free of fallacies.

    Theprobability of a set in the -algebra is defined as

    where the integration is with respect to the measure induced by

    Correlation Coefficient

    A correlation coefficient is a number between -1 and 1 which measures the

    degree to which two variables are linearly related. If there is perfect linear

    relationship with positive slope between the two variables, we have a correlation

    coefficient of 1; if there is positive correlation, whenever one variable has a high

    (low) value, so does the other. If there is a perfect linear relationship with

    negative slope between the two variables, we have a correlation coefficient of -1;

    if there is negative correlation, whenever one variable has a high (low) value, the

    other has a low (high) value. A correlation coefficient of 0 means that there is no

    linear relationship between the variables.

    There are a number of different correlation coefficients that might be appropriate

    depending on the kinds of variables being studied.

    See also Pearson's Product Moment Correlation Coefficient.

    See also Spearman Rank Correlation Coefficient.

    http://en.wikipedia.org/wiki/Cantor_distributionhttp://en.wikipedia.org/wiki/Cantor_distributionhttp://en.wikipedia.org/wiki/Measure_theoryhttp://en.wikipedia.org/wiki/Probability_spacehttp://en.wikipedia.org/wiki/Sigma-algebrahttp://en.wikipedia.org/wiki/Sigma-algebrahttp://en.wikipedia.org/wiki/Sigma-algebrahttp://en.wikipedia.org/wiki/Measure_(mathematics)http://en.wikipedia.org/wiki/Borel_algebrahttp://www.stats.gla.ac.uk/steps/glossary/paired_data.html#ppmcorrcoeffhttp://www.stats.gla.ac.uk/steps/glossary/paired_data.html#srcorrcoeffhttp://en.wikipedia.org/wiki/Cantor_distributionhttp://en.wikipedia.org/wiki/Measure_theoryhttp://en.wikipedia.org/wiki/Probability_spacehttp://en.wikipedia.org/wiki/Sigma-algebrahttp://en.wikipedia.org/wiki/Measure_(mathematics)http://en.wikipedia.org/wiki/Borel_algebrahttp://www.stats.gla.ac.uk/steps/glossary/paired_data.html#ppmcorrcoeffhttp://www.stats.gla.ac.uk/steps/glossary/paired_data.html#srcorrcoeff
  • 8/14/2019 Call Letter for Reliable Techno Designs

    29/31

    Pearson's Product Moment Correlation Coefficient

    Pearson's product moment correlation coefficient, usually denoted by r, is one

    example of a correlation coefficient. It is a measure of the linear association

    between two variables that have been measured on interval or ratio scales, such

    as the relationship between height in inches and weight in pounds. However, it

    can be misleadingly small when there is a relationship between the variables but

    it is a non-linear one.

    There are procedures, based on r, for making inferences about the population

    correlation coefficient. However, these make the implicit assumption that the two

    variables are jointly normally distributed. When this assumption is not justified, a

    non-parametric measure such as the Spearman Rank Correlation Coefficient

    might be more appropriate.

    See also correlation coefficient.

    Spearman Rank Correlation Coefficient

    The Spearman rank correlation coefficient is one example of a correlation

    coefficient. It is usually calculated on occasions when it is not convenient,

    economic, or even possible to give actual values to variables, but only to assign a

    rank order to instances of each variable. It may also be a better indicator that a

    relationship exists between two variables when the relationship is non-linear.

    Commonly used procedures, based on the Pearson's Product Moment

    Correlation Coefficient, for making inferences about the population correlation

    coefficient make the implicit assumption that the two variables are jointly normally

    distributed. When this assumption is not justified, a non-parametric measure such

    as the Spearman Rank Correlation Coefficient might be more appropriate.

    See also correlation coefficient.

    http://www.stats.gla.ac.uk/steps/glossary/paired_data.html#srcorrcoeffhttp://www.stats.gla.ac.uk/steps/glossary/paired_data.html#corrcoeffhttp://www.stats.gla.ac.uk/steps/glossary/paired_data.html#ppmcorrcoeffhttp://www.stats.gla.ac.uk/steps/glossary/paired_data.html#ppmcorrcoeffhttp://www.stats.gla.ac.uk/steps/glossary/paired_data.html#corrcoeffhttp://www.stats.gla.ac.uk/steps/glossary/paired_data.html#srcorrcoeffhttp://www.stats.gla.ac.uk/steps/glossary/paired_data.html#corrcoeffhttp://www.stats.gla.ac.uk/steps/glossary/paired_data.html#ppmcorrcoeffhttp://www.stats.gla.ac.uk/steps/glossary/paired_data.html#ppmcorrcoeffhttp://www.stats.gla.ac.uk/steps/glossary/paired_data.html#corrcoeff
  • 8/14/2019 Call Letter for Reliable Techno Designs

    30/31

    Regression Equation

    A regression equation allows us to express the relationship between two (or

    more) variables algebraically. It indicates the nature of the relationship between

    two (or more) variables. In particular, it indicates the extent to which you can

    predict some variables by knowing others, or the extent to which some are

    associated with others.

    A linear regression equation is usually written

    Y = a + bX + e

    where

    Y is the dependent variable

    a is the intercept

    b is the slope or regression coefficient

    X is the independent variable (or covariate)

    e is the error term

    The equation will specify the average magnitude of the expected change in Y

    given a change in X.

    The regression equation is often represented on a scatterplot by a regression

    line.

    Regression Line

    A regression line is a line drawn through the points on a scatterplot to summarisethe relationship between the variables being studied. When it slopes down (from

    top left to bottom right), this indicates a negative or inverse relationship between

    the variables; when it slopes up (from bottom right to top left), a positive or direct

    relationship is indicated.

    The regression line often represents the regression equation on a scatterplot.

    http://www.stats.gla.ac.uk/steps/glossary/paired_data.html#reglinehttp://www.stats.gla.ac.uk/steps/glossary/paired_data.html#reglinehttp://www.stats.gla.ac.uk/steps/glossary/paired_data.html#reglinehttp://www.stats.gla.ac.uk/steps/glossary/paired_data.html#regline
  • 8/14/2019 Call Letter for Reliable Techno Designs

    31/31

    Simple Linear Regression

    Simple linear regression aims to find a linear relationship between a response

    variable and a possible predictor variable by the method of least squares.

    Multiple Regression

    Multiple linear regression aims is to find a linear relationship between a response

    variable and several possible predictor variables.