Skip to main content

PROBABILITY

                                        PROBABILITY


                         INTRODUCTION
                                   
                              The word probability has two meaning:(1) a quantitative measure of uncertainty and (2) a measure of degree of belief in a particular or problem.

                               Probability and statistic are fundamentally interrelated. probability is often called the vehicle of statistics. the area of inferential statistics in which we are mainly concerned with drawing inferences from experiments or situations involving an element of uncertainty, leans heavily upon probability theory. Uncertainty is also an inherent part of statistics inference as inferences are based on a sample, and a sample being a small part of larger population, contains incomplete information. A similar type of uncertainty occurs when we toss a coin, draw a card or throw dice,etc. The uncertainty in all these cases is measured in terms of probability.

                             It is always clear what we mean when we make statement of the type that it is very likely to rain today or I have a fair chance of passing the annual examination or A will probably win a prize, etc. In each of these statements, the natural state of uncertainty is expressed, but on the basis of part evidence, we have some degree of personal belief in the truth of each statement.
                           
                             The foundations of probability were laid by french mathematician of the seventeenth century - Blaise pascal (1623-1662) and Pierre De Fermat (1601-1665) in connection with gambling problems. Later on it was developed by Jakob Bernoulli (1654-1605), Abraham De Moire (1667-1754) and Pierre Simon Laplace (1749-1827). The modern treatment of probability theory which consists of stating a few axioms and rules resulting from these axioms, was developed during the twenties and thirties of twentieth century.

                             Today the probability theory has a wider field of application and is used to make intelligent decisions in Economics, Management, operations Research Sociology, Astronomy, Physics, Engineering and Genetics where risk and uncertainty are involved.

                              
                              As the probability theory these days is best understood through the application of the modern set theory, therefore brief descriptions of the basic concepts, notation and operations of set theory that are relevant to probability,

 

Comments

Popular posts from this blog

SETS

                                             SETS       INTRODUCTION           We often speak some words in daily life which represent a collection of things, for example, a team of players, a bunch of flowers, a cluster of trees, a group of friends, a swarm of birds and so on. But in mathematics, we can use a single word "set" to denote all such type of words that show the collection of things such as, a set of players, a set of flowers, a set of trees, a set of friends and a set of birds, etc. Thus we can define a set as:     "A collection of distinct and well-defined objects is called a set. "  The objects of a set are called its members or elements.    ...

Banach Contraction Principle

Banach Contraction Principle Let (X,d) be a complete metric space, then each contraction map f : X → X has a unique fixed point. Let h be a contraction constant of the mapping f . We will explicitly construct a sequence converging to the fixed point. Let x0 be an arbitrary but fixed element in X. Define a sequence of iterates {xn} in X by xn = f(xn−1) (= f n(x0)), for all n ≥ 1. Since f is a contraction, we have d(xn, xn+1) = d(f(xn−1), f(xn)) ≤ hd(xn−1, xn), for any n ≥ 1. Thus, we obtain d(xn, xn+1) ≤ hnd(x0, x1), for all n ≥ 1. Hence, for any m > n, we have d(xn, xm) ≤  hn +hn+1 +···+hm−1 d(x0, x1) ≤ hn 1−h d(x0, x1). We deduce that {xn} is Cauchy sequence in a complete space X. Let xn → p ∈ X. Now using the  continuity of the map f , we get p = lim n→∞ xn = lim n→∞ f(xn−1) = f(p). Finally, to show f has at most one fixed point in X, let p and q be fixed points of f . Then, d(p,q) = d(f(p), f(q)) ≤ hd(p,q). Since h < 1, we must have p = q The proof of the BCP yields the ...