Introduction to Probability and Statistics
Oct 31, 2012
600 likes | 2.04k Views
Introduction to Probability and Statistics. Chapter 12. Topics. Types of Probability Fundamentals of Probability Statistical Independence and Dependence Expected Value The Normal Distribution. Sample Space and Event.
Share Presentation
- poisson distribution
- testing instrument
- 8-hour shift
Presentation Transcript
Introduction to Probability and Statistics Chapter 12
Topics • Types of Probability • Fundamentals of Probability • Statistical Independence and Dependence • Expected Value • The Normal Distribution
Sample Space and Event • Probability is associated with performing an experiment whose outcomes occur randomly • Sample space contains all the outcomes of an experiment • An event is a subset of sample space • Probability of an event is always greater than or equal to zero • Probabilities of all the events must sum to one • Events in an experiment are mutually exclusive if only one can occur at a time
Objective Probability • Objective Probability • Stated prior to the occurrence of the event • Based on the logic of the process producing the outcomes • Relative frequency is the more widely used definition of objective probability. • Subjective Probability • Based on personal belief, experience, or knowledge of a situation. • Frequently used in making business decisions. • Different people often arrive at different subjective probabilities.
Fundamentals of Probability Distributions • Frequency Distribution • organization of numerical data about the events • Probability Distribution • A list of corresponding probabilities for each event • Mutually Exclusive Events • If two or more events cannot occur at the same time • Probability that one or more events will occur is found by summing the individual probabilities of the events: • P(A or B) = P(A) + P(B)
Fundamentals of Probability A Frequency Distribution Example • Grades for past four years.
Fundamentals of Probability Non-Mutually Exclusive Events & Joint Probability • Probability that non-mutually exclusive events M and F or both will occur expressed as: • P(M or F) = P(M) + P(F) - P(MF) • A joint (intersection) probability, P(MF), is the probability that two or more events that are not mutually exclusive can occur simultaneously.
Fundamentals of Probability Cumulative Probability Distribution • Determined by adding the probability of an event to the sum of all previously listed probabilities • Probability that a student will get a grade of C or higher: • P(A or B or C) = P(A) + P(B) + P(C) = .10 + .20 + .50 = .80
Statistical Independence and Dependence Independent Events • Events that do not affect each other are independent. • Computed by multiplying the probabilities of each event. • P(AB) = P(A) P(B) • For coin tossed three consecutive times: Probability of getting head on first toss, tail on second, tail on third is: • P(HTT) = P(H) P(T) P(T) = (.5)(.5)(.5) = .125
Statistical Independence and Dependence Independent Events – Bernoulli Process Definition • Properties of a Bernoulli Process: • Two possible outcomes for each trial. • Probability of the outcome remains constant over time. • Outcomes of the trials are independent. • Number of trials is discrete and integer.
Binomial Distribution • Used to determine the probability of a number of successes in n trials. • where: p = probability of a success • q = 1- p = probability of a failure • n = number of trials • r = number of successes in n trials • Determine probability of getting exactly two tails in three tosses of a coin.
Example • Microchips are inspected at the quality control station • From every batch, four are selected and tested for defects • Given defective rate of 20%, what is the probability that each batch contains exactly two defectives
Binomial Distribution Example – Quality Control • What is probability that each batch will contain exactly two defectives? • What is probability of getting two or more defectives? • Probability of less than two defectives: • P(r<2) = P(r=0) + P(r=1) = 1.0 - [P(r=2) + P(r=3) + P(r=4)] • = 1.0 - .1808 = .8192
Dependent Events • If the occurrence of one event affects the probability of the occurrence of another event, the events are dependent. • Coin toss to select bucket, draw for blue ball. • If tail occurs, 1/6 chance of drawing blue ball from bucket 2; if head results, no possibility of drawing blue ball from bucket 1. • Probability of event “drawing a blue ball” dependent on event “flipping a coin.”
Dependent Events – Conditional Probabilities • Unconditional: P(H) = .5; P(T) = .5, must sum to one. • Conditional: P(RH) =.33, P(WH) = .67, P(RT) = .83, P(WT) = .17
Math Formulation of Conditional Probabilities • Given two dependent events A and B: • P(AB) = P(AB)/P(B) or P(AB) = P(A|B).P(B) • With data from previous example: • P(RH) = P(RH) P(H) = (.33)(.5) = .165 • P(WH) = P(WH) P(H) = (.67)(.5) = .335 • P(RT) = P(RT) P(T) = (.83)(.5) = .415 • P(WT) = P(WT) P(T) = (.17)(.5) = .085
Summary of Example Problem Probabilities
Bayesian Analysis • In Bayesian analysis, additional information is used to alter (improve) the marginal probability of the occurrence of an event. • Improved probability is called posterior probability • A posterior probability is the altered marginal probability of an event based on additional information. • Bayes’ Rule for two events, A and B, and third event, C, conditionally dependent on A and B:
Bayesian Analysis – Example (1 of 2) • Machine setup; if correct 10% chance of defective part; if incorrect, 40%. • 50% chance setup will be correct or incorrect. • What is probability that machine setup is incorrect if sample part is defective? • Solution: P(C) = .50, P(IC) = .50, P(DC) = .10, P(DIC) = .40 • where C = correct, IC = incorrect, D = defective
Statistical Independence and Dependence Bayesian Analysis – Example (2 of 2) • Previously, the manager knew that there was a 50% chance that the machine was set up incorrectly • Now, after testing the part, he knows that if it is defective, there is 0.8 probability that the machine was set up incorrectly
Expected Value Random Variables • When the values of variables occur in no particular order or sequence, the variables are referred to as random variables. • Random variables are represented by a letter x, y, z, etc. • Possible to assign a probability to the occurrence of possible values. Possible values of no. of heads are: Possible values of demand/week:
Expected Value Example (1 of 4) • Machines break down 0, 1, 2, 3, or 4 times per month. • Relative frequency of breakdowns , or a probability distribution:
Expected Value Example (2 of 4) • Computed by multiplying each possible value of the variable by its probability and summing these products. • The weighted average, or mean, of the probability distribution of the random variable. • Expected value of number of breakdowns per month: • E(x) = (0)(.10) + (1)(.20) + (2)(.30) + (3)(.25) + (4)(.15) • = 0 + .20 + .60 + .75 + .60 • = 2.15 breakdowns
Expected Value Example (3 of 4) • Variance is a measure of the dispersion of random variable values about the mean. • Variance computed as follows: • Square the difference between each value and the expected value. • Multiply resulting amounts by the probability of each value. • Sum the values compiled in step 2. • General formula: • 2 = [xi - E(xi)] 2 P(xi)
Expected Value Example (4 of 4) • Standard deviation computed by taking the square root of the variance. • For example data: • 2 = 1.425 breakdowns per month • standard deviation = = sqrt(1.425) • = 1.19 breakdowns per month
Based on the number of outcomes occurring during a given time interval or in a specified regions Examples # of accidents that occur on a given highway during a 1-week period # of customers coming to a bank during a 1-hour interval # of TVs sold at a department store during a given week # of breakdowns of a washing machine per month Poisson Distribution
Conditions • Consider the # of breakdowns of a washing machine per month example • Each breakdown is called an occurrence • Occurrences are random that they do not follow any pattern (unpredictable) • Occurrence is always considered with respect to an interval (one month)
X = number of counts in the interval Poisson random variable with > 0 PMF f(x)= x=0,1,2, Mean and Variance E[X] = , V (X) = The Probability Mass Distribution
If a bank gets on average = 6 bad checks per day, what are the probabilities that it will receive four bad checks on any given day?10 bad checks on any two consecutive days? Solution x = 4 and = 6, then f(4) = = 0.135 = 12 and x = 10, then f(10) = = 0.105 Example
The number of failures of a testing instrument from contamination particle on the product is a Poisson random variable with a mean of 0.02 failure per hour. What is the probability that the instrument does not fail in an 8-hour shift? What is the probability of at least one failure in one 24-hour day? Example
Let X denote the failure in 8 hours. Then, X has a Poisson distribution with =0.16 P(X=0)=0.8521 Let Y denote the number of failure in 24 hours. Then, Y has a Poisson distribution with =0.48 P(Y1) = 1-P(Y = 0) =0.3812 Solution
The Normal Distribution Continuous Random Variables • Continuous random variable can take on an infinite number of values within some interval. • Continuous random variables have values that are not countable • Cannot assign a unique probability to each value
The Normal Distribution Definition • The normal distribution is a continuous probability distribution that is symmetrical on both sides of the mean. • The center of a normal distribution is its mean . • The area under the normal curve represents probability, and total area under the curve sums to one.
The Normal Distribution Example (1 of 5) • Mean weekly carpet sales of 4,200 yards, with standard deviation of 1,400 yards. • What is probability of sales exceeding 6,000 yards? • = 4,200 yd; = 1,400 yd; probability that number of yards of carpet will be equal to or greater than 6,000 expressed as: P(x6,000).
The Normal Distribution Example (2 of 5) - -
The Normal Distribution Standard Normal Curve (1 of 2) • The area or probability under a normal curve is measured by determining the number of standard deviations from the mean. • Number of standard deviations a value is from the mean designated as Z. • Z = (x - )/
The Normal Distribution Standard Normal Curve (2 of 2)
The Normal Distribution Example (3 of 5) Z = (x - )/ = (6,000 - 4,200)/1,400 = 1.29 standard deviations P(x 6,000) = .5000 - .4015 = .0985
The Normal Distribution Example (4 of 5) • Determine probability that demand will be 5,000 yards or less. • Z = (x - )/ = (5,000 - 4,200)/1,400 = .57 standard deviations • P(x 5,000) = .5000 + .2157 = .7157
The Normal Distribution Example (5 of 5) • Determine probability that demand will be between 3,000 yards and 5,000 yards. • Z = (3,000 - 4,200)/1,400 = -1,200/1,400 = -.86 • P(3,000 x 5,000) = .2157 + .3051= .5208
Different Table • P(3,000 x 5,000)= • P((3,000 - 4,200)/1,400) z ((5,000 - 4,200)/1,400) • P(-0.86 z 0.57)= • P( z 0.57)- P( z -0.86)= • P( z 0.57)- P( z ≥0.86)= • P( z 0.57)- [1-P( z 0.86)]= • (0.7157)-[1-0.8051]=0.5208
The Normal Distribution Sample Mean and Variance • The population mean and variance are for the entire set of data being analyzed. • The sample mean and variance are derived from a subset of the population data and are used to make inferences about the population.
The Normal Distribution Computing the Sample Mean and Variance
The Normal Distribution Example Problem Re-Done Sample mean = 42,000/10 = 4,200 yd Sample variance = [(190,060,000) - (1,764,000,000/10)]/9 = 1,517,777 Sample std. dev. = sqrt(1,517,777) = 1,232 yd
The Normal Distribution Chi-Square Test for Normality (1 of 2) • It can never be simply assumed that data are normally distributed. • A statistical test must be performed to determine the exact distribution. • The Chi-square test is used to determine if a set of data fit a particular distribution. • It compares an observed frequency distribution with a theoretical frequency distribution that would be expected to occur if the data followed a particular distribution (testing the goodness-of-fit).
The Normal Distribution Chi-Square Test for Normality (2 of 2) • In the test, the actual number of frequencies in each range of frequency distribution is compared to the theoretical frequencies that should occur in each range if the data follow a particular distribution. • A Chi-square statistic is then calculated and compared to a number, called a critical value, from a chi-square table. • If the test statistic is greater than the critical value, the distribution does not follow the distribution being tested; if it is less, the distribution does exist. • Chi-square test is a form of hypothesis testing.
Statistical Analysis with Excel (1 of 3)
Statistical Analysis with Excel (2 of 3)
Statistical Analysis with Excel (3 of 3)
- More by User
Introduction to Probability and Statistics Thirteenth Edition
Introduction. Populations are described by their probability distributions and parameters.For quantitative populations, the location and shape are described by m and s.For a binomial populations, the location and shape are determined by p. If the values of parameters are unknown, we make inferenc
556 views • 40 slides
421 views • 23 slides
Introduction. Suppose that a pharmaceutical company is concerned that the mean potency m of an antibiotic meet the minimum government potency standards. They need to decide between two possibilities:. . The mean potency m does not exceed the mean allowable potency. The mean potency m exceeds the
676 views • 43 slides
Introduction to Probability and Statistics. Linear Regression and Correlation. Example. Let y be a student’s college achievement, measured by his/her GPA . This might be a function of several variables: x 1 = rank in high school class x 2 = high school’s overall rating
696 views • 39 slides
Statistics 300: Introduction to Probability and Statistics
Statistics 300: Introduction to Probability and Statistics. Section 2-2. Chapters 2 and 3. Describing data Exploring data Comparing data. Describe or Infer?. Descriptive statistics Inferential statistics. Descriptive Statistics.
492 views • 23 slides
Introduction to Probability and Statistics Thirteenth Edition. Chapter 1 Describing Data with Graphs. Variables and Data. A variable is a characteristic that changes or varies over time and/or for different individuals or objects under consideration.
462 views • 30 slides
Introduction to Probability and Statistics Thirteenth Edition. Chapter 5 Several Useful Discrete Distributions. Introduction. Discrete random variables take on only a finite or countably infinite number of values.
426 views • 28 slides
Introduction to Probability and Statistics Thirteenth Edition. Chapter 6 The Normal Probability Distribution. Continuous Random Variables. Continuous random variables can assume the infinitely many values corresponding to points on a line interval. Examples: Heights, weights
375 views • 18 slides
Introduction to Probability & Statistics
Introduction to Probability & Statistics. Dhon G. Dungca, M.Eng’g. STATISTICS. The branch of mathematics that deals with the systematic method of collecting, classifying, presenting, analyzing and interpreting quantitative or numerical data. IMPORTANCE. Research
368 views • 18 slides
Introduction to Probability and Statistics Fourteenth Edition
Introduction to Probability and Statistics Fourteenth Edition. Introduction Train Your Brain for Statistics. What is Statistics?. What does a statistician do?. Player Games Minutes Points Rebounds FG% Bob 34 32.7 24 7.6 .552
424 views • 17 slides
Introduction to Probability and Statistics. Chapter 6 The Normal Probability Distribution. Continuous Random Variables. Continuous random variables can assume infinitely many values corresponding to points on a line interval. Examples: Heights, Weights Lifetime of a particular product
636 views • 30 slides
Introduction to probability theory and statistics
Introduction to probability theory and statistics. HY 335 Presented by: George Fortetsanakis. Roadmap. Elements of probability theory Probability distributions Statistical estimation Reading plots. Terminology. Event: every possible outcome of an experiment.
599 views • 36 slides
Introduction to Probability & Statistics. Chapter # 1. Objectives.
3.18k views • 23 slides
Statistics 300: Introduction to Probability and Statistics. Section 1-2. More Definitions. Parameter a numerical characteristic of a population “population parameter”. More Definitions. Statistic a numerical characteristic of a sample “sample statistic”. Examples of “numerical
317 views • 26 slides
Introduction to Probability and Statistics Twelfth Edition
Introduction to Probability and Statistics Twelfth Edition. Chapter 3 Describing Bivariate Data. Bivariate Data. When two variables are measured on a single experimental unit, the resulting data are called bivariate data.
288 views • 26 slides
305 views • 28 slides
Statistics 300: Introduction to Probability and Statistics. Section 1-4. Design of Experiments. Definitions Five different approaches to sampling. Recap Scientific Method. Identify hypothesis and popul. Plan for collecting data Collect the data Analyze the data
300 views • 28 slides
Statistics 300: Introduction to Probability and Statistics. Section 1-1. What is this class all about? Definitions of “Statistics” as a “field of study”. Definitions of “Statistics”. George Box: “The art and science of making sense out of data”. Definitions of “Statistics”.
191 views • 17 slides
Introduction to Probability and Statistics. Chapter 5 Discrete Distributions. Discrete Random Variables. Discrete random variables take on only a finite or countable many of values. Number of heads in 1000 trials of coin tossing Number of cars that enter UNI in a certain day.
273 views • 25 slides
Statistics 300: Introduction to Probability and Statistics. Section 3-4. Exploratory Data Analysis. First stage in analysis Summaries that help us see Outliers Shape Comparisons. Exploratory Data Analysis. Outliers Observations that are not consistent with the rest of the data
94 views • 6 slides
IMAGES
VIDEO