• Tidak ada hasil yang ditemukan

probability konsep dan penerapannya

N/A
N/A
Aong Aong

Academic year: 2024

Membagikan "probability konsep dan penerapannya"

Copied!
57
0
0

Teks penuh

Adopting a probabilistic mindset is far more important than knowing “absolute truths” when it comes to success in data science. Finally, we focus on distributions, which are the “heart” of probability as used in data science. You may have heard of many of them, but this is the only place you'll find detailed information on many of the most common distributions.

Discrete distributions: Uniform distribution, Bernoulli distribution, binomial distribution (here you will see many of the combinations from previous parts), Poisson. You are here because you want to understand the basics of probability before you can dive into the world of statistics and machine learning. Understanding the driving forces behind key statistical features is essential to achieving your goal of mastering data science.

This way you will be able to extract important insights when analyzing data through supervised machine learning methods like regressions, but also understand the results that unsupervised or assisted ML gives you. Since many of these distributions have elegant relationships between certain outcomes and their probabilities of occurrence, knowing the key features of our data is extremely convenient and useful.

The Basics of Probability

Expected Values

In this case, the experimental probability of getting heads will be equal to the number of heads we record over the course of the 20 outcomes, over 20 (the total number of trials). The expected value can be numeric, boolean, categorical or other depending on the type of event we are interested in. For example, the expected value of the trial would be the more likely of the two outcomes, while the expected value of the experiment would be the number of times we expect to get heads or tails after the 20 trials.

Probability Frequency Distribution

Complements

Combinatorics

  • Permutations
  • Factorials
  • Variations
  • Combinations
    • Combinations with separate sample spaces
    • Winning the Lottery
    • Combinations with Repetition
    • Applications of Combination with Repetition
    • Pizza Example
  • Methodology
    • Pizza and Sequences
    • Always Ending in 0
    • Positions
    • The Final Step
  • Symmetry of Combinations

Variations represent the number of different possible ways we can select and arrange a number of elements. Combinations represent the number of different possible ways we can choose a number of elements. in the sample space Number of elements in us. you must select Properties of Combinations:. The option we choose for any element does not affect the number of options for other elements.

We use combinations with repetition when the events we are dealing with have sufficient amount of each of the distinct values ​​in the sample space. To get a better understanding of the number of combinations we have, let's examine a specific example. Since only 8 of the positions in the sequence can have a value of "1", the number of different pizzas will be the combination of 3 of the 8 positions.

Therefore, the number of pizzas we can get would be the number of combinations of choosing 3 items out of a set of 8. Before proceeding to the next lecture, let's do a quick summary of the algorithm and formula. We concluded that each unique sequence can be expressed as a combination of the positions of .

Let's look at the algebraic proof of the idea that choosing p-many elements from a set of n is the same as omitting n-p many elements.

Bayesian Notation

  • Multiple Events
    • Intersection
    • Union
    • Mutually Exclusive Sets
  • Independent and Dependent Events
  • Conditional Probability
  • Law of Total Probability
  • Additive Law
  • The Multiplication Rule
  • Bayes’ Law

The sets of outcomes satisfying two events A and B can interact in one of three ways. The intersection of two or more events expresses the set of outcomes that satisfy all events simultaneously. The union of two or more events expresses the set of outcomes that satisfy at least one of the events.

Dogs and cats are mutually exclusive groups, as no species is both a cat and a dog, but the two are not complementary, as there are other types of animals. If the probability of the occurrence of event A (P(A)) is affected by the occurrence of event B, then we say that A and B are dependent events. For any two events A and B, such that the probability of B occurring is greater than 0 (𝑃 𝐵 > 0), the conditional probability formula reads as follows.

Unlike union or crossover, changing the order of A and B in the conditional probability changes its meaning. The law of total probability dictates that for any set A, which is a union of many mutually exclusive sets 𝐵 , 𝐵. Since P(A) is the union of mutually exclusive sets, so it is equal to the sum of the individual sets.

The additive law calculates the probability of the union based on the probability of the individual sets it takes into account. The probability of each is simply its size relative to the size of the sample space. Bayes' law helps us understand the relationship between two events by calculating the various conditional probabilities.

Bayes' law is often used in medical or business analysis to determine which of two symptoms has a greater effect on the other.

Discrete Distributions

Important Notation for Distribution

Types of Distributions

  • Discrete Distributions
  • Uniform Distribution
  • Bernoulli Distribution
  • Binomial Distribution
  • Poisson Distribution

A distribution consisting of a single trial and only two possible outcomes – success or failure is called the Bernoulli Distribution. It is often used when trying to determine what we expect a single trial of an experiment to produce. It is often used when trying to predict how likely an event is to occur over a series of trials.

When we want to know the probability that a certain event will occur over a given interval of time or distance, we use a Poisson distribution. Used to determine how likely a particular outcome is, knowing how often the event usually occurs. Often incorporated into marketing analytics to determine whether above-average visits are unusual or not.

Continuous Distribution

  • Normal Distribution
  • Standardizing a Normal Distribution
  • Student T’s Distribution
  • Chi-Squared Distribution
  • Exponential Distribution
  • Logistic Distribution
  • Poisson Distribution: Expected Value and Variance

To standardize any normal distribution, we need to transform it so that the mean is 0 and the variance and standard deviation are 1. We can convert any normal distribution into a standard normal distribution using the transformation shown above. Convenient to use because of a table of known values ​​for the CDF, called the Z-score table, or simply the Z-table.

Often used in analysis when examining a small sample of data that usually follows a normal distribution. We often use the natural logarithm to transform the values ​​of such distributions since we do not have a table of known values ​​such as the Normal or Chi-squared. The Continuous Logistic Distribution is observed when trying to determine how continuous variable inputs can affect the probability of a binary outcome.

Often used in sports to anticipate how a player or team's performance could determine the outcome of the match. Suppose we have a random variable Y such that 𝑌 ~ 𝑃𝑜𝑖𝑠𝑠𝑜𝑛 (𝜆), then we can find its expected value and variance in the following way. Recall that the expected value for any discrete random variable is a sum of all possible values ​​multiplied by the probability of their occurrence P(y).

Additionally, we can divide the numerator and denominator by "y" since "y" will be non-zero in all cases. Since we have "y-1" in both numerator and denominator, and the sum starts from 1, this is equivalent to starting the sum from 0 and using y instead. Finally, since any value to the negative power is the same as 1 divided by the same value, then 𝑒−𝜆𝑒𝜆 = 1.

We will first express it in terms of expected values ​​and then apply an approach similar to the one we used for expected value. We start with the familiar relationship between expected value and variance, the variance is equal to the expected value of the variable squared, minus the expected value of the variable, squared. This next step may seem pretty counterintuitive, but we're just squaring the variable in a way that makes it easier to manipulate.

Setting Up Wolfram Alpha

Typing up the integral formula

Expressing Complex Formulas

Understanding the Solution

In such cases, we add "between a and b" at the end of the integral before the calculation. The expected value of a function is found by finding the sum of products for each possible outcome and its chance of occurring.

Normal Distribution 𝑬 𝒀 , 𝑽𝒂𝒓(𝒀 )

Learn the most in-demand data science skills from the best experts in the field.

Referensi

Dokumen terkait

In [9], we earlier obtained exponential upper bounds for tails of distributions of statistics of the form (1) based on a sample from the uniform distribution on [0, 1] and also

The probability distribution of this discrete random variable is called the binomial distribution , and its values will be denoted by b ( x ; n, p ) since they depend on the number

Pengantar Kerugian Acak dan Sifat Momen Distribusi Distribusi Poisson Distribusi Binomial Distribusi Geometrik Mixed and Mixture Distributions Kelas Distribusi (a, b, 0)