PDF Summary:Introduction to Statistics, by

Book Summary: Learn the key points in minutes.

Below is a preview of the Shortform book summary of Introduction to Statistics by Jim Frost. Read the full comprehensive summary at Shortform.

1-Page PDF Summary of Introduction to Statistics

Statistics is a powerful tool for understanding data, making decisions, and testing hypotheses. In Introduction to Statistics, Jim Frost elucidates the importance of statistical analysis across disciplines — from sciences and social studies to business.

The author provides an in-depth overview of descriptive and inferential statistics methods. He explores visualizing data with graphs, summarizing datasets, determining central tendencies and variability, and analyzing variable relationships. Frost further delves into probability distributions, statistical inference techniques like confidence intervals and hypothesis testing, and methods for integrating statistics into the scientific process.

(continued)...

Range Sensitive to Outliers; IQR and Standard Deviation More Robust Measures of Spread

The author defines the range as the gap between a dataset's maximum and minimum numbers, calling it the most straightforward variability measure. However, it's susceptible to extreme values, and its value is influenced by sample size. He suggests using it when dealing with limited sample sizes, where other measures can't be calculated accurately.

The author then introduces the interquartile range (IQR), which is the span between the first quartile (Q1) and third quartile (Q3), covering the central 50% of the data values. Like the median, the IQR is robust to outliers because it's not based on all values. It's a good measure of spread for skewed distributions, providing a practical alternative to standard deviation in these cases.

Practical Tips

  • Use a median-based approach when comparing your monthly expenses to avoid being misled by unusually high or low spends. Instead of just looking at the range between your most and least expensive months, calculate the median expense of each month over a year. This will give you a better sense of your typical monthly costs and help you budget more effectively.
  • Use range to compare daily variations in personal habits. Track something simple like your daily water intake or step count for a month. Note the lowest and highest values to find the range, which will give you a clear picture of your variability in that habit. This can help you identify patterns or outliers in your behavior that may need attention.
  • When comparing products or services with a wide range of reviews, use the IQR to determine the typical user experience. For instance, when looking at ratings for a hotel, calculate the IQR of the review scores to understand what the majority of guests experienced, minimizing the impact of extremely negative or positive reviews that might not represent the standard service level.
Variance Includes All Data but Is Challenging to Understand; Standard Deviation Uses Original Units

Frost delves into variance and deviation. Variance calculates the mean of the squared differences between the values and the mean, including all values in the computation. While valuable in statistical tests, the fact that it's measured in units squared hinders intuitive interpretation.

Conversely, the standard deviation, the square root of the variance, shows the usual gap between individual data values and the average, expressed in the original data's units. This makes standard deviation the most commonly used measure of variability. The ease of interpretation underscores its widespread use for conveying a data set's spread.

Practical Tips

  • Evaluate the performance diversity in your investment portfolio using variance. Record the weekly or monthly returns of your different investments. Calculate the variance to understand the volatility of each investment. This can inform your decisions on whether to rebalance your portfolio to either increase diversity and reduce risk or concentrate on higher-performing assets.
  • Create a simple game with friends where you guess the variance of everyday activities. Take an activity like the number of steps you walk in a day, and each person guesses the variance among the group. Track your steps with a pedometer or smartphone for a week, calculate the actual variance, and see who was closest. This game turns the abstract concept of variance into a fun, competitive challenge that can be understood without complex math.
  • Apply standard deviation to your fitness routine to assess consistency. Record your workout durations, distances, or intensities over time and calculate the standard deviation. A lower standard deviation indicates a consistent routine, while a higher one suggests variability that could be affecting your fitness goals.

Correlation Coefficients Gauge Relationship Strength and Orientation

Correlation coefficients quantify the direction and intensity of the linear relationship involving two continuous variables. Frost focuses on Pearson's r and provides examples to illustrate its interpretation.

Confounders Can Create Correlations, Not Cause and Effect Relationships

He emphasizes that a strong correlation between variables isn't proof of causation. Confounding factors may cause false connections. For example, while buying ice cream and getting bitten by sharks might be correlated, both are influenced by a third variable—the number of people at the beach—which is the actual causal factor.

Context

  • The mistaken belief that smoking causes weight loss was based on observed correlations, but further research showed that other factors, such as lifestyle and diet, were involved.
  • Techniques such as regression analysis, propensity score matching, and randomized controlled trials are used to minimize the impact of confounders.
  • Teaching the difference between correlation and causation is fundamental in statistics education to prevent common misconceptions and promote critical thinking.
Consider Study Context to Evaluate Correlation Strength

Furthermore, Frost explains that a "good" correlation is context-dependent. In social sciences, correlations weaker than +/-0.6 are common, while in physical processes with precise measurements, correlations may be close to +1 or -1. Correlation strength must be interpreted considering the particular subject matter.

Context

  • A coefficient of +1 indicates a perfect positive linear relationship, -1 indicates a perfect negative linear relationship, and 0 indicates no linear relationship.
  • The ability to replicate experiments with consistent results in physical sciences contributes to the expectation of strong correlations.
  • Smaller sample sizes can lead to less reliable correlation estimates. In fields where large samples are difficult to obtain, correlations might appear weaker.

Probability Distributions and Making Inferences About Populations

Discrete Distributions Model Counts and Two-Option Outcomes

This section focuses on probabilities and their applications for various data categories. Frost dives deep into both continuous and discrete distributions and the specific types within each category that are appropriate for particular data types.

Likelihoods for Events in Fixed-Trial Distributions

The author delves into discrete probability distributions—binomial, geometric, hypergeometric, and negative binomial—showing how they can model probabilities for binary data. He explains how each distribution answers a different question about the data. For example, the binomial distribution calculates the likelihood of an event happening a specific number of times over a fixed number of trials, like the probability of getting five heads in ten coin tosses.

The geometric distribution calculates the likelihood of the first occurrence of an event, like the probability of getting the first heads on the fifth coin toss. The negative binomial distribution determines the probability of observing a specified number of events within a certain number of trials, such as the probability of needing 15 coin tosses to get five heads. He uses die rolling and drawing candy examples to explain the principles and interpretation of every distribution.

Practical Tips

  • Create a simple game with binary outcomes to explore probability with friends. Design a game where players predict 'yes' or 'no' answers to future events, such as whether it will rain tomorrow. Keep score of predictions versus actual outcomes to see who has the best understanding of probability.
  • Incorporate geometric distribution into your personal finance strategy for irregular expenses. Track the frequency of unexpected expenses, like car repairs or medical bills, to estimate the probability of their occurrence. This can help you better prepare your emergency fund, ensuring you're financially ready for these events without overestimating the amount you need to set aside.
  • Track your habit formation progress with a custom spreadsheet. Create a spreadsheet where you log daily attempts at a new habit and the successes or setbacks you encounter. Use the principles of the negative binomial distribution to analyze the data over time, which can help you understand the probability of how many days it might take to firmly establish the habit based on your past performance.
  • Implement a 'probability-based' chore schedule for your household. Write down chores on cards and assign a probability to each based on how often they should be done. Each week, randomly draw a card to determine which chore gets done, adjusting the probabilities as necessary to ensure a fair distribution over time. This strategy helps you understand how probabilities influence outcomes in a practical, everyday context.
Useful in Predicting Binary or Discrete Outcomes

Frost demonstrates the practical use of these distributions by modeling flu vaccination results over decades. He considers the average rates of infection for individuals who have been vaccinated versus those who haven't, comparing scenarios of getting a yearly flu shot versus never getting vaccinated. He employs plots for probability distributions using the geometric and binomial models to demonstrate the cumulative effects of getting flu shots, showing how regular vaccination significantly decreases the likelihood of getting the flu over time.

Other Perspectives

  • The comparison of infection rates may not account for confounding variables such as age, pre-existing health conditions, or socioeconomic factors that could influence an individual's likelihood of getting the flu.
  • The scenarios compared by Frost may not reflect real-world complexities, such as individuals who get vaccinated some years but not others, rather than strictly "yearly" versus "never."
  • While Frost uses these models, it's possible that other statistical methods or distributions could provide more accurate or robust predictions for flu vaccination results.
  • Over time, the flu virus can mutate, potentially reducing the cumulative effectiveness of regular vaccination.

The Normal Curve Models Occurrences in Nature

This section explores the normal distribution, highlighting its importance in statistics due to its widespread use in modeling natural phenomena.

Empirical Rule: Relating the Variability to Proportion of Values From Mean

Frost explains the Empirical Rule, which dictates the percentage of data points that are located within specific distances from the mean in a normal distribution. This rule states that 68% of observations are within one standard deviation of the mean in either direction, 95% are within + /- two standard deviations, and 99.7% are within +/- three standard deviations. Using the example of pizza delivery times averaging 30 minutes with a 5-minute standard deviation, he illustrates how the rule helps understand the likelihood of delivery times falling within different time intervals.

Context

  • The rule assumes that the data follows a perfectly normal distribution. In real-world scenarios, data may not perfectly fit this model, so the rule provides an approximation.
  • The concept of the normal distribution and the Empirical Rule has been developed over centuries, with contributions from mathematicians like Carl Friedrich Gauss.
  • This rule is widely used in fields such as quality control, finance, and research to assess variability and predict outcomes.
  • The rarity of data points beyond three standard deviations can be used to identify statistically significant events or anomalies.
  • By analyzing delivery time data, companies can identify patterns and make adjustments to improve customer satisfaction and reduce wait times.
Z-Scores Standardize Data for Comparing Distributions and Estimating Probability

Frost introduces Z-scores or standardized scores, which show the number of standard deviations that a given data point is higher or lower than the average. Z-scores allow comparisons between datasets with different averages and variability. He uses a comparison of apple and orange weights to illustrate this, showing how standardizing the data allows us to evaluate a heavier-than-average apple to a light orange relative to their respective distributions.

He then explains how a z-table can be used to find the percentile of a data point using its Z-score. This table provides the region beneath the curve for specific intervals of Z-scores, allowing us to estimate probabilities for data that follow a normal distribution. He utilizes the apple weight example again, showing how the standardized score can be used to find its approximate weight percentile.

Practical Tips

  • Evaluate your daily productivity by using Z-scores to analyze tasks completed against your average output. Keep a log of tasks you accomplish each day and calculate the Z-score to see how productive you were relative to your normal productivity levels. This can highlight exceptionally productive days or slumps, prompting you to explore what factors contributed to these deviations.
  • Improve your fantasy sports team by analyzing players using z-scores. Calculate the z-score for each player's statistics to see how they compare to the league average. This can help you identify undervalued players to draft or trade for, as well as overvalued players to avoid or trade away.
  • Enhance your cooking skills by using Z-scores to determine which recipes are outliers in terms of preparation time or ingredient cost compared to your usual meals. Gather data on how long it takes you to prepare different dishes and how much each costs. Calculate the average and standard deviation, then find the Z-score for new recipes you try. This will help you understand which recipes are more time-consuming or expensive than what you typically cook, allowing you to better plan your meals according to your time and budget constraints.

Drawing Conclusions About Groups With Statistical Inference

Inferential statistics involve drawing conclusions about populations based on data from samples, using statistical techniques that account for sampling error.

Random Selection and Large Samples Needed to Estimate Population Precisely

Frost defines a population as all of the items being studied and a sample as a subset of the population. He emphasizes the importance of random sampling for obtaining unbiased parameter estimates of a population. He emphasizes that good estimates should be unbiased and reduce the gap between the estimated and actual value, stressing that random sampling and a greater number of samples contribute to these desired properties.

He further explains that having more samples is crucial for representing the population more accurately and minimizing the impact of unusual values or outliers. Frost illustrates this concept using a simulation of a repeated IQ study, where many samples of varying sizes are selected from a known population distribution. The simulation demonstrates that the distribution of sample means narrows as sample size grows, signifying a smaller margin of error for larger samples and hence greater precision in estimating the population mean.

Practical Tips

  • You can enhance your decision-making by treating everyday choices like mini-experiments, using your own experiences as samples to inform future decisions. Start by identifying a decision you face regularly, such as choosing what to eat for lunch. For one week, choose a different option each day and note your energy levels and satisfaction post-lunch. Analyze this 'sample' data to determine which choices yield the best results for you, and use this insight to inform your lunch decisions moving forward.
  • Use a random number generator to select participants for your next group project. By assigning each potential participant a number and using an online random number generator, you can ensure that your selection process is unbiased. This method can be particularly useful when you're looking to form a diverse team and want to avoid any subconscious biases that might influence your choices.
  • Implement a personal 'estimate diary' where you record your predictions for various aspects of your life and then track the actual results. Regularly review this diary to identify patterns where your estimates are consistently over or under the actual values, and adjust your estimation approach accordingly. This could be as simple as predicting how much you'll spend on groceries each week and then comparing it to your actual spending to improve your budgeting skills.
  • Enhance your understanding of public opinion by creating simple surveys and distributing them to different groups. If you're curious about community preferences for a local event, design a short questionnaire and share it with various local social media groups, ensuring you reach a diverse audience. Collecting responses from a wide range of participants can give you a more accurate picture of the community's interests.
  • Enhance your understanding of global issues by following a variety of international news sources. This mirrors the concept of needing more samples by exposing you to different perspectives and preventing a skewed view that might come from only following local or national news. Start by subscribing to news outlets from different continents or political leanings to get a broader picture of world events.
  • Enhance your understanding of a topic by reading articles from multiple sources. When researching a new subject, aim to read at least three articles from different publications or authors. This approach ensures that you're not relying on a single piece of information, which might be an outlier, and instead get a more balanced and comprehensive view of the subject.
  • Start a hobbyist data club with friends where each member collects data on a shared interest. If you and your friends enjoy a particular activity, such as bird watching or tracking fitness progress, each member can record their observations. Over time, compile the data to see how individual variations diminish as the group's collective data grows, illustrating the concept in a real-world context.
  • Apply the principle of larger samples to personal finance by tracking more data points. Instead of checking your bank balance weekly, do it daily, and record all your expenses meticulously. Over time, this larger 'sample' of financial data will give you a clearer picture of your spending habits and financial health, allowing for more precise budgeting and saving strategies.
Confidence Intervals Give Ranges for Population Parameters, Accounting for Sampling Error

Frost introduces CIs as a way to measure how precise sample estimates are. CIs offer a set of numbers that the true population parameter probably falls into, factoring in the uncertainty from sampling error. He demonstrates this using an IQ study example, where confidence intervals are computed for both small and large samples. A limited sample yields a wider confidence interval, indicating lower precision. The larger sample size results in a tighter confidence interval, suggesting greater precision and closer proximity of the sample estimate to the true population parameters.

Practical Tips

  • Incorporate confidence intervals into everyday predictions to set realistic expectations. If you're planning an outdoor event, use confidence intervals to predict the weather by looking at the range of temperatures provided by weather forecasts. This can help you prepare for the most likely scenarios, such as knowing whether to provide heaters or cooling fans for your guests.
  • Incorporate confidence intervals into your health tracking. If you're using a fitness app or device that estimates calories burned or steps taken, research the confidence intervals for these estimates. This will help you understand the potential margin of error and adjust your fitness goals accordingly. For instance, if your device has a wide confidence interval for calories burned, you might want to aim for a higher step count to ensure you're meeting your activity goals.
  • Conduct a taste test experiment with varying group sizes to understand precision in real life. Invite groups of friends in small, medium, and large sizes to taste different flavors of a homemade dish. Record their preferences and note how the smaller groups provide less precise consensus on the best flavor compared to larger groups, reflecting a broader confidence interval.
  • Apply the principle to everyday problem-solving by using broader sources of information. When troubleshooting issues like a tech glitch or a home repair, consult a wide range of forums, videos, and manuals. The more solutions you review, the more likely you are to find the most accurate fix for your problem, reflecting the increased precision of larger datasets.
  • Start a hobbyist project that requires measurement and tracking, like gardening or baking, and record your results meticulously. This hands-on approach will teach you the importance of precision in everyday activities. For instance, by measuring soil pH levels or ingredient weights precisely, you'll see a direct correlation between your accuracy and the success of your plants or recipes.

Distinguishing Descriptive and Inferential Statistics

Descriptive Statistics Condense and Visualize Data Without Generalizing to a Larger Population

This section emphasizes the distinction between inferential and descriptive statistics, which, while often using similar numerical measures, have different goals and methodologies.

Descriptive Statistics: Central Tendency, Variability, and Relationships Between Variables

The author defines descriptive statistics as those that describe a specific dataset for a chosen group without attempting to generalize to a broader population. These statistics summarize the data characteristics for that specific group. They include familiar measures like central tendency (average, middle value, mode), dispersion (range, standard deviation), skewness, and correlation to capture the relationship between pairs of variables.

Practical Tips

  • Use descriptive statistics to optimize your daily routines by logging time spent on various activities. For a week, note the time dedicated to work, leisure, chores, and sleep. Calculate the average time spent on each and compare it to your ideal time distribution to make informed adjustments for a more balanced lifestyle.
  • Organize your grocery shopping lists from the past few months and find the mode for commonly purchased items. This will reveal which items you buy most frequently, allowing you to potentially benefit from bulk purchasing or identifying alternatives for variety or cost-saving.
  • Use a budget tracking app to visualize your spending habits and identify correlations between different types of expenses. By inputting your daily expenditures, you can generate graphs and charts that show how certain spending categories may be related. For example, you might discover a correlation between dining out and transportation costs, indicating that when you eat out more, you also spend more on getting around.
Understanding Information Through Descriptive Analyses

Frost illustrates descriptive statistics with an example involving 30 students and their test results. He calculates and presents the average, standard deviation, and the percentage of students with acceptable scores for that class. These statistics are certain since they encompass data from the entire group. This example emphasizes how descriptive statistics give a clear picture of a specific dataset without making inferences beyond the measured group.

Practical Tips

  • Implement a feedback system for your personal projects or hobbies. After completing a project, like crafting or cooking, rate your performance and ask for ratings from friends or family. Over time, track these scores to calculate averages and identify trends in your performance, which can guide you in honing your skills.
  • Evaluate the effectiveness of your personal health regimen by tracking a wide array of health metrics, not just one or two. Use a spreadsheet or app to monitor your sleep, nutrition, physical activity, and mental well-being over a period of time to see the overall impact of your lifestyle on your health.

Using Inference to Make Generalizations Regarding Populations

In contrast to descriptive statistics, inferential statistics utilize a dataset to draw conclusions about the broader group that the sample reflects. This involves extending inferences from the observed data points to a broader group.

Inferential Methods Need Representative Sampling, Randomization, and Error Accounting

Frost highlights the crucial role of gathering a sample that's representative for valid inferences about the population. Randomly selecting samples is essential for avoiding bias and ensuring the sample reflects the characteristics of the population. He mentions different random selection techniques like basic random selection, stratified selection, and cluster selection. He revisits the example involving scores, but this time considers making inferences about all Pennsylvania eighth graders.

Other Perspectives

  • The concept of a "representative sample" can be subjective and dependent on the definition of what characteristics are considered important for the population, which may vary between researchers and studies.
  • Random sampling can be impractical or impossible in certain situations, such as when dealing with rare populations or when the population frame is not clearly defined.
  • Stratified and cluster selection require prior knowledge about the population to form strata or clusters, which may not always be available or may be based on flawed assumptions, potentially introducing bias.
Tests of Hypotheses Identify Population Relationships

He also introduces hypothesis tests, which help determine whether observed relationships in the data sample are likely to occur in the population, factoring in the random errors of sampling. Confidence intervals are presented as a way to quantify the potential error surrounding estimates derived from the data sample. These techniques help bridge the gap between limited sample data and the population, enabling researchers to draw conclusions about the broader group.

Other Perspectives

  • Hypothesis tests can suggest that an observed relationship in a sample is unlikely to be due to random variation, but they cannot confirm that the relationship will occur in the population.
  • The p-values used in hypothesis testing can be misinterpreted, leading to overemphasis on statistical significance rather than practical significance or effect size.
  • The choice of confidence level (e.g., 95%) is somewhat arbitrary and does not convey the probability that the specific interval calculated from the sample contains the population parameter.
  • The width of confidence intervals can be misleading, as larger samples produce narrower intervals, which might be interpreted as more precise, even if the underlying data is noisy or the sampling method is flawed.
  • These methods rely on the quality of the data collected; if the data is biased or flawed, the conclusions about the broader group will also be unreliable.

Integrating Statistics Into Science

The Scientific Process Relies on Experiments; Statistical Analysis Is Vital for Conclusions

This section elucidates how statistical analysis is integrated into the scientific method, playing a vital role in designing experiments, examining data, and making inferences.

Consider Confounders to Determine Causes, Not Just Correlations

Frost divides the research process into these five key steps: research (defining the problem, reviewing literature), operationalization (defining variables, measurement techniques, sample size), data collection, statistical analysis, and communication of results. He emphasizes that while statistical analysis occurs at the end of the process, each step must be executed carefully to ensure valid results. He then stresses the need to pinpoint causation, distinguishing it from mere correlation.

The author explains how confounders can create spurious correlations, misleading researchers regarding what's truly causing the observed effects. He points out that researchers employ techniques like randomization, pair matching, and statistical modeling (e.g., multiple linear regression) to manage confounder effects and strengthen the evidence for causality. He uses a hypothetical vitamin consumption example to show how confounders like healthy lifestyle habits can lead to erroneous findings if not appropriately controlled.

Practical Tips

  • Use a decision-making app that incorporates confounder checks. Look for or suggest the development of an app designed to help users make more informed decisions by prompting them to consider potential confounders. When faced with a decision based on observed correlations, the app could guide you through a checklist of common confounders to ensure a more thorough analysis.

Other Perspectives

  • The steps mentioned do not consider the ethical considerations and approvals that are often necessary before research can commence, especially when human subjects are involved.
  • The iterative nature of research means that sometimes, the steps are not linear and may need to be revisited, which can lead to a re-evaluation of the importance of each step as the research progresses.
  • Pair matching may introduce its own biases if the matching criteria are not well-chosen or if there are unmeasured variables that affect the outcome.
  • In some cases, what is considered a confounder might actually be a mediator in the causal pathway, and controlling for it could obscure the true relationship between the exposure and the outcome.
Criteria Like Hill's Causation, Data Reliability, Accuracy, and Experimentation Accuracy Determine if Causal Conclusions Are Warranted

Frost further discusses Hill's causation criteria, nine guidelines for evaluating how strong evidence is for causal relationships. These criteria encompass the strength of association, consistency of findings across studies, specificity of the link, temporality (cause preceding effect), biological gradient (dose-response relationship), plausibility, coherence with existing knowledge, experimental evidence, and analogy to similar cause-effect relationships. He also emphasizes the importance of data quality, discussing reliability (measurement consistency) and validity (measurement accuracy). Evaluating these elements provides a framework for determining the trustworthiness of experimental findings and the causal conclusions based on them.

Practical Tips

  • Optimize your learning by assessing the effectiveness of different study methods with a self-conducted experiment. Use two different study techniques for two similar topics or subjects. For instance, try active recall for one subject and spaced repetition for another. Track your retention and understanding over time through quizzes or practical application. This will help you determine which method is more valid and reliable for your learning style, allowing you to focus on the most effective approach for future learning endeavors.
  • Start a "Causal Journal" where you document daily occurrences and your hypotheses about their causes, then track subsequent related events to see if your hypotheses hold up. This practice will sharpen your ability to discern cause-and-effect relationships in real life. For instance, if you notice you feel more energized on days you drink a green smoothie, record this observation and track your energy levels on subsequent days with and without the smoothie to test the causal link.

Additional Materials

Want to learn the rest of Introduction to Statistics in 21 minutes?

Unlock the full book summary of Introduction to Statistics by signing up for Shortform .

Shortform summaries help you learn 10x faster by:

  • Being 100% comprehensive: you learn the most important points in the book
  • Cutting out the fluff: you don't spend your time wondering what the author's point is.
  • Interactive exercises: apply the book's ideas to your own life with our educators' guidance.

Here's a preview of the rest of Shortform's Introduction to Statistics PDF summary:

Read full PDF summary

What Our Readers Say

This is the best summary of Introduction to Statistics I've ever read. I learned all the main points in just 20 minutes.

Learn more about our summaries →

Why are Shortform Summaries the Best?

We're the most efficient way to learn the most useful ideas from a book.

Cuts Out the Fluff

Ever feel a book rambles on, giving anecdotes that aren't useful? Often get frustrated by an author who doesn't get to the point?

We cut out the fluff, keeping only the most useful examples and ideas. We also re-organize books for clarity, putting the most important principles first, so you can learn faster.

Always Comprehensive

Other summaries give you just a highlight of some of the ideas in a book. We find these too vague to be satisfying.

At Shortform, we want to cover every point worth knowing in the book. Learn nuances, key examples, and critical details on how to apply the ideas.

3 Different Levels of Detail

You want different levels of detail at different times. That's why every book is summarized in three lengths:

1) Paragraph to get the gist
2) 1-page summary, to get the main takeaways
3) Full comprehensive summary and analysis, containing every useful point and example