Or search by topic
A farmer has a very large farm which produces wheat. The yield of wheat per hectare is known to be normally distributed at 7.74 tonnes per hectare, with a standard deviation of 0.62.
Why can a normal distribution of a yield never be entirely accurate? Why, in this case, does it not matter?
It is nearly planting time and the farmer has future orders for 8000 tonnes of wheat. If he fails to produce enough wheat then he will have to pay a stiff penalty to the buyer; if he produces too much wheat then he will have to offload or destroy the surplus at a loss. In each case, the loss L can be modelled as follows, where A is the amount he plants:
\begin{align*} &\mbox{If } A < 8000, L = (8000 - A)\times 4\\ &\mbox{If } A > 8000, L = 0.5 \times (A-8000)\\\end{align*}
There are three possible levels of analysis of this problem
Level 1 - using confidence intervals
How much wheat would the farmer have to plant to have an expected yield of exactly 8000 tonnes? What distribution would the total yield have, and what would be the 95\% confidence interval for the actual yield?
In this case, what would be the 95\% confidence interval for the loss?
How would this alter if he instead planted 1000 hectares? Could you recommend an ideal amount of wheat to plant?
Level 2 - finding an expected loss
For your planting recommendation, what would be the expected loss?
Level 3 - minimising the expected loss (involves difficult, but fun, calculus)
How much should the farmer plant to minimise his expected loss? Make an initial considered estimate before performing a calculation.
The probability that a passenger books a flight and does not turn up is 0.05. For an aeroplane with 400 seats how many tickets can be sold so that only 1% of flights are over-booked?