What price do farmers get for their apples? random sample of 42 farming regions gave a sample mean of = $6.88 per
100 pounds of apples. Assume that σ is known to be $1.90 per 100 pounds of apples.
1. Find a 90% confidence interval for the mean price (per 100 pounds) that farmers in this region get for their apples (in dollars). What is the margin of error (in dollars)?
lower limit $
upper limit $
margin of error $
2.Find the sample size necessary for a 90% confidence level with maximal error of estimate E = 0.39 for the mean price per 100 pounds of apples. how many regions
3. A farm can bring 15 tons or 200 pounds per ton of apples to market. Find a 90% confidence interval for the population mean cash value of this crop. Also what is the margin of error ?