The standard error of the difference of sample means is 0.444
From the complete question, we have the following parameters
Canadians
Americans
The standard error of a sample is the quotient of the standard deviation and the square root of the sample size.
This is represented as:
[tex]SE = \frac{\sigma}{\sqrt n}[/tex]
The standard error of the Canadian sample is:
[tex]SE_1 = \frac{2.9}{\sqrt{50}}[/tex]
So, we have:
[tex]SE_1 = 0.41[/tex]
The standard error of the American sample is:
[tex]SE_2 = \frac{1.3}{\sqrt{60}}[/tex]
So, we have:
[tex]SE_2 = 0.17[/tex]
The standard error of the difference of sample means is then calculated as:
[tex]SE= \sqrt{SE_1^2 + SE_2^2}[/tex]
This gives
[tex]SE= \sqrt{0.41^2 + 0.17^2}[/tex]
[tex]SE= \sqrt{0.197}[/tex]
Take square roots
[tex]SE= 0.444[/tex]
Hence, the standard error of the difference of sample means is 0.444
Read more about standard errors at:
https://brainly.com/question/6851971