If an atom were to emit a photon whose wavelength was 5200 A, how much energy di
ID: 950432 • Letter: I
Question
If an atom were to emit a photon whose wavelength was 5200 A, how much energy did the atom lose. answer 3.8 x 10^-19. I got 3.8/ 3.6 x 10^-20 is it wrong ????
How much energy would a mole of these atoms lose? 230 kJ answer. Please explain
and
Cesium metal requires 4.78 x 10^-19 J of energy to remove an electron from its surface. What is the longest wavelength of radiation sufficient to eject such a photoelectron? Why is this a maximum wavelength? answer should be 4.16 x 10^-7 explain how to get answer.
Explanation / Answer
1)
Given that
wavelength = 5200 Ao = 5200 x 10-10 m ( 1 Ao = 10-10 m )
Energy E = hc /
h = planck's constant = 6.626 x 10-34 J.s
c = velocity of light = 3 x 108 m/s
Hence,
E = hc /
= (6.626 x 10-34 J.s) (3 x 10 8 m/s) / (5200 x 10-10 m)
= 3.8 x 10-19 J
E = 3.8 x 10-19 J
2) 1 mole = 6.023 x 1023
Energy lost by 1 mole of atoms = 6.023 x 1023 x 3.8 x 10-19 J = 230 kJ
3) Energy E = hc /
= hc/E
= (6.626 x 10-34 J.s) (3 x 10 8 m/s) / ( 4.78 x 10^-19 J)
= 4.16 x 10^-7 m
= 4.16 x 10^-7 m
Related Questions
drjack9650@gmail.com
Navigate
Integrity-first tutoring: explanations and feedback only — we do not complete graded work. Learn more.