contestada

A sinusoidal electromagnetic wave has an average intensity of 100 W/m2W/m2. By what factor would the electric-field amplitude of the wave have to be increased in order for the wave to have an average intensity of 8100 W/m2W/m2

Respuesta :

Answer:

The electric field is increase to 80 [tex]\frac{V}{m}[/tex]  so average intensity became 8100 [tex]\frac{W}{m^{2} }[/tex].

Explanation:

Given :

Initial average intensity  [tex]I_{1} = 100 \frac{W}{m^{2} }[/tex]

Final average intensity   [tex]I_{2} = 8100 \frac{W}{m^{2} }[/tex]

The average intensity is proportional to the square of the electric field.

∴  [tex]I = \frac{1}{2}c\epsilon_{o} E_{o}^{2}[/tex]

    [tex]I[/tex] ∝ [tex]E_{o^{2} }[/tex]

⇒  [tex]\sqrt{I_{1} } = E_{1}[/tex]

  [tex]\sqrt{100 } = E_{1}[/tex]

So initial electric field is [tex]E_{1} = 10 \frac{V}{m}[/tex]

Now intensity become [tex]8100\frac{W}{m^{2} }[/tex] so electric field is,

  [tex]E_{2} = \sqrt{8100}[/tex]

So final electric field  [tex]E_{2} = 90 \frac{V}{m}[/tex]

So electric field increase by 80 [tex]\frac{V}{m}[/tex]

ACCESS MORE