[Physics] Why does radiance remain constant along rays of light through empty space

opticsradiometryvisible-light

In Radiometry, radiance (L) is defined as flux density per unit area per unit solid angle. If we move further along a ray, away from a point, shouldn't the radiance decrease? I am unable to grasp an intuitive reason for the statement:

Radiance remain constant along rays of light through empty space.

Any explanation in this regard would be highly appreciated.

Best Answer

Imagine you start right next to a star. As you move away from the star, the intensity of the light $I$ (in $W/m^2$) goes down depending on distance $r$ following an inverse-square law:

$$I\ \alpha\ \frac{1}{r^2}$$

but the solid angle $\Omega$ (in $sr$) also decreases in the same proportion:

$$\Omega\ \alpha\ \frac{1}{r^2}$$

Therefore since radiance $L$ (in $W/m^2\cdot sr^{-1}$) is intensity per unit solid angle:

$$L\ \alpha\ \frac{I}{\Omega}$$

the $r^2$ cancels out and radiance does not depend on distance.