Abstract
This is the first of a series of three papers devoted to a theoretical study of the effects of electron correlations on cyclotron phase resonance. In this paper we solve the equations governing the microwave field in a diffuse-surface semi-infinite interacting electron gas in which there is a steady magnetic field normal to the surface. The short-range interaction between electrons is assumed to be described by Landau's theory of the Fermi liquid, subject to the condition that only A1 is nonzero. We find that the microwave field inside the gas reaches its maximum intensity near ωcω=1, in spite of the fact that the nonlocal conductivity, which in many other situations seems to describe the transmitted signal, reaches its maximum at ωcω=(1+A1)1.