This is a historical translation of the seminal paper by H. Gamo, originally published in *Oyo Buturi* (*Applied Physics*, a journal of The Japan Society of Applied Physics) Vol. 25, pp. 431–443, 1956. English translation by Kenji Yamazoe, with further editing by the translator and Anthony Yen.

Since optical systems have distinctive features as compared to electrical communication systems, some formulation should be prepared for the optical image in order to use it in information theory of optical systems. In this paper the following formula for the intensity distribution of the image by an optical system having a given aperture constant α in the absence of both aberration and defect in focusing is obtained by considering the nature of illumination, namely coherent, partially coherent, and incoherent:

where*u*

_{n}(y) = sin 2πα/λ (y − nλ/2α) / 2πα/λ (y − nλ/2α) and

*a*

_{nm}= (2α/λ)

^{2}∬ Γ

_{12}(

*x*

_{1}−

*x*

_{2}) E(

*x*

_{1})

*E*

^{*}(

*x*

_{2}) | A(

*x*

_{1}) ||

*A*

^{*}(

*x*

_{2}) |

*u*

_{n}(

*x*

_{1})

*u*

_{m}(

*x*

_{2})d

*x*

_{1}d

*x*

_{2}.

*I*(

*y*) is the intensity of the image at a point of coordinate

*y*, Γ

_{12}the phase coherence factor introduced by H. H. Hopkins et al.,

*E*(

*x*) the complex transmission coefficient of the object and

*A*(

*x*) the complex amplitude of the incident waves at the object, and the integration is taken over the object plane. The above expression has some interesting features, namely the “intensity matrix” composed of the element

*a*

_{nm}mentioned above is a positive-definite Hermitian matrix, and the diagonal elements are given by the intensities sampled at every point of the image plane separated by a distance λ / 2α, and the trace of the matrix or the sum of diagonal elements is equal to the total intensity integrated over the image plane. Since a Hermitian matrix can be reduced to diagonal form by a unitary transformation, the intensity distribution of the image can be expressed as

## 1.

## Introduction

Discussion of the relationship between the object and the image of an optical system using information theory is a recent topic. Some basic studies such as describing the optical system using the response function have been done. However, entropy or noise is not discussed enough, though it is important in information theory. Information theory has been successfully applied to electrical communication, but the way it is applied is not directly applicable to an optical system. Hence, we need to formulate the information theory taking specific properties of the optical system into account.

The properties specific to the optical system are the following ones. (1) Observable quantity is not the amplitude of the wave but its intensity represented by the square of the amplitude; the intensity must be positive. (2) When we regard the optical system as a spatial filter, the phase and amplitude of the response function defined in the optical system are independent physical quantities; in the electric circuit, they are restricted by causality because they change over time. (3) Information contained in the image depends on the coherence of the illumination; for example, it is possible to obtain the information of amplitude and phase of coherent illumination; however, the phase information cannot be obtained when the illumination is incoherent. (4) If we regard the optical system as a communication channel, noise is an important quantity in deriving its capacity; in the optical system, noise can be stray light, disturbed observation due to random movement of the medium, graininess of a film, perception, and so on; these factors will add additional intensity to the nominal intensity or may partially reduce the nominal intensity; the resultant intensity will, however, not be negative. (5) Optical system behaves like a multi-dimensional spatial filter; in most cases, it is sufficiently described by a two-dimensional spatial filter.

These properties should be considered when we formulate information theory for the optical imaging system. If we do not take them into account, we cannot finalize the formulation of the information theory for the optical imaging system and may only get an insight that might help us understand the optical system.

In what follows, the author examines an aberration-free optical imaging system taking into account the first and second properties described above. In particular, the author will discuss how information contained in the image changes with respect to the third property, i.e., the coherence of illumination, using the sampling theorem that is often used in communication theory. Distribution of the image intensity will be described by a positive-definite Hermitian matrix, termed “intensity matrix,” and the physical meaning of the image will be revealed from the properties of the matrix. For this analysis, “phase coherence factor” introduced by H. H. Hopkins et al. will be used.

## 2.

## Sampling Theorem

“The sampling theorem” used hereafter is briefly reviewed. We assume an optical imaging system in Fig. 1 consisting of a light source, an object, lens, and an image.

Let $A(x-z)$ be an incident wave from a point $P$ on the source to a point $Q$ on the object plane. When the complex transmission of the object is $E(x)$, the complex amplitude of the wave after object is $E(x)A(x-z)$. As shown in Fig. 1, if we define $\theta $ as the maximum half-angle of a light cone to be captured by the lens and $\lambda $ to be the wavelength, then $2\text{\hspace{0.17em}}\mathrm{sin}\text{\hspace{0.17em}}\theta /\lambda $ corresponds to the role of bandwidth of the electrical communication system. If we Fourier transform the wave $E(x)A(x-z)$, we obtain

In Eq. (1), $X=\mathrm{sin}\text{\hspace{0.17em}}\theta /\lambda $ represents the directional cosine of the plane wave; thus, $f(X)$ is the amplitude of the plane wave that propagates with the angle of $\mathrm{sin}\text{\hspace{0.17em}}\theta =X\lambda $. Here, $f(X)$ that forms the image is band-limited due to the aperture of the optical system. Since $2\text{\hspace{0.17em}}\mathrm{sin}\text{\hspace{0.17em}}\theta /\lambda $ behaves like the bandwidth of the optical system, as briefly noted above,

where $\alpha =\mathrm{sin}\text{\hspace{0.17em}}\theta $. Since the bandwidth of $f(X)$ is limited to $2\alpha /\lambda $, it can be expanded in a Fourier series as## Eq. (3′)

$${a}_{n}(z)=\frac{\lambda}{2\alpha}{\int}_{-\frac{\alpha}{\lambda}}^{\frac{\alpha}{\lambda}}f(X,z){e}^{2\pi in(\lambda /2\alpha )X}\mathrm{d}X$$The amplitude of the wave on the image plane $F(y)$ is given by the inverse Fourier transform of $f(X)$, which has the bandwidth of $2\alpha /\lambda $. Thus, from Eq. (3)

## Eq. (4)

$$F(y,z)={\int}_{-\frac{\alpha}{\lambda}}^{+\frac{\alpha}{\lambda}}f(X,z){e}^{+2\pi iXy}\mathrm{d}X$$## Eq. (4′)

$$=\sum _{n=-\infty}^{\infty}{a}_{\mathit{n}}(z)\frac{\mathrm{sin}\text{\hspace{0.17em}}\frac{2\pi \alpha}{\lambda}(y-\frac{n\lambda}{2\alpha})}{\pi (y-\frac{n\lambda}{2\alpha})}$$On comparing Eq. (3′) and Eq. (4), we notice that ${a}_{\mathit{n}}(z)$ is obtained by multiplying $\lambda /2\alpha $ to $F(n\lambda /2\alpha )$which is the image amplitude at $y=n\lambda /2\alpha $. Thus,

## Eq. (5)

$$F(y,z)=\sum _{n=-\infty}^{\infty}F(\frac{n\lambda}{2\alpha},z)\frac{\mathrm{sin}\text{\hspace{0.17em}}\frac{2\pi \alpha}{\lambda}(y-\frac{n\lambda}{2\alpha})}{\frac{2\pi \alpha}{\lambda}(y-\frac{n\lambda}{2\alpha})}$$Therefore, the image amplitude obtained through the optical system with the bandwidth of $2\alpha /\lambda $ due to its numerical aperture can be determined if the image amplitude is sampled at an interval of $\lambda /2\alpha $. The function set that appears in Eq. (5), i. e., $\{{u}_{n}=\mathrm{sin}[(2\pi \alpha /\lambda )(y-n\lambda /2\alpha )]/[(2\pi \alpha /\lambda )(y-n\lambda /2\alpha )]\}$ forms a complete orthogonal system. Hence, with any integer $n$ and $m$,

## Eq. (5′)

$${\int}_{-\infty}^{+\infty}{u}_{n}(y){u}_{m}(y)\mathrm{d}y=\frac{\lambda}{2\alpha}{\delta}_{nm},$$Above discussion holds when the illumination is a point source; in other words, it only holds when the illumination is coherent. Even when the illumination is coherent, its physical meaning is revealed only when we can measure the phase and amplitude independently by some appropriate way, such as taking the phase difference. This is due to the first property noted in Sec. 1, which demands that the direct measurement can be made to intensity only. In the following, a method to analyze the physical meaning of the image intensity will be introduced, which can be applied even when the illumination is partially coherent or incoherent.

## 3.

## Intensity Matrix Related to the Phase Coherence Factor

We are able to obtain the image intensity distribution by Eq. (5) in Sec. 2. Letting $I(y)$ be the image intensity at a point $R$ on the image plane with coordinate $y$,

## Eq. (6)

$$I(y)=F(y,z){F}^{*}(y,z)\phantom{\rule{0ex}{0ex}}=\sum _{n}\sum _{m}F(\frac{n\lambda}{2\alpha},z){F}^{*}(\frac{m\lambda}{2\alpha},z){u}_{n}(y){u}_{m}^{*}(y)$$We now consider the image intensity from a generalized source of a finite extent with an arbitrary intensity at each point. Letting $J(z)$ be the point source intensity at a point $P$ with coordinate $z$, the image intensity at a point $R$ is given by

## Eq. (7)

$$I(y)=\stackrel{+\infty}{\sum _{n=-\infty}}\stackrel{+\infty}{\sum _{m=-\infty}}{a}_{nm}{u}_{n}(y){u}_{m}^{*}(y)$$## Eq. (7′)

$${a}_{nm}=\int J(z)F(\frac{n\lambda}{2\alpha},z){F}^{*}(\frac{m\lambda}{2\alpha},z)\mathrm{d}z$$Since point sources are mutually incoherent, the sum of intensities formed by all point sources is the total intensity. This fact can be explained from the statistical view point as follows. Intensity is found by taking the time average of the square of the amplitude of the light wave. Since each point source has no correlation with all others, this time average of the square of superimposed light waves is equal to the sum of time average of the square of each light wave.

The main feature of this paper will be based on Eq. (7) which has a quadratic form of variables ${u}_{n}$ and ${u}_{m}$ with coefficients ${a}_{nm}$. Since the intensity cannot be negative, this is a positive quadratic form. The matrix with elements ${a}_{nm}$ is termed “intensity matrix” in this paper. (Intensity matrix has the similar property to the information matrix proposed for general systems by D. M. MacKay.^{7} Although N. Wiener^{8} proposed the coherency matrix by assuming time varying light waves, it is not applied to an optical imaging system.) Once this matrix is given, we can determine the image intensity distribution. In this sense, the intensity matrix contains all the information of the image intensity distribution. In what follows, the physical property of the intensity matrix will be examined.

The element of the intensity matrix ${a}_{nm}$ is expressed using the Hopkins’ phase coherence factor. The use of the phase coherence factor is beneficial for our purpose because it has been studied in detail. First, Eq. (7′) is changed to (See Sec. 8.)

## Eq. (8)

$${a}_{nm}=\iiint J(z)E({x}_{1}){E}^{*}({x}_{2})A({x}_{1}-z){A}^{*}({x}_{2}-z)u({x}_{1}-\frac{n\lambda}{2\alpha}){u}^{*}({x}_{2}-\frac{m\lambda}{2\alpha})\mathrm{d}{x}_{1}\text{\hspace{0.17em}}\mathrm{d}{x}_{2}\text{\hspace{0.17em}}\mathrm{d}z,$$*α*: numerical aperture) and having unit amplitude when it is at point ${Q}_{1}$ or ${Q}_{2}$ on the object plane. As we assume an aberration-free optical system, it is

## Eq. (8′)

$$u({x}_{1}-\frac{n\lambda}{2\alpha})=\mathrm{sin}\text{\hspace{0.17em}}\frac{2\pi \alpha}{\lambda}({x}_{1}-\frac{n\lambda}{2\alpha})/\pi ({x}_{1}-\frac{n\lambda}{2\alpha})$$Equation (8) shows the relationship of ${a}_{nm}$ to the phase coherence factor. Inside Eq. (8), we single out the integral with respect to the source coordinate $z$, and set it as $I({x}_{1},{x}_{2})$. Then

This value can be regarded as a correlation between light wave amplitudes at points ${x}_{1}$ and ${x}_{2}$ on the object plane. If the amplitude at each point generated by a point source at $\mathit{z}$ are ${U}_{1}$ and ${U}_{2}$, we obtain

When the intensity at ${x}_{1}$ is ${I}_{1}$ and at ${x}_{2}$ is ${I}_{2}$,

Here, $\mathrm{\Gamma}({x}_{1}-{x}_{2})$ is the phase coherence factor which is by definition given by

## Eq. (10)

$$\mathrm{\Gamma}({x}_{1}-{x}_{2})=\frac{1}{\sqrt{{I}_{1}{I}_{2}}}\int {U}_{1}{U}_{2}^{*}\mathrm{d}z$$When the light source $\mathrm{\Sigma}$ illuminates the object, the intensity ${I}_{i}$ ($i=1$, 2) is the square of the absolute value of complex amplitude of the incident light at ${x}_{i}$, which is represented as ${|A({x}_{i})|}^{2}$. Therefore,

## Eq. (11)

$${a}_{nm}=\iint \mathrm{\Gamma}({x}_{1}-{x}_{2})E({x}_{1}){E}^{*}({x}_{2})|A({x}_{1})||{A}^{*}({x}_{2})|u({x}_{1}-\frac{n\lambda}{2\alpha}){u}^{*}({x}_{2}-\frac{m\lambda}{2\alpha})\mathrm{d}{x}_{1}\text{\hspace{0.17em}}\mathrm{d}{x}_{2},$$## Eq. (12)

$$\mathrm{\Gamma}({X}_{1}-{X}_{2},{Y}_{1}-{Y}_{2})=\frac{1}{\sqrt{{I}_{1}{I}_{2}}}\iint J(x,y){e}^{i[x({X}_{1}-{X}_{2})+y({Y}_{1}-{Y}_{2})]}\mathrm{d}x\text{\hspace{0.17em}}\mathrm{d}y,$$Equation (11) is our first formulation. We discuss a few examples of the intensity matrix $||{a}_{nm}||$ in the next section.

## 4.

## Examples of the Intensity Matrix

## 4.1.

### Coherent Source

When the light source is coherent, the phase coherence factor $\mathrm{\Gamma}({x}_{1}-{x}_{2})$ is 1 regardless of ${x}_{1}$ or ${x}_{2}$.^{2} Therefore, by substituting $\mathrm{\Gamma}({x}_{1}-{x}_{2})=1$ into Eq. (11), the integral can be separated into a product of two integrals to yield

## Eq. (13)

$${a}_{nm}=F\left(\frac{n\lambda}{2\alpha}\right){F}^{*}\left(\frac{m\lambda}{2\alpha}\right)$$The matrix element is a product of the complex amplitude on the image plane at $n\lambda /2\alpha $ and $m\lambda /2\alpha $. This result simply produces Eq. (6) of Sec. 3, but analyzing the meaning of the above matrix will give us an important insight for the remaining part of this paper. In this coherent source case, a remarkable feature of the matrix is that its rank is 1 and its single eigenvalue determines $\mathrm{\Sigma}{a}_{nn}$. The rank of a matrix is said to be $r$ when the determinant of sub-matrix with the order lower than or equal to $r$ is non-zero but the determinant of all sub-matrices with the order higher than $r$ is 0. If the matrix element is given by Eq. (13), we may notice that the determinant of all sub-matrices with the order higher than 1 is 0. The characteristic equation that gives the eigenvalue is

which is equivalent to ${x}^{\mathrm{N}}-(\mathrm{\Sigma}{a}_{nn}){x}^{\mathrm{N}-1}=0$. The single non-zero eigenvalue is equal to the trace of the matrix.## 4.2.

### Incoherent Source

In this case, the phase coherence factor $\mathrm{\Gamma}$ is

When this $\mathrm{\Gamma}$ is substituted into Eq. (11), the element ${a}_{nm}$ of the intensity matrix for the incoherent source is given by the single integral as

## Eq. (14)

$${a}_{nm}={\int}_{-\infty}^{+\infty}I(x)u(x-\frac{n\lambda}{2\alpha}){u}^{*}(x-\frac{m\lambda}{2\alpha})\mathrm{d}x$$Equation (14) is connected to the equation that gives the image $I(y)$ from an incoherent source with intensity $I(x)$, which can be expressed as an integral by the incoherent imaging formula

According to Eq. (14′), a single integral yields the image whereas the expression by the intensity matrix seems to involve more integrations which gives us an impression of extra cost of calculation. This point is discussed at the end.

The simplest example is an object with uniform brightness. If we set $I(x)=A$ where $A$ is a constant

## Eq. (15)

$${a}_{nm}=A{\int}_{-\infty}^{+\infty}u(x-\frac{n\lambda}{2\alpha}){u}^{*}(x-\frac{m\lambda}{2\alpha})\mathrm{d}x\phantom{\rule{0ex}{0ex}}=\frac{2\alpha}{\lambda}A{\delta}_{nm}\phantom{\rule[-0.0ex]{1em}{0.0ex}}({\delta}_{nm}=1,n=m;{\delta}_{nm}=0,n\ne m)$$[The intensities *I*(*y*) in Eq. (14′) and *I*(*y*) from Eq. (15) must be the same (See Sec. 9).] Our next example is simple and basic, in which the object has a sinusoidal amplitude over the object plane. The brightness on the object plane is

In this case, the element of the intensity matrix is represented by the following integral

## Eq. (16)

$${a}_{nm}=\frac{1}{2}A{\int}_{-\infty}^{+\infty}(1+\mathrm{cos}\text{\hspace{0.17em}}2\omega x)\frac{\mathrm{sin}\text{\hspace{0.17em}}\frac{2\pi \alpha}{\lambda}(x-\frac{n\lambda}{2\alpha})\mathrm{sin}\frac{2\pi \alpha}{\lambda}(x-\frac{m\lambda}{2\alpha})}{\pi (x-\frac{n\lambda}{2\alpha})\pi (x-\frac{m\lambda}{2\alpha})}\mathrm{d}x$$We now change the valuables. Letting $\xi =2\pi \alpha x/\lambda $ and $p=\lambda \omega /2\pi \alpha $,

This type of definite integral will appear in the following but since the result may not be found in standard integral tables, the principal result together with its derivation is listed in Sec. 10. According to Sec. 10, when $0<p<1$, i.e., $0<\omega <2\pi \alpha /\lambda $,

## Eq. (16′)

$${a}_{nm}=A\frac{\alpha}{\lambda}\frac{\mathrm{sin}(n-m)\pi (1-p)}{(n-m)\pi}\mathrm{cos}\text{\hspace{0.17em}}(\mathit{n}+\mathit{m})\text{\hspace{0.17em}}\mathit{p}\mathit{\pi}\text{\hspace{0.17em}}\mathit{n}\ne \mathit{m}$$## Eq. (16″)

$${a}_{nm}=\frac{A}{2}[1+(1-p)\mathrm{cos}\text{\hspace{0.17em}}2np\pi ]\phantom{\rule[-0.0ex]{1em}{0.0ex}}n=m$$When $p>1$ or equivalently $\omega >2\pi \alpha /\lambda $, ${a}_{nm}$ is 0. Based on this result in which ${a}_{nm}$ is 0 above a certain spatial frequency threshold, the response function for the image intensity in Eq. (14) is of triangular shape^{1} with 0 value over the bandwidth of $4\pi \alpha /\lambda $.

We now consider the rank and eigenvalues of the intensity matrix. First, when the object brightness is uniform, the intensity matrix in Eq. (15) is a diagonal matrix with non-zero diagonal elements. Next, when the object brightness changes sinusoidally, especially when $p=1/2$ or equivalently $\omega =\alpha /\lambda $, the intensity matrix is also a diagonal matrix with the diagonal element of $(1/2)A(\lambda /2\alpha )(1+1/2{(-1)}^{n})$. These examples show that incoherent illumination gives the greatest matrix rank in contrast to coherent illumination which gives the minimum matrix rank of 1. (When the object size is finite, the object intensity or complex transmittance can be represented by a Fourier series; hence we may be able to apply the results of Secs. 4.2 and 4.3 to each Fourier term.)

## 4.3.

### Partially Coherent Source

We consider the one-dimensional case, under which the phase coherence factor is defined in Eq. (12) as

## Eq. (17)

$$\mathrm{\Gamma}({x}_{1}-{x}_{2})=\frac{\mathrm{sin}\text{\hspace{0.17em}}\frac{2\pi \alpha}{\lambda}S({x}_{1}-{x}_{2})}{\frac{2\pi \alpha}{\lambda}S({x}_{1}-{x}_{2})},$$Let us consider the intensity matrix element for an image formed by uniform object transmittance. Substituting $\mathrm{\Gamma}({x}_{1}-{x}_{2})$ in Eq. (17) into Eq. (11), and furthermore letting $E(x)A(x)=K$,

## Eq. (18)

$${a}_{nm}=\frac{{K}^{2}}{{\pi}^{2}}{\int}_{-\infty}^{+\infty}{\int}_{-\infty}^{+\infty}\frac{\mathrm{sin}\text{\hspace{0.17em}}S(\xi -\eta )}{S(\xi -\eta )}\frac{\mathrm{sin}(\xi -n\pi )}{(\xi -n\pi )}\frac{\mathrm{sin}(\eta -m\pi )}{(\eta -m\pi )}\mathrm{d}\xi \text{\hspace{0.17em}}\mathrm{d}\eta ,$$## Eq. (18″)

$$S<1\phantom{\rule[-0.0ex]{1em}{0.0ex}}{a}_{nm}={K}^{2}\frac{\mathrm{sin}\text{\hspace{0.17em}}S(n-m)\pi}{S(n-m)\pi}$$Next, we consider an object with a sinusoidal transmittance. Substituting $\mathrm{\Gamma}({x}_{1}-{x}_{2})$ in Eq. (17) into Eq. (11), and furthermore letting $E(x)A(x)=K\text{\hspace{0.17em}}\mathrm{cos}\text{\hspace{0.17em}}\omega x$,

## Eq. (19)

$${a}_{nm}=\frac{{K}^{2}}{{\pi}^{2}}{\int}_{-\infty}^{+\infty}{\int}_{-\infty}^{+\infty}\mathrm{cos}\text{\hspace{0.17em}}p\xi \text{\hspace{0.17em}}\mathrm{cos}\text{\hspace{0.17em}}p\eta \frac{\mathrm{sin}\text{\hspace{0.17em}}S(\xi -\eta )}{S(\xi -\eta )}\frac{\mathrm{sin}(\xi -n\pi )}{(\xi -n\pi )}\frac{\mathrm{sin}(\eta -m\pi )}{(\eta -m\pi )}\mathrm{d}\xi \text{\hspace{0.17em}}\mathrm{d}\eta ,$$## Eq. (20‐1)

$$(i)\text{\hspace{0.17em}\hspace{0.17em}}p>(1+S)\phantom{\rule[-0.0ex]{1em}{0.0ex}}and\phantom{\rule[-0.0ex]{1em}{0.0ex}}(S>1\text{\hspace{0.17em}\hspace{0.17em}}or\text{\hspace{0.17em}\hspace{0.17em}}S<1),\phantom{\rule[-0.0ex]{1em}{0.0ex}}{a}_{nm}=0$$## Eq. (20-2)

$$(ii)\text{\hspace{0.17em}\hspace{0.17em}}(1+S)<p<1\phantom{\rule[-0.0ex]{1em}{0.0ex}}and\phantom{\rule[-0.0ex]{1em}{0.0ex}}(S>1\text{\hspace{0.17em}\hspace{0.17em}}or\text{\hspace{0.17em}\hspace{0.17em}}S<1),\phantom{\rule[-0.0ex]{1em}{0.0ex}}\{\begin{array}{l}{a}_{nm}=\frac{{K}^{2}}{2S}\frac{A}{(n-m)\pi}\text{\hspace{0.17em}\hspace{0.17em}}(n\ne m)\\ {a}_{n\mathit{n}}=\frac{{K}^{2}}{2S}(1+S-p)\end{array}$$## Eq. (20-3)

$$1>p>(S-1)\text{\hspace{0.17em}\hspace{0.17em}}or\text{\hspace{0.17em}\hspace{0.17em}}(1-S)\phantom{\rule[-0.0ex]{1em}{0.0ex}}and\phantom{\rule[-0.0ex]{1em}{0.0ex}}(S>1\text{\hspace{0.17em}\hspace{0.17em}}or\text{\hspace{0.17em}\hspace{0.17em}}S<1),\phantom{\rule[-0.0ex]{1em}{0.0ex}}\{\begin{array}{l}{a}_{nm}=\frac{{K}^{2}}{2S}\frac{A+\mathit{B}}{(n-m)\pi}\text{\hspace{0.17em}\hspace{0.17em}}(n\ne m)\\ {a}_{n\mathit{n}}=\frac{{K}^{2}}{2S}[1+S-p+2(1-p)\mathrm{cos}\text{\hspace{0.17em}}2np\pi ]\end{array}$$## Eq. (20‐4)

$$S>1\phantom{\rule[-0.0ex]{1em}{0.0ex}}and\phantom{\rule[-0.0ex]{1em}{0.0ex}}1>p>0,\phantom{\rule[-0.0ex]{1em}{0.0ex}}\hspace{1em}\begin{array}{l}{a}_{nm}=\frac{{K}^{2}}{\mathrm{4}S}\frac{\mathrm{sin}(n-m)(1-p)\pi}{(n-m)\pi}\text{\hspace{0.17em}}\mathrm{cos}(n+m)p\pi \text{\hspace{0.17em}\hspace{0.17em}}(n\ne m)\\ {a}_{n\mathit{n}}=\frac{{K}^{2}}{4S}[1+(1-p)\mathrm{cos}\text{\hspace{0.17em}}2np\pi ]\end{array}\}$$## Eq. (20-5)

$$(iv)\text{\hspace{0.17em}\hspace{0.17em}}S<1\phantom{\rule[-0.0ex]{1em}{0.0ex}}and\phantom{\rule[-0.0ex]{1em}{0.0ex}}(1-S)>p>0,\hspace{1em}\begin{array}{l}{a}_{nm}={K}^{2}\text{\hspace{0.17em}}{\mathrm{cos}}^{2}\text{\hspace{0.17em}}p\eta \pi \frac{\mathrm{sin}\text{\hspace{0.17em}}S(n-m)\pi}{S(n-m)\pi}\text{\hspace{0.17em}\hspace{0.17em}}(n\ne m)\\ {a}_{n\mathit{n}}={K}^{2}\text{\hspace{0.17em}}{\mathrm{cos}}^{2}\text{\hspace{0.17em}}p\eta \pi \end{array}\}$$We may be aware of a few interesting points from the above example of partially coherent illumination. First, for incoherent illumination, ${a}_{nm}=0$ when $p>1$ or $\omega >2\pi \alpha /\lambda $, leading to no contribution to the image. Whereas for partially coherent illumination, ${a}_{nm}=0$ when $p>(1+\mathrm{S})$ or $\omega >(1+\mathrm{S})2\pi \alpha /\lambda $, meaning more non-zero matrix elements than incoherent illumination. Second, if $\mathrm{S}>1$ and $(\mathrm{S}-1)>p>1$ as in case (iii), we will obtain the same result as that of incoherent illumination. Lastly, if $\mathrm{S}<1$ and $(1-\mathrm{S})>p>0$ as in case (iv), the result reduces to the coherent case in the limit of $\mathrm{S}\to 0$. Deriving the rank of the intensity matrix with the elements ${a}_{nm}$ given above in order to calculate its eigenvalue is an extremely difficult problem, except in some special cases. This fact implies that the image formation with the intensity matrix may be limited in practice. However, since the intensity matrix itself has interesting general properties, it can be useful in clarifying and organizing the physical concept of optical image formation.

Up to this point, we have considered an object of sinusoidal transmittance or brightness. If the object has an arbitrary distribution of transmittance or brightness, its transmittance or brightness can be expressed as a sum of periodic terms using the Fourier integral. Since each term can be treated as is explained above, we can obtain the elements of the intensity matrix when the object has an arbitrary transmittance or brightness distribution.

## 5.

## General Properties of the Intensity Matrix

The intensity matrix was derived by squaring the amplitude obtained from the sampling theorem applied to a band-limited system. In electrical communication, since the square of the amplitude corresponds to the electric power, we may be able to derive an equation similar to the intensity matrix. However, this idea is not as crucial as the intensity matrix in an optical imaging system because the phase coherence factor explained in Sec. 1 is a unique feature only for an optical imaging system.

For detailed discussion in the following, Eq. (7) is repeated here. The intensity $I(y)$ is

## Eq. (7)

$$I(y)=\stackrel{+\infty}{\sum _{n=-\infty}\text{\hspace{0.17em}}}\stackrel{+\infty}{\sum _{m=-\infty}}{a}_{nm}{u}_{n}(y){u}_{m}^{*}(y),$$If ${u}_{1},{u}_{2},\dots ,{u}_{\mathit{n}},\dots $ are regarded as vectors in a multi-dimensional space, the linear sum $\mathrm{\Sigma}{a}_{\mathit{nm}}$ ${u}_{\mathit{n}\mathit{m}}={v}_{\mathit{n}}$ is another vector. The intensity is then given by the inner product of these two vectors

where $\varphi $ is a vector $\{{u}_{1},{u}_{2},\dots ,{u}_{\mathit{n}},\dots \}$ and $\mathbf{A}$ is a matrix with elements ${a}_{\mathit{nm}}$.Degrees of freedom for the image can be equated with the dimension number $N$ in the multi-dimensional space introduced above. Degrees of freedom $N$ is exactly the number of the sampling points used to express the image under consideration. If the area of the image is $S$, $N=4{\alpha}^{2}S/{\lambda}^{2}$. Toraldo di Francia defined the degrees of freedom when the source is coherent, which is $N$ defined above. However, Toraldo di Francia defined another definition when the source is incoherent. In comparison, it seems mathematically consistent, regardless of the coherence of the source, to define the degrees of freedom as the number of dimensions to be used to determine the image. The source coherence is contained in the intensity matrix. Therefore, one can separate the number of the dimensions determined by the numerical aperture of the imaging system from the source coherence.

Six important physical properties of the intensity matrix will be discussed next. The first physical property is related to the fact that the only observable quantity is the intensity, which must be real and positive. This fundamental fact leads to the following mathematical property:

If a matrix is a Hermitian matrix, its elements have the following property

which can be verified by putting $I(y)={\mathit{I}}^{*}(y)$ in Eq. (7). The matrix is positive-definite because Eq. (7) is always positive regardless of $y$. (To satisfy this condition, all the principal sub-matrices of matrix $||{a}_{nm}||$ need to have positive determinants.) We can now utilize the mathematical knowledge on positive-definite Hermitian matrices to investigate the intensity matrix for the image.Next, let us consider the direct relationship between the elements of the intensity matrix and the observable physical quantity. The first relationship is:

ii. Diagonal elements of the intensity matrix are equal to the intensities sampled at the interval of $\lambda /2\alpha $, and the trace of the intensity matrix is equal to the integrated image intensity over the image plane.

The first half of the statement can be understood from Eq. (7) in which ${u}_{\mathit{n}}(y)$ is 1 at the sampling point of $y=n\lambda /2\alpha $ and zero at other sampling points. For the second half of the statement, the integrated image is obtained by taking the integral all over the image plane; therefore it is proved by integrating Eq. (7) with reference to Eq. (5′),

## Eq. (23)

$${I}_{0}=\int I(y)\mathit{d}y=\frac{\lambda}{2\alpha}\left(\sum _{n=-\infty}^{+\infty}{a}_{n\mathit{n}}\right),$$The source coherence is contained in the property of the intensity matrix, which will be more explicitly expressed by matrix diagonalization with unitary transformation. If intensity $I$ is expressed as the inner product of vectors as in Eq. (21), let $\varphi $ be expressed by basis vectors ${\varphi}_{1},{\varphi}_{2},{\varphi}_{3},\dots $. These vectors are orthogonal, for example, ${\varphi}_{1}=(\mathrm{1,0},0,\dots )$, ${\varphi}_{2}=(\mathrm{0,1},0,\dots )$, and so on. With these basic vectors, vector $\varphi $ is given by

Now, orthogonal transformation of the basis will change Eq. (21) into the simplest form. The transformed vectors must be eigenvectors of the intensity matrix. That is, the eigenvectors ${\psi}_{1},{\psi}_{2},\dots {\psi}_{\mathrm{i}},\dots $ will satisfy the following condition

where ${\lambda}_{\mathit{i}}$ is a constant called an eigenvalue and it is proved that a Hermitian matrix always has real eigenvalues. Let the components of the eigenvector $\psi $ be (${\mathrm{S}}_{1},{\mathrm{S}}_{2},{\mathrm{S}}_{3},\dots $), the equation that determines the eigenvalue and eigenvector, $\mathbf{A}\psi =\lambda \psi $, is then written as## Eq. (26)

$$\begin{array}{l}({a}_{11}-\lambda ){S}_{1}+{a}_{12}{S}_{2}+{a}_{13}{S}_{3}+\cdots +{a}_{1n}{S}_{n}=0\\ {a}_{21}{S}_{1}+({a}_{22}-\lambda ){S}_{2}+{a}_{23}{S}_{3}+\cdots +{a}_{2n}{S}_{n}=0\\ \vdots \\ {a}_{n1}{S}_{1}+{a}_{n2}{S}_{2}+{a}_{n3}{S}_{3}+\cdots +({a}_{nn}-\lambda ){S}_{n}=0\end{array}\}$$For ${S}_{\mathrm{i}}$ to have non-trivial solutions, the determinant of the matrix formed by the coefficients has to be zero. The resulting equation is called the characteristic equation whose solutions are eigenvalues ${\lambda}_{1},{\lambda}_{2},\dots $. If each eigenvalue is substituted into Eq. (26), we will have a set of simultaneous equations whose solution is the corresponding eigenvector to the inserted eigenvalue. By letting the components of eigenvector ${\psi}_{\mathrm{i}}$ corresponding to eigenvalue ${\lambda}_{\mathit{i}}$ be (${\mathrm{S}}_{1\mathit{i}},{\mathrm{S}}_{2\mathit{i}},{\mathrm{S}}_{3\mathit{i}},\dots ,{\mathrm{S}}_{\mathit{ni}}$), the square of $|{\psi}_{\mathit{i}}|$ is given by $({\psi}_{\mathit{i}},{\psi}_{\mathit{i}})={S}_{1\mathit{i}}{S}_{1\mathit{i}}^{*}+{S}_{2\mathit{i}}{S}_{2\mathit{i}}^{*}+\dots +{S}_{\mathit{ni}}{S}_{\mathit{ni}}^{*}$. Hereinafter, we normalize the square of $|{\psi}_{\mathit{i}}|$ to be 1. In addition, it is proved that the eigenvectors are mutually orthogonal.^{5} Therefore, the orthonormal condition can be written as

## Eq. (27)

$$({\psi}_{i},{\psi}_{i})={S}_{1i}^{*}{S}_{1j}+{S}_{2i}^{*}{S}_{2j}+\cdots +{S}_{ni}^{*}{S}_{nj}={\delta}_{ij}$$Let us return to our original intention. Suppose that the original basis vectors ${\varphi}_{1},{\varphi}_{2},\dots ,{\varphi}_{\mathit{n}}$ are transformed to eigenvectors ${\psi}_{1},{\psi}_{2},\dots {\psi}_{\mathit{n}}$. Then, vector $\varphi =({u}_{1},{u}_{2},\dots ,{u}_{\mathit{n}})$ can be represented by a new vector $\psi =({\xi}_{1},{\xi}_{2},\dots {\xi}_{\mathit{n}})$, i.e.,

If we denote it by the vector components

we will obtain where matrix $\mathbf{S}$ has ${S}_{\mathit{ik}}$ as its $i$-th row $k$-th column element.The coefficients in Eq. (28), i.e., the vector components obtained by transforming vector $\varphi $ satisfy the following condition due to the orthogonality in Eq. (27)

If Eq. (31) is written explicitly,

where ${S}_{\mathit{ki}}^{*}$ is the*i*-th row and

*k*-th column element of the inverse transformation matrix, which is given by transposing matrix $\mathbf{S}$ in Eq. (29) followed by taking the complex conjugate of each component. Here, we denote the transpose of a matrix by the single quotation mark ’ and write the transforming matrix in Eq. (31) as ${{\mathbf{S}}^{\prime}}^{*}$. Therefore,

The square of vector $\varphi $ is

As vector ${u}_{i}$ is expressed in Eq. (29),

Because of the orthogonal property, we obtain $(\varphi ,\varphi )=\mathrm{\Sigma}{|{\xi}_{k}|}^{2}$. Thus, the matrix transformation used here preserves the square of the absolute value (norm) is not changed. This is proved by Eq. (27) in conjunction with Eq. (29). The constant norm property with Eq. (31′) proves

Equations (27) and (34) can be simplified with an identity matrix $\mathbf{E}$ as

In general, transformation that satisfies Eq. (35) is called a unitary transformation and matrix $\mathbf{S}$ is called a unitary matrix.

With the unitary transformation by matrix $\mathbf{S}$, we can examine how the image in Eqs. (7) or (21) is transformed. As in Eq. (28), vector $\varphi $ is written by the linear sum of transformed basic vectors ${\psi}_{\mathit{i}}$. Then,

If we use the property of eigenvalue and eigenvector in Eq. (25) and the orthogonality in Eq. (27), we obtain

Since Eqs. (31) or (31′) shows the explicit form of ${\xi}_{\mathit{i}}$,

## Eq. (37)

$$I=\sum {\lambda}_{i}{|({\psi}_{i},\varphi )|}^{2}=\sum _{i}{\lambda}_{i}{\left|\sum _{k}{S}_{ki}^{*}{u}_{k}\right|}^{2}$$If this result is written in the matrix form,

From Eq. (37), ${{\mathbf{S}}^{\prime}}^{*}\mathbf{AS}$ results in a diagonal matrix in which off-diagonal elements are zero. Letting the diagonal matrix be $\mathbf{D}$,

Diagonal elements of the diagonal matrix are eigenvalues ${\lambda}_{1},{\lambda}_{2},\dots ,{\lambda}_{\mathit{n}}$.

Next, because of Eqs. (35) and (38)

or if Eq. (39) is written using the matrix elementsAmong the results shown above, Eqs. (37) and (40) are useful in revealing the physical properties of the intensity matrix. Thus, Eqs. (37) and (40) will be examined in more detail.

iii. The image intensity is given by eigenvalues ${\lambda}_{\mathit{i}}$ and eigenfunctions ${\psi}_{\mathit{i}}$ (${S}_{1\mathit{i}},{S}_{2\mathit{i}},\dots ,{S}_{\mathit{ni}}$) of the intensity matrix $\mathbf{A}$ as

When the light source is coherent, only one eigenvalue is non-zero. Therefore

where $\sqrt{{\lambda}_{0}{S}_{k}}$ corresponds to $F(\mathit{k}\lambda /2\alpha )$ in Eq. (5). When the light source is partially coherent, there are more than one non-zero eigenvalues in Eq. (37), in which the eigenfunction associated with the greatest eigenvalue contributes to the image formation most significantly. When the light source is incoherent, each eigenfunction contributes almost equally to image formation. From Sec. 4.2, in special cases where the object brightness is uniform or object transmittance changes sinusoidally across the object plane, we can obtain Eq. (37) without the unitary transformation. When the object brightness is uniform, all eigenvalues have the same value.As a result, analyzing how the eigenvalue distributes reveals the degree of coherence of the light source, which is involved in image formation. When eigenvalues are ${\lambda}_{\mathit{i}}$, the sum of all eigenvalues ${\mathrm{I}}_{0}$ is equal to the trace of the intensity matrix and also equal to the integrated image intensity on the image plane. Let us consider the following quantity

If the light source is coherent, we always obtain $d=0$ and if the light source is incoherent with uniform intensity, $d$ takes the maximum value ${d}_{0}$

where $N$ is the degree of freedom of the image. For one-dimensional images with a length of $L$, $N=2\alpha L/\lambda $ and for two-dimensional images with an area of $S$, $N=4{\alpha}^{2}S/{\lambda}^{2}$. If the image area becomes infinitely large, $N$ will also be infinitely large. For this case, we may consider $d/{d}_{0}$ and take the limit with an infinitely large $N$. Therefore we define another quantity $\delta $ asThe value of $\delta $ changes continuously from 1 to 0 as the light source gradually changes its coherent state from coherent to incoherent. The quantity $\delta $ is an interesting quantity because we are able to measure coherence with $\delta $. Therefore, we define the degree of coherence using $\delta $. (The author noted in Ref. ^{9} that $\delta $ can be used for an object with uniform transmittance or brightness. However, according to Eq. (18′), $\delta $ from a light source with $\mathrm{S}>1$ is the same as $\delta $ from incoherent source. We may need to investigate this point more carefully.)

According to von Neumann,^{6} $-\sum {\lambda}_{n}\text{\hspace{0.17em}}\mathrm{log}\text{\hspace{0.17em}}{\lambda}_{n}$ is given with the matrix $\mathbf{A}$ as

## Eq. (43)

$$-\sum {\lambda}_{n}\text{\hspace{0.17em}}\mathrm{log}\text{\hspace{0.17em}}{\lambda}_{n}=-\text{Trace}(\mathbf{A}\text{\hspace{0.17em}}\mathrm{log}\text{\hspace{0.17em}}\mathbf{A}),$$Next, let us summarize the result of Eq. (40).

iv. A positive definite Hermite matrix of order $N$, in our case the element of intensity matrix ${a}_{\mathit{nm}}$, is given by $N$ positive real numbers ${\lambda}_{\mathit{i}}$ together with $N$ orthonormal vectors ${\psi}_{\mathit{i}}$ (${S}_{1\mathit{i}},{S}_{2\mathit{i}},\dots ,{S}_{\mathit{ni}}$) as

The number of the variables to define the intensity matrix is determined by the $N$ eigenvalues and $N$ eigenvectors. The vector component ${S}_{\mathit{ij}}$ is in general complex, so that it can be represented by ${S}_{\mathit{ij}}={r}_{\mathit{ij}}\text{\hspace{0.17em}}\mathrm{exp}({\theta}_{\mathit{ij}})$. The number of the variables that determines the vector components is $2{N}^{2}$. As a result, the number of the variables that determines the intensity matrix of order $N$ is $(2N+1)N$. Here, Eq. (27) defines the orthonormal condition, which is a set of

equations. Hence,Note that the simplest case for the intensity matrix is obtained when the light source is coherent because we may set the eigenvalues 0 except the single eigenvalue ${\lambda}_{0}$. In this case, only one vector decides the matrix elements, so that the number of independent variables is

The number of independent variables is a fundamental quantity of the intensity matrix that depends on the coherence of the light source. However, the number of independent variables is not always meaningful for image intensity as is discussed next. Rather, the image intensity is determined by $2N$ sampling points where $N$ is the number of sampling points for a coherent source. If the $2N$ sampled values could be arbitrary, the situation would be simpler; however, they need to be the sampled image intensity expressed in Eq. (7), which is obtained by the intensity matrix. Equation (7) introduces correlation among $2N$ sampled values. Thus, the property of the intensity matrix explained here is necessary to derive the amount of information of the image intensity, which will be explained elsewhere.

vi. Regardless of the coherence of the light source, image intensity $I(y)$ in Eq. (7) is determined by the sampled values with a sampling step of $\lambda /4\alpha $, and the following relationship between the sampled values and the intensity matrix is derived

where## Eq. (46)

$$I(y)=\sum _{k}I\left(\frac{k\lambda}{4\alpha}\right)\frac{\mathrm{sin}\text{\hspace{0.17em}}\frac{4\pi \alpha}{\lambda}(y-\frac{k\lambda}{4\alpha})}{\frac{4\pi \alpha}{\lambda}(y-\frac{k\lambda}{4\alpha})}$$$$I\left(\frac{k\lambda}{4\alpha}\right)=\sum _{n}\sum _{m}{a}_{nm}{u}_{u}\left(\frac{k\lambda}{4\alpha}\right){u}_{m}\left(\frac{k\lambda}{4\alpha}\right)$$$$k=2p:\text{\hspace{0.17em}\hspace{0.17em}}{u}_{u}(2p\lambda /4\alpha )=0,\phantom{\rule[-0.0ex]{1em}{0.0ex}}p\ne n\phantom{\rule{0ex}{0ex}}{u}_{u}(2p\lambda /4\alpha )=1,\phantom{\rule[-0.0ex]{1em}{0.0ex}}p=n\phantom{\rule{0ex}{0ex}}k=2p+1:\text{\hspace{0.17em}\hspace{0.17em}}{u}_{u}((2p+1)\lambda /4\alpha )={(-1)}^{p-n}/\{2(p-n)+1\}$$

The sampling step is derived by the Fourier transform of the image $I(y)$ in Eq. (7), in which the Fourier transform of ${u}_{\mathit{n}}(y){u}_{\mathit{m}}(y)$ has twice as wide as the bandwidth of the Fourier transform of ${u}_{\mathit{n}}(y)$. Outside of the bandwidth, the Fourier transform of $I(y)$ is 0. The equation is explicitly shown below by referring to the appendix

## Eq. (47)

$${I}_{nm}(\omega )={\int}_{-\infty}^{+\infty}\frac{\mathrm{sin}(\xi -n\pi )}{(\xi -n\pi )}\frac{\mathrm{sin}(\xi -m\pi )}{(\xi -m\pi )}{e}^{-\mathit{i}\omega \xi}\mathrm{d}\xi $$If $\omega >2$ or $\omega <-2$,

If $0<\omega <2$,

If $0>\omega >-2$,

When $n=m$, we will obtain the following.

If $|\omega |>2$,

If $0<\omega <2$,

If $0>\omega >-2$,

Since the bandwidth of $I(\omega )$ is limited to $4\alpha /\lambda $, the sampling theorem in Sec. 2 leads to the series expansion of Eq. (46).

In Eq. (46), rewriting the $k$-th sampled value as ${B}_{\mathrm{k}}$,

## Eq. (48)

$${B}_{2p}={a}_{pp},\phantom{\rule{0ex}{0ex}}{B}_{2p+1}=\sum _{k}\sum _{r}\frac{4}{{\pi}^{2}}{a}_{kr}{(-1)}^{k+r}/\{2(p-k)+1\}\{2(p-r)+1\}$$Also, the integrated intensity ${I}_{0}$ is rewritten using Eq. (23) as

## 6.

## Closing Remarks

The results presented in this paper are an expanded version of the presentation given at a symposium on “Application of the information theory to optics” organized by The Japan Society of Applied Physics on April 6, 1956. This paper explains the physical meaning of the intensity matrix introduced here. This work is just the beginning; we may need to work more on, for example, the application of the intensity matrix to calculating the amount of information, extension of the intensity matrix to two-dimensional imaging, evaluation of the change of the intensity matrix by phase difference, examination of the intensity matrix with aberration, and so on. For these cases, the author wishes to present at another opportunity.

In this paper, the discussion is limited to finite dimensions; however, for completeness of the study, it should be discussed with infinite dimensions. Therefore, we need to utilize the concept of Hilbert space, but the author was not able to reach this point in this short paper.

The author thanks Professor Hidetoshi Takahashi of the Department of Physics of the University of Tokyo and Associate Professor Kazuo Miyake of Tokyo University of Education for helpful discussions and Professor Hiroshi Kubota of the Institute of Industrial Science of the University of Tokyo, who encouraged the author to apply the information theory to optics and provided the related references.

## 7.

## Addendum

Since the submission of the manuscript, the author has noticed that D. Gabor of the United Kingdom independently proposed how to generally express the image by a Hermitian matrix. [Information Theory, Third London Symposium, edited by Colin Cherry (1956, Butterworths Scientific Publications) 4. Optical Transmission by D. Gabor, pp. 26-33.]

Although the concrete derivation differs from the one introduced in this paper, the author appreciates that he reached fundamentally the same conclusion. Those interested may read his paper together with this paper. The author mentioned in this paper that the intensity matrix for a generalized case would be a very difficult problem. However, since then, the author was able to conclude that the intensity matrix can be generalized by matrix transformation of the object or can be defined with the existence of aberration in the imaging optics, and report these findings in a physics seminar at the University of Tokyo. The details will be published elsewhere. These are not exactly pointed out by Gabor. However, our concepts are in agreement.

## 8.

## Appendix 1

Proof of Eq. (8).

Let the object complex transmittance be $E(x)$ and complex amplitude of the incident light be $A(x-z)$, the Fourier transform of the light amplitude at the object plane $E(x)A(x-z)$ is given by Eq. (1),

On the other hand, the light amplitude in the image plane $F(y)$ is given by Eq. (4)

## Eq. (4)

$$F(y,z)={\int}_{-\frac{\alpha}{\lambda}}^{+\frac{\alpha}{\lambda}}f(X,z){e}^{+2\pi iXy}\mathrm{d}X$$Substituting Eq. (1) into Eq. (4) followed by changing the order of integration gives

## Eq. (3)

$$F(y,z)={\int}_{-\infty}^{+\infty}E(x)A(x-z)\frac{\mathrm{sin}\text{\hspace{0.17em}}\frac{2\pi \alpha}{\lambda}(y-x)}{\pi (y-x)}\mathrm{d}x$$We obtain $F(\frac{n\lambda}{2\alpha},z)$ and $F(\frac{m\lambda}{2\alpha},z)$ by substituting $y=\frac{n\lambda}{2\alpha}$, $\frac{m\lambda}{2\alpha}$ into above equation, and insert them into Eq. (7′) to obtain Eq. (8).

## 9.

## Appendix 2

## 9.1.

### Incoherent Illumination

Substituting the intensity matrix of Eq. (15) into Eq. (7) gives

This result leads to the requirement of $\sum _{n=-\infty}^{+\infty}\frac{{\mathrm{sin}}^{2}(\eta -n\pi )}{{(\eta -n\pi )}^{2}}=1$. This relationship can be obtained by the series expansion of $1/{\mathrm{sin}}^{2}\text{\hspace{0.17em}}\eta $. [See, for example, Magnus, Overhettinger, Formeln und Sätze für die speziellen Functionen der mathematischen Physik (Springer 1948) p.215.] Similarly, if the object brightness changes periodically, substituting Eq. (16) into Eq. (7) yields

On the other hand, if $I(\xi )$ is derived by Eq. (14),

Since the above two equations are identical, we obtain the one we need.

## 9.2.

### Partially Coherent Illumination

If the object transmittance is uniform, the image intensity is obtained by putting ${a}_{\mathit{n}\mathit{m}}$ from Eqs. (18′) and (18″) into Eq. (7)

If the illumination is incoherent, we have Eq. (14′) and when the illumination is partially coherent, we have

If we set $E(x)A(x)=K$ and substitute $u(x-y)$ in Eq. (8′) and $\mathrm{\Gamma}({x}_{1}-{x}_{2})$ in Eq. (17) into above equation, we will obtain the following equation with a help of the integration in Eq. (18)

Therefore, we obtain

## 10.

## Appendix 3

## 10.1.

If we set a variable $\frac{2\pi \alpha}{\lambda}y=\xi $,

The integration can be decomposed into four terms each of which has the following result

To arrive at the above result, we have to choose an integration path such that $\mathrm{exp}(iax)$ is cancelled by a loop with an infinite radius. In addition, this integral has a pole on the real axis so that we have to take the Cauchy’s principal value. Note that if the residue on the real axis is ${R}_{0}$, the result of the complex integral is affected by $2\pi i(\mathrm{\Sigma}R+1/2\mathrm{\Sigma}{R}_{0})$, where $R$ is the residue inside the integration path. [Whittaker, Modern Analysis p.117 (1935).] With these procedures, we obtain the integration shown in the beginning.

## 10.2.

In Eq. (16), we need to calculate the following integral;

This integral can be carried out in the similar way as the one shown above, i.e., we have to evaluate complex integrals with an exponential function. As a result,

## 10.3.

In Eq. (18), the following integral results in

## 10.4.

Equation (18) has the following definite integral,

i. ($\mathrm{S}>1$ or $\mathrm{S}<1$) and $\mathit{p}>(1+\mathrm{S})$

ii. [$\mathrm{S}>1$ and $(1+\mathrm{S})>\mathit{p}>(\mathrm{S}-1)$] or [$\mathrm{S}<1$ and $(1+\mathrm{S})>\mathit{p}>(1-\mathrm{S})$]

iii. $\mathrm{S}>1$ and $(\mathrm{S}-1)>\mathit{p}>0$

iv. $\mathrm{S}<1$ and $(1-\mathrm{S})>\mathit{p}>0$

Above results reduces the integration of Eq. (19) to a single integral to give the results as follows

i. $\mathrm{S}\ge 1$ and $\mathit{p}>(1+\mathrm{S})$

ii. [$\mathrm{S}>1$ and $(1+\mathrm{S})>\mathit{p}>(\mathrm{S}-1)$] or [$\mathrm{S}<1$ and $(1+\mathrm{S})>\mathit{p}>(1-\mathrm{S})$]

iii. $\mathrm{S}>1$ and $(\mathrm{S}-1)>\mathit{p}>0$

iv. $\mathrm{S}<1$ and $(1-\mathrm{S})>\mathit{p}>0$

Among the above four results, (iii) and (iv) would be able to be calculated by the results of 10.2 and 10.4. For (ii),

For (b) and (c), we can use the formula shown already. However, we need some more calculation for (a) and (d).

In the end, the four integrals are

$(1+\mathrm{S})>p>1$

$1>p>(\mathrm{S}-1)$ or $1>p>(1-\mathrm{S})$

$(1-S)>p>1$

$1>p>(S-1)$ or $1>p>(1-S)$