We first reduce to the case when \(\spec(N) = \essspec(N)\text{.}\) Since \(N\) is a normal operator with finite spectrum, by the spectral theorem there is a finite rank perturbation \(N'\) of \(N\) for which \(N'\) is normal and \(\spec(N') = \essspec(N') = \essspec(N)\text{.}\) In particular, if \(P_{\lambda}\) are the spectral projections of \(N\) onto \(\{\lambda\}\text{,}\) and \(\lambda' \in \essspec(N)\) is a distinguished element, then we can choose
\begin{equation*}
N' := \lambda' P + \sum_{\lambda \in \essspec(N)} \lambda P_{\lambda}, \quad\text{where}\quad P = \sum_{\lambda \notin \essspec(N)} P_{\lambda}.
\end{equation*}
Since \(N'-N\) is finite rank, the diagonals of \(N'\) and \(N\) differ by an absolutely summable sequence, so \((d_n) \in \Lim(X)\) if and only if the diagonal of \(N'\) is in \(\Lim(X)\text{.}\) Moreover, the spectral projections of \(N\) and \(N'\) differ from one another by finite projections. Therefore, the spectral projections of \(N\) each differ from a diagonal projection by a Hilbert–Schmidt operator if and only if the same holds true for \(N'\text{.}\) By Theorem 3.4, \(N\) is diagonalizable by a unitary which is a Hilbert–Schmidt perturbation of the identity if and only if \(N'\) is as well. Therefore, by the above reduction, it suffices to prove the theorem with the added assumption that \(\spec(N) = \essspec(N)\text{.}\)
(Proof of \(\Rightarrow\)) Enumerate the elements of \(\essspec(N) = \spec(N)\) as \(\lambda_1,\ldots,\lambda_m\text{.}\) Let \(P_j\) denote the spectral projection corresponding to the eigenvalue \(\lambda_j\text{,}\) so that \(N = \sum_{j=1}^m \lambda_j P_j\text{.}\) Let \(\{e_n\}_{n=1}^{\infty}\) denote the orthonormal basis corresponding to the diagonal \((d_n)\text{.}\) Suppose \((d_n) \in \Lim(X)\text{,}\) and so there exist \(x_n \in X\) for which \((d_n - x_n) \in \ell^1\text{.}\) Let \(\Lambda_k := \{ n \in \mathbb{N} \mid x_n = \lambda_k\}\) be the index set where the sequence \((x_n)\) takes the value \(\lambda_k \in X\text{.}\)
The projections \(P_j\) sum to the identity, so for each \(n \in \mathbb{N}\text{,}\) \(\sum_{j=1}^m \angles{P_j e_n,e_n} = 1\) and therefore
\begin{equation*}
d_n = \sangles{Ne_n, e_n} = \sum_{j=1}^m \sangles{P_j e_n, e_n} \lambda_j
\end{equation*}
is a convex combination of the spectrum.
For \(\lambda_k \in X\text{,}\) let \(L_k\) be a line separating \(\lambda_k\) from the remaining elements of \(\essspec(N)\text{.}\) Such a line \(L_k\) exists because \(\lambda_k\) is an extreme point of the convex hull of \(\essspec(N)\text{,}\) and this is a finite set. Since \((d_n) \in \Lim (X)\) we know that \((d_n - \lambda_k)_{n \in \Lambda_k}\) is absolutely summable for every \(k\text{.}\) Therefore, for all but finitely many indices \(n \in \Lambda_k\text{,}\) the diagonal entry \(d_n\) lies on a line parallel to \(L_k\) separating \(\lambda_k\) from \(L_k\) and hence also \(\essspec(N) \setminus \{ \lambda_k \}\text{.}\)
By Lemma 4.1, for these indices \(n \in \Lambda_k\text{,}\)
\begin{equation}
\sum_{\substack{j=1 \\ j\not=k}}^m \sangles{P_j e_n, e_n} \le \frac{\abs{d_n - \lambda_k}}{\dist(\lambda_k,L_k)}.\label{eq-P_j-in-ell-1}\tag{4.1}
\end{equation}
Since this inequality holds for all but finitely many \(n \in \Lambda_k\text{,}\) and \(\dist(\lambda_k,L_k)\) is independent of \(n \in \Lambda_k\text{,}\) and \((d_n - \lambda_k)_{n \in \Lambda_k}\) is absolutely summable, (4.1) proves \(\big(\langle P_j e_n, e_n \rangle\big)_{n \in \Lambda_k}\) lies in \(\Lim (\{0\}) = \ell^1\) when \(j \not= k\text{.}\) If \(\lambda_j \in \essspec(N) \setminus X\text{,}\) by letting \(\lambda_k\) run through \(X\text{,}\) we find \(\big(\langle P_j e_n, e_n \rangle\big)_{n \in \mathbb{N}}\) is absolutely summable since \(\bigcup_{\lambda_k \in X} \Lambda_k = \mathbb{N}\text{.}\) This implies \(P_j\) is trace-class and hence a finite projection, contradicting the fact that \(\lambda_j \in \essspec(N)\text{.}\) Therefore \(X = \essspec(N)\text{.}\)
Now consider \(\lambda_j \in X = \essspec(N)\text{.}\) In analogy with the previous paragraph, using the fact that \(\big(\langle P_j e_n, e_n \rangle\big)_{n \in \Lambda_k} \in \ell^1\) when \(j \not= k\) and letting \(\lambda_k\) run through \(X \setminus \lambda_j\text{,}\) we find \(\big(\langle P_j e_n, e_n \rangle\big)_{n \notin \Lambda_j} \in \ell^1\text{.}\) Finally, for \(n \in \Lambda_j\text{,}\)
\begin{equation*}
1 - \sangles{P_j e_n, e_n} = \sum_{\substack{k=1 \\ k\not=j}}^m \sangles{P_k e_n, e_n},
\end{equation*}
and hence \(\big(1- \sangles{P_j e_n, e_n}\big)_{n \in \Lambda_k}\) is a finite sum of absolutely summable sequences, and is therefore absolutely summable. Thus \(\big(\sangles{P_j e_n, e_n}\big)_{n \in \Lambda_k} \in \Lim(\{1\})\text{,}\) so \(\big(\sangles{P_j e_n,e_n}\big) \in \Lim(\{0,1\})\text{.}\) Therefore, by Corollary 3.1, \(P_j\) differs from a diagonal projection by a Hilbert–Schmidt operator. Since this is true of all the spectral projections of \(N\text{,}\) we may apply Theorem 3.4 to conclude that \(N\) is a diagonalizable by a Hilbert–Schmidt perturbation of the identity.
(Proof of \(\Leftarrow\)) This implication is a direct corollary of Theorem 3.8. To see this, suppose \(\essspec(N) = X\) and \(N\) is diagonalizable by a unitary \(U\) which is a Hilbert–Schmidt perturbation of the identity. Thus \(UNU^{*} = \diag(x_n)\) for some sequence \(x_n \in \essspec(N) = X\text{.}\) Then by Theorem 3.8, \(E(N-UNU^{*})\) is trace-class. That is, \(\trace \big(E(N-UNU^{*})\big) = \sum_{n=1}^{\infty} (d_n-x_n)\) is an absolutely summable series, so \((d_n) \in \Lim(X)\text{.}\)