Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Dark current and single photon detection by 1550 nm avalanche photodiodes: dead time corrected probability distributions and entropy rates

Open Access Open Access

Abstract

Single photon detectors have dark count rates that depend strongly on the bias level for detector operation. In the case of weak light sources such as novel lasers or single-photon emitters, the rate of counts due to the light source can be comparable to that of the detector dark counts. In such cases, a characterization of the statistical properties of the dark counts is necessary. The dark counts are often assumed to follow a Poisson process that is statistically independent of the incident photon counts. This assumption must be validated for specific types of photodetectors. In this work, we focus on single-photon avalanche photodiodes (SPADs) made for 1550 nm. For the InGaAs detectors used, we find the measured distributions often differ significantly from Poisson due to the presence of dead time and afterpulsing with the difference increasing with the bias level used for obtaining higher quantum efficiencies. We find that when the dead time is increased to remove the effects of afterpulsing, it is necessary to correct the measured distributions for the effects of the dead time. To this end, we apply an iterative algorithm to remove dead time effects from the probability distribution for dark counts as well as for the case where light from an external weak laser source (known to be Poisson) is detected together with the dark counts. We believe this to be the first instance of the comprehensive application of this algorithm to real data and find that the dead time corrected probability distributions are Poisson distributions in both cases. We additionally use the Grassberger-Procaccia algorithm to estimate the entropy production rates of the dark count processes, which provides a single metric that characterizes the temporal correlations between dark counts as well as the shape of the distribution. We have thus developed a systematic procedure for taking data with 1550 nm SPADs and obtaining accurate photocount statistics to examine novel light sources.

© 2022 Optica Publishing Group under the terms of the Optica Open Access Publishing Agreement

1. Introduction

Single photon counting has been used for many decades to study the quantum properties of light and its interactions with atoms and molecules. Single photon detection was used to lay the foundations of coherent states emitted by laser sources and non-classical states of light fields [1,2]. More recently, it is an indispensable technique for the characterization of single photon emitters (SPEs) which are used widely in quantum information applications [3]. Traditionally, photomultiplier tubes were used to count photons [4] in the visible and ultraviolet spectrum, but with the development of fiber optic telecommunications, semiconductor-based detectors in the 1550nm region came into widespread use for weak signals at the single photon level. Geiger-mode single-photon avalanche diodes (SPADs) have found widespread use over the past two decades [58].

All single-photon detectors exhibit dark counts in the absence of light: in avalanche photodiodes, dark counts occur when avalanches are triggered by electrical carriers that are thermally generated or emitted by trapping levels in the semiconductor [9,10]. The rate at which these avalanches are triggered, called the dark count rate, is heavily dependent on detector settings such as the quantum efficiency and dead time [6,7,11]. Dark counts are a source of noise in any application and increase with the bias voltage needed to enhance the quantum efficiency of the detector. There have been several studies of dark current in avalanche photodiodes [5,12] and numerical models have been developed to compare with experimental measurements [13].

However, a complete statistical characterization of the dark counts in 1550 nm SPADs has not been reported. The purpose of our research is to explore the dark counts of SPADs by using time-tagged photodetector measurements that can be used to generate histograms of their interarrival time distributions, probability distributions, and entropy rates. A simplifying assumption often made about the statistics of dark counts is that they follow a Poisson distribution. We examine this assumption and show that it is not generally valid without proper processing of the data. This processing is essential when one wants to accurately characterize the statistics of single photon emitters or novel light sources.

It is long understood that Poisson statistics are not always an appropriate model for counts from photomultiplier tubes [4,14], image sensors such as CCD or CMOS sensors [15], general purpose digital pulse processing systems [16], and SPAD photocounts [13,17]. However, it is still often assumed that their dark counts follow Poisson statistics [9] whose probability distribution is given by eqn. 1 [18,19]

$$P(n) =\dfrac{(rT)^ne^{(rT)}}{n!},$$
where $P(n)$ is the probability of $n$ counts being detected in the time interval $T$, assuming $r$ is the average rate of detected counts. We will define $\lambda = rT$ as the average number of counts in the interval $T$. The detector quantum efficiency, $\eta$, relates the optical power to the observed photon count rate, $r$, as:
$$r = \eta (P/h\nu),$$
where $P$ is the average optical power incident on the detector, and $h\nu$ is the energy per photon.

In many experiments the dark count rates are relatively small in comparison with the photon count rates, therefore their effect on the distribution is negligible. However, there are other practical cases where the dark count rate is comparable to or only slightly less than the photon count rate. In these instances, we aim to quantitatively explore the dark count distributions and their influence on a weak attenuated coherent light source which is known to be Poisson [1,2].

We study the deviations of dark current statistics from Poisson statistics due to dead time and afterpulsing [4,20] and propose a method for correcting the histograms. First, we recommend extending the detector dead time to eliminate the effects of afterpulsing as observed from the interarrival distributions. Second, we suggest calculating the adjusted occurrence rate of dark counts from the measured rate by using a standard formula. This enables us to determine the correct slope for the interarrival time distribution of a Poisson process. Lastly, we will show that a further step is necessary to correct the probability distributions for dead time which involves implementing an iterative algorithm [21,22].

2. Instrumentation

Single-photon avalanche diodes are named as such because when the reverse bias of its p-n junction is raised above the breakdown voltage, just a single carrier can trigger an electrical avalanche process, leading to a measurable current [9,10]. To detect a subsequent photon, the bias voltage must be reduced to near or below the breakdown value. Restoring the SPAD to its operative level, a process called quenching, is achieved by a quenching circuit that introduces a finite recovery time, known as dead time, during which the device cannot respond to another incident photon [23]. There are two main quenching modes: passive quenching (PQ) and active quenching (AQ). PQ SPADs are paralyzable detectors where photons arriving during the dead time are not counted and the dead time is extended [13]. Alternatively, AQ SPADs, a type of nonparalyzable detector used in our experiments, will not count photons arriving during the dead time nor will the dead time be extended by the quenching circuit. If a carrier is triggered by photon absorption, the generated current will precisely mark the photon arrival time. However, avalanches can also be triggered by dark current, thereby marking the time of avalanche generation. Another phenomenon observed in SPADs is afterpulsing, a type of correlated noise found in real, non-ideal detectors where more than one electric pulse is generated per event due to traps holding extra charge carriers [12,18,24]. Figure 1 depicts the detection behavior of AQ SPADs in the presence of dead time. Figure 1(a) shows the SPADs electrical response (black curve) to an electron (blue circle) which may be triggered by a photon or dark current. The additional fluctuations of the curve after the initial response are the afterpulsing. After each response, there is a pre-determined dead time, represented by the shaded blue boxes, during which the detector cannot detect any incoming electrons. In some detectors, the dead time $\tau$ is a parameter that can be controlled by the user, hence the shaded boxes are of different lengths. Figure 1(b) shows the electrical signal that is sent from the output of the detector after the SPAD response passes through a discriminator.

 figure: Fig. 1.

Fig. 1. A schematic depicting the detector response to an electron carrier, which may be triggered by a photon, dark current, etc., in the presence of dead time. (a) The avalanche photodiode’s electrical response (black curve) to a triggered electron (blue circle) including afterpulsing, the additional fluctuations seen after the initial response. After each response, there is a dead time (shaded blue boxes) when the detector cannot detect any incoming electrons. The dead time, $\tau$, is adjustable by the user. (b) Electrical signal sent from the output of the detector after the SPAD response passes through a discriminator.

Download Full Size | PDF

All our experimental data is taken using InGaAs Geiger-mode avalanche photodetectors from Aurea Technologies [25] operated in continuous mode and designed for 1550 nm wavelengths. Our SPADs have a range of selectable dead times from 1 $\mu$s to 999 $\mu$s and three quantum efficiency settings, 10%, 20%, and 30%, which correspond to the percentage of incident photons that are detected and depend on the bias voltage applied to the APD. We observed dark current using two chosen dead time values: 20 $\mu$s and 500 $\mu$s. The first is the shortest dead time that can be used across all efficiency settings while in continuous mode and the second proved to be sufficiently long to remove the effects of afterpulsing.

To experimentally measure dark current, detected events were recorded when no light was incident on the detector and the input was sealed and covered to minimize external light leakage (Fig. 2- switch open). The precise arrival of events was determined using time-tagging measurements with a ChronoXea [26], a time correlator device from Aurea Technologies. This instrument accepts the electrical output of the detector and records the occurrence time with respect to an internal clock with a resolution of 13 ps. Afterward, the time-tagged data is post-processed on a computer.

 figure: Fig. 2.

Fig. 2. Switch Open: Experimental setup used for measuring dark current. A SPAD set in continuous mode with no light entering it has its electrical output connected to the input of the time-tagging correlator. The time-tagged data is then sent to the computer to be post-processed. Switch Closed: Laser data experimental setup. A 1550 nm distributed feedback (DFB) laser is connected through single-mode (SM) fiber to 110 dB of fixed attenuation and up to 30 dB of variable attenuation. Photons are then sent through SM fiber to the SPAD. The electrical output of the SPAD is ultimately sent to the input of the time-correlator. The data is then sent to a computer to be post-processed.

Download Full Size | PDF

While using a dead time of 20 $\mu$s, around 500 dark counts per second were observed at 10% efficiency; 5000 dark counts per second at 20% efficiency, and 50,000 per second at 30% efficiency. In this paper, we restrict ourselves to the analysis of dark counts at the 10% and 20% efficiency levels because at 30% efficiency the dark current nears the saturation limit of the detector which makes it practical to use only while in gated mode.

3. Interarrival distributions and adjusted occurrence rate

The interarrival time distribution for a Poisson process is expected to be an exponential distribution shifted by the dead time [27]. Dead time hinders our ability to observe the true nature of the dark current distributions. One common way to account for dead time, $\tau$, is to calculate the adjusted occurrence rate, $r_a$, of counts if there were no dead time. This can be found using the experimentally measured occurrence rate, $r_m$, taken with dead time $\tau$ and the following equation [2831]:

$$r_a = r_m / (1 - (r_m\tau)).$$

This adjusted occurrence rate is often used to find fits for both the interarrival distributions and probability distributions. We calculated and plotted on a semi-log plot the interarrival times of the dark counts taken at 10% and 20% efficiency, both with a 20 $\mu$s dead time (Figs. 3(a),(b)) and used Eq. (4) to fit an exponential curve to the data and compare the slope of best fit, $r$, to the measured slope $r_m$ and adjusted slope $r_a$.

$$y(t) = A\exp[{-}r(t-\tau)]$$

 figure: Fig. 3.

Fig. 3. Dark count data with SPAD settings of 20 microsecond dead time. (a) A histogram (grey) of the interarrival times between consecutive dark counts taken at 10% efficiency. An exponential curve fit excluding the first 500 microseconds is shown in red. An inset image of the interarrival histogram with the first 2 milliseconds enlarged is included. (b) Interarrival time histogram at 20% efficiency.

Download Full Size | PDF

In Eq. (4), $A$ is the frequency of counts in the first time bin, $r$ is the slope representing the occurrence rate of events per time $t$, and $\tau$ is the dead time. The distribution for both data sets is exponential at longer interarrival times, but there is a large accumulation of dark counts present at very short interarrival times, seen in the inset figures of Fig. 3(a) and Fig. 3(b). This indicates the presence of afterpulsing which occurs on a time scale that is larger than the dead time used for these data sets. $r_a$ and $r_m$ are very similar in value because of the small dead time but are both much larger than the best fit, $r$. In the case of short dead times, the correction using Eq. (3) resulting in $r_a$ does not result in a better fit because the distribution is skewed by afterpulsing, not dead time. The only way to obtain the best fit is to remove the afterpulsing effect which can either be done through post-processing, such as analyzing only the bins after the first 500 microseconds as shown in the red curve of Figs. 3(a),(b), or experimentally by extending the dead time which we demonstrate next.

In order to experimentally eliminate the effect of afterpulsing, we increased our dead time to 500 $\mu$s which is well beyond the duration of the effect. The interarrival distributions for dark current at 10% and 20% quantum efficiency with a 500 $\mu$s dead time are shown in Figs. 4(a),(b). The afterpulsing effect is now successfully omitted in the interarrival distributions and they now closely resemble an exponential distribution. It should be noted that in Fig. 4(a), the interarrival bins bend downwards in the last two milliseconds because they approach the maximum interarrival time measured by the Chronoxea, which in this case is 10 milliseconds. Therefore, the curve fit was only applied to data up to 8 milliseconds. The rate, $r$, determined by the curve fit in these figures is now very close to the calculated $r_a$ value. This demonstrates that Eq. (3) works better for fitting interarrival distributions with higher dead times.

 figure: Fig. 4.

Fig. 4. Dark count data with SPAD settings of 500 microsecond dead time. (a) A histogram (grey) of the interarrival times between consecutive dark counts taken at 10% efficiency. An exponential curve fit is shown in red. (b) Interarrival time histogram at 20% efficiency.

Download Full Size | PDF

4. Probability distributions and iterative correction for dead time

We have shown that afterpulsing distorts the interarrival distributions and that it is insufficient to simply ignore interarrival times smaller than 500 $\mu$s during post-processing. Instead, the electronic dead time must be extended in order to effectively suppress afterpulsing. In this regime of long dead time, using Eq. (3) to find $r_a$ provides a sufficient correction for dead time to fit the interarrival distribution of the data. However, next, we will demonstrate that this is not an effective correction to fit a Poisson distribution to the probability histogram, which expresses the probability of a given number of dark count occurrences within a fixed counting interval. Rather, a more rigorous method is required.

Finite dead time is an experimental feature in photon counting instrumentation that we would like to utilize and then correct by using a method of post-processing distributions. Mandel and Srinivas [21,22] offer a mathematical formula for the probability of photon counting in a counting interval which corrects for the effects of dead time in nonparalyzable single-photon detectors. It is an iterative method that allows for the counting probability of an ideal detector without dead time to be derived from the measured probability distributions observed with a real detector. Equation (5) gives the probability, $p(n,T,0)$, that $n$ counts are registered in a time interval $T$ when the detector has no dead time in terms of probabilities that can be directly measured using a detector with dead time $\tau$. It assumes that the condition $T>n\tau$ is met throughout its use. When originally written, the algorithm required measurements for several different counting intervals ranging from $T$ to $T+n_0\tau$, where $n_0$ is the largest value of $n$ for which $p(n)$ is not negligible. However, with the common availability of time-tagging devices, one long measurement of time-tagged arrivals can be recorded. Afterwards, during post-processing, the arrivals can be grouped into counting intervals of duration $T+n\tau$ from which the probability $p(n,T+n\tau,\tau )$ can be estimated from the data, allowing us to solve for $p(n,T,0)$. We implement this more practical method rather than the multiple measurements suggested by Mandel. His equation is presented below:

$$\begin{aligned} p(n,T,0) =& p(n,T+n\tau,\tau)\\ & +\sum_{r=0}^{n-1} [p(r,T+n\tau,\tau) - p(r,T+(n-1)\tau,\tau)], \quad n\geq 1,\\ p(0,T,0) =& p(0,T,\tau). \end{aligned}$$

This iterative algorithm was applied to our data resulting in dead-time corrected probability distributions. While an approximate algorithm (correct to third-order in $\tau /T$) for dead-time correction has previously been applied to photon counting data with photomultiplier tubes [32], to our knowledge, this work demonstrates the first comprehensive application of this exact algorithm to real SPAD data. Note that the Mandel algorithm corrects only for dead time, not afterpulsing, and that afterpulsing effects were removed by extending the dead time. Figures 5(a),(b) depict the algorithm as applied to the 500 $\mu$s dead time data. Both figures include the original histogram (light blue) and the dead-time-corrected histogram (purple) plotted with a counting interval of T = 0.015 seconds. These histograms are compared to several different theoretical Poisson distributions, which for numerical stability and to avoid computing errors were calculated using:

$$P(k,\lambda) = \exp[k\ln\lambda-\lambda-\ln\Gamma(k+1)],$$
where $k = \lfloor \delta t/T\rfloor$ is the bin number with $\delta t$ as the interarrival time, $\Gamma$ is the Gamma function, and $\lambda = RT$ is the average number of counts with $R$ representing the occurrence rate of dark counts and $T$ representing the integration time. For all plots, three different Poisson distributions are plotted to compare to the original and corrected histograms. The first, shown in solid black, uses $\lambda _m = r_mT$ where $R = r_m$ is the measured occurrence rate. The second, shown in dotted red, uses $\lambda _a = r_aT$ where $R = r_a$ is the adjusted occurrence rate. The third, shown in dashed orange, uses $\lambda _c = r_cT$ where $R = r_c$ is the corrected occurrence rate calculated from the dead time corrected histogram.

 figure: Fig. 5.

Fig. 5. Data with 500$\mu$s dead time. (a) A histogram of the normalized probability distribution (light blue) of dark counts which arrived in the time interval T = 0.015 seconds at 10% efficiency along with the dead-time corrected histogram (solid purple) generated from the iterative algorithm. Theoretical Poisson curves are included with varying values of $\lambda$, namely $\lambda _m$ (solid black), $\lambda _a$ (dotted red), and $\lambda _c$ (dashed orange). (b) Probability distribution for data taken at 20% efficiency with the corresponding dead-time correction and several theoretical Poisson curves.

Download Full Size | PDF

Examining the probability distributions for data with detector settings of 500$\mu$s (Figs. 5(a),(b)), we see they are not closely aligned with the Poisson distributions using $\lambda _m$ or $\lambda _a$. The longer dead time affects the shape of the statistical distribution and simulates the effect of anti-bunching on the dark counts. The histograms are negatively skewed and sub-poissonian, meaning the statistical spread is narrower than a theoretical Poisson distribution of the same $\lambda$.

The long dead time makes this data set a good candidate for the iterative dead time correction algorithm and applying it appreciably shifts the histograms of the 10% and 20% efficiency data to more closely resemble the corresponding corrected Poisson distributions using $\lambda _c$. Here, the average number of counts lost due to the long dead time is high, allowing the Mandel algorithm to be effective. As described in section 2, the dark counts increase with increasing quantum efficiency. This means that the dark count rate is higher with the 20% efficiency setting, and therefore more counts are lost due to the dead time than in the 10% efficiency case. The dead time correction restores these lost counts, which is why the dead time corrected Poisson distribution has a higher count rate than the original histogram in the 20% efficiency case. After applying the algorithm, the corrected distribution (purple) almost perfectly aligns with the Poisson distribution (orange) for the same $\lambda$ value in both cases. In the regime of long dead time, the effects of afterpulsing on the distributions are now negligible and the effects of dead time are dominant. Now, the Mandel algorithm can be used to correct for the effects of dead time. The iterative algorithm provides a correction to the actual histogram that can be used to derive a corrected $\lambda$ value that gives a perfectly fitting Poisson distribution. This correction method is more rigorous and much more effective than using Eq. (3) to adjust the $\lambda$ value.

5. Dead time correction with attenuated laser source

Typically, SPADs are used to perform photon counting in systems using a light source. Below we analyze the photon distributions generated from a laser source and demonstrate that though the effects of dead time are still present, they can be adjusted for by using the dead time correction algorithm to find the corrected $\lambda$ value. In addition to the counts corresponding to the light source, dark counts occur as before. To produce light at the single photon level, a distributed feedback laser with a wavelength of 1550nm operated at 1 mW was attenuated by 140 dB using a combination of both fixed and variable attenuators connected by fiber. The detector was set to 10% efficiency with a 500$\mu$s dead time which corresponds to approximately 225 dark counts per second. To simulate a scenario where the photon counts are on the same order of magnitude as the dark counts, we chose a laser attenuation level with a photon rate about twice as high as the dark counts, giving a total average of 525 measured counts per second. Just as in the experimental setup measuring dark counts, the electrical output of the SPAD was connected to the input of the time-tagging correlator, shown in Fig. 2 (switch closed), which recorded 10 million data points for each trial. The photon interarrival and probability distributions are shown in Figs. 6(a),(b) along with the dead time corrected histogram and several Poisson plots.

 figure: Fig. 6.

Fig. 6. Laser data (10 million points) with an average of 650 photons per second (adjusted for dead time) taken with detector settings of 10% efficiency and 500 microsecond dead time. (a) Interarrival times on a log-linear plot (grey) with an exponential curve fit (red) based on the true photon counts detected per second, but excluding the first 500 microseconds. (b) The original histogram (light blue) of the frequency of photon arrivals per time interval T. The dead time corrected histogram (purple) can be compared to the theoretical Poisson fit (dashed orange) for the corrected lambda value. Poisson fits are plotted using both the $\lambda _m$ (black) and $\lambda _a$ (dotted red)

Download Full Size | PDF

Using this long dead time, the effects of afterpulsing are effectively eliminated (Fig. 6(a)). The probability distribution histogram (light blue) (Fig. 6(b)) again does not fit well with the Poisson distributions computed with $\lambda _m$ (black) and $\lambda _a$ (red) values and is sub-poissonian due to the long dead time. Applying the dead-time correction algorithm in Fig. 6(b) proves to correct for this poor fit and produces a corrected probability distribution (purple) that almost perfectly aligns with the Poisson distribution computed with its corresponding $\lambda _c$ value. This extension of the dead time past the regime of afterpulsing paired with the dead-time correction produces an accurate picture of the true distribution. It again demonstrates that a correction beyond adjusting the counts using Eq. (3) is necessary in order to generate a proper statistical description of the system data.

6. Entropy rate of dark counts

In the previous sections, we presented measurements of the dark counts in SPADs at 1550nm. All of these measurements consisted of histograms, which do not consider the temporal correlations or dynamical behavior of the measurements. In this section, we consider the $\epsilon$-entropy rate of the SPAD dark counts, which relates to the entropy generation rate of a system as a function of its measurement resolution $\epsilon$. The $\epsilon$-entropy rate is a single metric that can simultaneously quantify both the shape of the interarrival time histogram and temporal correlations in the data. As we will show, the effects of the dead time and afterpulsing in SPAD dark count measurements are both captured using the $\epsilon$-entropy.

Considering the entropy rate as a function of the measurement resolution is useful for characterizing the information production of stochastic and deterministic processes in general [3335] and in particular, has important implications for random number generation based on physical entropy sources [36]. The detection of single photons is perhaps the best-known optical technique for generating entropy for random bit generation [37].

In this section, we perform an analysis of the rate at which entropy can be harvested from the interarrival times of dark counts in our SPAD setup described above. In particular, we study how the entropy rate changes as a function of the interarrival time resolution $\Delta t$. While this is typically called the $\epsilon$-entropy, here we refer to it as the $\Delta t$-entropy to emphasize that we are studying the entropy production rate as a function of $\Delta t$. We will show that this single metric provides information about both the shape of the histogram as well as the temporal correlations present in the data.

The amount of information produced by a system per unit of time is given by its Shannon entropy rate. More precisely, the Shannon entropy rate is the average amount of new information obtained by measuring one additional sample with resolution $\epsilon$ given the history of all previous samples [34]. Stochastic systems have an infinite Shannon entropy rate [33]. However, information can only be extracted from the system at a finite rate, which is determined in part by the resolution of the measurement [33,34]. This motivated the definition of the so-called $\epsilon$-entropy by Shannon [38]. The $\epsilon$-entropy has been studied for a variety of systems in theory and simulations [33,34] and experiments [35,36]. Since here we are considering interarrival time measurements, for which the resolution is in terms of time, we denote the resolution of the interarrival time measurement as $\Delta t$.

In principle, one could compute the $\Delta t$-entropy directly from Shannon’s definition $h_1(\Delta t)$ (see Eq. (7)). The first step is to generate a list of points in $d$-dimensional space using $d$ consecutive interarrival times. These vectors can be regarded as samples of a $d$-dimensional probability distribution. The entropy of this probability distribution is sometimes referred to as the pattern entropy for patterns of length $d$. Computing the entropy in this way often requires an impractically large amount of data, especially as $d$ increases.

Instead, we will use the correlation entropy rate $h_2(\Delta t;d)$, which we estimate using the Grassberger-Procaccia algorithm [39]. The correlation entropy rate is a lower bound on the information entropy rate, and in many cases is numerically very close to the information entropy rate [39]. A detailed discussion of the Shannon entropy rate and the correlation entropy rate and their analytic solutions for interarrival time measurements of Poisson processes can be found in the Appendix. For the SPAD system considered here, the only expected temporal correlations are due to afterpulsling, which should not create correlations between more than 2 or 3 consecutive events. Therefore, it is sufficient to consider only pattern lengths up to $d=3$.

When the dead time is large, one expects there to be no intersample correlations since counts due to afterpulsing are eliminated. In this case, one might expect dark count measurements to be well-described by a Poisson process (Eq. (12)). Figure 7 shows that this is the case for the detectors used here for both the 10% (Fig. 7(a)) and 20% (Fig. 7(b)) quantum efficiency settings when the dead time is 500$\mu$s. The correlation entropy rate $h_2(\Delta t,d=1)$ estimated using the Grassberger-Procaccia algorithm agrees almost perfectly with the analytic result for the correlation entropy rate of a Poisson process with the same rate. This means that the distribution of interarrival times is extremely close to the exponential distribution expected for a Poisson process. Additionally, $h_2(\Delta t,d)$ is independent of $d$, as would be expected for a memoryless process. While essentially no intersample correlations are detected, $\lambda _d$ is also very low, so the entropy generation rate per unit time is quite low. Figure 7 also shows that, in this case, $h_2$ is an excellent approximation for $h_1$, as expected from Eqs. (10) and (12).

 figure: Fig. 7.

Fig. 7. Entropy rate analysis for dark count measurements with 500 $\mu$s dead time and (a) 10% quantum efficiency and (b) 20% quantum efficiency. The analytic solution for $h_1$ (Eq. (10)) and $h_2$ (Eq. (11)) are shown as black and dotted green lines, respectively. The Grassberger-Procaccia estimates of $h_2(\Delta t,d)$ do not depend on $d$, implying that there are no intersample correlations. The sharp decrease as $\Delta t$ decreases for small $\Delta t$ in the $d=3$ curve (red) is due to the finite size of the dataset used. Further, the Grassberger-Procaccia $h_2$ estimates agree excellently with the analytic solution for $h_2$, suggesting that the dark count measurements are well-modeled by a Poisson process. All entropy rate estimates were computed using 2 million recorded interarrival times.

Download Full Size | PDF

For short enough dead times, on the other hand, one might expect to observe effects from afterpulsing. These effects might appear as intersample correlations as well as a probability distribution that is concentrated at lower interarrival times. Figure 8 shows the Grassberger-Procaccia entropy rate estimate for dark count interarrival time measurements on our detectors with quantum efficiency 10% (Fig. 8(a)) and 20% (Fig. 8(b)). The striking disagreement between $h_2$ for a Poisson process (green dashed line) and the 1-$d$ Grassberger-Procaccia entropy rate (solid magenta line) estimate for both quantum efficiencies shows that these interarrival time distributions are significantly more peaked than the expected exponential distribution, as was shown in Figs. 3(a),(b). Further, the Grassberger-Procaccia entropy rate estimate decreases as $d$ increases from 1 to 2, indicating intersample correlations, which we attribute to afterpulsing.

 figure: Fig. 8.

Fig. 8. Entropy rate analysis for dark count measurements with 20 $\mu$s dead time and (a) 10% quantum efficiency and (b) 20% quantum efficiency. The analytic solution for $h_1$ (Eq. (10)) and $h_2$ (Eq. (11)) are shown as black and dotted green lines, respectively. The Grassberger-Procaccia estimates of $h_2(\Delta t,d)$ decrease with increasing $d$, implying that there are intersample correlations. The reduction with $d$ is stronger for the 20% efficiency case, suggesting that there are stronger intersample correlations in this case, which is expected for afterpulsing. Further, for both efficiencies the Grassberger-Procaccia $h_2$ estimate severely underestimates the analytic solution for $h_2$ even for $d=1$, reflecting that the dark count interarrival time histogram is much more strongly peaked than the exponential distribution expected for a Poisson process. All entropy rate estimates were computed using 2 million recorded interarrival times.

Download Full Size | PDF

7. Conclusion

Throughout our study, we find that both dead time and afterpulsing affect the probability distributions of dark counts and interarrival times and we quantitatively assess the assumption that dark current follows a Poisson distribution. We realize that the dead time of the detection circuitry must be adjusted to eliminate afterpulsing effects. With dead times that are non-negligible compared to the counting time intervals, a correction algorithm needs to be implemented. This algorithm [21,22] is iterative in nature and can be applied conveniently to time-tagged events measured by your detectors. As a result of these correction techniques, we are able to utilize SPADs for accurate determinations of the statistics of novel light sources in the 1550 nm range.

8. Appendix: Entropy rates of time-tagged photon arrivals

In this Appendix, we define the information entropy rate and the Grassberger-Procaccia entropy rate, and we present the analytic solutions for these entropy rates for a Poisson process.

For a single variable system like the single photon interarrival time measurements considered here, a $d$-sequence is defined as $\mathbf {x}_d[n]=\{x[1],\ldots,x[d]\}$. The information entropy $H_1$ can be defined as a function of time-tagging resolution $\Delta t$ and sequence length $d$ [33]:

$$H_1(\Delta t,d)={-}\sum_{k=0}^\infty p_k(\Delta t;d)\log_2(p_k(\Delta t;d))=\langle\log_2(p(\Delta t,d))\rangle,$$
where $p_k(\Delta t;d)$ is the probability that a sequence of length $d$ is in the $k^{th}$ $d$-dimensional box of size $\Delta t$ and $\langle \rangle$ indicates the average over all boxes with non-zero probability. The information entropy rate per sample can be defined as a function of $\Delta t$ and $d$: $h_1(\Delta t;d)= [H_1(\Delta t;d) - H_1(\Delta t;d-1)]$; this is the additional amount of information obtained from the the next measurement given the previous $d-1$ measurements.

The probability $p_i(\Delta t;d)$ can be estimated from data using [33]

$$p_i(\Delta t,d) = \frac{1}{N_d}\{\mathrm{number\, of\, }d\mathrm{-sequences}\: \mathbf{x}_j \: \mathrm{s.t.} \, |\textbf{x}_i-\textbf{x}_j|<\Delta t\},$$
where $N_d = N-d+1$ is the number of sequences of length $d$ in the data set, and $|\cdot |$ indicates a distance. This method of estimating the entropy rate has been used with some success for entropy rate generation for stochastic and/or chaotic systems [33,35,36,40], but has the well-known problem that the estimation of $p_i(\Delta t;d)$ can require impractically large amounts of data when $\Delta t$ is small and $d$ is large.

The correlation entropy rate $h_2$, which is easier to estimate from experimental data, has been suggested as a substitute for the information entropy $h_1$ [39,41]. The correlation entropy rate is a lower bound on the information entropy rate, and in many cases is numerically very close to the information entropy rate [39]. Therefore, we use the correlation entropy rate $h_2$ as an estimate for $h_1$ (recognizing that $h_2\leq h_1$). The correlation $d-$block entropy is defined as:

$$H_2(\Delta t,d) = \log_2\bigg(\sum_ip_i^2(\Delta t;d)\bigg)=\log_2\langle p^2(\Delta t;d)\rangle,$$
with $h_2(\Delta t;d)= [H_2(\Delta t;d) - H_2(\Delta t;d-1)]$ in units of bits per sample.

Equation (8) leads to $\sum p_i^2(\Delta t;d)=C(\Delta t;d)$, the well-known correlation sum [39]. Equations (8) and (9) are typically known as the Grassberger-Procaccia algorithm. We choose to use the infinity norm as was done by Takens [41]. For each Grassberger-Procaccia entropy rate estimate, we use a data set of 2 million interarrival times.

For a white noise process, $h_2(\Delta t,d)$ increases linearly as the logarithm of $\Delta t$ decreases, and is independent of $d$ [33]. On the other hand, for colored noise processes, $h_2(d)$ decreases as $d$ increases when $\Delta t$ is sufficiently small to resolve the time scale of the correlations [33]. The entropy rate additionally contains some information about the shape of the probability distribution: all else being equal, a broad distribution will have a greater entropy rate than a narrower distribution. This effect of the shape of the distribution is revealed explicitly by $h_2(\Delta t;d=1)$, which does not consider intersample correlations.

We now turn to the specific case of dark counts. It is often assumed that dark counts can be modeled by a Poisson process. The analytic solution for the information entropy generated per event $H_1(\Delta t;d)$ for a Poisson process is well-known to be:

$$H(\Delta t)=\frac{(1-p_0)\log_2(1-p_0)}{p_0}+\log_2(p_0),$$
where $p_0\equiv 1-\exp {[-r_t\Delta t]}$. We note that this is independent of $d$ because Poisson processes are memory-less so there are no intersample correlations. The information entropy rate is then $h_1=r_m H_1$, where $r_m$ is the mean number of events measured per unit time.

Of course, single photon detectors have a dead time. For non-paralyzable detectors, the dead time $\tau$ does not affect the shape of the PDF; it only shifts it by $\tau$ [27,36]. This does not affect $H$, the entropy per photon. The dead time does, however, affect the average rate of photons that are detected: $r_m=r_t/(1+r_t\tau )$ [27]. Thus the entropy rate becomes $h_1=r_mH$ in units of bits per unit time.

Further, if the probability of more than one event occurring in a single time bin is small ($r_t\Delta t<<1$), the information entropy rate for photon time-of-arrival measurements can be approximated as

$$h_1(\Delta t)={-}r_m\log_2(\frac{r_t\Delta t}{e}).$$

The correlation entropy per event for a Poisson process is independent of $d$ and can be computed directly from Eq. (9) using $p_k=\exp {[-kr_t\Delta t]}(1-\exp {[-r_t\Delta t]})$:

$$H_2(\Delta t) = 2\log_2(1-e^{{-}r_t\Delta t})-\log_2(1-e^{{-}2r_t\Delta t}).$$

In the limit that $r_t\Delta t<<1$, we can approximate the correlation entropy rate for a Poisson process with non-paralyzable dead time as

$$h_2(\Delta t)={-}r_m\log_2(\frac{r_t\Delta t}{2})$$
in units of bits per second. We note that this is less than, but close to, $h_1$.

Funding

Office of Naval Research (N00142012139).

Acknowledgments

N.M. and R.R. thank the Office of Naval Research (Grant Number: N00142012139) (R.R.). We thank the experts at Aurea Technologies for their help with questions regarding the instrumentation.

Disclosures

The authors declare no conflicts of interest.

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

References

1. L. Mandel and E. Wolf, Optical Coherence and Quantum Optics (Cambridge University, 1995).

2. R. Loudon, The Quantum Theory of Light (Oxford University, 2000), 3rd ed.

3. M. A. Nielsen and I. L. Chuang, Quantum Computation and Quantum Information: 10th Anniversary Edition (Cambridge University, 2010), 10th ed.

4. B. E. A. Saleh, Photoelectron Statistics: with Applications to Spectroscopy and Optical Communication (Springer, 1978).

5. M. A. Itzler, U. Krishnamachari, M. Entwistle, X. Jiang, M. Owens, and K. Slomkowski, “Dark count statistics in geiger-mode avalanche photodiode cameras for 3-d imaging LADAR,” IEEE J. Sel. Top. Quantum Electron. 20(6), 318–328 (2014). [CrossRef]  

6. X. Jiang, M. A. Itzler, R. Ben-Michael, and K. Slomkowski, “InGaAsP–InP avalanche photodiodes for single photon detection,” IEEE J. Sel. Top. Quantum Electron. 13(4), 895–905 (2007). [CrossRef]  

7. J. Zhang, R. Thew, J.-D. Gautier, N. Gisin, and H. Zbinden, “Comprehensive characterization of InGaAs–InP avalanche photodiodes at 1550 nm with an active quenching ASIC,” IEEE J. Quantum Electron. 45(7), 792–799 (2009). Conference Name: IEEE Journal of Quantum Electronics. [CrossRef]  

8. L. N. Hillesheim, “The photon counting histogram in fluorescence fluctuation spectroscopy with non-ideal photodetectors,” Biophys. J. 85(3), 1948–1958 (2003). [CrossRef]  

9. Y. Kang, H. Lu, Y.-H. Lo, D. Bethune, and W. Risk, “Dark count probability and quantum efficiency of avalanche photodiodes for single-photon detection,” Appl. Phys. Lett. 83(14), 2955–2957 (2003). [CrossRef]  

10. A. Spinelli and A. Lacaita, “Physics and numerical simulation of single photon avalanche diodes,” IEEE Trans. Electron Devices 44(11), 1931–1943 (1997). Conference Name: IEEE Transactions on Electron Devices. [CrossRef]  

11. J. Wen, W. J. Wang, X. R. Chen, N. Li, X. S. Chen, and W. Lu, “Origin of large dark current increase in InGaAs/InP avalanche photodiode,” J. Appl. Phys. 123(16), 161530 (2018). [CrossRef]  

12. M. Fishburn, “Fundamentals of CMOS single-photon avalanche diodes,” Ph.D. thesis (Delft University of Technology, 2012).

13. E. Sarbazi, M. Safari, and H. Haas, “Statistical modeling of single-photon avalanche diode receivers for optical wireless communications,” IEEE Trans. Commun. Technol. 66(9), 4043–4058 (2018). [CrossRef]  

14. R. W. Engstrom, Photomultiplier Handbook (RCA Corporation, 1980).

15. R. L. Baer, “A model for dark current characterization and simulation,” in Sensors, Cameras, and Systems for Scientific/Industrial Applications VII, vol. 6068V. W. Xavier, ed. (SPIE, 2006), p. 606805.

16. L. Abbene and G. Gerardi, “High-rate dead-time corrections in a general purpose digital pulse processing system,” J. Synchrotron Radiat. 22(5), 1190–1201 (2015). [CrossRef]  

17. A. Gallivanoni, I. Rech, and M. Ghioni, “Progress in quenching circuits for single photon avalanche diodes,” IEEE Trans. Nucl. Sci. 57, 3815–3826 (2010). Conference Name: IEEE Transactions on Nuclear Science. [CrossRef]  

18. B.-W. Tzou, J.-Y. Wu, Y.-S. Lee, and S.-D. Lin, “Method to evaluate afterpulsing probability in single-photon avalanche diodes,” Opt. Lett. 40(16), 3774 (2015). [CrossRef]  

19. S. Vinogradov, T. Vinogradova, V. Shubin, D. Shushakov, and K. Sitarsky, “Probability distribution and noise factor of solid state photomultiplier signals with cross-talk and afterpulsing,” in Nuclear Science Symposium Conference Record (NSS/MIC) (IEEE, 2009), pp. 1496–1500.

20. M. C. Teich and W. J. McGill, “Neural counting and photon counting in the presence of dead time,” Phys. Rev. Lett. 36(13), 754–758 (1976). [CrossRef]  

21. L. Mandel, “Inversion problem in photon counting with dead time,” J. Opt. Soc. Am. 70(7), 873 (1980). [CrossRef]  

22. M. Srinivas, “Dead time corrections to photon counting statistics i: Classical theory,” Pramana 17(3), 203–216 (1981). [CrossRef]  

23. S. Cova, M. Ghioni, A. Lacaita, C. Samori, and F. Zappa, “Avalanche photodiodes and quenching circuits for single-photon detection,” Appl. Opt. 35(12), 1956–1976 (1996). Optica Publishing Group. [CrossRef]  

24. A. W. Ziarkash, S. K. Joshi, M. Stipcevic, and R. Ursin, “Comparative study of afterpulsing behavior and models in single photon counting avalanche photo diode detectors,” Sci. Rep. 8(1), 5076 (2018). [CrossRef]  

25. “NIR single photon detector user manual, SPD_oem_nir v1.0,” (2019). 18 Rue Alain Savary, 25000 Besançon, France.

26. “Time correlator user manual, CHRONOXEA v2.1,” (2020). 18 Rue Alain Savary, 25000 Besançon, France.

27. M. Wahl, M. Leifgen, M. Berlin, T. Röhlicke, H.-J. Rahn, and O. Benson, “An ultrafast quantum random number generator with provably bounded output bias based on photon arrival time measurements,” Appl. Phys. Lett. 98(17), 171105 (2011). [CrossRef]  

28. M. L. Larsen and A. B. Kostinski, “Simple dead-time corrections for discrete time series of non-poisson data,” Meas. Sci. Technol. 20(9), 095101 (2009). [CrossRef]  

29. S. H. Lee and R. P. Gardner, “A new g–m counter dead time model,” Appl. Radiat. Isot. 53(4-5), 731–737 (2000). [CrossRef]  

30. L. Neri, S. Tudisco, F. Musumeci, A. Scordino, G. Fallica, M. Mazzillo, and M. Zimbone, “Note: Dead time causes and correction method for single photon avalanche diode devices,” Rev. Sci. Instrum. 81(8), 086102 (2010). [CrossRef]  

31. R. L. Lucke, “Counting statistics for nonnegligible dead time corrections,” Rev. Sci. Instrum. 47(6), 766–767 (1976). [CrossRef]  

32. K. A. O’Donnell, “Correction of dead-time effects in photoelectric-counting distributions,” J. Opt. Soc. Am. A 3(1), 113–115 (1986). [CrossRef]  

33. P. Gaspard and X.-J. Wang, “Noise, chaos, and (ɛ, τ)-entropy per unit time,” Phys. Rep. 235(6), 291–343 (1993). [CrossRef]  

34. G. Boffetta, M. Cencini, M. Falcioni, and A. Vulpiani, “Predictability: a way to characterize complexity,” Phys. Rep. 356(6), 367–474 (2002). [CrossRef]  

35. A. M. Hagerstrom, T. E. Murphy, and R. Roy, “Harvesting entropy and quantifying the transition from noise to chaos in a photon-counting feedback loop,” Proc. Natl. Acad. Sci. 112(30), 9258–9263 (2015). [CrossRef]  

36. J. D. Hart, Y. Terashima, A. Uchida, G. B. Baumgartner, T. E. Murphy, and R. Roy, “Recommendations and illustrations for the evaluation of photonic random number generators,” APL Photonics 2(9), 090901 (2017). [CrossRef]  

37. M. Herrero-Collantes and J. C. Garcia-Escartin, “Quantum random number generators,” Rev. Mod. Phys. 89(1), 015004 (2017). [CrossRef]  

38. C. E. Shannon, “A mathematical theory of communication,” The Bell Syst. Tech. J. 27(3), 379–423 (1948). [CrossRef]  

39. P. Grassberger and I. Procaccia, “Estimation of the kolmogorov entropy from a chaotic signal,” Phys. Rev. A 28(4), 2591–2593 (1983). [CrossRef]  

40. Y. Kawaguchi, T. Okuma, K. Kanno, and A. Uchida, “Entropy rate of chaos in an optically injected semiconductor laser for physical random number generation,” Opt. Express 29(2), 2442–2457 (2021). [CrossRef]  

41. F. Takens, “Invariants related to dimension and entropy,” Proc. Thirteen. Coloquio Brasileiro de Matematica13, 353–359 (1983).

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (8)

Fig. 1.
Fig. 1. A schematic depicting the detector response to an electron carrier, which may be triggered by a photon, dark current, etc., in the presence of dead time. (a) The avalanche photodiode’s electrical response (black curve) to a triggered electron (blue circle) including afterpulsing, the additional fluctuations seen after the initial response. After each response, there is a dead time (shaded blue boxes) when the detector cannot detect any incoming electrons. The dead time, $\tau$, is adjustable by the user. (b) Electrical signal sent from the output of the detector after the SPAD response passes through a discriminator.
Fig. 2.
Fig. 2. Switch Open: Experimental setup used for measuring dark current. A SPAD set in continuous mode with no light entering it has its electrical output connected to the input of the time-tagging correlator. The time-tagged data is then sent to the computer to be post-processed. Switch Closed: Laser data experimental setup. A 1550 nm distributed feedback (DFB) laser is connected through single-mode (SM) fiber to 110 dB of fixed attenuation and up to 30 dB of variable attenuation. Photons are then sent through SM fiber to the SPAD. The electrical output of the SPAD is ultimately sent to the input of the time-correlator. The data is then sent to a computer to be post-processed.
Fig. 3.
Fig. 3. Dark count data with SPAD settings of 20 microsecond dead time. (a) A histogram (grey) of the interarrival times between consecutive dark counts taken at 10% efficiency. An exponential curve fit excluding the first 500 microseconds is shown in red. An inset image of the interarrival histogram with the first 2 milliseconds enlarged is included. (b) Interarrival time histogram at 20% efficiency.
Fig. 4.
Fig. 4. Dark count data with SPAD settings of 500 microsecond dead time. (a) A histogram (grey) of the interarrival times between consecutive dark counts taken at 10% efficiency. An exponential curve fit is shown in red. (b) Interarrival time histogram at 20% efficiency.
Fig. 5.
Fig. 5. Data with 500$\mu$s dead time. (a) A histogram of the normalized probability distribution (light blue) of dark counts which arrived in the time interval T = 0.015 seconds at 10% efficiency along with the dead-time corrected histogram (solid purple) generated from the iterative algorithm. Theoretical Poisson curves are included with varying values of $\lambda$, namely $\lambda _m$ (solid black), $\lambda _a$ (dotted red), and $\lambda _c$ (dashed orange). (b) Probability distribution for data taken at 20% efficiency with the corresponding dead-time correction and several theoretical Poisson curves.
Fig. 6.
Fig. 6. Laser data (10 million points) with an average of 650 photons per second (adjusted for dead time) taken with detector settings of 10% efficiency and 500 microsecond dead time. (a) Interarrival times on a log-linear plot (grey) with an exponential curve fit (red) based on the true photon counts detected per second, but excluding the first 500 microseconds. (b) The original histogram (light blue) of the frequency of photon arrivals per time interval T. The dead time corrected histogram (purple) can be compared to the theoretical Poisson fit (dashed orange) for the corrected lambda value. Poisson fits are plotted using both the $\lambda _m$ (black) and $\lambda _a$ (dotted red)
Fig. 7.
Fig. 7. Entropy rate analysis for dark count measurements with 500 $\mu$s dead time and (a) 10% quantum efficiency and (b) 20% quantum efficiency. The analytic solution for $h_1$ (Eq. (10)) and $h_2$ (Eq. (11)) are shown as black and dotted green lines, respectively. The Grassberger-Procaccia estimates of $h_2(\Delta t,d)$ do not depend on $d$, implying that there are no intersample correlations. The sharp decrease as $\Delta t$ decreases for small $\Delta t$ in the $d=3$ curve (red) is due to the finite size of the dataset used. Further, the Grassberger-Procaccia $h_2$ estimates agree excellently with the analytic solution for $h_2$, suggesting that the dark count measurements are well-modeled by a Poisson process. All entropy rate estimates were computed using 2 million recorded interarrival times.
Fig. 8.
Fig. 8. Entropy rate analysis for dark count measurements with 20 $\mu$s dead time and (a) 10% quantum efficiency and (b) 20% quantum efficiency. The analytic solution for $h_1$ (Eq. (10)) and $h_2$ (Eq. (11)) are shown as black and dotted green lines, respectively. The Grassberger-Procaccia estimates of $h_2(\Delta t,d)$ decrease with increasing $d$, implying that there are intersample correlations. The reduction with $d$ is stronger for the 20% efficiency case, suggesting that there are stronger intersample correlations in this case, which is expected for afterpulsing. Further, for both efficiencies the Grassberger-Procaccia $h_2$ estimate severely underestimates the analytic solution for $h_2$ even for $d=1$, reflecting that the dark count interarrival time histogram is much more strongly peaked than the exponential distribution expected for a Poisson process. All entropy rate estimates were computed using 2 million recorded interarrival times.

Equations (13)

Equations on this page are rendered with MathJax. Learn more.

P ( n ) = ( r T ) n e ( r T ) n ! ,
r = η ( P / h ν ) ,
r a = r m / ( 1 ( r m τ ) ) .
y ( t ) = A exp [ r ( t τ ) ]
p ( n , T , 0 ) = p ( n , T + n τ , τ ) + r = 0 n 1 [ p ( r , T + n τ , τ ) p ( r , T + ( n 1 ) τ , τ ) ] , n 1 , p ( 0 , T , 0 ) = p ( 0 , T , τ ) .
P ( k , λ ) = exp [ k ln λ λ ln Γ ( k + 1 ) ] ,
H 1 ( Δ t , d ) = k = 0 p k ( Δ t ; d ) log 2 ( p k ( Δ t ; d ) ) = log 2 ( p ( Δ t , d ) ) ,
p i ( Δ t , d ) = 1 N d { n u m b e r o f d s e q u e n c e s x j s . t . | x i x j | < Δ t } ,
H 2 ( Δ t , d ) = log 2 ( i p i 2 ( Δ t ; d ) ) = log 2 p 2 ( Δ t ; d ) ,
H ( Δ t ) = ( 1 p 0 ) log 2 ( 1 p 0 ) p 0 + log 2 ( p 0 ) ,
h 1 ( Δ t ) = r m log 2 ( r t Δ t e ) .
H 2 ( Δ t ) = 2 log 2 ( 1 e r t Δ t ) log 2 ( 1 e 2 r t Δ t ) .
h 2 ( Δ t ) = r m log 2 ( r t Δ t 2 )
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.