Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Layout pattern analysis and coverage evaluation in computational lithography

Open Access Open Access

Abstract

In advanced semiconductor technology nodes, the model accuracy of optical proximity correction (OPC) is the key for integrated circuit (IC) chip mask tape out, yield ramp up, and product time-to-market. An accurate model means a small prediction error for the full chip layout. As the full chip layout usually has large pattern variety, an optimal pattern set with good coverage is desired during the model calibration process. Currently, no existing solutions can provide the effective metrics to evaluate the coverage sufficiency of the selected pattern set before a real mask tape out, which may potentially cause higher re-tape out cost and product time-to-market delay due to the multiple rounds of model calibration. In this paper, we construct the metrics to evaluate the pattern coverage before any metrology data is obtained. The metrics are based on either the pattern’s intrinsic, numerical feature representation, or its potential model simulation behavior. Experimental results show a positive correlation between these metrics and lithographic model accuracy. An incremental selection method is also proposed based on the pattern simulation error. It reduces up to 53% of the model’s verification error range. These pattern coverage evaluation methods can improve the efficiency of OPC model building, and are, in turn, beneficial to the whole OPC recipe development process.

© 2023 Optica Publishing Group under the terms of the Optica Open Access Publishing Agreement

1. Introduction

During the semiconductor manufacturing process, lithography is used to transfer the patterns from an integrated circuit (IC) design inlayed on a photomask to the semiconductor wafer. As the dimension of the design shrinks to be smaller than the wavelength of the light source, lithography printing suffers from optical proximity effects and cannot yield the sufficient printing resolution without modifications to the raw designs. This modification process that compensates for these optical proximity effects is called optical proximity correction (OPC) [14]. This design-to-mask transformative process requires an accurate lithography model in order to simulate the whole lithography process (Fig. 1) [3] and to drive the design correction. Typically, a lithography model takes a predefined mathematical form with multiple open parameters. The model form may be derived from optical/physical/chemical processes [58] or constructed by a neural network [912]; while the open parameters associated with the model form need to be determined through a calibration process [13].

 figure: Fig. 1.

Fig. 1. Mathematical illustration of lithography process simulation.

Download Full Size | PDF

As technology nodes shrink in size, the error budget for the lithography process deceases, and the lithography model’s accuracy targets become increasingly challenging to achieve. An over-fitted model, which yields good performance on the calibration pattern set, may predict poorly on different patterns that show up in real chip designs. An inaccurate model tends to result in wafer defects that are not predicted by OPC verification, thus leading to lower-than-expected wafer yield once it goes into high-volume manufacturing.

A pattern set used for calibration with sufficient pattern coverage is an area that becomes increasingly important as model forms become more and more complicated. A common practice is to set up the calibration patterns based on geometrical properties according to an engineer’s experience. Generally, regular line/space or contact array patterns that have a certain range of critical dimension (CD) are selected with the intent to cover a large portion of the pattern space. Patterns selected from real chips are also added to provide additional design information for model calibration, while such manual selection suffers from high labor-intensity and low consistency. To address these issues, an image-based pattern selection (IBPS) [14] method has been demonstrated to automatically and smartly select patterns from a full real chip using feature vectors.

Here the feature vector (FV) is defined as an abstracted, numerical representation of the patterns, which is formed from the patterns’ aerial images. Feature vectorization is a widely-used technique in the computational lithography field for pattern selection [1517] and hotspot detection [1820]. Typical FVs which originate from the design layouts or the pattern clips after OPC generally contain only geometrical information [21,22]. In the IBPS method, FVs are converted from aerial images of pattern clips. In this way, all optical effects in the lithography model are considered when constructing the feature space with better physical correspondence.

Although IBPS method has been demonstrated in previous literature to improve the pattern coverage compared to a random selection, we still need to have methods to assess and quantify the pattern coverage in order for industry users to put together a set of calibration patterns that can best represent the design as the eventual goal. In this paper, we introduce the FV-based and simulation-based methods to analyze the coverage of a sub pattern set and evaluate their correlation to a model prediction’s performance. Feature vectors can enable one to check the relationship among patterns which then allows one to evaluate the pattern coverage given the overlapping status of a selected pattern subset to the full set of patterns within the feature space. This statistical evaluation method specifically focuses on the diversity of a selected pattern set. Its results suggest a positive correlation between the FV based metrics and the average model accuracy across a full pattern set. However, model accuracy can also be sensitive to a few critical patterns which could severely damage the model’s prediction accuracy. This phenomenon is difficult to be well captured by FV based metric alone, so the simulation based metric is introduced in this paper for complementation. The simulation-based method constructs the metrics by performing a model inaccuracy evaluation using a simulation-based flow mimicking the model calibration. Its metrics focus on the patterns with high sensitivity and criticality to the lithography model calibration process. As a result, a strong positive correlation is exhibited between these metrics and the model’s measured error range. It is also shown that the minor critical patterns can strongly deteriorate these metrics, leading to a poor coverage score. On the contrary, guided by these metrics, one can optimize the calibration pattern set by adding back into it the most critical patterns via an incremental selection methodology. Incremental selection simply picks the top ranking patterns that limit the simulation-based coverage score. This is done prior to the cycle of mask making, wafer exposure and metrology data collection, which can significantly reduce the overall turnaround time. The result verifies that incremental selection improves the final lithography model’s performance on those remaining critical patterns without sacrificing the accuracy of the pattern majority. It should be noted that FV-based method is recommended in the first round of OPC iterations as it is decoupled with resist model calibration, has short runtime, and high stability; while the simulation-based method, albeit with slow runtime, is more preferable and less expensive when being adopted in the later rounds as there already obtained a baseline model from previous rounds. It is expected that reasonably using the pattern selection and coverage evaluation methods at different stages in the model calibration flow, as shown in Fig. 2, can benefit the lithography model accuracy. In the flowchart, green blocks indicate the two pattern coverage evaluation methods, and the incremental pattern selection method introduced in this paper.

 figure: Fig. 2.

Fig. 2. Flow chart of pattern selection and coverage evaluation methods in the model calibration flow.

Download Full Size | PDF

2. Pattern coverage evaluation methods

2.1 FV based pattern coverage evaluation

Pattern diversity within a calibration pattern set is very important for model accuracy. When the calibration pattern set cannot represent all the pattern types in a real chip, the model prediction power will decrease significantly on certain patterns. The FV based method evaluates the pattern coverage in the feature space and focuses on pattern diversity. Three primary procedures which are used within this methodology include FV extraction, pattern clustering on the full set of patterns, and the pattern coverage calculation.

2.1.1 Feature vector extraction

In computational lithography, aerial images are generally obtained by applying an optical model to a mask layer. The intensity of an aerial image in the lithography system is given by [22]:

$$I({\hat{x},\hat{y}} )= \mathop {\int\!\!\!\int }\limits_{ - \infty }^{ + \infty } J({\hat{f},\hat{g}} )\left[ {{{\left|{\mathop {\int\!\!\!\int }\limits_{ - \infty }^{ + \infty } H({\hat{f} + {{\hat{f}}^{\prime}},\,\hat{g} + \hat{g}^{\prime}} )O({\hat{f}^{\prime},\hat{g}^{\prime}} ){e^{ - i2\pi ({\hat{f}^{\prime}\hat{x} + \hat{g}^{\prime}\hat{y}} )}}d\hat{f}^{\prime}d\hat{g}^{\prime}} \right|}^2}} \right]d\hat{f}d\hat{g}$$
where $({\hat{x},\hat{y}} )$ is the coordinate at the aerial image plane, $({\hat{f},\hat{g}} )$ is the coordinate at the pupil plane, $J({\hat{f},\hat{g}} )$ is the source’s mutual intensity function, $H({\hat{f},\hat{g}} )$ is the light propagation function, and $O({\hat{f},\hat{g}} )$ is the mask spectrum.

Since aerial images simultaneously contain both optical and geometric signatures, they are used instead of polygons to extract FVs through an unsupervised autoencoder model [2325] as shown in Fig. 3. By doing so, both geometric and optical information are encoded into the FVs.

 figure: Fig. 3.

Fig. 3. Illustration of FV extraction.

Download Full Size | PDF

An unsupervised autoencoder is used in our FV based method to construct the FV space. With a proper design, the autoencoder model can help to find the proper low dimensional manifold embedded within the high-dimension space [26]. This is needed to ensure that the data points maintain their structured similarity between the output’s low dimensional FV space and the original aerial image in high-dimensional space. The loss function in the autoencoder model, the major component of which is mean square loss, can be mathematically described by the following equation:

$$Loss = \; ||\textrm{Mask}(f({g(x )} )- x)||_2^2$$
where a customized Mask() function is added into our autoencoder’s loss function to highlight the aerial image’s centroid region or the physically guided components as shown in Fig. 4. In order to avoid any overfitting risk, the unsupervised autoencoder model is trained with aerial images from all patterns that are extracted from the full layout.

 figure: Fig. 4.

Fig. 4. Autoencoder training. Mask() is the customized function of autoencoder training loss function in our methodology.

Download Full Size | PDF

Every pattern’s FV is output after applying the autoencoder model. The resultant FV dimension is ∼1/25 of the raw image’s with minimal information loss, which significantly improves the runtime of this FV based method. With certain neural network training to ignore insignificant data by constraining the input/output of network to achieve high structural similarity while have sufficient reduction in the output FV dimension, the significant features in the original aerial images can be enhanced which may benefit the result of the following modules [26]. An example of the encoder training performance in the test case of this paper is shown in Fig. 5. For pattern with maximum encoder loss, the mean pixel-to-pixel difference between input image and recovered image is 0.008 while AI intensity contrast in it is ∼0.27, which means average pixel distortion is < 3%. For pattern with minimum encoder loss, the mean pixel-to-pixel difference between input image and recovered image is 0.002 while AI intensity contrast in it is ∼0.28, which means average pixel distortion is < 1%.

 figure: Fig. 5.

Fig. 5. Examples of encoder training performance.

Download Full Size | PDF

In the scope of this paper, a Euclidean distance is used to quantify the pattern differences and to generate the relationships between patterns in the FV space. Patterns or images which have higher similarities tend to have smaller FV distances as shown in Fig. 6. In the FV space, a pattern is considered covered if the distance from itself to any selected pattern is less than a distance threshold. The pattern clustering module is designed to provide a solution to define the distance threshold for checking covered and uncovered patterns.

 figure: Fig. 6.

Fig. 6. Illustration of pattern similarity vs FV distance: (a) reference pattern with 96 × 96 pixels, (b)-(e) are context patterns with similarity decreasing, and the FV distance to reference pattern is 0.12, 0.46, 0.88, and 1.45, respectively.

Download Full Size | PDF

2.1.2 Pattern clustering on a full layout

Based on the extracted FVs, pattern clustering is applied on a full pattern set in the way that the separation between each pair of patterns in a cluster is less than a distance threshold. An illustration of pattern clustering is shown in Fig. 7, where principal component analysis (PCA), a dimensionality-reduction method, is applied for visualization purposes.

 figure: Fig. 7.

Fig. 7. A 2D illustration of a clustered pattern set in design space. Patterns within the blue circle are in one cluster. The blue circle radius indicates the distance threshold.

Download Full Size | PDF

Since similar images have small FV distance, the characteristic of pattern similarity in the full pattern set can be recognized with the nearest neighbor distance of the full pattern set. By statistically analyzing the nearest neighbor (NN) distance distribution, e.g., calculating the mean value of the NN distances, a distance threshold of this known space is defined to cluster patterns into different groups. A pair of patterns that are nearer than this distance threshold in the FV space is eligible to be clustered together. The final clustering result is determined by identifying a minimum count of clusters containing all of the patterns. Denoting the number of cluster as N and the maximum pairwise distance of two patterns appearing in the same cluster as R, the clustering behavior can be mathematically described with the following optimization statement:

$$Minimize\; N,\; subject\; to\; R < distance\_threshod$$

This optimization problem can be solved by a greedy algorithm [27,28] to generate the clustering result. The detailed algorithm of greedy algo is listed as following:

Greedy (F, C) { // F is full pattern set, and C are representative patterns
 Initialize U = F, C = Ø
 while (U is not empty) {
 N[i] = loop U to collect neighbor in U < distance_threshold
 Pick i with largest count(N[i])
 add i to C
 remove N[i] from U
 }
 return C
 }

In real production tape-out process, greedy algo can be too expensive to adopt on a large data set from full chip layout. Thus, distributive computing of greedy algo could be introduced to achieve the result within acceptable runtime.

From the clustering result, one representative pattern set associated with the distance threshold can be obtained for a full pattern set by picking up the centroid pattern from each individual cluster. They can be repeatedly used to evaluate pattern coverage of different selected pattern sets against the full pattern set within the same FV space. It reduces the information redundancy and increases the computational efficiency. And the distance threshold can be used as a criterion to recognize covered and uncovered patterns within the pattern calculation module.

2.1.3 Pattern coverage calculation

A FV based pattern coverage index is calculated by comparing representative patterns to selected patterns. As shown in Fig. 8, a covered pattern means its minimum distance to any selected patterns is less than the distance threshold. The FV based pattern coverage ratio (FV-PCR) is calculated as

$$FV\_PCR = \; \frac{{Covered\,Pattern\,Count}}{{Total\; Representative\; Pattern\,Count}}$$

 figure: Fig. 8.

Fig. 8. A 2D illustration of covered patterns in FV space.

Download Full Size | PDF

The larger ratio indicates better pattern coverage. The FV-PCR’s key performance index (KPI) deteriorates when more and more patterns are found to be outside of the distance threshold from that of all of the selected patterns.

One advantage of our FV based pattern coverage evaluation method is that it has the capacity to support full chip calculations with a low runtime cost. In addition, it is flexible and convenient enough to be implemented with different use cases which can make use of the application’s KPI.

2.2 Simulation based pattern coverage evaluation

Similar to FV based pattern coverage evaluation method, the simulation based pattern coverage evaluation method could also save the turnaround time and the tape-out cost by improving the pattern coverage to enhance model accuracy and reduce re-calibration iterations. The pattern coverage simulation based method mimics a neural network’s model training process. It uses simulated data generated from a baseline physical model, which has an error RMS of 1.99 nm, an error Range of 14.69 nm, and a 3 Sigma of 7.32) as input. In this way, this method can be applied without the need for real metrology data. This is a benefit since gettering the real, ground truth data comes associated with a rather high portion of the cost when training a model. The overall methodology, metrics design and incremental selection will be introduced in the following sections.

2.2.1 Methodology

The metrics of our simulation-based method is designed based on model error severity estimation and analysis. Model error severity is estimated by disagreement between the trained neural network model and the baseline physical model from which the neural network model is trained. The flow is shown in Fig. 9(a). Firstly, the existing baseline physical model is simulated on the already selected pattern set to prepare the ground-truth data. Then the neural network model is trained with such data generated from the baseline model’s simulation results. By further doing simulation on a full pattern set using both models, the model disagreement errors and associated model severity can be extracted which is used to calculate the simulation-based metrics and drive the incremental selection.

 figure: Fig. 9.

Fig. 9. Simulation based pattern coverage evaluation method: (a) flow, (b) illustration of model disagreement error.

Download Full Size | PDF

Generally, the model disagreement error can be described as the sum of resist image mismatch between the deep learning model prediction and the baseline model prediction. In practical application, due to specific features derived from resist image such as critical dimension variation (CDV) is a key point, a reduced KPI of CDV is constructed for ranking in this work:

$$pattern\; severity\; = \; rank(\sum CDV({DL\,model,\,baseline\,model} ){|_{critical\,locations}})$$

As the constructed neural network is a deep learning network containing > 20 layers and > 5000 neurons, it has stronger fitting power than physical models [29,11]. This means, it can have good predictions not only on those selected patterns that are used for training the neural network but also on those unused but well covered patterns in the full pattern set. This means that larger absolute disagreement errors between the two models applied to the full pattern set can highlight uncovered patterns by the selected patterns. Metrics from the simulation based method are constructed with the simulated model error result, i.e., model disagreement as described in the following section.

2.2.2 Metrics design

The metrics of the simulation-based method are constructed by statistical analysis of the estimated model errors, including error range and standard deviation of the full pattern set. The error range (Err-R) indicates the maximum positive and minimum negative model disagreement error. The error standard deviation (Err-STD) shows the trend of the overall model disagreement error.

$$Err\_R = \max error - \min error$$
$$Err\_STD\; = \sqrt {\frac{{\sum {{({error - \overline {error} } )}^2}}}{N}} $$

Since the neural network model has its original fitting power, we infer that there is no perfect model to have zero model disagreement error even on the selected training pattern set. The residual error, also known as neural network model training error on the selected pattern set, should be taken into account in the pattern coverage metrics design. Thus, the following metrics are designed in this paper as the simulation-based pattern coverage evaluation KPIs.

  • 1. Error range difference ratio (ERDR):
    $$ERDR = \frac{{Full\,Pattern\,Set\,Err\_R - Selected\,Pattern\,Set\,Err\_R}}{{Full\,Pattern\,Set\,Err\_R}}$$
  • (2) Error standard deviation ratio (ESDR):
    $$ESDR = \frac{{Selected\,Pattern\,Err\_STD}}{{Full\,Pattern\,Err\_STD}}$$

In principle, with a smaller error range difference ratio, more unused patterns in the full pattern set are better predicted by the neural network model, thus indicating better pattern coverage. The error distribution reflects the disagreement between the neural network model and the physical model on the overall behaviors of the patterns. As the error standard deviation ratio gets closer to 1, the predication result of the trained neural network model and the baseline physical model on the selected patterns and the full patterns become more similar, indicating better coverage of the selected patterns on the full pattern set.

2.2.3 Incremental selection

Incremental selection optimizes the calibration pattern set by adding those top-ranking patterns in the severity ranking of the full pattern set as shown in Fig. 10. The top-ranking patterns are critical, unused and uncovered patterns in the original selected set. The originally selected pattern set is optimized by adding these critical patterns, thus approaching better coverage.

 figure: Fig. 10.

Fig. 10. Conceptual illustration of incremental selection to optimize a selected pattern set.

Download Full Size | PDF

The optimization is done before metrology. Incremental selection is proven in the result to improve OPC model accuracy with a shortened turnaround time.

3. Pattern selection and the experimental design

With the pattern coverage evaluation methods, an efficient model calibration flow is shown in Fig. 11. The experiment is designed according to this flow.

 figure: Fig. 11.

Fig. 11. Model calibration flow.

Download Full Size | PDF

3.1 Image based pattern selection

As the initial step of the model calibration flow, pattern selection is desired to be applied on a full layout to ensure the model’s coverage of a full layout pattern set. Compared to random selection, IBPS automatically selects patterns from a full layout with larger pattern diversity, thus achieving higher pattern coverage. In addition, IBPS reduces the efforts of manual pattern review and thus increases the engineering efficiency [14].

IBPS is a machine learning technique to select patterns that considers FV diversity. The initial input of IBPS is a full chip layout. The output of IBPS is the selected patterns that can represent the pattern diversity of the full chip layout. IBPS undergoes three primary steps as shown in Fig. 12, unique patterns extraction, FV extraction, pattern grouping and down sampling.

 figure: Fig. 12.

Fig. 12. Image based pattern selection flow.

Download Full Size | PDF

Unique patterns refer to geometrically unique patterns when extracted within a defined region which considers the proximity effect. Geometrically unique patterns mean the exact matching of each polygon and vertex. In general, unique patterns are extracted from a full chip layout to reduce the pattern redundancy and increase the efficiency. Feature vectors are extracted from aerial images using an unsupervised auto-encoder method. The pattern grouping module arranges patterns with similar FVs into one group meanwhile separate patterns with different FVs. In the down sampling module, certain number of patterns from all groups are selected to ensure better coverage on pattern geometry, lithography optical behavior and other model sensitive features.

As shown in Fig. 13, which schematically illustrate the distribution of 300 selected patterns in design space, IBPS generally has an even selection of patterns regardless of pattern density in the FV space while random selection is prone to select more patterns in the high population region. Under such a circumstance, IBPS can achieve a more diverse selection and thus better pattern coverage than random selection.

 figure: Fig. 13.

Fig. 13. A 2D illustration of the distribution of 300 selected patterns in design space: (a) IBPS shows a uniform selection thus achieving a higher FV coverage index, (b) random selection selects patterns in high populated regions. The blue circle radius visualizes the distance threshold used in the FV pattern coverage evaluation method.

Download Full Size | PDF

3.2 Experimental design

In this paper, we demonstrate two methods to evaluate pattern coverage of a calibration pattern set in a DUV memory case with an NA of 1.3, the feature size of which ranging from 80 nm to 120 nm. In order to prove the validity of the pattern coverage evaluation method and the corresponding pattern coverage metrics, random pattern selection and IBPS methods are implemented on a real full chip layout using a contact layer with different selected counts to prepare the experimental data as shown in Table 1. All the unique patterns in the real full chip are arranged as a full pattern set for this pattern coverage evaluation.

Tables Icon

Table 1. Pattern selection cases for experiment preparation

An incremental selection is applied on IBPS 100, IBPS 300 cases to add the 10 most critical patterns identified by the simulation based method from the unselected pool and optimize them for cases Incremental 110 and Incremental 310. After the pattern coverage evaluation, metrology data for each selected pattern set is collected from an exposed wafer, and one real lithography model is calibrated and verified for each set.

4. Results

The FV based and simulation-based pattern coverage evaluation methods are applied on the experimental cases. Physical models are calibrated with the metrology data of the selected patterns in each case. Ideally, the full pattern set should be used as a verification pattern set. However, there are over 400,000 unique patterns in the layout which are not affordable for metrology. Given a reasonable metrology budget, a 3,000-pattern subset is extracted from all of the unique patterns and are used as the verification pattern set for the experiment.

Model prediction accuracy is used to verify the effectiveness of the two pattern coverage evaluation methods. To quantify the model prediction accuracy, two KPIs (differences between the model predicted contour and the measured contour on wafer) are constructed by statistical analysis of the physical model’s verification errors. The first KPI is the root mean square (RMS) of the model verification error, indicating the overall prediction accuracy across the verification patterns. The second KPI is the model verification error range (max error – min error) which shows whether there are particular patterns for which the current model’s predication accuracy is sufficient or not. The smaller KPI resultant values obtained are indicative of us achieving better model accuracy.

4.1 FV pattern coverage metric correlates with model accuracy

In general, better design diversity of the calibration pattern set could help achieve a lower error RMS for the calibrated model. As shown in Fig. 14, the FV pattern coverage index shows a positive correlation to the model verification error RMS and range. An increasing FV pattern coverage index corresponds to a reduction in the model verification error RMS, and thus improves the corresponding model accuracy. The full set calibration results, which have an error RMS of 1.61 nm and an error Range of 17.57 nm, largely follows the result, indicating the efficacy of FV pattern coverage metric.

 figure: Fig. 14.

Fig. 14. Model accuracy vs FV pattern coverage index: (a) model verification error RMS vs coverage index, (b) model verification error range vs coverage index.

Download Full Size | PDF

However, the error range is also sensitive to the minor critical patterns since the FV pattern coverage index of case Incremental 110/310 (index value is 0.351/0.447) is very similar to IBPS 100/300 (index value is 0.351/0.449) while model error range is greatly reduced in the incremental selection cases. These phenomena can only be captured by the simulation-based pattern coverage metrics, the result of which will be shown in the following section.

4.2 Simulation pattern coverage metrics correlate with improved model accuracy

The simulation pattern coverage result of each case is shown in Table 2. The lower range difference ratio (Range Diff Ratio) as well as the higher standard deviation ratio (Std Ratio) indicates better pattern coverage.

Tables Icon

Table 2. Simulation-based pattern coverage result

As shown in Fig. 15, the simulation-based pattern coverage shows a strong positive correlation with the model verification error RMS and range. Under the same pattern count, the model verification error RMS/range of the IBPS set is smaller than the random selection set. This is consistent to the pattern coverage result where IBPS has better pattern coverage than random selection according to Table 2. Among all the cases, incremental pattern selection has the best pattern coverage and the smallest model verification error RMS/range. Specifically, incremental 110 has a 16.9% reduction in error range while incremental 310 reaches as large as a 53% reduction compared to IBPS 100 and IBPS 300, respectively. The results suggest that the incremental pattern selection has the capability to improve the model verification performance by further suppressing the error range whilst considering the minor critical patterns. In addition, incremental selection based on an initial selected set with larger pattern diversity, i.e., IBPS 300, tends to have more significant reduction in error range. This indicates that the diversity of patterns selected from IBPS during the initial selection and the severity based incremental selection complements each other very well.

 figure: Fig. 15.

Fig. 15. Model accuracy in different cases: (a) model verification error RMS of 100 patterns, (b) model verification error RMS of 300 patterns, (c) model verification error range of 100 patterns, (d) model verification error range of 300 patterns.

Download Full Size | PDF

The model verification error distributions of each experiment case are shown in Fig. 16. According to the distribution, every model has a very good prediction on major patterns thus model verification error RMS are within or around 2 nm. The model verification error range is dominated by a few particular critical patterns with a very large verification error. The model calibrated using selected patterns with the better pattern coverage KPI shows a better prediction on these patterns, thus improving model accuracy.

 figure: Fig. 16.

Fig. 16. Model verification error distribution: (a) 100 patterns, (b) 300 patterns, (c) full set. The large (negative) CD errors that occur on a single location are the patterns originated from the incremental selection. The upper and lower red line indicated the error range from full set.

Download Full Size | PDF

Figure 17 shows the measured contour and the model predicted contour of two critical patterns: (a) the critical pattern found by incremental selection (referred as “Pattern A”), (b) a pattern only found in the verification set (referred as “Pattern B”). It is demonstrated that incremental selection improves the model predication accuracy not only on those used patterns but also on those unused critical patterns. Further manual analysis on the critical patterns indicates that they are much smaller in CD (critical dimension) than the majority patterns in this test case.

 figure: Fig. 17.

Fig. 17. Pattern design, measured and model simulated contours: (a) the critical pattern incrementally selected in ‘Incremental 310’ case, (b) a critical pattern only in verification set. The two SEM images are extracted from the region highlighted by the two blue boxes.

Download Full Size | PDF

From Table 3, the wafer CD of Pattern A and Pattern B in the horizontal direction are 44.25 nm and 53.93 nm, respectively, while the majority patterns’ wafer CD are in the range of 75 to 130 nm. From the two SEM images in Fig. 17, it is shown that the model contour extracted from Incremental 310 case follows better to the wafer contour in both Pattern A and Pattern B compared to Random 300 and IBPS 300 cases. The model CDs of different cases are 26.53 nm (Random 300), 30.5 nm (IBPS 300), 39.9 nm (Incremental 310) for Pattern A, and 43.35 nm (Random 300), 44.66 nm (IBPS 300), and 52.05 nm (Incremental 310) for Pattern B, respectively. The much-reduced model error in Incremental 310 (Pattern A: -4.36 nm, Pattern B: -1.87 nm) compared to Random 300 (Pattern A: -17.73 nm, Pattern B: -10.58 nm) and IBPS 300 (Pattern A: -13.75 nm, Pattern B: -9.27 nm) can be ascribed to the inclusion of 10 critical patterns in Incremental 310 case. The model results from the vertical direction, albeit is worse than that of horizontal direction, still show similar trend among all three selection. Without the inclusion of critical patterns in the calibration set, there remains some difficulty for the model to predict them well.

Tables Icon

Table 3. Wafer and data of two critical patterns in three cases: Random 300, IBPS 300, and Incremental 310.

5. Conclusion

Pattern selection techniques and pattern coverage evaluation methods are playing increasingly important roles in lithography model calibration since they can help to reduce model calibration turn-around time and to enhance the model’s accuracy. The FV based pattern coverage evaluation and simulation-based methods have been illustrated in this paper to solve the pattern coverage evaluation problem. The FV based method shows a positive relationship to model prediction accuracy with short runtime and a relatively simple flow. The simulation based method shows a strong positive correlation with the model prediction accuracy. A comparison on the characteristic of these two methods is summarized in Table 4. The simulation-based pattern coverage evaluation also enables an incremental selection method before mask tape-out and wafer metrology. It can supplement an initial selection method, e.g., IBPS, and optimize the final pattern set, which can dramatically reduce model’s verification error range. Table 5 summarizes the differences between incremental selection method and the other two selection methods developed prior to this work. With both the pattern coverage evaluation and the pattern selection optimization, the OPC model calibration efficiency can be improved by reducing metrology time and the overall calibration turn-around time.

Tables Icon

Table 4. Summary of the advantages and disadvantages of the two pattern coverage evaluation methods

Tables Icon

Table 5. Comparison of the three different pattern selection methods

It is worth conducting further research to enhance the correlation between FV coverage and model prediction accuracy and thus enables the qualification and optimization of selected patterns using this FV based method alone. Doing so can further reduce the turnaround time of the whole flow. Lastly, it is valuable to use the simulation-based method to verify an existing and fixed model for a new layout and to identify patterns that need special monitoring.

Funding

National Natural Science Foundation of China (52130504).

Acknowledgments

The authors would like to express their thanks to Meng Liu, Gengxin Liu, Quan Chen, Jingang Xie, Jun He, Haidong Zhang, Fei Yan, Pingan Li, Liang Liu, and Yuting Cao in ASML Brion.

Disclosures

The authors declare the following competing interest: Jiaman Liu, Zhiqin Li, and Feng Yang have financial interests in Brion Technologies (Shenzhen) Co., Ltd., a subsidiary company of ASML that is focused on computational lithography and holds patents related to technologies for lithography modelling and pattern selection.

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

References

1. F. M. Schellenberg, O. Toublan, L. Capodieci, and B. Socha, “Adoption of OPC and the impact on design and layout,” in Proceedings of the 38th Design Automation Conference (2021), pp. 89–92.

2. A. Erdmann, Optical and EUV Lithography a Modeling Perspective (SPIE Press, 2021).

3. Z. Wang, X. Ma, G. R. Arce, and J. Garcia-Frias, “Fast optical proximity correction method based on nonlinear compressive sensing,” Opt. Express 26(13), 16736–16751 (2018). [CrossRef]  

4. E. Y. Lam and A. K. K. Wong, “Computation lithography: Virtual reality and virtual virtuality,” Opt. Express 17(15), 12259–12268 (2009). [CrossRef]  

5. C. Mack, Fundamental Principles of Optical Lithography: The Science of Microfabrication (John Wiley & Sons, 2008), Chap. 2.

6. C. A. Mack, J. J. Biafore, and M. D. Smith, MicroLithography (CRC Press. 2020).

7. P. Liu, L.-W. Zheng, M. Ma, Q. Zhao, Y.-F. Fan, Q. Zhang, M. Feng, X. Guo, T. Wallow, K. Gronlund, R. Goossens, G. Zhang, and Y.-W. Lu, “A physical resist shrinkage model for full-chip lithography simulations,” Proc. SPIE 9779, 97790Y (2016). [CrossRef]  

8. U. K. Klostermann, T. Mülders, D. Ponomarenco, T. Schmoeller, J. V. de Kerkhove, and P. D. Bisschop, “Calibration of physical resist models: methods, usability, and predictive power,” J. Micro/Nanolith. MEMS MOEMS 8(3), 033005 (2009). [CrossRef]  

9. K. Adam, S. Ganjugunte, C. Moyroud, K. Shchehlik, M. Lam, A. Burbine, G. Fenger, and Y. Granik, “Using machine learning in the physical modeling of lithography processes,” Proc. SPIE 10962, 109620F (2019). [CrossRef]  

10. X. Ma, Q. Zhao, H. Zhang, Z.-Q. Wang, and G. R. Arce, “Model-driven convolution neural network for inverse lithography,” Opt. Express 26(25), 32565–32584 (2018). [CrossRef]  

11. Y.-S. Kim, S. Lee, Z.-Y. Hou, et al., “OPC Model accuracy study using high volume contour based gauges and deep learning on memory device,” Proc. SPIE 10959, 1095913 (2019). [CrossRef]  

12. Z.-Q. Li, L.-S. Dong, X.Y. Jing, X. Ma, and Y.-Y. Wei, “High-precision lithography thick-mask model based on a decomposition machine learning method,” Opt. Express 30(11), 17680–17697 (2022). [CrossRef]  

13. K. Lai, “Review of computational lithography modeling: focusing on extending optical lithography and design-technology co-optimization,” Adv. Opt. Technologies 1(4), 249–267 (2012). [CrossRef]  

14. R.-C. Sun, D.-K. Kang, C. Jia, M. Liu, D.-B. Shao, Y.-S. Kim, J.-H. Shin, M. Simmons, Q. Zhao, M. Feng, Y.-Q. Zhao, S.-B. Wang, S.-H. Kim, S.-W. Ko, S.-Y. Kim, J.-S. Choi, and C.-H. Park, “Enhancing model accuracy and calibration efficiency with image-based pattern selection using machine learning techniques,” Proc. SPIE 11613, 116130Y (2021). [CrossRef]  

15. L.-F. Liao, S.-K. Li, X.-Z. Wang, L.-B. Zhang, P.-Z. Gao, Y.-Y. Wei, and W.-J. Shi, “Critical pattern selection method for full-chip source and mask optimization,” Opt. Express 28(14), 20748–20763 (2020). [CrossRef]  

16. M.-C. Tsai, S. Hsu, L.-Q. Chen, Y.-W. Lu, J.-W. Li, F. Chen, H. Chen, J. Tao, B.-D. Chen, H.-Y. Feng, W. Wong, W. Yuan, X.-Y. Li, Z.-P. Li, L. Li, R. Dover, H.-Y. Liu, and J. Koonmen, “Full-chip source and mask optimization,” Proc. SPIE 7973, 79730A (2011). [CrossRef]  

17. D.-Q Zhang, G.-S. Chua, Y.-M. Foong, et al., “Source mask optimization methodology (SMO) and application to real full chip optical proximity correction,” Proc. SPIE 8326, 83261 V (2012). [CrossRef]  

18. W. Yuan, Y.-F. Lu, M. Li, B.-Y. Pan, Y. Gao, Y. Tian, Z.-Q. Li, L. Ji, Y. Huang, H. Chen, Y.-L. Yao, and S. Park, “Machine learning hotspot prediction significantly improve capture rate on wafer,” in International Workshop on Advanced Patterning Solutions (2020), pp. 1–4.

19. Y.-B. Lin, X.-Q. Xu, J.-J. Ou, and D. Z. Pan, “Machine learning for mask/wafer hotspot detection and mask synthesis,” Proc. SPIE 10451, 104510A (2017). [CrossRef]  

20. T. Matsunawa, S. Nojima, and T. Kotani, “Automatic layout feature extraction for lithography hotspot detection based on deep neural network,” Proc. SPIE 9781, 97810 H (2016). [CrossRef]  

21. X.-L. Shi, Y.-H. Zhao, S.-M. Cheng, M. Li, W. Yuan, L. Yao, W.-H. Zhao, Y.-J. Xiao, X.-H. Kang, and A. Li, “Optimal feature vector design for computational lithography,” Proc. SPIE 10961, 109610O (2019). [CrossRef]  

22. A. K. Wong, Optical Imaging in Projection Microlithography (SPIE, 2005), Chap. 4.

23. S. J. Wetzel, “Unsupervised learning of phase transitions: From principal component analysis to variational autoencoders,” Phys. Rev. E 96(2), 022140 (2017). [CrossRef]  

24. M. Maggipinto, C. Masiero, A. Beghi, and G. A. Susto, “A convolutional autoencoder approach for feature extraction in virtual metrology,” Procedia Manuf. 17, 126–133 (2018). [CrossRef]  

25. X.-Y. Liu, G.-Z. Shao, Y.-W. Liu, X.-W. Liu, J.-Y. Li, X.-L. Chen, and Y.-K. Chen, “Deep classified autoencoder for lithofacies identification,” IEEE Trans. Geosci. Electron. 60, 5909914 (2022). [CrossRef]  

26. D. Holden, J. Saito, T. Komura, and T. Joyce, “Learning motion manifolds with convolutional autoencoders,” in SIGGRAPH Asia Technical Briefs (2015), pp. 1–4.

27. V. Chvatal, “A Greedy Heuristic for the Set-Covering Problem,” Math. Oper. Res. 4(3), 233–235 (1979). [CrossRef]  

28. P. Slavík, “A Tight Analysis of the Greedy Algorithm for Set Cover,” J. Algorithms 25(2), 237–254 (1997). [CrossRef]  

29. F. X. Zach, “Neural-network-based approach to resist modeling and OPC,” Proc. SPIE 5377, 670–679 (2004). [CrossRef]  

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (17)

Fig. 1.
Fig. 1. Mathematical illustration of lithography process simulation.
Fig. 2.
Fig. 2. Flow chart of pattern selection and coverage evaluation methods in the model calibration flow.
Fig. 3.
Fig. 3. Illustration of FV extraction.
Fig. 4.
Fig. 4. Autoencoder training. Mask() is the customized function of autoencoder training loss function in our methodology.
Fig. 5.
Fig. 5. Examples of encoder training performance.
Fig. 6.
Fig. 6. Illustration of pattern similarity vs FV distance: (a) reference pattern with 96 × 96 pixels, (b)-(e) are context patterns with similarity decreasing, and the FV distance to reference pattern is 0.12, 0.46, 0.88, and 1.45, respectively.
Fig. 7.
Fig. 7. A 2D illustration of a clustered pattern set in design space. Patterns within the blue circle are in one cluster. The blue circle radius indicates the distance threshold.
Fig. 8.
Fig. 8. A 2D illustration of covered patterns in FV space.
Fig. 9.
Fig. 9. Simulation based pattern coverage evaluation method: (a) flow, (b) illustration of model disagreement error.
Fig. 10.
Fig. 10. Conceptual illustration of incremental selection to optimize a selected pattern set.
Fig. 11.
Fig. 11. Model calibration flow.
Fig. 12.
Fig. 12. Image based pattern selection flow.
Fig. 13.
Fig. 13. A 2D illustration of the distribution of 300 selected patterns in design space: (a) IBPS shows a uniform selection thus achieving a higher FV coverage index, (b) random selection selects patterns in high populated regions. The blue circle radius visualizes the distance threshold used in the FV pattern coverage evaluation method.
Fig. 14.
Fig. 14. Model accuracy vs FV pattern coverage index: (a) model verification error RMS vs coverage index, (b) model verification error range vs coverage index.
Fig. 15.
Fig. 15. Model accuracy in different cases: (a) model verification error RMS of 100 patterns, (b) model verification error RMS of 300 patterns, (c) model verification error range of 100 patterns, (d) model verification error range of 300 patterns.
Fig. 16.
Fig. 16. Model verification error distribution: (a) 100 patterns, (b) 300 patterns, (c) full set. The large (negative) CD errors that occur on a single location are the patterns originated from the incremental selection. The upper and lower red line indicated the error range from full set.
Fig. 17.
Fig. 17. Pattern design, measured and model simulated contours: (a) the critical pattern incrementally selected in ‘Incremental 310’ case, (b) a critical pattern only in verification set. The two SEM images are extracted from the region highlighted by the two blue boxes.

Tables (5)

Tables Icon

Table 1. Pattern selection cases for experiment preparation

Tables Icon

Table 2. Simulation-based pattern coverage result

Tables Icon

Table 3. Wafer and data of two critical patterns in three cases: Random 300, IBPS 300, and Incremental 310.

Tables Icon

Table 4. Summary of the advantages and disadvantages of the two pattern coverage evaluation methods

Tables Icon

Table 5. Comparison of the three different pattern selection methods

Equations (9)

Equations on this page are rendered with MathJax. Learn more.

I ( x ^ , y ^ ) = + J ( f ^ , g ^ ) [ | + H ( f ^ + f ^ , g ^ + g ^ ) O ( f ^ , g ^ ) e i 2 π ( f ^ x ^ + g ^ y ^ ) d f ^ d g ^ | 2 ] d f ^ d g ^
L o s s = | | Mask ( f ( g ( x ) ) x ) | | 2 2
M i n i m i z e N , s u b j e c t t o R < d i s t a n c e _ t h r e s h o d
F V _ P C R = C o v e r e d P a t t e r n C o u n t T o t a l R e p r e s e n t a t i v e P a t t e r n C o u n t
p a t t e r n s e v e r i t y = r a n k ( C D V ( D L m o d e l , b a s e l i n e m o d e l ) | c r i t i c a l l o c a t i o n s )
E r r _ R = max e r r o r min e r r o r
E r r _ S T D = ( e r r o r e r r o r ¯ ) 2 N
E R D R = F u l l P a t t e r n S e t E r r _ R S e l e c t e d P a t t e r n S e t E r r _ R F u l l P a t t e r n S e t E r r _ R
E S D R = S e l e c t e d P a t t e r n E r r _ S T D F u l l P a t t e r n E r r _ S T D
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.