Solving Vocabulary Problems with Interactive Query Expansion

曾元顯
Yuen-Hsien Tseng
Associate Professor
Department of Library and Information Science
Fu Jen Catholic University
Email: tseng@blue.lins.fju.edu.tw
http://blue.lins.fju.edu.tw/~tseng/papers/solvoc/solvoc.htm
June. 15, 1998
Accessed times


Abstract

One of the major causes of search failures in information retrieval systems is vocabulary mismatch. This paper presents a solution to the vocabulary problem through two strategies known as term suggestion (TS) and term relevance feedback (TRF). In TS, collection-specific terms are extracted from the text collection. These terms and their frequencies constitute the keyword database for suggesting terms in response to users' queries. One effect of this term suggestion is that it functions as a dynamic directory if the query is a general term that contains broad meaning. In term relevance feedback, terms extracted from the top-ranked documents retrieved from the previous query are shown to users for relevance feedback. This kind of local TRF expands users' search vocabularies and guides users in search directions closer to their goals. In our experiment, interactive TS provides very high precision rate while achieving similar recall rate as n-gram matching. Local TRF achieves improvement in both precision and recall rate in full-text News database and degrades slightly in recall rate in bibliographic database due to the very limited source of information for feedback. In terms of Rijsbergen's combined measure of recall and precision, both TS and TRF achieve better performance than n-gram matching, which implies that the greater improvement in precision rate compensates the slightly degradation in recall rate for TS and TRF. We conclude that both TS and TRF provide users with richer guidance and more predictable results than n-gram matching alone.

Keywords: interactive query expansion, relevance feedback, term suggestion, performance evaluation, information retrieval

1. Introduction

Users of today's information retrieval systems often expect that information can be found faster and easier. However they may not have the skill to search effectively. Although advanced retrieval models such as vector space model or probability mode with document ranking capability have made searching more easier, librarians or experienced users used to Boolean searching have seen these systems more of a black box, with the system, not the user, doing most of the work [1]. Novice or casual users may desire a retrieval system with rich user guidance, while domain experts or experienced users may wish a reliable and predictable retrieval system to work with. Suggestions for further searching and facilities for providing users more control over the search process may be what is really needed, especially when search failures occur.

The large volume of ever-increasing information poses another problem. Excessive results often return from a simple query. Users seem have little patience in inspecting the results in the next page, partly because of another waiting for a "less relevant" page. Thus the challenge is to design a retrieval system with high precision rate while maintaining sufficient recall rate for the benefits of most users.

In this paper, we seek to provide solutions to the above problems observed in our past experience with a retrieval system we have been developing. The information retrieval system is based on n-gram indexing model and vector space retrieval model. So far it supports natural language-like queries, fuzzy matching, document ranking, term suggestion and term relevance feedback. It has been applied since December 1997 to the Online Public Access Catalog (OPAC) of the library of Fu Jen Catholic University [2]. The retrieval system has also been applied to a full-text collection, where documents come from the online news articles from Central Daily News [3] and China Times [4].

The distinct features in this system are term suggestion and term relevance feedback. In term suggestion, users are shown a ranked list of terms "suggested" by the system based on the matching of the initial query with a keyword database. Users can then select appropriate terms to query the document database. In term relevance feedback, terms extracted from the top-ranked documents retrieved from the previous query are shown to users for relevance feedback. This paper will illustrate their applications and describe an evaluation of their retrieval effectiveness.

Another features in our retrieval system are relevance ranking and syntactically approximate matching, or fuzzy matching in short. Relevance ranking ranks retrieved documents according to some similarity measurement between queries and documents. High-similarity documents are placed at the top of the result set in an attempt to help users locate relevant documents more easily. Such ranking is enabled by the fuzzy matching which retrieves documents containing string patterns syntactically approximate to the query string. Fuzzy matching is a powerful function to free users from worrying about search failure due to inexact query expression. However, it is also the main cause that baffles most users for its complex similarity measurement. As we will show later, fuzzy matching alone may produce unexpected results. It is better used in the first-half phase of term suggestion or term relevance feedback.

The rest of the paper is organized as follows: In the next section, our approach to term suggestion and term relevance feedback is described. A number of retrieval examples is illustrated. In Section 3, we describe the retrieval experiments and the results. Finally Section 4 concludes the paper.

2. Term Suggestion and Term Relevance Feedback

One of the major causes of search failures in information retrieval systems is vocabulary mismatch [5]. Users often input queries containing terms that do not match the terms used in the relevant documents. Retrieval failures of such cases can be effectively alleviated through strategies known as relevance feedback or query expansion [6-14]. In these strategies, additional query terms or re-weighted terms are added to the original query to retrieve more relevant documents. This modification of the original query can be done by the retrieval system in corporation with the user interactively.

Relevance feedback can take the form of document relevance feedback (DRF) or term relevance feedback (TRF). In DRF, users are asked to judge relevant documents retrieved from the previous query and resubmit the selected documents to the system. The system then extracts terms from these documents for further searching. This kind of relevance feedback is easy to implement. However, users in this process have no control over which terms are more appropriate than the others. Besides, the requirement for judging a document relevant or irrelevant may place an extra burden on users, especially when documents are in lengthy full text. Thus in our design, we do not use document relevance feedback.

In TRF, terms extracted from retrieved documents are shown in certain order of listing for user selection. Users then inspect the term list and select those terms related to their goals for another round of searching. The primary difficulty in this process is that the system should be able to extract meaningful terms suitable for perusal. Because Chinese texts have no explicit word boundaries, TRF for Chinese is not possible without a term extraction module. To overcome this problem, we used a fast keyword extraction algorithm [15-16] to extract terms from the search results. Retrieved documents in the same result page are analyzed. Keywords or key-phrases whose frequencies exceed a certain threshold (usually 1 or 2) are extracted and presented to users at the bottom of the same result page. An example from the results of querying the bibliographic database is illustrated in Fig. 1, where the query term is "artificial intelligence". Another example from the full-text News database is in Fig. 2, where the query term is "禽性流行性感冒" ("bird flu"). This kind of local TRF provides users more search terms which are semantically related to (not necessarily syntactically similar to) the query terms under the same topic. An example is the term "expert systems" in Fig. 1. More examples are in Fig. 2, such as "香港" (Hong Kong), "禽流感病毒" (bird flu virus), and "H5N1". Those who care about how serious the tourism has been affected might choose the terms "旅遊" (tourism), "航空" (air service), "退費" (refund), and "旅客" (traveler) for further searching. Thus TRF has the effects of expanding users' search vocabulary and of guiding users in search directions closer to their goals.

Figure 1. An example from bibliographic database for term relevance feedback.

Figure 2. An example from News database for term relevance feedback.

However, there is a problem with relevance feedback: If there is no relevant documents retrieved from the initial query, there may not have relevant terms (or documents) for feedback. Thus it is important to be able to suggest terms before queries are submitted to the document database. This would require the system to have a global term list or keyword database to present vocabularies related to the initial queries. Again we applied the fast keyword extraction algorithm for term extraction. From the bibliographic database of 356000 titles, we have obtained a keyword database of over 110,000 terms. From the full-text News database, 65,500 terms are extracted from 13,035 news articles.

With these collection-specific keyword resources, users can search the document databases directly or search the keyword databases first and then the document databases indirectly. By searching the keyword databases first, the system prompts a ranked list of related terms and their estimated term frequencies. With the understanding of the real terms stored in the document database and of their frequency distributions, users may choose appropriate terms for a more effective searching.

A search example is illustrated in Fig. 3, where the query is "中經院", an abbreviation of "中華經濟研究院" (Chung-Hua Institution for Economic Research, CHIER). Terms related to the query are retrieved with fuzzy matching based on vector space retrieval model and n-gram indexing model. They are sorted by match score (the first column) and then by term frequency (the third column). Note a relevant item, namely the full name of CHIER, has been retrieved with very low match score (the highest score is 1000, standing for exact match). Such low-score matching may only be possible for these keyword databases, because each record contains only a few characters or words so that any character hit or word hit is useful to the query. This result prompts users more complete terms for searching the document database.

The keyword database has an additional effect that if the query term is too broad, the query results function as a dynamic directory. That is, more specific terms under the category specified by the query are present in the results. From this dynamic directory, users are able to understand the distribution of the collection by inspecting the term frequencies and then choose more appropriate terms for searching, avoiding excessive results by submitting the original query to the document database. The example in Fig. 4 with the query term "history of literature" illustrates this effect.


Figure 3. An example for term suggestion.

Figure 4. An example illustrating the effect of "dynamic directory".

Our system is based on 1-gram and 2-gram indexing and vector space model. This way of n-gram indexing provides fuzzy matching capability and thus helps locate low-score terms in keyword databases, such as those found in Fig. 3. If this matching strategy is applied to the document database with the terms selected from the keyword database, the results might be a little unexpected. Again, take Fig. 3 as an example. After selecting the full name "中華經濟研究院" and the abbreviation "中經院" of CHIER for searching the document database, the user may expect that all the results are related to CHIER. However, Fig. 5 shows that between the documents containing "中華經濟研究院" and those containing "中經院", there are documents in between containing "經濟研究院" (Institute for Economy Research), "臺灣經濟研究院" (Taiwan Institute for Economy Research),"經濟研究" (economy research), and "中央研究院" (Academia Sinica), because of their higher match scores or even match scores. Relevant documents are interfered with irrelevant ones due to the n-gram matching strategy in this case. To overcome this problem, making the results more predicable, we adopt "term matching strategy" for TS and TRF. That is, only documents containing those terms selected from TS and TRF are retrieved. This might degrade the recall rate, however. The remedy is to add enough related terms to balance between precision and recall rates, while maintaining a predictable result set.

In the next section, we will evaluate the performance among these three retrieval approaches.

Figure 5. An unexpected effect of n-gram matching


3. Experimental Results

Three retrieval approaches were compared for their retrieval effectiveness, namely original query with n-gram matching (the baseline), term suggestion (TS) with term matching, and term relevance feedback (TRF) with term matching. There were two test collections in this experiment. One was an OPAC bibliographic database with 356,000 titles, the other was a full-text News database from 13,035 online news articles, as mentioned previously. To form two sets of queries for each test collection, seven MS students majored in library and information science were invited to make up queries. Each of them created 5 queries for each test collection. A number of queries were removed from the 70 queries, some because they were redundant, others because there were too little (less than 5) relevant documents to match the queries. This left 30 queries for each of the test collection, as shown in Appendix 1. One example of the queries was "飛碟協會斂財事件" (The swindle event of UFO Association).

The relevance judgement were made by the seven students. Only the first N (N=50 in this experiment) top-ranked documents were judged, in an attempt to reflect the patterns that users behave when using a ranking retrieval system. Because the total relevant records for each query is unknown due to the large volume of the collections, the total number of relevant records was assumed to be the largest one found by any of these three retrieval approaches. Appendix 2 lists the precision and recall rates for each query of the two test databases. In Table 1, average precision and recall rates, and performance improvement over the baseline approach are listed.

It can be seen that TS with term matching achieves very high precision rate in both test collections while maintaining similar recall rate as n-gram matching. For TRF with term matching, it achieves very high precision rate but slightly lower recall rate in bibliographic database, and both slightly higher precision and recall rates in News database. The reason that local TRF with term matching degrades slightly in recall rate in bibliographic database is due to the relatively few texts contained in a page of search results, which includes at most 20 titles. In average, there are only 9.2 terms for relevance feedback. However, in the full-text News database, the average number of terms for feedback is 57, a richer set of search terms and thus a better performance in both precision and recall rates.

Bibliographic Collection
News Collection
Precision
Recall
Precision
Recall
N-gram matching

(baseline)
0.43 0.79 0.43 0.79
TS with term matching

increase over baseline
0.73

+69.2%

0.80

+2.4%

0.55

+28.9%

0.78

-1.1%

TRF with term matching

increase over baseline
0.69

+61.3%

0.73

-6.6%

0.45

+6.1%

0.83

+5.0%

Table 1. Average precision and recall rates, and performance improvement over the baseline approach.

The experiment can also be evaluated with a combined measure of precision and recall, E, developed by van Rijsbergen [17]. The evaluation measure E is defined as:


where P stands for precision, R stands for recall, and b is a measure of the relative importance, to a user, of recall and precision. For example, b levels of 0.5 indicates that a user was twice as interested in precision as recall, b levels of 1, indicates that a user was interested in precision and recall evenly, and b levels of 2 indicates that a user was twice as interested in recall as precision.

The range of the E measure is between 0 and 1, inclusively. The smaller the value E, the better the performance in precision and in recall. Because the value of this E measure is inversely proportional to the performance, we modified the measure into


such that 0 ME 1, and that ME is positively proportional to the performance. (This is more widely known as F measure.) With this modification, Table 2 lists the ME values for the three retrieval approaches with respect to b values at 0.5, 1, and 2.
Bibliographic Collection
News Collection
b
0.5
1.0
2.0
0.5
1.0
2.0
N-gram matching

(baseline)
0.47 0.55 0.67 0.47 0.55 0.68
TS with term matching

increase over baseline
0.74

+57.4%

0.76

+38.2%

0.79

+17.9%

0.58

+23.4%

0.65

+18.2%

0.72

+5.9%

TRF with term matching

increase over baseline
0.70

+48.9%

0.71

+29.1%

0.72

+7.5%

0.50

+6.4%

0.59

+7.3%

0.71

+4.4%

Table 2. ME measures and performance improvement over the baseline approach.

The result in Table 2 shows that both TS and TRF perform better than n-gram matching alone in terms of Rijsbergen's combined measure. This means that the much greater improvement in precision rate compensates the slightly degradation in recall rate for both TS and TRF as observed in Table 1.

4. Conclusions

An approach to term suggestion (TS) and term relevance feedback (TRF) for interactive query expansion is illustrated. Collection-specific terms are extracted from the text collection. These terms and their term frequencies constitute the keyword database for term suggestion. One effect of this term suggestion is that it functions as a dynamic directory if the query is a general term that contains broad meaning.

While suggested terms are syntactically similar to query terms, terms extracted from retrieved documents are semantically related to the query terms. This kind of local TRF expands users' search vocabularies and guides users in search directions closer to their goals. Thus both TS and TRF with term matching provide richer user guidance and more predictable results than n-gram matching alone.

In our performance evaluation experiment, interactive TS provides high precision rate while achieving similar recall rate as n-gram matching. Local TRF achieves improvement in both precision and recall rates in News collection and degrades slightly in recall rate in bibliographic database due to the very limited source of information for feedback. However, in terms of combined measure of precision and recall, both TS and TRF with term matching perform better than n-gram matching alone. This means that the much greater improvement in precision rate compensates the slightly degradation in recall rate for both TS and TRF. Nevertheless, It is anticipated that a thesaurus with global term-term relationships and with term matching capability may achieve greater performance improvement in both precision rate and recall rate, while meeting users' expectations for predicable results and for user guidance.

References

  1. Carol Tenopir, "Generations of Online Searching," Library Journal, September, pp.128-130, 1996
  2. "Online Public Access Catalog, Library of Fu Jen Catholic University," http://xlib.fju.edu.tw/
  3. "Central Daily News" http://www.cdn.com.tw/
  4. "China Times" http://www.chinatimes.com.tw/
  5. W. Bruce Croft, "What Do People Want from Information Retrieval?" D-Lib Magazine, Nov., 1995. Http://mirrored.ukoln.ac.uk/lis-journal/dlib/dlib/dlib/novermber95/11croft.html.
  6. Gerard Salton, "Automatic Text Processing: The Transformation, Analysis, and Retrieval of Information by Computer" Addison-Wesley, 1989.
  7. Donna Harman, "Relevance Feedback and Other Query Modification Techniques" in Information Retrieval: Data Structure and Algorithm, edited by William B. Frakes and Ricardo Baeza-Yates, Prentice-Hall, 1992.
  8. Donna Harman, "Towards Interactive Query Expansion." Paper presented at ACM Conference on Research and Development in Information Retrieval, Grenoble, France, 1988.
  9. Hancock-Beaulieu, "Interactive Query Expansion in an OPAC: Interface and Retrieval Issues", Journal of Document & Text Management, 2(3)1994, 172-185. http://www.aslib.co.uk/caa/abstracts/open/95-0980.html
  10. Padmini Srinivasan, "Query Expansion and MEDLINE" Information Processing & Management, Vol. 32, No. 4, 1996, pp. 431-443.
  11. Morris Hirsch, David Aronow, "Suggesting Terms for Query Expansion in a Medical Information Retrieval System", Proceedings of the 19th Annual Symposium on Computer Applications in Medical (SCAMC). JAMIA, 1995, p. 965. Also available at http://ciir.cs.umass.edu/info/psfiles/irpubs/ir-63.ps.gz
  12. Bruce R. Schatz and Hsinchun Chen, "Interactive Term Suggestion for Users of Digital Libraries: Using Subject Thesauri and Co-occurrence Lists for Information Retrieval", ACM Digital Library Conference, 1996, http://dli.grainger.uiuc.edu/schatzDL96.htm
  13. Xu, Jinxi and Croft, W. Bruce. "Query Expansion Using Local and Global Document Analysis," Proceedings of the 19th Annual ACM-SIGIR Conference, 1996, pp.4-11
  14. James W. Cooper and Roy J. Byrd, "Lexical Navigation: Visually Prompted Query Expansion and Refinement," Proceedings of Digital Libraries, Philadelphia, USA, pp.237-246, 1997.
  15. Yuen-Hsien Tseng, "Fast Keyword Extraction of Chinese Documents in a Web Environment," Information Retrieval Workshop for Asia Languages - 1997, Oct. 8-9, Japan, pp. 81-87.
  16. Yuen-Hsien Tseng, "Multilingual Keyword Extraction for Term Suggestion," to appear in 21st International ACM SIGIR Conference on Research and Development in Information Retrieval - SIGIR '98, Aug. 24-28, Australia, 1998
  17. W. B. Frakes, "Introduction to Information Storage and Retrieval Systems" in Information Retrieval: Data Structure and Algorithm, edited by William B. Frakes and Ricardo Baeza-Yates, Prentice-Hall, 1992.




Appendix 1: Query sets for the test collections

The set of queries for the bibliographic database:
  1. 台灣經濟發展之研究
  1. 如何投資理財
  1. 紅樓夢對中國文學的影響
  1. 休閒活動之研究
  1. 新聞媒體對兒童之影響
  1. 社會秩序之重建
  1. 美國留學資訊
  1. 兒童發展心理學
  1. 張愛玲的小說
  1. 微生物學實驗
  1. 中國歷代婦女在法律上的地位
  1. 人力資源管理實務
  1. 台商在大陸投資的情況
  1. 中國通史教學
  1. 電視廣告對兒童之影響
  1. 台灣的日本語教材
  1. 青少年犯罪與家庭之關係
  1. 輔大學生社團活動與學習成就之關係
  1. 周代祖先祭祀制度
  1. 演進中的電影語言
  1. 技術職業教育
  1. 廣告內容分析研究
  1. 台灣的原住民
  1. 運動員應如何避免運動傷害之研究
  1. 兒童發展與心理學
  1. 數位影像處理
  1. 追尋人生的意義
  1. 人工智慧系統之開發
  1. 讀書的方法
  1. 邊疆民族之風俗

The set of queries for the News database:
  1. 戴安那王妃死亡之旅
  1. 政府戒急用忍的大陸政策
  1. 拜耳公司投資案
  1. 縣市長選舉的結果分析
  1. 飛碟協會斂財事件
  1. 進軍南科的廠商
  1. 南韓總統大選的結果
  1. 日本山一事件
  1. 林肯大郡災變後對建商之影響
  1. 黛妃猝逝
  1. 陳進興挾持南非武官事件
  1. 德蕾莎修女病逝
  1. 連副總統的破兵之旅
  1. 有線電視系統業者與頻道業者之爭
  1. 周休二日何處去
  1. 從林肯大郡事件談山坡地開發
  1. 東南亞金融風暴對台灣的衝擊
  1. 青少年犯罪問題面面觀
  1. 南非武官等三人槍傷送醫
  1. 宋楚瑜請辭風波對台灣政壇的影響
  1. 李總統過境檀島將獲最佳禮遇
  1. 從白案談媒體自律
  1. 亞太營運中心企業界高度期待
  1. 職棒投手的相關報導
  1. 南非與我國斷交
  1. 公安事件之相關報導
  1. 亞洲金融風暴對我國的影響
  1. 青少年犯罪事件
  1. 和信、東森爭奪有線市場換約問題引發斷訊風波
  1. 白曉燕案及陳進興挾持南非武官事件

Appendix 2: Precision and Recall Rates for the Test Databases

Tables below are the precision and recall rates for the two test databases. The meaning of each column is described as follows:

No : the query identification corresponding to those in Appendix 1.

Ca : No. of retrieved documents by n-gram matching

Cb : No. of relevant documents in the result set by n-gram matching

Da : No. of retrieved documents by TS with term matching

Db : No. of relevant documents in the result set by TS with term matching

Ea : No. of retrieved documents by TRF with term matching

Eb : No. of relevant documents in the result set by TRF with term matching

C-P : precision rate by n-gram matching

D-P : precision rate by TS with term matching

E-P : precision rate by TRF with term matching

C-R : recall rate by n-gram matching

D-R : recall rate by TS with term matching

E-R : recall rate by TRF with term matching

Precision and recall rates of the three retrieval approaches for bibliographic database:
No Ca Cb Da Db Ea Eb C-P D-P E-P C-R D-R E-R
1 50 25 32 28 50 47 0.50 0.88 0.94 0.53 0.60 1.00
2 50 3 50 14 50 17 0.06 0.28 0.34 0.18 0.82 1.00
3 50 8 5 2 50 19 0.16 0.40 0.38 0.42 0.11 1.00
4 50 19 21 16 50 6 0.38 0.76 0.12 1.00 0.84 0.32
5 50 39 49 49 49 49 0.78 1.00 1.00 0.80 1.00 1.00
6 50 9 9 5 45 8 0.18 0.56 0.18 1.00 0.56 0.89
7 50 19 32 18 23 12 0.38 0.56 0.52 1.00 0.95 0.63
8 50 6 19 5 19 5 0.12 0.26 0.26 1.00 0.83 0.83
9 50 15 50 12 50 11 0.30 0.24 0.22 1.00 0.80 0.73
10 14 6 5 5 7 5 0.43 1.00 0.71 1.00 0.83 0.83
11 50 22 29 29 17 17 0.44 1.00 1.00 0.76 1.00 0.59
12 50 26 37 37 7 7 0.52 1.00 1.00 0.70 1.00 0.19
13 50 16 50 48 50 48 0.32 0.96 0.96 0.33 1.00 1.00
14 50 24 11 11 13 13 0.48 1.00 1.00 1.00 0.46 0.54
15 50 15 20 20 8 8 0.30 1.00 1.00 0.75 1.00 0.40
16 50 14 9 9 12 12 0.28 1.00 1.00 1.00 0.64 0.86
17 50 31 37 37 23 23 0.62 1.00 1.00 0.84 1.00 0.62
18 43 21 9 9 9 9 0.49 1.00 1.00 1.00 0.43 0.43
19 50 13 50 30 50 30 0.26 0.60 0.60 0.43 1.00 1.00
20 50 21 21 20 10 10 0.42 0.95 1.00 1.00 0.95 0.48
21 50 34 50 40 28 28 0.68 0.80 1.00 0.85 1.00 0.70
22 50 45 39 39 50 47 0.90 1.00 0.94 0.96 0.83 1.00
23 50 48 50 50 50 33 0.96 1.00 0.66 0.96 1.00 0.66
24 50 7 50 13 50 5 0.14 0.26 0.10 0.54 1.00 0.38
25 50 28 50 39 50 18 0.56 0.78 0.36 0.72 1.00 0.46
26 50 7 50 13 50 9 0.14 0.26 0.18 0.54 1.00 0.69
27 50 12 50 11 50 17 0.24 0.22 0.34 0.71 0.65 1.00
28 50 35 27 27 27 27 0.70 1.00 1.00 1.00 0.77 0.77
29 50 40 50 35 50 46 0.80 0.70 0.92 0.87 0.76 1.00
30 28 9 14 4 13 13 0.32 0.29 1.00 0.69 0.31 1.00

Precision and recall rates of the three retrieval approaches for News database:
No Ca Cb Da Db Ea Eb C-P D-P E-P C-R D-R E-R
1 31 16 47 27 50 27 0.52 0.57 0.54 0.59 1.00 1.00
2 50 47 50 48 50 50 0.94 0.96 1.00 0.94 0.96 1.00
3 50 24 50 34 50 42 0.48 0.68 0.84 0.57 0.81 1.00
4 50 24 32 13 50 21 0.48 0.41 0.42 1.00 0.54 0.88
5 50 8 50 8 50 11 0.16 0.16 0.22 0.73 0.73 1.00
6 50 33 50 33 50 35 0.66 0.66 0.70 0.94 0.94 1.00
7 50 9 50 8 50 14 0.18 0.16 0.28 0.64 0.57 1.00
8 50 13 50 19 50 13 0.26 0.38 0.26 0.68 1.00 0.68
9 50 22 50 20 50 15 0.44 0.40 0.30 1.00 0.91 0.68
10 50 16 50 37 50 17 0.32 0.74 0.34 0.43 1.00 0.46
11 50 19 18 18 50 12 0.38 1.00 0.24 1.00 0.95 0.63
12 50 6 50 4 50 6 0.12 0.08 0.12 1.00 0.67 1.00
13 50 18 22 9 50 19 0.36 0.41 0.38 0.95 0.47 1.00
14 50 32 31 8 50 17 0.64 0.26 0.34 1.00 0.25 0.53
15 18 12 5 5 50 14 0.67 1.00 0.28 0.86 0.36 1.00
16 50 25 50 22 50 21 0.50 0.44 0.42 1.00 0.88 0.84
17 50 18 50 32 50 38 0.36 0.64 0.76 0.47 0.84 1.00
18 50 11 50 11 50 15 0.22 0.22 0.30 0.73 0.73 1.00
19 50 15 2 2 21 15 0.30 1.00 0.71 1.00 0.13 1.00
20 32 27 11 11 41 36 0.84 1.00 0.88 0.75 0.31 1.00
21 13 9 9 9 9 7 0.69 1.00 0.78 1.00 1.00 0.78
22 50 19 50 18 50 18 0.38 0.36 0.36 1.00 0.95 0.95
23 50 25 50 27 50 16 0.50 0.54 0.32 0.93 1.00 0.59
24 50 20 50 21 50 15 0.40 0.42 0.30 0.95 1.00 0.71
25 50 18 50 30 50 19 0.36 0.60 0.38 0.60 1.00 0.63
26 50 12 50 14 50 19 0.24 0.28 0.38 0.63 0.74 1.00
27 50 9 50 26 50 8 0.18 0.52 0.16 0.35 1.00 0.31
28 50 6 50 11 50 4 0.12 0.22 0.08 0.55 1.00 0.36
29 50 6 23 10 22 12 0.12 0.43 0.55 0.50 0.83 1.00
30 50 47 50 45 50 45 0.94 0.90 0.90 1.00 0.96 0.96