Search (40 results, page 1 of 2)

  • × theme_ss:"Semantisches Umfeld in Indexierung u. Retrieval"
  1. Rekabsaz, N. et al.: Toward optimized multimodal concept indexing (2016) 0.08
    0.0771009 = product of:
      0.1542018 = sum of:
        0.1542018 = sum of:
          0.09291325 = weight(_text_:n in 2751) [ClassicSimilarity], result of:
            0.09291325 = score(doc=2751,freq=2.0), product of:
              0.19504215 = queryWeight, product of:
                4.3116565 = idf(docFreq=1611, maxDocs=44218)
                0.045236014 = queryNorm
              0.47637522 = fieldWeight in 2751, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.3116565 = idf(docFreq=1611, maxDocs=44218)
                0.078125 = fieldNorm(doc=2751)
          0.061288558 = weight(_text_:22 in 2751) [ClassicSimilarity], result of:
            0.061288558 = score(doc=2751,freq=2.0), product of:
              0.15840882 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.045236014 = queryNorm
              0.38690117 = fieldWeight in 2751, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.078125 = fieldNorm(doc=2751)
      0.5 = coord(1/2)
    
    Date
    1. 2.2016 18:25:22
  2. Klas, C.-P.; Fuhr, N.; Schaefer, A.: Evaluating strategic support for information access in the DAFFODIL system (2004) 0.05
    0.046260543 = product of:
      0.09252109 = sum of:
        0.09252109 = sum of:
          0.05574795 = weight(_text_:n in 2419) [ClassicSimilarity], result of:
            0.05574795 = score(doc=2419,freq=2.0), product of:
              0.19504215 = queryWeight, product of:
                4.3116565 = idf(docFreq=1611, maxDocs=44218)
                0.045236014 = queryNorm
              0.28582513 = fieldWeight in 2419, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.3116565 = idf(docFreq=1611, maxDocs=44218)
                0.046875 = fieldNorm(doc=2419)
          0.036773134 = weight(_text_:22 in 2419) [ClassicSimilarity], result of:
            0.036773134 = score(doc=2419,freq=2.0), product of:
              0.15840882 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.045236014 = queryNorm
              0.23214069 = fieldWeight in 2419, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.046875 = fieldNorm(doc=2419)
      0.5 = coord(1/2)
    
    Date
    16.11.2008 16:22:48
  3. Robertson, A.M.; Willett, P.: Applications of n-grams in textual information systems (1998) 0.05
    0.04551801 = product of:
      0.09103602 = sum of:
        0.09103602 = product of:
          0.18207204 = sum of:
            0.18207204 = weight(_text_:n in 4715) [ClassicSimilarity], result of:
              0.18207204 = score(doc=4715,freq=12.0), product of:
                0.19504215 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.045236014 = queryNorm
                0.933501 = fieldWeight in 4715, product of:
                  3.4641016 = tf(freq=12.0), with freq of:
                    12.0 = termFreq=12.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.0625 = fieldNorm(doc=4715)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Provides an introduction to the use of n-grams in textual information systems, where an n-gram is a string of n, usually adjacent, characters, extracted from a section of continuous text. Applications that can be implemented efficiently and effectively using sets of n-grams include spelling errors detection and correction, query expansion, information retrieval with serial, inverted and signature files, dictionary look up, text compression, and language identification
    Object
    n-grams
  4. Brandão, W.C.; Santos, R.L.T.; Ziviani, N.; Moura, E.S. de; Silva, A.S. da: Learning to expand queries using entities (2014) 0.04
    0.03855045 = product of:
      0.0771009 = sum of:
        0.0771009 = sum of:
          0.046456624 = weight(_text_:n in 1343) [ClassicSimilarity], result of:
            0.046456624 = score(doc=1343,freq=2.0), product of:
              0.19504215 = queryWeight, product of:
                4.3116565 = idf(docFreq=1611, maxDocs=44218)
                0.045236014 = queryNorm
              0.23818761 = fieldWeight in 1343, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                4.3116565 = idf(docFreq=1611, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1343)
          0.030644279 = weight(_text_:22 in 1343) [ClassicSimilarity], result of:
            0.030644279 = score(doc=1343,freq=2.0), product of:
              0.15840882 = queryWeight, product of:
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.045236014 = queryNorm
              0.19345059 = fieldWeight in 1343, product of:
                1.4142135 = tf(freq=2.0), with freq of:
                  2.0 = termFreq=2.0
                3.5018296 = idf(docFreq=3622, maxDocs=44218)
                0.0390625 = fieldNorm(doc=1343)
      0.5 = coord(1/2)
    
    Date
    22. 8.2014 17:07:50
  5. Stojanovic, N.: On the query refinement in the ontology-based searching for information (2005) 0.02
    0.023228312 = product of:
      0.046456624 = sum of:
        0.046456624 = product of:
          0.09291325 = sum of:
            0.09291325 = weight(_text_:n in 2907) [ClassicSimilarity], result of:
              0.09291325 = score(doc=2907,freq=2.0), product of:
                0.19504215 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.045236014 = queryNorm
                0.47637522 = fieldWeight in 2907, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2907)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  6. Boyack, K.W.; Wylie,B.N.; Davidson, G.S.: Information Visualization, Human-Computer Interaction, and Cognitive Psychology : Domain Visualizations (2002) 0.02
    0.021668777 = product of:
      0.043337554 = sum of:
        0.043337554 = product of:
          0.08667511 = sum of:
            0.08667511 = weight(_text_:22 in 1352) [ClassicSimilarity], result of:
              0.08667511 = score(doc=1352,freq=4.0), product of:
                0.15840882 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045236014 = queryNorm
                0.54716086 = fieldWeight in 1352, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=1352)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 2.2003 17:25:39
    22. 2.2003 18:17:40
  7. Smeaton, A.F.; Rijsbergen, C.J. van: ¬The retrieval effects of query expansion on a feedback document retrieval system (1983) 0.02
    0.021450995 = product of:
      0.04290199 = sum of:
        0.04290199 = product of:
          0.08580398 = sum of:
            0.08580398 = weight(_text_:22 in 2134) [ClassicSimilarity], result of:
              0.08580398 = score(doc=2134,freq=2.0), product of:
                0.15840882 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045236014 = queryNorm
                0.5416616 = fieldWeight in 2134, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.109375 = fieldNorm(doc=2134)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    30. 3.2001 13:32:22
  8. Bettencourt, N.; Silva, N.; Barroso, J.: Semantically enhancing recommender systems (2016) 0.02
    0.019709876 = product of:
      0.03941975 = sum of:
        0.03941975 = product of:
          0.0788395 = sum of:
            0.0788395 = weight(_text_:n in 3374) [ClassicSimilarity], result of:
              0.0788395 = score(doc=3374,freq=4.0), product of:
                0.19504215 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.045236014 = queryNorm
                0.40421778 = fieldWeight in 3374, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3374)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  9. Tseng, Y.-H.: Solving vocabulary problems with interactive query expansion (1998) 0.02
    0.016424898 = product of:
      0.032849796 = sum of:
        0.032849796 = product of:
          0.06569959 = sum of:
            0.06569959 = weight(_text_:n in 5159) [ClassicSimilarity], result of:
              0.06569959 = score(doc=5159,freq=4.0), product of:
                0.19504215 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.045236014 = queryNorm
                0.33684817 = fieldWeight in 5159, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=5159)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    One of the major causes of search failures in information retrieval systems is vocabulary mismatch. Presents a solution to the vocabulary problem through 2 strategies known as term suggestion (TS) and term relevance feedback (TRF). In TS, collection specific terms are extracted from the text collection. These terms and their frequencies constitute the keyword database for suggesting terms in response to users' queries. One effect of this term suggestion is that it functions as a dynamic directory if the query is a general term that contains broad meaning. In term relevance feedback, terms extracted from the top ranked documents retrieved from the previous query are shown to users for relevance feedback. In the experiment, interactive TS provides very high precision rates while achieving similar recall rates as n-gram matching. Local TRF achieves improvement in both precision and recall rate in a full text news database and degrades slightly in recall rate in bibliographic databases due to the very limited source of information for feedback. In terms of Rijsbergen's combined measure of recall and precision, both TS and TRF achieve better performance than n-gram matching, which implies that the greater improvement in precision rate compensates the slight degradation in recall rate for TS and TRF
  10. Morato, J.; Llorens, J.; Genova, G.; Moreiro, J.A.: Experiments in discourse analysis impact on information classification and retrieval algorithms (2003) 0.02
    0.016424898 = product of:
      0.032849796 = sum of:
        0.032849796 = product of:
          0.06569959 = sum of:
            0.06569959 = weight(_text_:n in 1083) [ClassicSimilarity], result of:
              0.06569959 = score(doc=1083,freq=4.0), product of:
                0.19504215 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.045236014 = queryNorm
                0.33684817 = fieldWeight in 1083, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=1083)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Researchers in indexing and retrieval systems have been advocating the inclusion of more contextual information to improve results. The proliferation of full-text databases and advances in computer storage capacity have made it possible to carry out text analysis by means of linguistic and extra-linguistic knowledge. Since the mid 80s, research has tended to pay more attention to context, giving discourse analysis a more central role. The research presented in this paper aims to check whether discourse variables have an impact on modern information retrieval and classification algorithms. In order to evaluate this hypothesis, a functional framework for information analysis in an automated environment has been proposed, where the n-grams (filtering) and the k-means and Chen's classification algorithms have been tested against sub-collections of documents based on the following discourse variables: "Genre", "Register", "Domain terminology", and "Document structure". The results obtained with the algorithms for the different sub-collections were compared to the MeSH information structure. These demonstrate that n-grams does not appear to have a clear dependence on discourse variables, though the k-means classification algorithm does, but only on domain terminology and document structure, and finally Chen's algorithm has a clear dependence on all of the discourse variables. This information could be used to design better classification algorithms, where discourse variables should be taken into account. Other minor conclusions drawn from these results are also presented.
  11. Bayer, O.; Höhfeld, S.; Josbächer, F.; Kimm, N.; Kradepohl, I.; Kwiatkowski, M.; Puschmann, C.; Sabbagh, M.; Werner, N.; Vollmer, U.: Evaluation of an ontology-based knowledge-management-system : a case study of Convera RetrievalWare 8.0 (2005) 0.02
    0.016424898 = product of:
      0.032849796 = sum of:
        0.032849796 = product of:
          0.06569959 = sum of:
            0.06569959 = weight(_text_:n in 624) [ClassicSimilarity], result of:
              0.06569959 = score(doc=624,freq=4.0), product of:
                0.19504215 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.045236014 = queryNorm
                0.33684817 = fieldWeight in 624, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=624)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  12. Roy, R.S.; Agarwal, S.; Ganguly, N.; Choudhury, M.: Syntactic complexity of Web search queries through the lenses of language models, networks and users (2016) 0.02
    0.016424898 = product of:
      0.032849796 = sum of:
        0.032849796 = product of:
          0.06569959 = sum of:
            0.06569959 = weight(_text_:n in 3188) [ClassicSimilarity], result of:
              0.06569959 = score(doc=3188,freq=4.0), product of:
                0.19504215 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.045236014 = queryNorm
                0.33684817 = fieldWeight in 3188, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=3188)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    Across the world, millions of users interact with search engines every day to satisfy their information needs. As the Web grows bigger over time, such information needs, manifested through user search queries, also become more complex. However, there has been no systematic study that quantifies the structural complexity of Web search queries. In this research, we make an attempt towards understanding and characterizing the syntactic complexity of search queries using a multi-pronged approach. We use traditional statistical language modeling techniques to quantify and compare the perplexity of queries with natural language (NL). We then use complex network analysis for a comparative analysis of the topological properties of queries issued by real Web users and those generated by statistical models. Finally, we conduct experiments to study whether search engine users are able to identify real queries, when presented along with model-generated ones. The three complementary studies show that the syntactic structure of Web queries is more complex than what n-grams can capture, but simpler than NL. Queries, thus, seem to represent an intermediate stage between syntactic and non-syntactic communication.
  13. Kozikowski, P. et al.: Support of part-whole relations in query answering (2016) 0.02
    0.0153221395 = product of:
      0.030644279 = sum of:
        0.030644279 = product of:
          0.061288558 = sum of:
            0.061288558 = weight(_text_:22 in 2754) [ClassicSimilarity], result of:
              0.061288558 = score(doc=2754,freq=2.0), product of:
                0.15840882 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045236014 = queryNorm
                0.38690117 = fieldWeight in 2754, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=2754)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    1. 2.2016 18:25:22
  14. Marx, E. et al.: Exploring term networks for semantic search over RDF knowledge graphs (2016) 0.02
    0.0153221395 = product of:
      0.030644279 = sum of:
        0.030644279 = product of:
          0.061288558 = sum of:
            0.061288558 = weight(_text_:22 in 3279) [ClassicSimilarity], result of:
              0.061288558 = score(doc=3279,freq=2.0), product of:
                0.15840882 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045236014 = queryNorm
                0.38690117 = fieldWeight in 3279, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3279)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Metadata and semantics research: 10th International Conference, MTSR 2016, Göttingen, Germany, November 22-25, 2016, Proceedings. Eds.: E. Garoufallou
  15. Kopácsi, S. et al.: Development of a classification server to support metadata harmonization in a long term preservation system (2016) 0.02
    0.0153221395 = product of:
      0.030644279 = sum of:
        0.030644279 = product of:
          0.061288558 = sum of:
            0.061288558 = weight(_text_:22 in 3280) [ClassicSimilarity], result of:
              0.061288558 = score(doc=3280,freq=2.0), product of:
                0.15840882 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045236014 = queryNorm
                0.38690117 = fieldWeight in 3280, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.078125 = fieldNorm(doc=3280)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Source
    Metadata and semantics research: 10th International Conference, MTSR 2016, Göttingen, Germany, November 22-25, 2016, Proceedings. Eds.: E. Garoufallou
  16. Sacco, G.M.: Dynamic taxonomies and guided searches (2006) 0.02
    0.015168144 = product of:
      0.030336289 = sum of:
        0.030336289 = product of:
          0.060672577 = sum of:
            0.060672577 = weight(_text_:22 in 5295) [ClassicSimilarity], result of:
              0.060672577 = score(doc=5295,freq=4.0), product of:
                0.15840882 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045236014 = queryNorm
                0.38301262 = fieldWeight in 5295, product of:
                  2.0 = tf(freq=4.0), with freq of:
                    4.0 = termFreq=4.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0546875 = fieldNorm(doc=5295)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    22. 7.2006 17:56:22
  17. Li, N.; Sun, J.: Improving Chinese term association from the linguistic perspective (2017) 0.01
    0.013936987 = product of:
      0.027873974 = sum of:
        0.027873974 = product of:
          0.05574795 = sum of:
            0.05574795 = weight(_text_:n in 3381) [ClassicSimilarity], result of:
              0.05574795 = score(doc=3381,freq=2.0), product of:
                0.19504215 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.045236014 = queryNorm
                0.28582513 = fieldWeight in 3381, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.046875 = fieldNorm(doc=3381)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
  18. Goslin, K.; Hofmann, M.: ¬A Wikipedia powered state-based approach to automatic search query enhancement (2018) 0.01
    0.013936987 = product of:
      0.027873974 = sum of:
        0.027873974 = product of:
          0.05574795 = sum of:
            0.05574795 = weight(_text_:n in 5083) [ClassicSimilarity], result of:
              0.05574795 = score(doc=5083,freq=2.0), product of:
                0.19504215 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.045236014 = queryNorm
                0.28582513 = fieldWeight in 5083, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.046875 = fieldNorm(doc=5083)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Abstract
    This paper describes the development and testing of a novel Automatic Search Query Enhancement (ASQE) algorithm, the Wikipedia N Sub-state Algorithm (WNSSA), which utilises Wikipedia as the sole data source for prior knowledge. This algorithm is built upon the concept of iterative states and sub-states, harnessing the power of Wikipedia's data set and link information to identify and utilise reoccurring terms to aid term selection and weighting during enhancement. This algorithm is designed to prevent query drift by making callbacks to the user's original search intent by persisting the original query between internal states with additional selected enhancement terms. The developed algorithm has shown to improve both short and long queries by providing a better understanding of the query and available data. The proposed algorithm was compared against five existing ASQE algorithms that utilise Wikipedia as the sole data source, showing an average Mean Average Precision (MAP) improvement of 0.273 over the tested existing ASQE algorithms.
  19. Efthimiadis, E.N.: End-users' understanding of thesaural knowledge structures in interactive query expansion (1994) 0.01
    0.012257711 = product of:
      0.024515422 = sum of:
        0.024515422 = product of:
          0.049030844 = sum of:
            0.049030844 = weight(_text_:22 in 5693) [ClassicSimilarity], result of:
              0.049030844 = score(doc=5693,freq=2.0), product of:
                0.15840882 = queryWeight, product of:
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.045236014 = queryNorm
                0.30952093 = fieldWeight in 5693, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  3.5018296 = idf(docFreq=3622, maxDocs=44218)
                  0.0625 = fieldNorm(doc=5693)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    
    Date
    30. 3.2001 13:35:22
  20. Schaefer, A.; Jordan, M.; Klas, C.-P.; Fuhr, N.: Active support for query formulation in virtual digital libraries : a case study with DAFFODIL (2005) 0.01
    0.011614156 = product of:
      0.023228312 = sum of:
        0.023228312 = product of:
          0.046456624 = sum of:
            0.046456624 = weight(_text_:n in 4296) [ClassicSimilarity], result of:
              0.046456624 = score(doc=4296,freq=2.0), product of:
                0.19504215 = queryWeight, product of:
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.045236014 = queryNorm
                0.23818761 = fieldWeight in 4296, product of:
                  1.4142135 = tf(freq=2.0), with freq of:
                    2.0 = termFreq=2.0
                  4.3116565 = idf(docFreq=1611, maxDocs=44218)
                  0.0390625 = fieldNorm(doc=4296)
          0.5 = coord(1/2)
      0.5 = coord(1/2)
    

Years

Languages

  • e 33
  • d 6
  • chi 1
  • More… Less…

Types