-
Yi, K.: Automatic text classification using library classification schemes : trends, issues and challenges (2007)
0.08
0.07699507 = product of:
0.15399013 = sum of:
0.15399013 = sum of:
0.11154386 = weight(_text_:2007 in 2560) [ClassicSimilarity], result of:
0.11154386 = score(doc=2560,freq=5.0), product of:
0.20205033 = queryWeight, product of:
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.044755515 = queryNorm
0.55205977 = fieldWeight in 2560, product of:
2.236068 = tf(freq=5.0), with freq of:
5.0 = termFreq=5.0
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.0546875 = fieldNorm(doc=2560)
0.042446278 = weight(_text_:22 in 2560) [ClassicSimilarity], result of:
0.042446278 = score(doc=2560,freq=2.0), product of:
0.15672618 = queryWeight, product of:
3.5018296 = idf(docFreq=3622, maxDocs=44218)
0.044755515 = queryNorm
0.2708308 = fieldWeight in 2560, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.5018296 = idf(docFreq=3622, maxDocs=44218)
0.0546875 = fieldNorm(doc=2560)
0.5 = coord(1/2)
- Date
- 22. 9.2008 18:31:54
- Source
- International cataloguing and bibliographic control. 36(2007) no.4, S.78-82
- Year
- 2007
-
Hotho, A.; Bloehdorn, S.: Data Mining 2004 : Text classification by boosting weak learners based on terms and concepts (2004)
0.07
0.071503974 = sum of:
0.053312715 = product of:
0.21325086 = sum of:
0.21325086 = weight(_text_:3a in 562) [ClassicSimilarity], result of:
0.21325086 = score(doc=562,freq=2.0), product of:
0.37943774 = queryWeight, product of:
8.478011 = idf(docFreq=24, maxDocs=44218)
0.044755515 = queryNorm
0.56201804 = fieldWeight in 562, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
8.478011 = idf(docFreq=24, maxDocs=44218)
0.046875 = fieldNorm(doc=562)
0.25 = coord(1/4)
0.018191261 = product of:
0.036382522 = sum of:
0.036382522 = weight(_text_:22 in 562) [ClassicSimilarity], result of:
0.036382522 = score(doc=562,freq=2.0), product of:
0.15672618 = queryWeight, product of:
3.5018296 = idf(docFreq=3622, maxDocs=44218)
0.044755515 = queryNorm
0.23214069 = fieldWeight in 562, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.5018296 = idf(docFreq=3622, maxDocs=44218)
0.046875 = fieldNorm(doc=562)
0.5 = coord(1/2)
- Content
- Vgl.: http://www.google.de/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&ved=0CEAQFjAA&url=http%3A%2F%2Fciteseerx.ist.psu.edu%2Fviewdoc%2Fdownload%3Fdoi%3D10.1.1.91.4940%26rep%3Drep1%26type%3Dpdf&ei=dOXrUMeIDYHDtQahsIGACg&usg=AFQjCNHFWVh6gNPvnOrOS9R3rkrXCNVD-A&sig2=5I2F5evRfMnsttSgFF9g7Q&bvm=bv.1357316858,d.Yms.
- Date
- 8. 1.2013 10:22:32
-
Pfeffer, M.: Automatische Vergabe von RVK-Notationen mittels fallbasiertem Schließen (2009)
0.06
0.060948916 = product of:
0.12189783 = sum of:
0.12189783 = sum of:
0.085515305 = weight(_text_:2007 in 3051) [ClassicSimilarity], result of:
0.085515305 = score(doc=3051,freq=4.0), product of:
0.20205033 = queryWeight, product of:
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.044755515 = queryNorm
0.42323765 = fieldWeight in 3051, product of:
2.0 = tf(freq=4.0), with freq of:
4.0 = termFreq=4.0
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.046875 = fieldNorm(doc=3051)
0.036382522 = weight(_text_:22 in 3051) [ClassicSimilarity], result of:
0.036382522 = score(doc=3051,freq=2.0), product of:
0.15672618 = queryWeight, product of:
3.5018296 = idf(docFreq=3622, maxDocs=44218)
0.044755515 = queryNorm
0.23214069 = fieldWeight in 3051, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.5018296 = idf(docFreq=3622, maxDocs=44218)
0.046875 = fieldNorm(doc=3051)
0.5 = coord(1/2)
- Date
- 22. 8.2009 19:51:28
- Footnote
- Vgl. auch die Präsentationen unter: http://www.bibliothek.uni-regensburg.de/Systematik/pdf/Anw2008_PPT1.pdf. http://blog.bib.uni-mannheim.de/Classification/wp-content/uploads/2007/10/hu-berlin-2007-2.pdf. Volltexte unter:
-
Reiner, U.: Automatische DDC-Klassifizierung bibliografischer Titeldatensätze der Deutschen Nationalbibliografie (2009)
0.04
0.04063261 = product of:
0.08126522 = sum of:
0.08126522 = sum of:
0.057010204 = weight(_text_:2007 in 3284) [ClassicSimilarity], result of:
0.057010204 = score(doc=3284,freq=4.0), product of:
0.20205033 = queryWeight, product of:
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.044755515 = queryNorm
0.28215843 = fieldWeight in 3284, product of:
2.0 = tf(freq=4.0), with freq of:
4.0 = termFreq=4.0
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.03125 = fieldNorm(doc=3284)
0.024255017 = weight(_text_:22 in 3284) [ClassicSimilarity], result of:
0.024255017 = score(doc=3284,freq=2.0), product of:
0.15672618 = queryWeight, product of:
3.5018296 = idf(docFreq=3622, maxDocs=44218)
0.044755515 = queryNorm
0.15476047 = fieldWeight in 3284, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.5018296 = idf(docFreq=3622, maxDocs=44218)
0.03125 = fieldNorm(doc=3284)
0.5 = coord(1/2)
- Abstract
- Das Klassifizieren von Objekten (z. B. Fauna, Flora, Texte) ist ein Verfahren, das auf menschlicher Intelligenz basiert. In der Informatik - insbesondere im Gebiet der Künstlichen Intelligenz (KI) - wird u. a. untersucht, inweit Verfahren, die menschliche Intelligenz benötigen, automatisiert werden können. Hierbei hat sich herausgestellt, dass die Lösung von Alltagsproblemen eine größere Herausforderung darstellt, als die Lösung von Spezialproblemen, wie z. B. das Erstellen eines Schachcomputers. So ist "Rybka" der seit Juni 2007 amtierende Computerschach-Weltmeistern. Inwieweit Alltagsprobleme mit Methoden der Künstlichen Intelligenz gelöst werden können, ist eine - für den allgemeinen Fall - noch offene Frage. Beim Lösen von Alltagsproblemen spielt die Verarbeitung der natürlichen Sprache, wie z. B. das Verstehen, eine wesentliche Rolle. Den "gesunden Menschenverstand" als Maschine (in der Cyc-Wissensbasis in Form von Fakten und Regeln) zu realisieren, ist Lenat's Ziel seit 1984. Bezüglich des KI-Paradeprojektes "Cyc" gibt es CycOptimisten und Cyc-Pessimisten. Das Verstehen der natürlichen Sprache (z. B. Werktitel, Zusammenfassung, Vorwort, Inhalt) ist auch beim intellektuellen Klassifizieren von bibliografischen Titeldatensätzen oder Netzpublikationen notwendig, um diese Textobjekte korrekt klassifizieren zu können. Seit dem Jahr 2007 werden von der Deutschen Nationalbibliothek nahezu alle Veröffentlichungen mit der Dewey Dezimalklassifikation (DDC) intellektuell klassifiziert.
- Date
- 22. 1.2010 14:41:24
-
Reiner, U.: Automatic analysis of DDC notations (2007)
0.04
0.03702921 = product of:
0.07405842 = sum of:
0.07405842 = product of:
0.14811684 = sum of:
0.14811684 = weight(_text_:2007 in 118) [ClassicSimilarity], result of:
0.14811684 = score(doc=118,freq=3.0), product of:
0.20205033 = queryWeight, product of:
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.044755515 = queryNorm
0.73306906 = fieldWeight in 118, product of:
1.7320508 = tf(freq=3.0), with freq of:
3.0 = termFreq=3.0
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.09375 = fieldNorm(doc=118)
0.5 = coord(1/2)
0.5 = coord(1/2)
- Year
- 2007
-
Hung, C.-M.; Chien, L.-F.: Web-based text classification in the absence of manually labeled training documents (2007)
0.02
0.023902256 = product of:
0.047804512 = sum of:
0.047804512 = product of:
0.095609024 = sum of:
0.095609024 = weight(_text_:2007 in 87) [ClassicSimilarity], result of:
0.095609024 = score(doc=87,freq=5.0), product of:
0.20205033 = queryWeight, product of:
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.044755515 = queryNorm
0.47319412 = fieldWeight in 87, product of:
2.236068 = tf(freq=5.0), with freq of:
5.0 = termFreq=5.0
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.046875 = fieldNorm(doc=87)
0.5 = coord(1/2)
0.5 = coord(1/2)
- Source
- Journal of the American Society for Information Science and Technology. 58(2007) no.1, S.88-96
- Year
- 2007
-
Liu, R.-L.: Dynamic category profiling for text filtering and classification (2007)
0.02
0.023902256 = product of:
0.047804512 = sum of:
0.047804512 = product of:
0.095609024 = sum of:
0.095609024 = weight(_text_:2007 in 900) [ClassicSimilarity], result of:
0.095609024 = score(doc=900,freq=5.0), product of:
0.20205033 = queryWeight, product of:
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.044755515 = queryNorm
0.47319412 = fieldWeight in 900, product of:
2.236068 = tf(freq=5.0), with freq of:
5.0 = termFreq=5.0
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.046875 = fieldNorm(doc=900)
0.5 = coord(1/2)
0.5 = coord(1/2)
- Source
- Information processing and management. 43(2007) no.1, S.154-168
- Year
- 2007
-
Yoon, Y.; Lee, G.G.: Efficient implementation of associative classifiers for document classification (2007)
0.02
0.023902256 = product of:
0.047804512 = sum of:
0.047804512 = product of:
0.095609024 = sum of:
0.095609024 = weight(_text_:2007 in 909) [ClassicSimilarity], result of:
0.095609024 = score(doc=909,freq=5.0), product of:
0.20205033 = queryWeight, product of:
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.044755515 = queryNorm
0.47319412 = fieldWeight in 909, product of:
2.236068 = tf(freq=5.0), with freq of:
5.0 = termFreq=5.0
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.046875 = fieldNorm(doc=909)
0.5 = coord(1/2)
0.5 = coord(1/2)
- Source
- Information processing and management. 43(2007) no.2, S.393-405
- Year
- 2007
-
Hagedorn, K.; Chapman, S.; Newman, D.: Enhancing search and browse using automated clustering of subject metadata (2007)
0.02
0.023902256 = product of:
0.047804512 = sum of:
0.047804512 = product of:
0.095609024 = sum of:
0.095609024 = weight(_text_:2007 in 1168) [ClassicSimilarity], result of:
0.095609024 = score(doc=1168,freq=5.0), product of:
0.20205033 = queryWeight, product of:
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.044755515 = queryNorm
0.47319412 = fieldWeight in 1168, product of:
2.236068 = tf(freq=5.0), with freq of:
5.0 = termFreq=5.0
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.046875 = fieldNorm(doc=1168)
0.5 = coord(1/2)
0.5 = coord(1/2)
- Source
- D-Lib magazine. 13(2007) nos.7/8, x S
- Year
- 2007
-
Golub, K.; Hamon, T.; Ardö, A.: Automated classification of textual documents based on a controlled vocabulary in engineering (2007)
0.02
0.023902256 = product of:
0.047804512 = sum of:
0.047804512 = product of:
0.095609024 = sum of:
0.095609024 = weight(_text_:2007 in 1461) [ClassicSimilarity], result of:
0.095609024 = score(doc=1461,freq=5.0), product of:
0.20205033 = queryWeight, product of:
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.044755515 = queryNorm
0.47319412 = fieldWeight in 1461, product of:
2.236068 = tf(freq=5.0), with freq of:
5.0 = termFreq=5.0
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.046875 = fieldNorm(doc=1461)
0.5 = coord(1/2)
0.5 = coord(1/2)
- Source
- Knowledge organization. 34(2007) no.4, S.247-263
- Year
- 2007
-
Li, T.; Zhu, S.; Ogihara, M.: Hierarchical document classification using automatically generated hierarchy (2007)
0.02
0.023902256 = product of:
0.047804512 = sum of:
0.047804512 = product of:
0.095609024 = sum of:
0.095609024 = weight(_text_:2007 in 4797) [ClassicSimilarity], result of:
0.095609024 = score(doc=4797,freq=5.0), product of:
0.20205033 = queryWeight, product of:
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.044755515 = queryNorm
0.47319412 = fieldWeight in 4797, product of:
2.236068 = tf(freq=5.0), with freq of:
5.0 = termFreq=5.0
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.046875 = fieldNorm(doc=4797)
0.5 = coord(1/2)
0.5 = coord(1/2)
- Source
- Journal of intelligent information systems. 29(2007) no.2, S.211-230
- Year
- 2007
-
Cathey, R.J.; Jensen, E.C.; Beitzel, S.M.; Frieder, O.; Grossman, D.: Exploiting parallelism to support scalable hierarchical clustering (2007)
0.02
0.019918546 = product of:
0.039837092 = sum of:
0.039837092 = product of:
0.079674184 = sum of:
0.079674184 = weight(_text_:2007 in 448) [ClassicSimilarity], result of:
0.079674184 = score(doc=448,freq=5.0), product of:
0.20205033 = queryWeight, product of:
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.044755515 = queryNorm
0.39432842 = fieldWeight in 448, product of:
2.236068 = tf(freq=5.0), with freq of:
5.0 = termFreq=5.0
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.0390625 = fieldNorm(doc=448)
0.5 = coord(1/2)
0.5 = coord(1/2)
- Source
- Journal of the American Society for Information Science and Technology. 58(2007) no.8, S.1207-1221
- Year
- 2007
-
Peng, F.; Huang, X.: Machine learning for Asian language text classification (2007)
0.02
0.019918546 = product of:
0.039837092 = sum of:
0.039837092 = product of:
0.079674184 = sum of:
0.079674184 = weight(_text_:2007 in 831) [ClassicSimilarity], result of:
0.079674184 = score(doc=831,freq=5.0), product of:
0.20205033 = queryWeight, product of:
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.044755515 = queryNorm
0.39432842 = fieldWeight in 831, product of:
2.236068 = tf(freq=5.0), with freq of:
5.0 = termFreq=5.0
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.0390625 = fieldNorm(doc=831)
0.5 = coord(1/2)
0.5 = coord(1/2)
- Source
- Journal of documentation. 63(2007) no.3, S.378-397
- Year
- 2007
-
Pfeffer, M.: Automatische Vergabe von RVK-Notationen anhand von bibliografischen Daten mittels fallbasiertem Schließen (2007)
0.02
0.018514605 = product of:
0.03702921 = sum of:
0.03702921 = product of:
0.07405842 = sum of:
0.07405842 = weight(_text_:2007 in 558) [ClassicSimilarity], result of:
0.07405842 = score(doc=558,freq=3.0), product of:
0.20205033 = queryWeight, product of:
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.044755515 = queryNorm
0.36653453 = fieldWeight in 558, product of:
1.7320508 = tf(freq=3.0), with freq of:
3.0 = termFreq=3.0
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.046875 = fieldNorm(doc=558)
0.5 = coord(1/2)
0.5 = coord(1/2)
- Year
- 2007
-
Helmbrecht-Schaar, A.: Entwicklung eines Verfahrens der automatischen Klassifizierung für Textdokumente aus dem Fachbereich Informatik mithilfe eines fachspezifischen Klassifikationssystems (2007)
0.02
0.018514605 = product of:
0.03702921 = sum of:
0.03702921 = product of:
0.07405842 = sum of:
0.07405842 = weight(_text_:2007 in 1410) [ClassicSimilarity], result of:
0.07405842 = score(doc=1410,freq=3.0), product of:
0.20205033 = queryWeight, product of:
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.044755515 = queryNorm
0.36653453 = fieldWeight in 1410, product of:
1.7320508 = tf(freq=3.0), with freq of:
3.0 = termFreq=3.0
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.046875 = fieldNorm(doc=1410)
0.5 = coord(1/2)
0.5 = coord(1/2)
- Year
- 2007
-
Puzicha, J.: Informationen finden! : Intelligente Suchmaschinentechnologie & automatische Kategorisierung (2007)
0.02
0.018514605 = product of:
0.03702921 = sum of:
0.03702921 = product of:
0.07405842 = sum of:
0.07405842 = weight(_text_:2007 in 2817) [ClassicSimilarity], result of:
0.07405842 = score(doc=2817,freq=3.0), product of:
0.20205033 = queryWeight, product of:
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.044755515 = queryNorm
0.36653453 = fieldWeight in 2817, product of:
1.7320508 = tf(freq=3.0), with freq of:
3.0 = termFreq=3.0
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.046875 = fieldNorm(doc=2817)
0.5 = coord(1/2)
0.5 = coord(1/2)
- Year
- 2007
-
Subramanian, S.; Shafer, K.E.: Clustering (2001)
0.02
0.018191261 = product of:
0.036382522 = sum of:
0.036382522 = product of:
0.072765045 = sum of:
0.072765045 = weight(_text_:22 in 1046) [ClassicSimilarity], result of:
0.072765045 = score(doc=1046,freq=2.0), product of:
0.15672618 = queryWeight, product of:
3.5018296 = idf(docFreq=3622, maxDocs=44218)
0.044755515 = queryNorm
0.46428138 = fieldWeight in 1046, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.5018296 = idf(docFreq=3622, maxDocs=44218)
0.09375 = fieldNorm(doc=1046)
0.5 = coord(1/2)
0.5 = coord(1/2)
- Date
- 5. 5.2003 14:17:22
-
Reiner, U.: Automatische DDC-Klassifizierung von bibliografischen Titeldatensätzen (2009)
0.02
0.015159386 = product of:
0.030318772 = sum of:
0.030318772 = product of:
0.060637545 = sum of:
0.060637545 = weight(_text_:22 in 611) [ClassicSimilarity], result of:
0.060637545 = score(doc=611,freq=2.0), product of:
0.15672618 = queryWeight, product of:
3.5018296 = idf(docFreq=3622, maxDocs=44218)
0.044755515 = queryNorm
0.38690117 = fieldWeight in 611, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.5018296 = idf(docFreq=3622, maxDocs=44218)
0.078125 = fieldNorm(doc=611)
0.5 = coord(1/2)
0.5 = coord(1/2)
- Date
- 22. 8.2009 12:54:24
-
HaCohen-Kerner, Y. et al.: Classification using various machine learning methods and combinations of key-phrases and visual features (2016)
0.02
0.015159386 = product of:
0.030318772 = sum of:
0.030318772 = product of:
0.060637545 = sum of:
0.060637545 = weight(_text_:22 in 2748) [ClassicSimilarity], result of:
0.060637545 = score(doc=2748,freq=2.0), product of:
0.15672618 = queryWeight, product of:
3.5018296 = idf(docFreq=3622, maxDocs=44218)
0.044755515 = queryNorm
0.38690117 = fieldWeight in 2748, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
3.5018296 = idf(docFreq=3622, maxDocs=44218)
0.078125 = fieldNorm(doc=2748)
0.5 = coord(1/2)
0.5 = coord(1/2)
- Date
- 1. 2.2016 18:25:22
-
Kasprzik, A.: Automatisierte und semiautomatisierte Klassifizierung : eine Analyse aktueller Projekte (2014)
0.02
0.015117113 = product of:
0.030234225 = sum of:
0.030234225 = product of:
0.06046845 = sum of:
0.06046845 = weight(_text_:2007 in 2470) [ClassicSimilarity], result of:
0.06046845 = score(doc=2470,freq=2.0), product of:
0.20205033 = queryWeight, product of:
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.044755515 = queryNorm
0.2992742 = fieldWeight in 2470, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
4.514535 = idf(docFreq=1315, maxDocs=44218)
0.046875 = fieldNorm(doc=2470)
0.5 = coord(1/2)
0.5 = coord(1/2)
- Abstract
- Das sprunghafte Anwachsen der Menge digital verfügbarer Dokumente gepaart mit dem Zeit- und Personalmangel an wissenschaftlichen Bibliotheken legt den Einsatz von halb- oder vollautomatischen Verfahren für die verbale und klassifikatorische Inhaltserschließung nahe. Nach einer kurzen allgemeinen Einführung in die gängige Methodik beleuchtet dieser Artikel eine Reihe von Projekten zur automatisierten Klassifizierung aus dem Zeitraum 2007-2012 und aus dem deutschsprachigen Raum. Ein Großteil der vorgestellten Projekte verwendet Methoden des Maschinellen Lernens aus der Künstlichen Intelligenz, arbeitet meist mit angepassten Versionen einer kommerziellen Software und bezieht sich in der Regel auf die Dewey Decimal Classification (DDC). Als Datengrundlage dienen Metadatensätze, Abstracs, Inhaltsverzeichnisse und Volltexte in diversen Datenformaten. Die abschließende Analyse enthält eine Anordnung der Projekte nach einer Reihe von verschiedenen Kriterien und eine Zusammenfassung der aktuellen Lage und der größten Herausfordungen für automatisierte Klassifizierungsverfahren.