-
Vlachidis, A.; Binding, C.; Tudhope, D.; May, K.: Excavating grey literature : a case study on the rich indexing of archaeological documents via natural language-processing techniques and knowledge-based resources (2010)
0.04
0.03754375 = product of:
0.150175 = sum of:
0.150175 = weight(_text_:java in 935) [ClassicSimilarity], result of:
0.150175 = score(doc=935,freq=2.0), product of:
0.48216656 = queryWeight, product of:
7.0475073 = idf(docFreq=104, maxDocs=44421)
0.06841661 = queryNorm
0.31145877 = fieldWeight in 935, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
7.0475073 = idf(docFreq=104, maxDocs=44421)
0.03125 = fieldNorm(doc=935)
0.25 = coord(1/4)
- Abstract
- Purpose - This paper sets out to discuss the use of information extraction (IE), a natural language-processing (NLP) technique to assist "rich" semantic indexing of diverse archaeological text resources. The focus of the research is to direct a semantic-aware "rich" indexing of diverse natural language resources with properties capable of satisfying information retrieval from online publications and datasets associated with the Semantic Technologies for Archaeological Resources (STAR) project. Design/methodology/approach - The paper proposes use of the English Heritage extension (CRM-EH) of the standard core ontology in cultural heritage, CIDOC CRM, and exploitation of domain thesauri resources for driving and enhancing an Ontology-Oriented Information Extraction process. The process of semantic indexing is based on a rule-based Information Extraction technique, which is facilitated by the General Architecture of Text Engineering (GATE) toolkit and expressed by Java Annotation Pattern Engine (JAPE) rules. Findings - Initial results suggest that the combination of information extraction with knowledge resources and standard conceptual models is capable of supporting semantic-aware term indexing. Additional efforts are required for further exploitation of the technique and adoption of formal evaluation methods for assessing the performance of the method in measurable terms. Originality/value - The value of the paper lies in the semantic indexing of 535 unpublished online documents often referred to as "Grey Literature", from the Archaeological Data Service OASIS corpus (Online AccesS to the Index of archaeological investigationS), with respect to the CRM ontological concepts E49.Time Appellation and P19.Physical Object.
-
Kreymer, O.: ¬An evaluation of help mechanisms in natural language information retrieval systems (2002)
0.04
0.035480015 = product of:
0.14192006 = sum of:
0.14192006 = weight(_text_:help in 3557) [ClassicSimilarity], result of:
0.14192006 = score(doc=3557,freq=4.0), product of:
0.32182297 = queryWeight, product of:
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.06841661 = queryNorm
0.44098797 = fieldWeight in 3557, product of:
2.0 = tf(freq=4.0), with freq of:
4.0 = termFreq=4.0
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.046875 = fieldNorm(doc=3557)
0.25 = coord(1/4)
- Abstract
- The field of natural language processing (NLP) demonstrates rapid changes in the design of information retrieval systems and human-computer interaction. While natural language is being looked on as the most effective tool for information retrieval in a contemporary information environment, the systems using it are only beginning to emerge. This study attempts to evaluate the current state of NLP information retrieval systems from the user's point of view: what techniques are used by these systems to guide their users through the search process? The analysis focused on the structure and components of the systems' help mechanisms. Results of the study demonstrated that systems which claimed to be using natural language searching in fact used a wide range of information retrieval techniques from real natural language processing to Boolean searching. As a result, the user assistance mechanisms of these systems also varied. While pseudo-NLP systems would suit a more traditional method of instruction, real NLP systems primarily utilised the methods of explanation and user-system dialogue.
-
Rosemblat, G.; Resnick, M.P.; Auston, I.; Shin, D.; Sneiderman, C.; Fizsman, M.; Rindflesch, T.C.: Extending SemRep to the public health domain (2013)
0.04
0.035480015 = product of:
0.14192006 = sum of:
0.14192006 = weight(_text_:help in 3096) [ClassicSimilarity], result of:
0.14192006 = score(doc=3096,freq=4.0), product of:
0.32182297 = queryWeight, product of:
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.06841661 = queryNorm
0.44098797 = fieldWeight in 3096, product of:
2.0 = tf(freq=4.0), with freq of:
4.0 = termFreq=4.0
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.046875 = fieldNorm(doc=3096)
0.25 = coord(1/4)
- Abstract
- We describe the use of a domain-independent method to extend a natural language processing (NLP) application, SemRep (Rindflesch, Fiszman, & Libbus, 2005), based on the knowledge sources afforded by the Unified Medical Language System (UMLS®; Humphreys, Lindberg, Schoolman, & Barnett, 1998) to support the area of health promotion within the public health domain. Public health professionals require good information about successful health promotion policies and programs that might be considered for application within their own communities. Our effort seeks to improve access to relevant information for the public health profession, to help those in the field remain an information-savvy workforce. Natural language processing and semantic techniques hold promise to help public health professionals navigate the growing ocean of information by organizing and structuring this knowledge into a focused public health framework paired with a user-friendly visualization application as a way to summarize results of PubMed® searches in this field of knowledge.
-
Ruge, G.; Schwarz, C.: Linguistically based term associations : a new semantic component for a hyperterm system (1990)
0.03
0.03345088 = product of:
0.13380352 = sum of:
0.13380352 = weight(_text_:help in 5543) [ClassicSimilarity], result of:
0.13380352 = score(doc=5543,freq=2.0), product of:
0.32182297 = queryWeight, product of:
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.06841661 = queryNorm
0.41576743 = fieldWeight in 5543, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.0625 = fieldNorm(doc=5543)
0.25 = coord(1/4)
- Abstract
- REALIST (Retrieval Aids by Linguistics and Statistics) is a tool which supplies the user of free text information retrieval systems with information about the terms in the databases. The resulting tables of terms show term relations according to their meaning in the database and form a kind of 'road map' of the database to give the user orientation help
-
Jacquemin, C.: What is the tree that we see through the window : a linguistic approach to windowing and term variation (1996)
0.03
0.029269518 = product of:
0.11707807 = sum of:
0.11707807 = weight(_text_:help in 5646) [ClassicSimilarity], result of:
0.11707807 = score(doc=5646,freq=2.0), product of:
0.32182297 = queryWeight, product of:
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.06841661 = queryNorm
0.3637965 = fieldWeight in 5646, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.0546875 = fieldNorm(doc=5646)
0.25 = coord(1/4)
- Abstract
- Provides a linguistic approach to text windowing through an extraction of term variants with the help of a partial parser. The syntactic grounding of the method ensures ehat words observed within restricted spans are lexically related and that spurious word cooccurrences are rules out with a good level of confidence. The system is computationally tractable on large corpora and large lists of terms. Gives illustrative examples of term variation from a large medical corpus. An experimental evaluation of the method shows that only a small proportion of co-occuring words are lexically related and motivates the call for natural language parsing techniques in text windowing
-
Chen, K.-H.: Evaluating Chinese text retrieval with multilingual queries (2002)
0.03
0.029269518 = product of:
0.11707807 = sum of:
0.11707807 = weight(_text_:help in 2851) [ClassicSimilarity], result of:
0.11707807 = score(doc=2851,freq=2.0), product of:
0.32182297 = queryWeight, product of:
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.06841661 = queryNorm
0.3637965 = fieldWeight in 2851, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.0546875 = fieldNorm(doc=2851)
0.25 = coord(1/4)
- Abstract
- This paper reports the design of a Chinese test collection with multilingual queries and the application of this test collection to evaluate information retrieval Systems. The effective indexing units, IR models, translation techniques, and query expansion for Chinese text retrieval are identified. The collaboration of East Asian countries for construction of test collections for cross-language multilingual text retrieval is also discussed in this paper. As well, a tool is designed to help assessors judge relevante and gather the events of relevante judgment. The log file created by this tool will be used to analyze the behaviors of assessors in the future.
-
Semantik, Lexikographie und Computeranwendungen : Workshop ... (Bonn) : 1995.01.27-28 (1996)
0.03
0.028651804 = product of:
0.114607215 = sum of:
0.114607215 = weight(_text_:und in 1190) [ClassicSimilarity], result of:
0.114607215 = score(doc=1190,freq=76.0), product of:
0.15174113 = queryWeight, product of:
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.06841661 = queryNorm
0.75528115 = fieldWeight in 1190, product of:
8.717798 = tf(freq=76.0), with freq of:
76.0 = termFreq=76.0
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.0390625 = fieldNorm(doc=1190)
0.25 = coord(1/4)
- BK
- 18.00 Einzelne Sprachen und Literaturen allgemein
- Classification
- ES 940 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Spezialbereiche der allgemeinen Sprachwissenschaft / Datenverarbeitung und Sprachwissenschaft. Computerlinguistik / Maschinelle Sprachanalyse
ET 400 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Einzelgebiete der Sprachwissenschaft, Sprachbeschreibung / Semantik und Lexikologie / Allgemeines
ES 945 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Spezialbereiche der allgemeinen Sprachwissenschaft / Datenverarbeitung und Sprachwissenschaft. Computerlinguistik / Spracherkennung
ET 580 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Einzelgebiete der Sprachwissenschaft, Sprachbeschreibung / Semantik und Lexikologie / Lexikologie (diachrone und synchrone) / Lexikographie
18.00 Einzelne Sprachen und Literaturen allgemein
- RVK
- ES 940 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Spezialbereiche der allgemeinen Sprachwissenschaft / Datenverarbeitung und Sprachwissenschaft. Computerlinguistik / Maschinelle Sprachanalyse
ET 400 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Einzelgebiete der Sprachwissenschaft, Sprachbeschreibung / Semantik und Lexikologie / Allgemeines
ES 945 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Spezialbereiche der allgemeinen Sprachwissenschaft / Datenverarbeitung und Sprachwissenschaft. Computerlinguistik / Spracherkennung
ET 580 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Einzelgebiete der Sprachwissenschaft, Sprachbeschreibung / Semantik und Lexikologie / Lexikologie (diachrone und synchrone) / Lexikographie
- Series
- Sprache und Information ; 33
-
Schank, R.C.: Computer, elementare Aktionen und linguistische Theorien (1977)
0.03
0.02602844 = product of:
0.10411376 = sum of:
0.10411376 = weight(_text_:und in 6141) [ClassicSimilarity], result of:
0.10411376 = score(doc=6141,freq=8.0), product of:
0.15174113 = queryWeight, product of:
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.06841661 = queryNorm
0.6861275 = fieldWeight in 6141, product of:
2.828427 = tf(freq=8.0), with freq of:
8.0 = termFreq=8.0
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.109375 = fieldNorm(doc=6141)
0.25 = coord(1/4)
- Series
- Grundlagen der Kommunikation und Kognition
- Source
- Semantik und künstliche Intelligenz: Beiträge zur automatischen Sprachbearbeitung II. Hrsg. und eingeleitet von P. Eisenberg
-
Weber, N.: ¬Die Semantik von Bedeutungsexplikationen (1999)
0.03
0.025559423 = product of:
0.102237694 = sum of:
0.102237694 = weight(_text_:und in 2399) [ClassicSimilarity], result of:
0.102237694 = score(doc=2399,freq=42.0), product of:
0.15174113 = queryWeight, product of:
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.06841661 = queryNorm
0.6737639 = fieldWeight in 2399, product of:
6.4807405 = tf(freq=42.0), with freq of:
42.0 = termFreq=42.0
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.046875 = fieldNorm(doc=2399)
0.25 = coord(1/4)
- BK
- 18.00 Einzelne Sprachen und Literaturen allgemein
- Classification
- ES 950 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Spezialbereiche der allgemeinen Sprachwissenschaft / Datenverarbeitung und Sprachwissenschaft. Computerlinguistik / Maschinelle Redeanlyse
ES 965 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Spezialbereiche der allgemeinen Sprachwissenschaft / Datenverarbeitung und Sprachwissenschaft. Computerlinguistik / Datenverarbeitung und Einzelgebiete der Grammatik
18.00 Einzelne Sprachen und Literaturen allgemein
- RVK
- ES 950 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Spezialbereiche der allgemeinen Sprachwissenschaft / Datenverarbeitung und Sprachwissenschaft. Computerlinguistik / Maschinelle Redeanlyse
ES 965 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Spezialbereiche der allgemeinen Sprachwissenschaft / Datenverarbeitung und Sprachwissenschaft. Computerlinguistik / Datenverarbeitung und Einzelgebiete der Grammatik
- Series
- Sprache, Sprechen und Computer ; 3
-
Lonsdale, D.; Mitamura, T.; Nyberg, E.: Acquisition of large lexicons for practical knowledge-based MT (1994/95)
0.03
0.025088158 = product of:
0.10035263 = sum of:
0.10035263 = weight(_text_:help in 478) [ClassicSimilarity], result of:
0.10035263 = score(doc=478,freq=2.0), product of:
0.32182297 = queryWeight, product of:
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.06841661 = queryNorm
0.31182557 = fieldWeight in 478, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.046875 = fieldNorm(doc=478)
0.25 = coord(1/4)
- Abstract
- Although knowledge based MT systems have the potential to achieve high translation accuracy, each successful application system requires a large amount of hand coded lexical knowledge. Systems like KBMT-89 and its descendants have demonstarted how knowledge based translation can produce good results in technical domains with tractable domain semantics. Nevertheless, the magnitude of the development task for large scale applications with 10s of 1000s of of domain concepts precludes a purely hand crafted approach. The current challenge for the next generation of knowledge based MT systems is to utilize online textual resources and corpus analysis software in order to automate the most laborious aspects of the knowledge acquisition process. This partial automation can in turn maximize the productivity of human knowledge engineers and help to make large scale applications of knowledge based MT an viable approach. Discusses the corpus based knowledge acquisition methodology used in KANT, a knowledge based translation system for multilingual document production. This methodology can be generalized beyond the KANT interlinhua approach for use with any system that requires similar kinds of knowledge
-
French, J.C.; Powell, A.L.; Schulman, E.: Using clustering strategies for creating authority files (2000)
0.03
0.025088158 = product of:
0.10035263 = sum of:
0.10035263 = weight(_text_:help in 5811) [ClassicSimilarity], result of:
0.10035263 = score(doc=5811,freq=2.0), product of:
0.32182297 = queryWeight, product of:
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.06841661 = queryNorm
0.31182557 = fieldWeight in 5811, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.046875 = fieldNorm(doc=5811)
0.25 = coord(1/4)
- Abstract
- As more online databases are integrated into digital libraries, the issue of quality control of the data becomes increasingly important, especially as it relates to the effective retrieval of information. Authority work, the need to discover and reconcile variant forms of strings in bibliographical entries, will become more critical in the future. Spelling variants, misspellings, and transliteration differences will all increase the difficulty of retrieving information. We investigate a number of approximate string matching techniques that have traditionally been used to help with this problem. We then introduce the notion of approximate word matching and show how it can be used to improve detection and categorization of variant forms. We demonstrate the utility of these approaches using data from the Astrophysics Data System and show how we can reduce the human effort involved in the creation of authority files
-
Dunning, T.: Statistical identification of language (1994)
0.03
0.025088158 = product of:
0.10035263 = sum of:
0.10035263 = weight(_text_:help in 4627) [ClassicSimilarity], result of:
0.10035263 = score(doc=4627,freq=2.0), product of:
0.32182297 = queryWeight, product of:
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.06841661 = queryNorm
0.31182557 = fieldWeight in 4627, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.046875 = fieldNorm(doc=4627)
0.25 = coord(1/4)
- Abstract
- A statistically based program has been written which learns to distinguish between languages. The amount of training text that such a program needs is surprisingly small, and the amount of text needed to make an identification is also quite small. The program incorporates no linguistic presuppositions other than the assumption that text can be encoded as a string of bytes. Such a program can be used to determine which language small bits of text are in. It also shows a potential for what might be called 'statistical philology' in that it may be applied directly to phonetic transcriptions to help elucidate family trees among language dialects. A variant of this program has been shown to be useful as a quality control in biochemistry. In this application, genetic sequences are assumed to be expressions in a language peculiar to the organism from which the sequence is taken. Thus language identification becomes species identification.
-
Oard, D.W.; He, D.; Wang, J.: User-assisted query translation for interactive cross-language information retrieval (2008)
0.03
0.025088158 = product of:
0.10035263 = sum of:
0.10035263 = weight(_text_:help in 3030) [ClassicSimilarity], result of:
0.10035263 = score(doc=3030,freq=2.0), product of:
0.32182297 = queryWeight, product of:
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.06841661 = queryNorm
0.31182557 = fieldWeight in 3030, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.046875 = fieldNorm(doc=3030)
0.25 = coord(1/4)
- Abstract
- Interactive Cross-Language Information Retrieval (CLIR), a process in which searcher and system collaborate to find documents that satisfy an information need regardless of the language in which those documents are written, calls for designs in which synergies between searcher and system can be leveraged so that the strengths of one can cover weaknesses of the other. This paper describes an approach that employs user-assisted query translation to help searchers better understand the system's operation. Supporting interaction and interface designs are introduced, and results from three user studies are presented. The results indicate that experienced searchers presented with this new system evolve new search strategies that make effective use of the new capabilities, that they achieve retrieval effectiveness comparable to results obtained using fully automatic techniques, and that reported satisfaction with support for cross-language searching increased. The paper concludes with a description of a freely available interactive CLIR system that incorporates lessons learned from this research.
-
Fernández, R.T.; Losada, D.E.: Effective sentence retrieval based on query-independent evidence (2012)
0.03
0.025088158 = product of:
0.10035263 = sum of:
0.10035263 = weight(_text_:help in 3728) [ClassicSimilarity], result of:
0.10035263 = score(doc=3728,freq=2.0), product of:
0.32182297 = queryWeight, product of:
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.06841661 = queryNorm
0.31182557 = fieldWeight in 3728, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.046875 = fieldNorm(doc=3728)
0.25 = coord(1/4)
- Abstract
- In this paper we propose an effective sentence retrieval method that consists of incorporating query-independent features into standard sentence retrieval models. To meet this aim, we apply a formal methodology and consider different query-independent features. In particular, we show that opinion-based features are promising. Opinion mining is an increasingly important research topic but little is known about how to improve retrieval algorithms with opinion-based components. In this respect, we consider here different kinds of opinion-based features to act as query-independent evidence and study whether this incorporation improves retrieval performance. On the other hand, information needs are usually related to people, locations or organizations. We hypothesize here that using these named entities as query-independent features may also improve the sentence relevance estimation. Finally, the length of the retrieval unit has been shown to be an important component in different retrieval scenarios. We therefore include length-based features in our study. Our evaluation demonstrates that, either in isolation or in combination, these query-independent features help to improve substantially the performance of state-of-the-art sentence retrieval methods.
-
Seelbach, D.: Computerlinguistik und Dokumentation : keyphrases in Dokumentationsprozessen (1975)
0.02
0.02431186 = product of:
0.09724744 = sum of:
0.09724744 = weight(_text_:und in 1299) [ClassicSimilarity], result of:
0.09724744 = score(doc=1299,freq=38.0), product of:
0.15174113 = queryWeight, product of:
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.06841661 = queryNorm
0.64087725 = fieldWeight in 1299, product of:
6.164414 = tf(freq=38.0), with freq of:
38.0 = termFreq=38.0
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.046875 = fieldNorm(doc=1299)
0.25 = coord(1/4)
- Classification
- ES 950 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Spezialbereiche der allgemeinen Sprachwissenschaft / Datenverarbeitung und Sprachwissenschaft. Computerlinguistik / Maschinelle Redeanlyse
ES 955 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Spezialbereiche der allgemeinen Sprachwissenschaft / Datenverarbeitung und Sprachwissenschaft. Computerlinguistik / Maschinelle Referatherstellung, linguistische Dokumentation und Information
- RVK
- ES 950 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Spezialbereiche der allgemeinen Sprachwissenschaft / Datenverarbeitung und Sprachwissenschaft. Computerlinguistik / Maschinelle Redeanlyse
ES 955 Allgemeine und vergleichende Sprach- und Literaturwissenschaft. Indogermanistik. Außereuropäische Sprachen und Literaturen / Spezialbereiche der allgemeinen Sprachwissenschaft / Datenverarbeitung und Sprachwissenschaft. Computerlinguistik / Maschinelle Referatherstellung, linguistische Dokumentation und Information
-
Aydin, Ö.; Karaarslan, E.: OpenAI ChatGPT generated literature review: : digital twin in healthcare (2022)
0.02
0.023653343 = product of:
0.09461337 = sum of:
0.09461337 = weight(_text_:help in 1852) [ClassicSimilarity], result of:
0.09461337 = score(doc=1852,freq=4.0), product of:
0.32182297 = queryWeight, product of:
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.06841661 = queryNorm
0.29399198 = fieldWeight in 1852, product of:
2.0 = tf(freq=4.0), with freq of:
4.0 = termFreq=4.0
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.03125 = fieldNorm(doc=1852)
0.25 = coord(1/4)
- Abstract
- Literature review articles are essential to summarize the related work in the selected field. However, covering all related studies takes too much time and effort. This study questions how Artificial Intelligence can be used in this process. We used ChatGPT to create a literature review article to show the stage of the OpenAI ChatGPT artificial intelligence application. As the subject, the applications of Digital Twin in the health field were chosen. Abstracts of the last three years (2020, 2021 and 2022) papers were obtained from the keyword "Digital twin in healthcare" search results on Google Scholar and paraphrased by ChatGPT. Later on, we asked ChatGPT questions. The results are promising; however, the paraphrased parts had significant matches when checked with the Ithenticate tool. This article is the first attempt to show the compilation and expression of knowledge will be accelerated with the help of artificial intelligence. We are still at the beginning of such advances. The future academic publishing process will require less human effort, which in turn will allow academics to focus on their studies. In future studies, we will monitor citations to this study to evaluate the academic validity of the content produced by the ChatGPT. 1. Introduction OpenAI ChatGPT (ChatGPT, 2022) is a chatbot based on the OpenAI GPT-3 language model. It is designed to generate human-like text responses to user input in a conversational context. OpenAI ChatGPT is trained on a large dataset of human conversations and can be used to create responses to a wide range of topics and prompts. The chatbot can be used for customer service, content creation, and language translation tasks, creating replies in multiple languages. OpenAI ChatGPT is available through the OpenAI API, which allows developers to access and integrate the chatbot into their applications and systems. OpenAI ChatGPT is a variant of the GPT (Generative Pre-trained Transformer) language model developed by OpenAI. It is designed to generate human-like text, allowing it to engage in conversation with users naturally and intuitively. OpenAI ChatGPT is trained on a large dataset of human conversations, allowing it to understand and respond to a wide range of topics and contexts. It can be used in various applications, such as chatbots, customer service agents, and language translation systems. OpenAI ChatGPT is a state-of-the-art language model able to generate coherent and natural text that can be indistinguishable from text written by a human. As an artificial intelligence, ChatGPT may need help to change academic writing practices. However, it can provide information and guidance on ways to improve people's academic writing skills.
-
Neumann, H.: Inszenierung und Metabotschaften eines periodisch getakteten Fernsehauftritts : Die Neujahrsansprachen der Bundeskanzler Helmut Kohl und Gerhard Schröder im Vergleich (2003)
0.02
0.021581644 = product of:
0.08632658 = sum of:
0.08632658 = weight(_text_:und in 2632) [ClassicSimilarity], result of:
0.08632658 = score(doc=2632,freq=22.0), product of:
0.15174113 = queryWeight, product of:
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.06841661 = queryNorm
0.5689069 = fieldWeight in 2632, product of:
4.690416 = tf(freq=22.0), with freq of:
22.0 = termFreq=22.0
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.0546875 = fieldNorm(doc=2632)
0.25 = coord(1/4)
- Abstract
- Herstellen der gleichen Wellenlänge zwischen Sender und Empfänger entscheidet über den kommunikativen Erfolg -gerade auch im politischen Bereich. Unter politikwissenschaftlicher als auch unter kommunikationswissenschaftlicher Fragestellung werden in der vorliegenden Arbeit acht Neujahrsansprachen von 1994 bis 2001 der Bundeskanzler Helmut Kohl und Gerhard Schröder einer systematischen Analyse unterzogen. Es findet eine Untersuchung der Sach- und Beziehungsebene statt. Verbale und visuelle Rhetorik beider Bundeskanzler werden miteinander verglichen und decodiert. Die Arbeit gibt zum einen Aufschluss über die Metabotschaften und das Corporate Design beider Bundeskanzler und diskutiert zum anderen Vor- und Nachteile der Kommunikationsstrategien zweier Kommunikationstypen, die unterschiedlicher nicht sein können.
- Source
- Information - Wissenschaft und Praxis. 54(2003) H.5, S.261-272
-
Semantik und künstliche Intelligenz : Beiträge zur automatischen Sprachbearbeitung II (1977)
0.02
0.021034157 = product of:
0.08413663 = sum of:
0.08413663 = weight(_text_:und in 6129) [ClassicSimilarity], result of:
0.08413663 = score(doc=6129,freq=4.0), product of:
0.15174113 = queryWeight, product of:
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.06841661 = queryNorm
0.5544748 = fieldWeight in 6129, product of:
2.0 = tf(freq=4.0), with freq of:
4.0 = termFreq=4.0
2.217899 = idf(docFreq=13141, maxDocs=44421)
0.125 = fieldNorm(doc=6129)
0.25 = coord(1/4)
- Series
- Grundlagen der Kommunikation und Kognition
-
Wang, F.L.; Yang, C.C.: Mining Web data for Chinese segmentation (2007)
0.02
0.020906799 = product of:
0.083627194 = sum of:
0.083627194 = weight(_text_:help in 1604) [ClassicSimilarity], result of:
0.083627194 = score(doc=1604,freq=2.0), product of:
0.32182297 = queryWeight, product of:
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.06841661 = queryNorm
0.25985464 = fieldWeight in 1604, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.0390625 = fieldNorm(doc=1604)
0.25 = coord(1/4)
- Abstract
- Modern information retrieval systems use keywords within documents as indexing terms for search of relevant documents. As Chinese is an ideographic character-based language, the words in the texts are not delimited by white spaces. Indexing of Chinese documents is impossible without a proper segmentation algorithm. Many Chinese segmentation algorithms have been proposed in the past. Traditional segmentation algorithms cannot operate without a large dictionary or a large corpus of training data. Nowadays, the Web has become the largest corpus that is ideal for Chinese segmentation. Although most search engines have problems in segmenting texts into proper words, they maintain huge databases of documents and frequencies of character sequences in the documents. Their databases are important potential resources for segmentation. In this paper, we propose a segmentation algorithm by mining Web data with the help of search engines. On the other hand, the Romanized pinyin of Chinese language indicates boundaries of words in the text. Our algorithm is the first to utilize the Romanized pinyin to segmentation. It is the first unified segmentation algorithm for the Chinese language from different geographical areas, and it is also domain independent because of the nature of the Web. Experiments have been conducted on the datasets of a recent Chinese segmentation competition. The results show that our algorithm outperforms the traditional algorithms in terms of precision and recall. Moreover, our algorithm can effectively deal with the problems of segmentation ambiguity, new word (unknown word) detection, and stop words.
-
Malo, P.; Sinha, A.; Korhonen, P.; Wallenius, J.; Takala, P.: Good debt or bad debt : detecting semantic orientations in economic texts (2014)
0.02
0.020906799 = product of:
0.083627194 = sum of:
0.083627194 = weight(_text_:help in 2226) [ClassicSimilarity], result of:
0.083627194 = score(doc=2226,freq=2.0), product of:
0.32182297 = queryWeight, product of:
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.06841661 = queryNorm
0.25985464 = fieldWeight in 2226, product of:
1.4142135 = tf(freq=2.0), with freq of:
2.0 = termFreq=2.0
4.7038717 = idf(docFreq=1093, maxDocs=44421)
0.0390625 = fieldNorm(doc=2226)
0.25 = coord(1/4)
- Abstract
- The use of robo-readers to analyze news texts is an emerging technology trend in computational finance. Recent research has developed sophisticated financial polarity lexicons for investigating how financial sentiments relate to future company performance. However, based on experience from fields that commonly analyze sentiment, it is well known that the overall semantic orientation of a sentence may differ from that of individual words. This article investigates how semantic orientations can be better detected in financial and economic news by accommodating the overall phrase-structure information and domain-specific use of language. Our three main contributions are the following: (a) a human-annotated finance phrase bank that can be used for training and evaluating alternative models; (b) a technique to enhance financial lexicons with attributes that help to identify expected direction of events that affect sentiment; and (c) a linearized phrase-structure model for detecting contextual semantic orientations in economic texts. The relevance of the newly added lexicon features and the benefit of using the proposed learning algorithm are demonstrated in a comparative study against general sentiment models as well as the popular word frequency models used in recent financial studies. The proposed framework is parsimonious and avoids the explosion in feature space caused by the use of conventional n-gram features.