Search (1293 results, page 7 of 65)

  • × language_ss:"e"
  1. Huang, T.; Nie, R.; Zhao, Y.: Archival knowledge in the field of personal archiving : an exploratory study based on grounded theory (2021) 0.04
    0.04483224 = product of:
      0.17932896 = sum of:
        0.17932896 = weight(_text_:huang in 1174) [ClassicSimilarity], result of:
          0.17932896 = score(doc=1174,freq=2.0), product of:
            0.456803 = queryWeight, product of:
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.06428098 = queryNorm
            0.39257395 = fieldWeight in 1174, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1174)
      0.25 = coord(1/4)
    
  2. Xiang, R.; Chersoni, E.; Lu, Q.; Huang, C.-R.; Li, W.; Long, Y.: Lexical data augmentation for sentiment analysis (2021) 0.04
    0.04483224 = product of:
      0.17932896 = sum of:
        0.17932896 = weight(_text_:huang in 1393) [ClassicSimilarity], result of:
          0.17932896 = score(doc=1393,freq=2.0), product of:
            0.456803 = queryWeight, product of:
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.06428098 = queryNorm
            0.39257395 = fieldWeight in 1393, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1393)
      0.25 = coord(1/4)
    
  3. Lindau, S.T.; Makelarski, J.A.; Abramsohn, E.M.; Beiser, D.G.; Boyd, K.; Huang, E.S.; Paradise, K.; Tung, E.L.: Sharing information about health-related resources : observations from a community resource referral intervention trial in a predominantly African American/Black community (2022) 0.04
    0.04483224 = product of:
      0.17932896 = sum of:
        0.17932896 = weight(_text_:huang in 1503) [ClassicSimilarity], result of:
          0.17932896 = score(doc=1503,freq=2.0), product of:
            0.456803 = queryWeight, product of:
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.06428098 = queryNorm
            0.39257395 = fieldWeight in 1503, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1503)
      0.25 = coord(1/4)
    
  4. Kulczycki, E.; Huang, Y.; Zuccala, A.A.; Engels, T.C.E.; Ferrara, A.; Guns, R.; Pölönen, J.; Sivertsen, G.; Taskin, Z.; Zhang, L.: Uses of the Journal Impact Factor in national journal rankings in China and Europe (2022) 0.04
    0.04483224 = product of:
      0.17932896 = sum of:
        0.17932896 = weight(_text_:huang in 1770) [ClassicSimilarity], result of:
          0.17932896 = score(doc=1770,freq=2.0), product of:
            0.456803 = queryWeight, product of:
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.06428098 = queryNorm
            0.39257395 = fieldWeight in 1770, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1770)
      0.25 = coord(1/4)
    
  5. Jiang, Y.; Meng, R.; Huang, Y.; Lu, W.; Liu, J.: Generating keyphrases for readers : a controllable keyphrase generation framework (2023) 0.04
    0.04483224 = product of:
      0.17932896 = sum of:
        0.17932896 = weight(_text_:huang in 2014) [ClassicSimilarity], result of:
          0.17932896 = score(doc=2014,freq=2.0), product of:
            0.456803 = queryWeight, product of:
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.06428098 = queryNorm
            0.39257395 = fieldWeight in 2014, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.0390625 = fieldNorm(doc=2014)
      0.25 = coord(1/4)
    
  6. Zhang, L.; Gou, Z.; Fang, Z.; Sivertsen, G.; Huang, Y.: Who tweets scientific publications? : a large-scale study of tweeting audiences in all areas of research (2023) 0.04
    0.04483224 = product of:
      0.17932896 = sum of:
        0.17932896 = weight(_text_:huang in 2191) [ClassicSimilarity], result of:
          0.17932896 = score(doc=2191,freq=2.0), product of:
            0.456803 = queryWeight, product of:
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.06428098 = queryNorm
            0.39257395 = fieldWeight in 2191, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.0390625 = fieldNorm(doc=2191)
      0.25 = coord(1/4)
    
  7. Usbeck, R.; Yan, X.; Perevalov, A.; Jiang, L.; Schulz, J.; Kraft, A.; Möller, C.; Huang, J.; Reineke, J.; Ngonga Ngomo, A.-C.; Saleem, M.; Both, A.: QALD-10 - The 10th challenge on question answering over linked data: : shifting from DBpedia to Wikidata as a KG for KGQA (2023) 0.04
    0.04483224 = product of:
      0.17932896 = sum of:
        0.17932896 = weight(_text_:huang in 2350) [ClassicSimilarity], result of:
          0.17932896 = score(doc=2350,freq=2.0), product of:
            0.456803 = queryWeight, product of:
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.06428098 = queryNorm
            0.39257395 = fieldWeight in 2350, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.0390625 = fieldNorm(doc=2350)
      0.25 = coord(1/4)
    
  8. Zhou, H.; Sun, B.; Guns, R.; Engels, T.C.E.; Huang, Y.; Zhang, L.: How do life sciences cite social sciences? : Characterizing the volume and trajectory of citations (2024) 0.04
    0.04483224 = product of:
      0.17932896 = sum of:
        0.17932896 = weight(_text_:huang in 2376) [ClassicSimilarity], result of:
          0.17932896 = score(doc=2376,freq=2.0), product of:
            0.456803 = queryWeight, product of:
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.06428098 = queryNorm
            0.39257395 = fieldWeight in 2376, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.0390625 = fieldNorm(doc=2376)
      0.25 = coord(1/4)
    
  9. Chen, H.; Chung, Y.-M.; Ramsey, M.; Yang, C.C.: ¬A smart itsy bitsy spider for the Web (1998) 0.04
    0.044092894 = product of:
      0.17637157 = sum of:
        0.17637157 = weight(_text_:java in 1871) [ClassicSimilarity], result of:
          0.17637157 = score(doc=1871,freq=2.0), product of:
            0.45302066 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06428098 = queryNorm
            0.38932347 = fieldWeight in 1871, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0390625 = fieldNorm(doc=1871)
      0.25 = coord(1/4)
    
    Abstract
    As part of the ongoing Illinois Digital Library Initiative project, this research proposes an intelligent agent approach to Web searching. In this experiment, we developed 2 Web personal spiders based on best first search and genetic algorithm techniques, respectively. These personal spiders can dynamically take a user's selected starting homepages and search for the most closely related homepages in the Web, based on the links and keyword indexing. A graphical, dynamic, Jav-based interface was developed and is available for Web access. A system architecture for implementing such an agent-spider is presented, followed by deteiled discussions of benchmark testing and user evaluation results. In benchmark testing, although the genetic algorithm spider did not outperform the best first search spider, we found both results to be comparable and complementary. In user evaluation, the genetic algorithm spider obtained significantly higher recall value than that of the best first search spider. However, their precision values were not statistically different. The mutation process introduced in genetic algorithms allows users to find other potential relevant homepages that cannot be explored via a conventional local search process. In addition, we found the Java-based interface to be a necessary component for design of a truly interactive and dynamic Web agent
  10. Chen, C.: CiteSpace II : detecting and visualizing emerging trends and transient patterns in scientific literature (2006) 0.04
    0.044092894 = product of:
      0.17637157 = sum of:
        0.17637157 = weight(_text_:java in 272) [ClassicSimilarity], result of:
          0.17637157 = score(doc=272,freq=2.0), product of:
            0.45302066 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06428098 = queryNorm
            0.38932347 = fieldWeight in 272, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0390625 = fieldNorm(doc=272)
      0.25 = coord(1/4)
    
    Abstract
    This article describes the latest development of a generic approach to detecting and visualizing emerging trends and transient patterns in scientific literature. The work makes substantial theoretical and methodological contributions to progressive knowledge domain visualization. A specialty is conceptualized and visualized as a time-variant duality between two fundamental concepts in information science: research fronts and intellectual bases. A research front is defined as an emergent and transient grouping of concepts and underlying research issues. The intellectual base of a research front is its citation and co-citation footprint in scientific literature - an evolving network of scientific publications cited by research-front concepts. Kleinberg's (2002) burst-detection algorithm is adapted to identify emergent research-front concepts. Freeman's (1979) betweenness centrality metric is used to highlight potential pivotal points of paradigm shift over time. Two complementary visualization views are designed and implemented: cluster views and time-zone views. The contributions of the approach are that (a) the nature of an intellectual base is algorithmically and temporally identified by emergent research-front terms, (b) the value of a co-citation cluster is explicitly interpreted in terms of research-front concepts, and (c) visually prominent and algorithmically detected pivotal points substantially reduce the complexity of a visualized network. The modeling and visualization process is implemented in CiteSpace II, a Java application, and applied to the analysis of two research fields: mass extinction (1981-2004) and terrorism (1990-2003). Prominent trends and pivotal points in visualized networks were verified in collaboration with domain experts, who are the authors of pivotal-point articles. Practical implications of the work are discussed. A number of challenges and opportunities for future studies are identified.
  11. Eddings, J.: How the Internet works (1994) 0.04
    0.044092894 = product of:
      0.17637157 = sum of:
        0.17637157 = weight(_text_:java in 2514) [ClassicSimilarity], result of:
          0.17637157 = score(doc=2514,freq=2.0), product of:
            0.45302066 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06428098 = queryNorm
            0.38932347 = fieldWeight in 2514, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0390625 = fieldNorm(doc=2514)
      0.25 = coord(1/4)
    
    Abstract
    How the Internet Works promises "an exciting visual journey down the highways and byways of the Internet," and it delivers. The book's high quality graphics and simple, succinct text make it the ideal book for beginners; however it still has much to offer for Net vets. This book is jam- packed with cool ways to visualize how the Net works. The first section visually explores how TCP/IP, Winsock, and other Net connectivity mysteries work. This section also helps you understand how e-mail addresses and domains work, what file types mean, and how information travels across the Net. Part 2 unravels the Net's underlying architecture, including good information on how routers work and what is meant by client/server architecture. The third section covers your own connection to the Net through an Internet Service Provider (ISP), and how ISDN, cable modems, and Web TV work. Part 4 discusses e-mail, spam, newsgroups, Internet Relay Chat (IRC), and Net phone calls. In part 5, you'll find out how other Net tools, such as gopher, telnet, WAIS, and FTP, can enhance your Net experience. The sixth section takes on the World Wide Web, including everything from how HTML works to image maps and forms. Part 7 looks at other Web features such as push technology, Java, ActiveX, and CGI scripting, while part 8 deals with multimedia on the Net. Part 9 shows you what intranets are and covers groupware, and shopping and searching the Net. The book wraps up with part 10, a chapter on Net security that covers firewalls, viruses, cookies, and other Web tracking devices, plus cryptography and parental controls.
  12. Wu, D.; Shi, J.: Classical music recording ontology used in a library catalog (2016) 0.04
    0.044092894 = product of:
      0.17637157 = sum of:
        0.17637157 = weight(_text_:java in 4179) [ClassicSimilarity], result of:
          0.17637157 = score(doc=4179,freq=2.0), product of:
            0.45302066 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06428098 = queryNorm
            0.38932347 = fieldWeight in 4179, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.0390625 = fieldNorm(doc=4179)
      0.25 = coord(1/4)
    
    Abstract
    In order to improve the organization of classical music information resources, we constructed a classical music recording ontology, on top of which we then designed an online classical music catalog. Our construction of the classical music recording ontology consisted of three steps: identifying the purpose, analyzing the ontology, and encoding the ontology. We identified the main classes and properties of the domain by investigating classical music recording resources and users' information needs. We implemented the ontology in the Web Ontology Language (OWL) using five steps: transforming the properties, encoding the transformed properties, defining ranges of the properties, constructing individuals, and standardizing the ontology. In constructing the online catalog, we first designed the structure and functions of the catalog based on investigations into users' information needs and information-seeking behaviors. Then we extracted classes and properties of the ontology using the Apache Jena application programming interface (API), and constructed a catalog in the Java environment. The catalog provides a hierarchical main page (built using the Functional Requirements for Bibliographic Records (FRBR) model), a classical music information network and integrated information service; this combination of features greatly eases the task of finding classical music recordings and more information about classical music.
  13. Hoenkamp, E.; Bruza, P.D.; Song, D.; Huang, Q.: ¬An effective approach to verbose queries using a limited dependencies language model (2009) 0.04
    0.03586579 = product of:
      0.14346316 = sum of:
        0.14346316 = weight(_text_:huang in 3122) [ClassicSimilarity], result of:
          0.14346316 = score(doc=3122,freq=2.0), product of:
            0.456803 = queryWeight, product of:
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.06428098 = queryNorm
            0.31405917 = fieldWeight in 3122, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.03125 = fieldNorm(doc=3122)
      0.25 = coord(1/4)
    
  14. Ayadi, H.; Torjmen-Khemakhem, M.; Daoud, M.; Xiangji Huang, J.; Ben Jemaa, M.: MF-Re-Rank : a modality feature-based re-ranking model for medical image retrieval (2018) 0.04
    0.03586579 = product of:
      0.14346316 = sum of:
        0.14346316 = weight(_text_:huang in 459) [ClassicSimilarity], result of:
          0.14346316 = score(doc=459,freq=2.0), product of:
            0.456803 = queryWeight, product of:
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.06428098 = queryNorm
            0.31405917 = fieldWeight in 459, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.03125 = fieldNorm(doc=459)
      0.25 = coord(1/4)
    
  15. Noerr, P.: ¬The Digital Library Tool Kit (2001) 0.04
    0.035274316 = product of:
      0.14109726 = sum of:
        0.14109726 = weight(_text_:java in 774) [ClassicSimilarity], result of:
          0.14109726 = score(doc=774,freq=2.0), product of:
            0.45302066 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06428098 = queryNorm
            0.31145877 = fieldWeight in 774, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.03125 = fieldNorm(doc=774)
      0.25 = coord(1/4)
    
    Footnote
    This Digital Library Tool Kit was sponsored by Sun Microsystems, Inc. to address some of the leading questions that academic institutions, public libraries, government agencies, and museums face in trying to develop, manage, and distribute digital content. The evolution of Java programming, digital object standards, Internet access, electronic commerce, and digital media management models is causing educators, CIOs, and librarians to rethink many of their traditional goals and modes of operation. New audiences, continuous access to collections, and enhanced services to user communities are enabled. As one of the leading technology providers to education and library communities, Sun is pleased to present this comprehensive introduction to digital libraries
  16. Herrero-Solana, V.; Moya Anegón, F. de: Graphical Table of Contents (GTOC) for library collections : the application of UDC codes for the subject maps (2003) 0.04
    0.035274316 = product of:
      0.14109726 = sum of:
        0.14109726 = weight(_text_:java in 3758) [ClassicSimilarity], result of:
          0.14109726 = score(doc=3758,freq=2.0), product of:
            0.45302066 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06428098 = queryNorm
            0.31145877 = fieldWeight in 3758, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.03125 = fieldNorm(doc=3758)
      0.25 = coord(1/4)
    
    Abstract
    The representation of information contents by graphical maps is an extended ongoing research topic. In this paper we introduce the application of UDC codes for the subject maps development. We use the following graphic representation methodologies: 1) Multidimensional scaling (MDS), 2) Cluster analysis, 3) Neural networks (Self Organizing Map - SOM). Finally, we conclude about the application viability of every kind of map. 1. Introduction Advanced techniques for Information Retrieval (IR) currently make up one of the most active areas for research in the field of library and information science. New models representing document content are replacing the classic systems in which the search terms supplied by the user were compared against the indexing terms existing in the inverted files of a database. One of the topics most often studied in the last years is bibliographic browsing, a good complement to querying strategies. Since the 80's, many authors have treated this topic. For example, Ellis establishes that browsing is based an three different types of tasks: identification, familiarization and differentiation (Ellis, 1989). On the other hand, Cove indicates three different browsing types: searching browsing, general purpose browsing and serendipity browsing (Cove, 1988). Marcia Bates presents six different types (Bates, 1989), although the classification of Bawden is the one that really interests us: 1) similarity comparison, 2) structure driven, 3) global vision (Bawden, 1993). The global vision browsing implies the use of graphic representations, which we will call map displays, that allow the user to get a global idea of the nature and structure of the information in the database. In the 90's, several authors worked an this research line, developing different types of maps. One of the most active was Xia Lin what introduced the concept of Graphical Table of Contents (GTOC), comparing the maps to true table of contents based an graphic representations (Lin 1996). Lin applies the algorithm SOM to his own personal bibliography, analyzed in function of the words of the title and abstract fields, and represented in a two-dimensional map (Lin 1997). Later on, Lin applied this type of maps to create websites GTOCs, through a Java application.
  17. Vlachidis, A.; Binding, C.; Tudhope, D.; May, K.: Excavating grey literature : a case study on the rich indexing of archaeological documents via natural language-processing techniques and knowledge-based resources (2010) 0.04
    0.035274316 = product of:
      0.14109726 = sum of:
        0.14109726 = weight(_text_:java in 935) [ClassicSimilarity], result of:
          0.14109726 = score(doc=935,freq=2.0), product of:
            0.45302066 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06428098 = queryNorm
            0.31145877 = fieldWeight in 935, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.03125 = fieldNorm(doc=935)
      0.25 = coord(1/4)
    
    Abstract
    Purpose - This paper sets out to discuss the use of information extraction (IE), a natural language-processing (NLP) technique to assist "rich" semantic indexing of diverse archaeological text resources. The focus of the research is to direct a semantic-aware "rich" indexing of diverse natural language resources with properties capable of satisfying information retrieval from online publications and datasets associated with the Semantic Technologies for Archaeological Resources (STAR) project. Design/methodology/approach - The paper proposes use of the English Heritage extension (CRM-EH) of the standard core ontology in cultural heritage, CIDOC CRM, and exploitation of domain thesauri resources for driving and enhancing an Ontology-Oriented Information Extraction process. The process of semantic indexing is based on a rule-based Information Extraction technique, which is facilitated by the General Architecture of Text Engineering (GATE) toolkit and expressed by Java Annotation Pattern Engine (JAPE) rules. Findings - Initial results suggest that the combination of information extraction with knowledge resources and standard conceptual models is capable of supporting semantic-aware term indexing. Additional efforts are required for further exploitation of the technique and adoption of formal evaluation methods for assessing the performance of the method in measurable terms. Originality/value - The value of the paper lies in the semantic indexing of 535 unpublished online documents often referred to as "Grey Literature", from the Archaeological Data Service OASIS corpus (Online AccesS to the Index of archaeological investigationS), with respect to the CRM ontological concepts E49.Time Appellation and P19.Physical Object.
  18. Radhakrishnan, A.: Swoogle : an engine for the Semantic Web (2007) 0.04
    0.035274316 = product of:
      0.14109726 = sum of:
        0.14109726 = weight(_text_:java in 709) [ClassicSimilarity], result of:
          0.14109726 = score(doc=709,freq=2.0), product of:
            0.45302066 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06428098 = queryNorm
            0.31145877 = fieldWeight in 709, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.03125 = fieldNorm(doc=709)
      0.25 = coord(1/4)
    
    Content
    "Swoogle, the Semantic web search engine, is a research project carried out by the ebiquity research group in the Computer Science and Electrical Engineering Department at the University of Maryland. It's an engine tailored towards finding documents on the semantic web. The whole research paper is available here. Semantic web is touted as the next generation of online content representation where the web documents are represented in a language that is not only easy for humans but is machine readable (easing the integration of data as never thought possible) as well. And the main elements of the semantic web include data model description formats such as Resource Description Framework (RDF), a variety of data interchange formats (e.g. RDF/XML, Turtle, N-Triples), and notations such as RDF Schema (RDFS), the Web Ontology Language (OWL), all of which are intended to provide a formal description of concepts, terms, and relationships within a given knowledge domain (Wikipedia). And Swoogle is an attempt to mine and index this new set of web documents. The engine performs crawling of semantic documents like most web search engines and the search is available as web service too. The engine is primarily written in Java with the PHP used for the front-end and MySQL for database. Swoogle is capable of searching over 10,000 ontologies and indexes more that 1.3 million web documents. It also computes the importance of a Semantic Web document. The techniques used for indexing are the more google-type page ranking and also mining the documents for inter-relationships that are the basis for the semantic web. For more information on how the RDF framework can be used to relate documents, read the link here. Being a research project, and with a non-commercial motive, there is not much hype around Swoogle. However, the approach to indexing of Semantic web documents is an approach that most engines will have to take at some point of time. When the Internet debuted, there were no specific engines available for indexing or searching. The Search domain only picked up as more and more content became available. One fundamental question that I've always wondered about it is - provided that the search engines return very relevant results for a query - how to ascertain that the documents are indeed the most relevant ones available. There is always an inherent delay in indexing of document. Its here that the new semantic documents search engines can close delay. Experimenting with the concept of Search in the semantic web can only bore well for the future of search technology."
  19. Piros, A.: Automatic interpretation of complex UDC numbers : towards support for library systems (2015) 0.04
    0.035274316 = product of:
      0.14109726 = sum of:
        0.14109726 = weight(_text_:java in 3301) [ClassicSimilarity], result of:
          0.14109726 = score(doc=3301,freq=2.0), product of:
            0.45302066 = queryWeight, product of:
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.06428098 = queryNorm
            0.31145877 = fieldWeight in 3301, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.0475073 = idf(docFreq=104, maxDocs=44421)
              0.03125 = fieldNorm(doc=3301)
      0.25 = coord(1/4)
    
    Abstract
    Analytico-synthetic and faceted classifications, such as Universal Decimal Classification (UDC) express content of documents with complex, pre-combined classification codes. Without classification authority control that would help manage and access structured notations, the use of UDC codes in searching and browsing is limited. Existing UDC parsing solutions are usually created for a particular database system or a specific task and are not widely applicable. The approach described in this paper provides a solution by which the analysis and interpretation of UDC notations would be stored into an intermediate format (in this case, in XML) by automatic means without any data or information loss. Due to its richness, the output file can be converted into different formats, such as standard mark-up and data exchange formats or simple lists of the recommended entry points of a UDC number. The program can also be used to create authority records containing complex UDC numbers which can be comprehensively analysed in order to be retrieved effectively. The Java program, as well as the corresponding schema definition it employs, is under continuous development. The current version of the interpreter software is now available online for testing purposes at the following web site: http://interpreter-eto.rhcloud.com. The future plan is to implement conversion methods for standard formats and to create standard online interfaces in order to make it possible to use the features of software as a service. This would result in the algorithm being able to be employed both in existing and future library systems to analyse UDC numbers without any significant programming effort.
  20. Rasmussen, E.M.: Indexing and retrieval for the Web (2002) 0.03
    0.03138257 = product of:
      0.12553027 = sum of:
        0.12553027 = weight(_text_:huang in 5285) [ClassicSimilarity], result of:
          0.12553027 = score(doc=5285,freq=2.0), product of:
            0.456803 = queryWeight, product of:
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.06428098 = queryNorm
            0.27480176 = fieldWeight in 5285, product of:
              1.4142135 = tf(freq=2.0), with freq of:
                2.0 = termFreq=2.0
              7.1063476 = idf(docFreq=98, maxDocs=44421)
              0.02734375 = fieldNorm(doc=5285)
      0.25 = coord(1/4)
    
    Abstract
    Techniques for automated indexing and information retrieval (IR) have been developed, tested, and refined over the past 40 years, and are well documented (see, for example, Agosti & Smeaton, 1996; BaezaYates & Ribeiro-Neto, 1999a; Frakes & Baeza-Yates, 1992; Korfhage, 1997; Salton, 1989; Witten, Moffat, & Bell, 1999). With the introduction of the Web, and the capability to index and retrieve via search engines, these techniques have been extended to a new environment. They have been adopted, altered, and in some Gases extended to include new methods. "In short, search engines are indispensable for searching the Web, they employ a variety of relatively advanced IR techniques, and there are some peculiar aspects of search engines that make searching the Web different than more conventional information retrieval" (Gordon & Pathak, 1999, p. 145). The environment for information retrieval an the World Wide Web differs from that of "conventional" information retrieval in a number of fundamental ways. The collection is very large and changes continuously, with pages being added, deleted, and altered. Wide variability between the size, structure, focus, quality, and usefulness of documents makes Web documents much more heterogeneous than a typical electronic document collection. The wide variety of document types includes images, video, audio, and scripts, as well as many different document languages. Duplication of documents and sites is common. Documents are interconnected through networks of hyperlinks. Because of the size and dynamic nature of the Web, preprocessing all documents requires considerable resources and is often not feasible, certainly not an the frequent basis required to ensure currency. Query length is usually much shorter than in other environments-only a few words-and user behavior differs from that in other environments. These differences make the Web a novel environment for information retrieval (Baeza-Yates & Ribeiro-Neto, 1999b; Bharat & Henzinger, 1998; Huang, 2000).

Languages

  • d 32
  • m 3
  • nl 1
  • More… Less…

Types

  • a 854
  • m 310
  • el 106
  • s 94
  • i 21
  • n 17
  • x 12
  • r 10
  • b 7
  • ? 1
  • v 1
  • More… Less…

Themes

Subjects

Classifications