Gálvez, Carmen . Standardization of Terms Applying Finite-State Transducers (FST)., 2009 In: Handbook of Research on Digital Libraries: Design, Development and Impact. School of Communication and Information, Nanyang Technological University (Singapore) & Idea Group Inc., pp. 102-112. [Book chapter]
Preview |
PDF
Handbook-Galvez.pdf Download (2MB) | Preview |
English abstract
This chapter presents the different standardization methods of terms at the two basic approaches, non-linguistic and linguistic techniques, and to justify the application of processes based on Finite-State Transducers (FST). Standardization of terms is the procedure of matching and grouping together variants of the same term that are semantically equivalent. A term variant is a text occurrence that is conceptually related to an original term and can be used to search for information in text database. The uniterm and multiterm variants can be considered equivalent units for the purposes of automatic indexing. This chapter describes the computational and linguistic base of the finite-state approach, with emphasis on the influence of the formal language theory in the standardization process of uniterms and multiterms. The lemmatization and the use of syntactic pattern-matching, through equivalence relations represented in FST, are emerging methods for the standardization of terms.
Item type: | Book chapter |
---|---|
Keywords: | Finite-State Transduces; Term Conflation; Automatic Indexing |
Subjects: | L. Information technology and library technology > LL. Automated language processing. |
Depositing user: | Carmen Galvez |
Date deposited: | 09 Feb 2009 |
Last modified: | 02 Oct 2014 12:13 |
URI: | http://hdl.handle.net/10760/12780 |
References
Downloads
Downloads per month over past year
Actions (login required)
![]() |
View Item |