Please use this identifier to cite or link to this item: https://www.um.edu.mt/library/oar/handle/123456789/78384
Title: Don't let's try to break this down : teasing apart lexical chunks
Authors: Bogart, Zoë (2011)
Keywords: Computational linguistics
Natural language processing (Computer science)
Lexicology -- Data processing
Issue Date: 2011
Citation: Bogart, Z. (2011). Don't let's try to break this down: teasing apart lexical chunks (Master's dissertation).
Abstract: Chunk acquisition from the BNC data resulted in chunk lists of hundreds of thousands and occasionally over 1 million chunks. Because of the impracticability of evaluating such a large number of chunks for each measure, evaluation was performed in the following manner: an evaluation text was selected, and all 'good' chunks in that text were determined through a combination of human judgments and comparison with dictionaries, as described later. The chunk databases found by each statistical measure were then used by a program which went through the article and extracted all lexical chunks that occurred in the database. The final list of chunks extracted in this way was then compared to the pre-determined 'good' chunks for each measure, and values of precision, recall, and f-measure were calculated.
Description: M.SC.ARTIFICIAL INTELLIGENCE
URI: https://www.um.edu.mt/library/oar/handle/123456789/78384
Appears in Collections:Dissertations - FacICT - 2011
Dissertations - FacICTAI - 2002-2014

Files in This Item:
File Description SizeFormat 
M.SC.LANG.SCIENCE_TECH._Bogart_Zoë_2011.pdf
  Restricted Access
8.23 MBAdobe PDFView/Open Request a copy


Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.