Please use this identifier to cite or link to this item: https://hdl.handle.net/2440/66745
Full metadata record
DC FieldValueLanguage
dc.contributor.authorShi, Q.en
dc.contributor.authorAltun, Y.en
dc.contributor.authorSmola, A.en
dc.contributor.authorVishwanathan, S.en
dc.date.issued2007en
dc.identifier.citationProceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, 2007: pp.640-648en
dc.identifier.urihttp://hdl.handle.net/2440/66745-
dc.description.abstractIn this paper, we study the problem of automatically segmenting written text into paragraphs. This is inherently a sequence labelling problem, however, previous approaches ignore this dependency. We propose a novel approach for automatic paragraph segmentation, namely training Semi-Markov models discriminatively using a Max-Margin method. This method allows us to model the sequential nature of the problem and to incorporate features of a whole paragraph, such as paragraph coherence which cannot be used in previous models. Experimental evaluation on four text corpora shows improvement over the previous state-of-the art method on this task.en
dc.description.statementofresponsibilityQinfeng Shi, Yasemin Altun, Alex Smola and S. V. N. Vishwanathanen
dc.language.isoenen
dc.publisherAssociation for Computational Linguisticsen
dc.rightsCopyright 2007 Association for Computational Linguisticsen
dc.source.urihttp://www.eprints.pascal-network.org/archive/00003986/en
dc.titleSemi-Markov models for sequence segmentationen
dc.typeConference paperen
dc.contributor.conferenceJoint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (2007 : Prague, Czech Republic)en
dc.publisher.placeUnited Statesen
pubs.publication-statusPublisheden
dc.identifier.orcidShi, Q. [0000-0002-9126-2107]en
Appears in Collections:Aurora harvest
Computer Science publications

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.