[Author Login]
[Home]
ROA:980
Title:Some correct error-driven versions of the Constraint Demotion algorithm
Authors:Paul Boersma
Comment:
Length:20
Abstract:This paper supersedes ROA 953: a proof of correct convergence is now included. This paper shows that Error-Driven Constraint Demotion (EDCD), an error-driven learning algorithm proposed by Tesar (1995) for Prince and Smolensky's (1993) version of Optimality Theory, can fail to converge to a correct totally ranked hierarchy of constraints, unlike the earlier non-error-driven learning algorithms proposed by Tesar and Smolensky (1993). The cause of the problem is found in Tesar's use of 'mark-pooling ties', indicating that EDCD can be repaired by assuming Anttila's (1997) 'permuting ties' instead. Proofs show, and simulations confirm, that totally ranked hierarchies can indeed be found by both this repaired version of EDCD and Boersma's (1998) Minimal Gradual Learning Algorithm.
Type:Paper/tech report
Area/Keywords:Learnability,Language Acquisition,Computation
Article:Version 1