[developers] training parse selection models using the fine system

Yi Zhang yzhang at CoLi.Uni-SB.DE
Tue Apr 11 11:53:02 CEST 2006

Hi Stephan,

Tim, Valia and I are still trying some experiments with the fine system 
and ERG. We have treebanked with different versions of the grammar and now 
I am trying to train the disambiguation models. I started with the fine 
system from http://lingo.stanford.edu/ftp/builds/2006-03-05/ , following 
the guide on the wiki. But when I click the Trees/Train, I got error 
podium-loop(): Symbol TRAIN does not have a function definition.

I also tried to follow the file `load' and `fc.lisp' in $HINOKI, using the 
LOGON tree. The feature caching seems to work, so that I got big `fc.abt' 
files afterwards. But when I try Trees/Train:

[11:07:26] operate-on-profiles(): reading `jan-06/jh0'
operate-on-profiles(): caching `jan-06/jh0' [11 - 511|.
[11:07:26] open-fc(): new BTree `fc.bdb'.
podium-loop(): Attempt to call #("db_open" 11440496 0 2 11440496) for
                which the definition has not yet been (or is no longer)

So my questions go: 1) how do I train the parse selection model with 
either DELPHIN tree or LOGON tree?
2) can I only cache some basic features for a model used by pet? I tried 
to set:
(setf tsdb::*feature-grandparenting* 0)
(setf tsdb::*feature-use-preterminal-types-p* nil)
(setf tsdb::*feature-lexicalization-p* 0)
(setf tsdb::*feature-active-edges-p* nil)
(setf tsdb::*feature-ngram-size* 0)
(setf tsdb::*feature-ngram-tag* :type)
(setf tsdb::*feature-ngram-back-off-p* nil)
but the fc.abt is still huge, and it still takes a long time.


More information about the developers mailing list