;;; -*- Mode: TDL; Coding: utf-8 -*- 

;;; Author: 
;;; 2009 Berthold Crysmann

encoding := utf-8.

:include "global".
:include "mrs".

;; name of type containing quick check structure
qc-structure-unif := $qc_unif_set_pack.
qc-structure-subs := $qc_subs_set_pack.

;spanning-only-rules := $head-filler-get-rpunct-msgpred-rule $imp-get-rpunct-msgpred $coord-get-rpunct-msgpred-rule.

packing := true.



packing-restrictor :=  RELS HCONS.  

;;
;; Tokenizer settings
;;

case-sensitive.               ; be case sensitive in string and yy tokenizer
trivial-tokenizer.            ; for Lingo-style tokenizer "string"

sm := "hag.mem".

;preprocessor := "preprocessor.fsr".

vpm := "semi".

;ersatz-carg-path := "SYNSEM.LKEYS.KEYTAG".


;; name of type to get info about grammar into (if at all)
;grammar-info := grammar_info.


;; MRS 
postload-lisp-files := "lkb/mrsglobals.lisp".



;; type of a valid parse
;start-symbols := $root.
start-symbols := $root $root_spelling_supra_robust $root_spelling_robust $root_supra_robust.



;;; start-symbols := $root $root_robust.

;; names of attributes not to pass from daughter to mother in parsing
deleted-daughters := ARGS HEAD-DTR NON-HEAD-DTR DTR.

;; prefix of list-valued path where to unify the affix
;affixation-path := "ARGS.FIRST".

;; path into the MRS semantics
;mrs-path := "SYNSEM.LOCAL.CONT".

;:include "mrs".

;
;; (BK 2004-04-06) path into the MRSs CFROM, CTO, and CARG
;;
;mrs-cfrom-path := "SYNSEM.LOCAL.CONT.RELS.LIST.CFROM".
;mrs-cto-path   := "SYNSEM.LOCAL.CONT.RELS.LIST.CTO".
;mrs-carg-path  := "SYNSEM.LKEYS.KEYREL.CARG".

;;
;; path to LABEL for relation-to-word mapping, absolute and within a relation
;;
;label-path := "SYNSEM.LOCAL.CONT.KEY.WLINK".
;label-path-tail := "WLINK".



;; list of characters for the tokenizer to ignore, default is
;; "\t?!.:;,()-+*$\n"

;; removed `-' because of `I-C-E' and `Berlin-Zehlendorf'
punctuation-characters := ".,;!?".

;; allow creation of lexical entries to fail
lex-entries-can-fail.


orthographemics-maximum-chain-depth := 10.
orthographemics-minimum-stem-length := 1.



;;;
;;; following are a number of settings for the new (as of late 2008) token
;;; mapping and lexical filtering support in PET.
;;;

;;
;; first, the general format of chart mapping rules, much like MRS transfer.
;;
chart-mapping-context-path  := "+CONTEXT".
chart-mapping-input-path    := "+INPUT".
chart-mapping-output-path   := "+OUTPUT".
chart-mapping-position-path := "+POSITION".
;;
;; in lexical instatiation, the list of tokens activating a lexical entry (be
;; it native or generic) are unified into the lexical entry under this path.
;;
lexicon-tokens-path := "TOKENS.+LIST".
lexicon-last-token-path := "TOKENS.+LAST".
;;
;; furthermore, for the various input formats, we need to declare how parts of
;; input descriptions correspond to the grammar-internal feature geometry; in
;; the YY input format, for example, token feature structures (aka input items
;; PET-internally) are created from various parts of the token description.
;; 
token-form-path     := "+FORM".       ; [required] string for lexical lookup
token-id-path       := "+ID".         ; [optional] list of external ids
token-from-path     := "+FROM".       ; [optional] surface start position
token-to-path       := "+TO".         ; [optional] surface end position
token-postags-path  := "+TNT.+TAGS".  ; [optional] list of POS tags
token-posprobs-path := "+TNT.+PRBS".  ; [optional] list of POS probabilities
;;
;; finally, declare TDL status values for the various new entity types
;;
;generic-lexentry-status-values := generic-lex-entry.
token-mapping-rule-status-values := token-mapping-rule.
;lexical-filtering-rule-status-values := lexical-filtering-rule.