X-Git-Url: http://matita.cs.unibo.it/gitweb/?a=blobdiff_plain;f=helm%2Focaml%2Fcic_disambiguation%2FcicTextualLexer2.ml;h=ba70d2745c8f73126733eff03c32664e15a1fc75;hb=0c6a5aadb1a7746681a8e26fc0b009f847c10557;hp=be2a034ac7f6c48a08828cc7685202c441416b9c;hpb=57603e5f5f5507c21ec7cb823e5bef708441b0d0;p=helm.git diff --git a/helm/ocaml/cic_disambiguation/cicTextualLexer2.ml b/helm/ocaml/cic_disambiguation/cicTextualLexer2.ml index be2a034ac..ba70d2745 100644 --- a/helm/ocaml/cic_disambiguation/cicTextualLexer2.ml +++ b/helm/ocaml/cic_disambiguation/cicTextualLexer2.ml @@ -35,7 +35,7 @@ let regexp blanks = blank+ let regexp num = digit+ let regexp tex_token = '\\' alpha+ let regexp symbol = [^ 'a' - 'z' 'A' - 'Z' '0' - '9' ' ' '\t' '\n' ] -let regexp ident_cont = alpha | num | '_' +let regexp ident_cont = alpha | num | '_' | '\'' let regexp ident_cont' = ident_cont | tex_token let regexp ident = (alpha ident_cont*) | ('_' ident_cont+) let regexp ident' = ((alpha | tex_token) ident_cont'*) | ('_' ident_cont'+) @@ -51,14 +51,22 @@ let regexp uri = let keywords = Hashtbl.create 17 let _ = List.iter (fun keyword -> Hashtbl.add keywords keyword ("", keyword)) - [ "Prop"; "Type"; "Set"; "let"; "rec"; "using"; "match"; "with" ] + [ "Prop"; "Type"; "Set"; "let"; "Let"; "rec"; "using"; "match"; "with" ] let error lexbuf msg = raise (Error (Ulexing.lexeme_start lexbuf, Ulexing.lexeme_end lexbuf, msg)) let error_at_end lexbuf msg = raise (Error (Ulexing.lexeme_end lexbuf, Ulexing.lexeme_end lexbuf, msg)) -let return lexbuf token = (token, Ulexing.loc lexbuf) +let return lexbuf token = + let flocation_begin = + { Lexing.pos_fname = ""; Lexing.pos_lnum = -1; Lexing.pos_bol = -1; + Lexing.pos_cnum = Ulexing.lexeme_start lexbuf } + in + let flocation_end = + { flocation_begin with Lexing.pos_cnum = Ulexing.lexeme_end lexbuf } + in + (token, (flocation_begin, flocation_end)) (* let parse_ext_ident ident = @@ -119,7 +127,7 @@ let rec token = lexer let tok_func stream = let lexbuf = Ulexing.from_utf8_stream stream in - Token.make_stream_and_location + Token.make_stream_and_flocation (fun () -> try token lexbuf