Merge branch 'rinderknecht-dev' into 'dev'
Refactoring of the front-end See merge request ligolang/ligo!332
This commit is contained in:
commit
7828b57636
@ -99,9 +99,9 @@ ligo: lexer error: Negative byte sequence.
|
||||
|
||||
run_ligo_bad [ "compile-contract" ; "../../test/lexer/reserved_name.ligo" ; "main" ] ;
|
||||
[%expect {|
|
||||
ligo: lexer error: Reserved name: args.
|
||||
ligo: lexer error: Reserved name: arguments.
|
||||
Hint: Change the name.
|
||||
{"parser_loc":"in file \"reserved_name.ligo\", line 1, characters 4-8"}
|
||||
{"parser_loc":"in file \"reserved_name.ligo\", line 1, characters 4-13"}
|
||||
|
||||
|
||||
If you're not sure how to fix this error, you can
|
||||
|
@ -137,17 +137,17 @@ let parsify_string = fun (syntax : v_syntax) source_filename ->
|
||||
let pretty_print_pascaligo = fun source ->
|
||||
let%bind ast = Parser.Pascaligo.parse_file source in
|
||||
let buffer = Buffer.create 59 in
|
||||
let state = Parser.Pascaligo.ParserLog.mk_state
|
||||
let state = Parser_pascaligo.ParserLog.mk_state
|
||||
~offsets:true
|
||||
~mode:`Byte
|
||||
~buffer in
|
||||
Parser.Pascaligo.ParserLog.pp_ast state ast;
|
||||
Parser_pascaligo.ParserLog.pp_ast state ast;
|
||||
ok buffer
|
||||
|
||||
let pretty_print_cameligo = fun source ->
|
||||
let%bind ast = Parser.Cameligo.parse_file source in
|
||||
let buffer = Buffer.create 59 in
|
||||
let state = Parser.Cameligo.ParserLog.mk_state
|
||||
let state = Parser_cameligo.ParserLog.mk_state
|
||||
~offsets:true
|
||||
~mode:`Byte
|
||||
~buffer in
|
||||
|
@ -18,7 +18,7 @@ module Errors = struct
|
||||
] in
|
||||
error ~data title message
|
||||
|
||||
let parser_error source (start: Lexing.position) (end_: Lexing.position) lexbuf =
|
||||
let parser_error source (start: Lexing.position) (stop: Lexing.position) lexbuf =
|
||||
let title () = "parser error" in
|
||||
let file = if source = "" then
|
||||
""
|
||||
@ -29,18 +29,18 @@ module Errors = struct
|
||||
"Parse error at \"%s\" from (%d, %d) to (%d, %d). %s\n"
|
||||
(Lexing.lexeme lexbuf)
|
||||
start.pos_lnum (start.pos_cnum - start.pos_bol)
|
||||
end_.pos_lnum (end_.pos_cnum - end_.pos_bol)
|
||||
stop.pos_lnum (stop.pos_cnum - stop.pos_bol)
|
||||
file
|
||||
in
|
||||
let message () = str in
|
||||
let loc = if start.pos_cnum = -1 then
|
||||
Region.make
|
||||
~start: Pos.min
|
||||
~stop:(Pos.from_byte end_)
|
||||
~start:(Pos.min ~file:source)
|
||||
~stop:(Pos.from_byte stop)
|
||||
else
|
||||
Region.make
|
||||
~start:(Pos.from_byte start)
|
||||
~stop:(Pos.from_byte end_)
|
||||
~stop:(Pos.from_byte stop)
|
||||
in
|
||||
let data =
|
||||
[
|
||||
@ -51,7 +51,7 @@ module Errors = struct
|
||||
in
|
||||
error ~data title message
|
||||
|
||||
let unrecognized_error source (start: Lexing.position) (end_: Lexing.position) lexbuf =
|
||||
let unrecognized_error source (start: Lexing.position) (stop: Lexing.position) lexbuf =
|
||||
let title () = "unrecognized error" in
|
||||
let file = if source = "" then
|
||||
""
|
||||
@ -62,13 +62,13 @@ module Errors = struct
|
||||
"Parse error at \"%s\" from (%d, %d) to (%d, %d). %s\n"
|
||||
(Lexing.lexeme lexbuf)
|
||||
start.pos_lnum (start.pos_cnum - start.pos_bol)
|
||||
end_.pos_lnum (end_.pos_cnum - end_.pos_bol)
|
||||
stop.pos_lnum (stop.pos_cnum - stop.pos_bol)
|
||||
file
|
||||
in
|
||||
let message () = str in
|
||||
let loc = Region.make
|
||||
~start:(Pos.from_byte start)
|
||||
~stop:(Pos.from_byte end_)
|
||||
~stop:(Pos.from_byte stop)
|
||||
in
|
||||
let data = [
|
||||
("unrecognized_loc",
|
||||
@ -91,15 +91,15 @@ let parse (parser: 'a parser) source lexbuf =
|
||||
with
|
||||
| Parser.Error ->
|
||||
let start = Lexing.lexeme_start_p lexbuf in
|
||||
let end_ = Lexing.lexeme_end_p lexbuf in
|
||||
fail @@ (parser_error source start end_ lexbuf)
|
||||
let stop = Lexing.lexeme_end_p lexbuf in
|
||||
fail @@ (parser_error source start stop lexbuf)
|
||||
| Lexer.Error e ->
|
||||
fail @@ (lexer_error e)
|
||||
| _ ->
|
||||
let _ = Printexc.print_backtrace Pervasives.stdout in
|
||||
let start = Lexing.lexeme_start_p lexbuf in
|
||||
let end_ = Lexing.lexeme_end_p lexbuf in
|
||||
fail @@ (unrecognized_error source start end_ lexbuf)
|
||||
let stop = Lexing.lexeme_end_p lexbuf in
|
||||
fail @@ (unrecognized_error source start stop lexbuf)
|
||||
in
|
||||
close ();
|
||||
result
|
||||
@ -122,8 +122,8 @@ let parse_file (source: string) : AST.t result =
|
||||
|
||||
let parse_string (s:string) : AST.t result =
|
||||
let lexbuf = Lexing.from_string s in
|
||||
parse (Parser.contract) "" lexbuf
|
||||
parse Parser.contract "" lexbuf
|
||||
|
||||
let parse_expression (s:string) : AST.expr result =
|
||||
let lexbuf = Lexing.from_string s in
|
||||
parse (Parser.interactive_expr) "" lexbuf
|
||||
parse Parser.interactive_expr "" lexbuf
|
||||
|
@ -1,5 +1,4 @@
|
||||
$HOME/git/OCaml-build/Makefile
|
||||
$HOME/git/OCaml-build/Makefile.cfg
|
||||
$HOME/git/ligo/vendors/ligo-utils/simple-utils/pos.mli
|
||||
$HOME/git/ligo/vendors/ligo-utils/simple-utils/pos.ml
|
||||
$HOME/git/ligo/vendors/ligo-utils/simple-utils/region.mli
|
||||
|
@ -21,15 +21,6 @@ open Utils
|
||||
|
||||
type 'a reg = 'a Region.reg
|
||||
|
||||
let rec last to_region = function
|
||||
[] -> Region.ghost
|
||||
| [x] -> to_region x
|
||||
| _::t -> last to_region t
|
||||
|
||||
let nsepseq_to_region to_region (hd,tl) =
|
||||
let reg (_, item) = to_region item in
|
||||
Region.cover (to_region hd) (last reg tl)
|
||||
|
||||
(* Keywords of OCaml *)
|
||||
|
||||
type keyword = Region.t
|
||||
@ -321,6 +312,7 @@ and comp_expr =
|
||||
| Neq of neq bin_op reg
|
||||
|
||||
and record = field_assign reg ne_injection
|
||||
|
||||
and projection = {
|
||||
struct_name : variable;
|
||||
selector : dot;
|
||||
@ -344,6 +336,7 @@ and update = {
|
||||
updates : record reg;
|
||||
rbrace : rbrace;
|
||||
}
|
||||
|
||||
and path =
|
||||
Name of variable
|
||||
| Path of projection reg
|
||||
@ -387,7 +380,16 @@ and cond_expr = {
|
||||
ifnot : expr
|
||||
}
|
||||
|
||||
(* Projecting regions of the input source code *)
|
||||
(* Projecting regions from some nodes of the AST *)
|
||||
|
||||
let rec last to_region = function
|
||||
[] -> Region.ghost
|
||||
| [x] -> to_region x
|
||||
| _::t -> last to_region t
|
||||
|
||||
let nsepseq_to_region to_region (hd,tl) =
|
||||
let reg (_, item) = to_region item in
|
||||
Region.cover (to_region hd) (last reg tl)
|
||||
|
||||
let type_expr_to_region = function
|
||||
TProd {region; _}
|
||||
|
@ -85,7 +85,7 @@ type t =
|
||||
| Mutez of (string * Z.t) Region.reg
|
||||
| String of string Region.reg
|
||||
| Bytes of (string * Hex.t) Region.reg
|
||||
| Attr2 of string Region.reg
|
||||
| Attr of string Region.reg
|
||||
|
||||
(* Keywords *)
|
||||
|
||||
@ -150,8 +150,7 @@ val mk_kwd : lexeme -> Region.t -> (token, kwd_err) result
|
||||
val mk_string : lexeme -> Region.t -> token
|
||||
val mk_bytes : lexeme -> Region.t -> token
|
||||
val mk_constr : lexeme -> Region.t -> token
|
||||
val mk_attr : lexeme -> Region.t -> (token, attr_err) result
|
||||
val mk_attr2 : lexeme -> Region.t -> (token, attr_err) result
|
||||
val mk_attr : string -> lexeme -> Region.t -> (token, attr_err) result
|
||||
val eof : Region.t -> token
|
||||
|
||||
(* Predicates *)
|
||||
|
@ -69,7 +69,7 @@ type t =
|
||||
| Mutez of (string * Z.t) Region.reg
|
||||
| String of string Region.reg
|
||||
| Bytes of (string * Hex.t) Region.reg
|
||||
| Attr2 of string Region.reg
|
||||
| Attr of string Region.reg
|
||||
|
||||
(* Keywords *)
|
||||
|
||||
@ -147,6 +147,8 @@ let proj_token = function
|
||||
region,
|
||||
sprintf "Bytes (\"%s\", \"0x%s\")"
|
||||
s (Hex.show b)
|
||||
| Attr Region.{region; value} ->
|
||||
region, sprintf "Attr \"%s\"" value
|
||||
| Begin region -> region, "Begin"
|
||||
| Else region -> region, "Else"
|
||||
| End region -> region, "End"
|
||||
@ -166,7 +168,6 @@ let proj_token = function
|
||||
| With region -> region, "With"
|
||||
| C_None region -> region, "C_None"
|
||||
| C_Some region -> region, "C_Some"
|
||||
| Attr2 Region.{region; value} -> region, sprintf "Attr2 %s" value
|
||||
| EOF region -> region, "EOF"
|
||||
|
||||
let to_lexeme = function
|
||||
@ -205,6 +206,7 @@ let to_lexeme = function
|
||||
| Mutez i -> fst i.Region.value
|
||||
| String s -> String.escaped s.Region.value
|
||||
| Bytes b -> fst b.Region.value
|
||||
| Attr a -> a.Region.value
|
||||
|
||||
| Begin _ -> "begin"
|
||||
| Else _ -> "else"
|
||||
@ -226,7 +228,7 @@ let to_lexeme = function
|
||||
|
||||
| C_None _ -> "None"
|
||||
| C_Some _ -> "Some"
|
||||
| Attr2 a -> a.Region.value
|
||||
|
||||
| EOF _ -> ""
|
||||
|
||||
let to_string token ?(offsets=true) mode =
|
||||
@ -469,11 +471,10 @@ let mk_constr lexeme region =
|
||||
|
||||
(* Attributes *)
|
||||
|
||||
let mk_attr _lexeme _region =
|
||||
let mk_attr header lexeme region =
|
||||
if header = "[@" then
|
||||
Error Invalid_attribute
|
||||
|
||||
let mk_attr2 lexeme region =
|
||||
Ok (Attr2 { value = lexeme; region })
|
||||
else Ok (Attr Region.{value=lexeme; region})
|
||||
|
||||
(* Predicates *)
|
||||
|
||||
|
5
src/passes/1-parser/cameligo/Makefile.cfg
Normal file
5
src/passes/1-parser/cameligo/Makefile.cfg
Normal file
@ -0,0 +1,5 @@
|
||||
SHELL := dash
|
||||
BFLAGS := -strict-sequence -w +A-48-4 -g
|
||||
|
||||
clean::
|
||||
> rm -f Parser.msg Parser.msg.map Parser.msg.states Version.ml
|
@ -46,7 +46,7 @@ let message =
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 33 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 460 ->
|
||||
| 478 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 27 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
@ -68,9 +68,13 @@ let message =
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 133 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 373 ->
|
||||
| 379 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 375 ->
|
||||
| 381 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 472 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 169 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 134 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
@ -80,7 +84,7 @@ let message =
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 153 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 374 ->
|
||||
| 380 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 63 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
@ -144,137 +148,141 @@ let message =
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 156 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 463 ->
|
||||
| 481 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 465 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 217 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 242 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 219 ->
|
||||
| 483 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 221 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 215 ->
|
||||
| 246 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 226 ->
|
||||
| 223 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 255 ->
|
||||
| 225 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 256 ->
|
||||
| 219 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 243 ->
|
||||
| 230 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 264 ->
|
||||
| 259 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 228 ->
|
||||
| 260 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 257 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 258 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 266 ->
|
||||
| 247 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 268 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 232 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 261 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 262 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 270 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 272 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 274 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 192 ->
|
||||
| 276 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 259 ->
|
||||
| 278 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 285 ->
|
||||
| 195 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 288 ->
|
||||
| 263 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 245 ->
|
||||
| 289 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 293 ->
|
||||
| 292 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 262 ->
|
||||
| 249 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 297 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 266 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 160 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 164 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 429 ->
|
||||
| 445 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 332 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 313 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 431 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 315 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 316 ->
|
||||
| 337 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 317 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 432 ->
|
||||
| 447 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 445 ->
|
||||
| 319 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 446 ->
|
||||
| 320 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 433 ->
|
||||
| 321 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 434 ->
|
||||
| 448 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 435 ->
|
||||
| 462 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 436 ->
|
||||
| 463 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 437 ->
|
||||
| 449 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 438 ->
|
||||
| 450 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 440 ->
|
||||
| 452 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 328 ->
|
||||
| 451 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 330 ->
|
||||
| 453 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 454 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 455 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 457 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 333 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 335 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 339 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 336 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 334 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 331 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 329 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 340 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 341 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 342 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 343 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 344 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 345 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 367 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 346 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 348 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 441 ->
|
||||
| 347 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 443 ->
|
||||
| 349 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 447 ->
|
||||
| 350 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 430 ->
|
||||
| 351 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 312 ->
|
||||
| 373 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 428 ->
|
||||
| 352 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 354 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 458 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 460 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 464 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 446 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 316 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 444 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 165 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
@ -282,65 +290,71 @@ let message =
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 168 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 169 ->
|
||||
| 172 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 171 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 163 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 448 ->
|
||||
| 465 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 450 ->
|
||||
| 467 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 451 ->
|
||||
| 468 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 166 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 235 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 236 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 239 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 240 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 425 ->
|
||||
| 243 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 170 ->
|
||||
| 244 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 171 ->
|
||||
| 441 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 172 ->
|
||||
| 173 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 418 ->
|
||||
| 428 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 419 ->
|
||||
| 429 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 174 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 175 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 434 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 435 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 438 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 439 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 427 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 421 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 422 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 423 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 174 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 304 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 305 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 405 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 412 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 404 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 306 ->
|
||||
| 177 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 308 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 320 ->
|
||||
| 309 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 321 ->
|
||||
| 412 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 322 ->
|
||||
| 419 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 323 ->
|
||||
| 411 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 310 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 312 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 324 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
@ -350,67 +364,79 @@ let message =
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 327 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 378 ->
|
||||
| 329 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 379 ->
|
||||
| 328 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 381 ->
|
||||
| 330 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 335 ->
|
||||
| 331 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 310 ->
|
||||
| 332 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 307 ->
|
||||
| 384 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 395 ->
|
||||
| 385 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 396 ->
|
||||
| 387 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 397 ->
|
||||
| 340 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 398 ->
|
||||
| 314 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 399 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 400 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 408 ->
|
||||
| 311 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 401 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 402 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 404 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 403 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 175 ->
|
||||
| 405 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 176 ->
|
||||
| 406 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 407 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 415 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 408 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 410 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 178 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 179 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 180 ->
|
||||
| 182 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 183 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 302 ->
|
||||
| 186 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 300 ->
|
||||
| 306 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 185 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 187 ->
|
||||
| 304 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 188 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 189 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 190 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 195 ->
|
||||
| 191 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 192 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 193 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 198 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 218 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 197 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 214 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 194 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 210 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| _ ->
|
||||
raise Not_found
|
||||
|
@ -12,7 +12,7 @@
|
||||
%token <(string * Z.t) Region.reg> Mutez "<mutez>"
|
||||
%token <string Region.reg> Ident "<ident>"
|
||||
%token <string Region.reg> Constr "<constr>"
|
||||
%token <string Region.reg> Attr2 "<attr>"
|
||||
%token <string Region.reg> Attr "<attr>"
|
||||
|
||||
(* Symbols *)
|
||||
|
||||
|
@ -119,6 +119,7 @@ declaration:
|
||||
|
||||
type_decl:
|
||||
"type" type_name "=" type_expr {
|
||||
Scoping.check_reserved_name $2;
|
||||
let region = cover $1 (type_expr_to_region $4) in
|
||||
let value = {
|
||||
kwd_type = $1;
|
||||
@ -128,23 +129,23 @@ type_decl:
|
||||
in {region; value} }
|
||||
|
||||
type_expr:
|
||||
cartesian | sum_type | record_type { $1 }
|
||||
|
||||
cartesian:
|
||||
fun_type { $1 }
|
||||
| fun_type "*" nsepseq(fun_type,"*") {
|
||||
let value = Utils.nsepseq_cons $1 $2 $3 in
|
||||
let region = nsepseq_to_region type_expr_to_region value
|
||||
in TProd {region; value} }
|
||||
fun_type | sum_type | record_type { $1 }
|
||||
|
||||
fun_type:
|
||||
core_type { $1 }
|
||||
| core_type "->" fun_type {
|
||||
cartesian { $1 }
|
||||
| cartesian "->" fun_type {
|
||||
let start = type_expr_to_region $1
|
||||
and stop = type_expr_to_region $3 in
|
||||
let region = cover start stop in
|
||||
TFun {region; value=$1,$2,$3} }
|
||||
|
||||
cartesian:
|
||||
core_type { $1 }
|
||||
| core_type "*" nsepseq(core_type,"*") {
|
||||
let value = Utils.nsepseq_cons $1 $2 $3 in
|
||||
let region = nsepseq_to_region type_expr_to_region value
|
||||
in TProd {region; value} }
|
||||
|
||||
core_type:
|
||||
type_name { TVar $1 }
|
||||
| par(type_expr) { TPar $1 }
|
||||
@ -175,6 +176,7 @@ type_tuple:
|
||||
|
||||
sum_type:
|
||||
ioption("|") nsepseq(variant,"|") {
|
||||
Scoping.check_variants (Utils.nsepseq_to_list $2);
|
||||
let region = nsepseq_to_region (fun x -> x.region) $2
|
||||
in TSum {region; value=$2} }
|
||||
|
||||
@ -188,6 +190,8 @@ variant:
|
||||
record_type:
|
||||
"{" sep_or_term_list(field_decl,";") "}" {
|
||||
let ne_elements, terminator = $2 in
|
||||
let () = Utils.nsepseq_to_list ne_elements
|
||||
|> Scoping.check_fields in
|
||||
let region = cover $1 $3
|
||||
and value = {compound = Braces ($1,$3); ne_elements; terminator}
|
||||
in TRecord {region; value} }
|
||||
@ -202,7 +206,7 @@ field_decl:
|
||||
(* Top-level non-recursive definitions *)
|
||||
|
||||
let_declaration:
|
||||
"let" let_binding seq(Attr2) {
|
||||
"let" let_binding seq(Attr) {
|
||||
let kwd_let = $1 in
|
||||
let attributes = $3 in
|
||||
let binding = $2 in
|
||||
@ -214,9 +218,11 @@ let_declaration:
|
||||
let_binding:
|
||||
"<ident>" nseq(sub_irrefutable) type_annotation? "=" expr {
|
||||
let binders = Utils.nseq_cons (PVar $1) $2 in
|
||||
Utils.nseq_iter Scoping.check_pattern binders;
|
||||
{binders; lhs_type=$3; eq=$4; let_rhs=$5}
|
||||
}
|
||||
| irrefutable type_annotation? "=" expr {
|
||||
Scoping.check_pattern $1;
|
||||
{binders=$1,[]; lhs_type=$2; eq=$3; let_rhs=$4} }
|
||||
|
||||
type_annotation:
|
||||
@ -441,10 +447,12 @@ cases(right_expr):
|
||||
in fst_case, ($2,snd_case)::others }
|
||||
|
||||
case_clause(right_expr):
|
||||
pattern "->" right_expr { {pattern=$1; arrow=$2; rhs=$3} }
|
||||
pattern "->" right_expr {
|
||||
Scoping.check_pattern $1;
|
||||
{pattern=$1; arrow=$2; rhs=$3} }
|
||||
|
||||
let_expr(right_expr):
|
||||
"let" let_binding seq(Attr2) "in" right_expr {
|
||||
"let" let_binding seq(Attr) "in" right_expr {
|
||||
let kwd_let = $1
|
||||
and binding = $2
|
||||
and attributes = $3
|
||||
|
@ -131,7 +131,7 @@ let rec print_tokens state {decl;eof} =
|
||||
and print_attributes state attributes =
|
||||
List.iter (
|
||||
fun ({value = attribute; region}) ->
|
||||
let attribute_formatted = sprintf "[@%s]" attribute in
|
||||
let attribute_formatted = sprintf "[@@%s]" attribute in
|
||||
print_token state region attribute_formatted
|
||||
) attributes
|
||||
|
||||
@ -610,31 +610,41 @@ let rec pp_ast state {decl; _} =
|
||||
List.iteri (List.length decls |> apply) decls
|
||||
|
||||
and pp_declaration state = function
|
||||
Let {value = (_, let_binding, _); region} ->
|
||||
Let {value = (_, let_binding, attr); region} ->
|
||||
pp_loc_node state "Let" region;
|
||||
pp_let_binding state let_binding
|
||||
pp_let_binding state let_binding attr;
|
||||
| TypeDecl {value; region} ->
|
||||
pp_loc_node state "TypeDecl" region;
|
||||
pp_type_decl state value
|
||||
|
||||
and pp_let_binding state node =
|
||||
and pp_let_binding state node attr =
|
||||
let {binders; lhs_type; let_rhs; _} = node in
|
||||
let fields = if lhs_type = None then 2 else 3 in
|
||||
let () =
|
||||
let fields = if attr = [] then fields else fields+1 in
|
||||
let arity =
|
||||
let state = state#pad fields 0 in
|
||||
pp_node state "<binders>";
|
||||
pp_binders state binders in
|
||||
let () =
|
||||
pp_binders state binders; 0 in
|
||||
let arity =
|
||||
match lhs_type with
|
||||
None -> ()
|
||||
None -> arity
|
||||
| Some (_, type_expr) ->
|
||||
let state = state#pad fields 1 in
|
||||
let state = state#pad fields (arity+1) in
|
||||
pp_node state "<lhs type>";
|
||||
pp_type_expr (state#pad 1 0) type_expr in
|
||||
let () =
|
||||
let state = state#pad fields (fields - 1) in
|
||||
pp_type_expr (state#pad 1 0) type_expr;
|
||||
arity+1 in
|
||||
let arity =
|
||||
let state = state#pad fields (arity+1) in
|
||||
pp_node state "<rhs>";
|
||||
pp_expr (state#pad 1 0) let_rhs
|
||||
pp_expr (state#pad 1 0) let_rhs;
|
||||
arity+1 in
|
||||
let () =
|
||||
if attr <> [] then
|
||||
let state = state#pad fields (arity+1) in
|
||||
pp_node state "<attributes>";
|
||||
let length = List.length attr in
|
||||
let apply len rank = pp_ident (state#pad len rank)
|
||||
in List.iteri (apply length) attr
|
||||
in ()
|
||||
|
||||
and pp_type_decl state decl =
|
||||
@ -838,28 +848,39 @@ and pp_fun_expr state node =
|
||||
in ()
|
||||
|
||||
and pp_let_in state node =
|
||||
let {binding; body; _} = node in
|
||||
let {binding; body; attributes; _} = node in
|
||||
let {binders; lhs_type; let_rhs; _} = binding in
|
||||
let fields = if lhs_type = None then 3 else 4 in
|
||||
let () =
|
||||
let fields = if attributes = [] then fields else fields+1 in
|
||||
let arity =
|
||||
let state = state#pad fields 0 in
|
||||
pp_node state "<binders>";
|
||||
pp_binders state binders in
|
||||
let () =
|
||||
pp_binders state binders; 0 in
|
||||
let arity =
|
||||
match lhs_type with
|
||||
None -> ()
|
||||
None -> arity
|
||||
| Some (_, type_expr) ->
|
||||
let state = state#pad fields 1 in
|
||||
let state = state#pad fields (arity+1) in
|
||||
pp_node state "<lhs type>";
|
||||
pp_type_expr (state#pad 1 0) type_expr in
|
||||
let () =
|
||||
let state = state#pad fields (fields - 2) in
|
||||
pp_type_expr (state#pad 1 0) type_expr;
|
||||
arity+1 in
|
||||
let arity =
|
||||
let state = state#pad fields (arity+1) in
|
||||
pp_node state "<rhs>";
|
||||
pp_expr (state#pad 1 0) let_rhs in
|
||||
let () =
|
||||
let state = state#pad fields (fields - 1) in
|
||||
pp_expr (state#pad 1 0) let_rhs;
|
||||
arity+1 in
|
||||
let arity =
|
||||
let state = state#pad fields (arity+1) in
|
||||
pp_node state "<body>";
|
||||
pp_expr (state#pad 1 0) body
|
||||
pp_expr (state#pad 1 0) body;
|
||||
arity+1 in
|
||||
let () =
|
||||
if attributes <> [] then
|
||||
let state = state#pad fields (arity+1) in
|
||||
pp_node state "<attributes>";
|
||||
let length = List.length attributes in
|
||||
let apply len rank = pp_ident (state#pad len rank)
|
||||
in List.iteri (apply length) attributes
|
||||
in ()
|
||||
|
||||
and pp_tuple_expr state {value; _} =
|
||||
|
@ -25,6 +25,7 @@ val pattern_to_string :
|
||||
val expr_to_string :
|
||||
offsets:bool -> mode:[`Point|`Byte] -> AST.expr -> string
|
||||
|
||||
(** {1 Pretty-printing of the AST} *)
|
||||
(** {1 Pretty-printing of AST nodes} *)
|
||||
|
||||
val pp_ast : state -> AST.t -> unit
|
||||
val pp_expr : state -> AST.expr -> unit
|
||||
|
@ -6,22 +6,86 @@ module IO =
|
||||
let options = EvalOpt.read "CameLIGO" ext
|
||||
end
|
||||
|
||||
module ExtParser =
|
||||
module Parser =
|
||||
struct
|
||||
type ast = AST.t
|
||||
type expr = AST.expr
|
||||
include Parser
|
||||
end
|
||||
|
||||
module ExtParserLog =
|
||||
module ParserLog =
|
||||
struct
|
||||
type ast = AST.t
|
||||
type expr = AST.expr
|
||||
include ParserLog
|
||||
end
|
||||
|
||||
module MyLexer = Lexer.Make (LexToken)
|
||||
module Lexer = Lexer.Make (LexToken)
|
||||
|
||||
module Unit =
|
||||
ParserUnit.Make (IO)(MyLexer)(AST)(ExtParser)(ParErr)(ExtParserLog)
|
||||
ParserUnit.Make (Lexer)(AST)(Parser)(ParErr)(ParserLog)(IO)
|
||||
|
||||
let () = Unit.run ()
|
||||
(* Main *)
|
||||
|
||||
let issue_error point =
|
||||
let error = Unit.format_error ~offsets:IO.options#offsets
|
||||
IO.options#mode point
|
||||
in Stdlib.Error error
|
||||
|
||||
let parse parser : ('a,string) Stdlib.result =
|
||||
try parser () with
|
||||
(* Scoping errors *)
|
||||
|
||||
| Scoping.Error (Scoping.Reserved_name name) ->
|
||||
let token =
|
||||
Lexer.Token.mk_ident name.Region.value name.Region.region in
|
||||
(match token with
|
||||
(* Cannot fail because [name] is a not a
|
||||
reserved name for the lexer. *)
|
||||
Stdlib.Error _ -> assert false
|
||||
| Ok invalid ->
|
||||
issue_error
|
||||
("Reserved name.\nHint: Change the name.\n", None, invalid))
|
||||
|
||||
| Scoping.Error (Scoping.Duplicate_variant name) ->
|
||||
let token =
|
||||
Lexer.Token.mk_constr name.Region.value name.Region.region in
|
||||
let point = "Duplicate constructor in this sum type declaration.\n\
|
||||
Hint: Change the constructor.\n",
|
||||
None, token
|
||||
in issue_error point
|
||||
|
||||
| Scoping.Error (Scoping.Non_linear_pattern var) ->
|
||||
let token =
|
||||
Lexer.Token.mk_ident var.Region.value var.Region.region in
|
||||
(match token with
|
||||
(* Cannot fail because [var] is a not a
|
||||
reserved name for the lexer. *)
|
||||
Stdlib.Error _ -> assert false
|
||||
| Ok invalid ->
|
||||
let point = "Repeated variable in this pattern.\n\
|
||||
Hint: Change the name.\n",
|
||||
None, invalid
|
||||
in issue_error point)
|
||||
|
||||
| Scoping.Error (Scoping.Duplicate_field name) ->
|
||||
let token =
|
||||
Lexer.Token.mk_ident name.Region.value name.Region.region in
|
||||
(match token with
|
||||
(* Cannot fail because [name] is a not a
|
||||
reserved name for the lexer. *)
|
||||
Stdlib.Error _ -> assert false
|
||||
| Ok invalid ->
|
||||
let point = "Duplicate field name in this record declaration.\n\
|
||||
Hint: Change the name.\n",
|
||||
None, invalid
|
||||
in issue_error point)
|
||||
|
||||
let () =
|
||||
if IO.options#expr
|
||||
then match parse (fun () -> Unit.parse Unit.parse_expr) with
|
||||
Stdlib.Ok _ -> ()
|
||||
| Error msg -> Printf.eprintf "\027[31m%s\027[0m%!" msg
|
||||
else match parse (fun () -> Unit.parse Unit.parse_contract) with
|
||||
Stdlib.Ok _ -> ()
|
||||
| Error msg -> Printf.eprintf "\027[31m%s\027[0m%!" msg
|
||||
|
132
src/passes/1-parser/cameligo/Scoping.ml
Normal file
132
src/passes/1-parser/cameligo/Scoping.ml
Normal file
@ -0,0 +1,132 @@
|
||||
[@@@warning "-42"]
|
||||
|
||||
|
||||
type t =
|
||||
Reserved_name of AST.variable
|
||||
| Duplicate_variant of AST.variable
|
||||
| Non_linear_pattern of AST.variable
|
||||
| Duplicate_field of AST.variable
|
||||
|
||||
type error = t
|
||||
|
||||
exception Error of t
|
||||
|
||||
open Region
|
||||
|
||||
(* Useful modules *)
|
||||
|
||||
module SSet = Utils.String.Set
|
||||
|
||||
module Ord =
|
||||
struct
|
||||
type t = AST.variable
|
||||
let compare v1 v2 =
|
||||
compare v1.value v2.value
|
||||
end
|
||||
|
||||
module VarSet = Set.Make (Ord)
|
||||
|
||||
(* Checking the definition of reserved names (shadowing) *)
|
||||
|
||||
let reserved =
|
||||
let open SSet in
|
||||
empty
|
||||
|> add "assert"
|
||||
|> add "balance"
|
||||
|> add "time"
|
||||
|> add "amount"
|
||||
|> add "gas"
|
||||
|> add "sender"
|
||||
|> add "source"
|
||||
|> add "failwith"
|
||||
|> add "continue"
|
||||
|> add "stop"
|
||||
|> add "int"
|
||||
|> add "abs"
|
||||
|> add "unit"
|
||||
|
||||
let check_reserved_names vars =
|
||||
let is_reserved elt = SSet.mem elt.value reserved in
|
||||
let inter = VarSet.filter is_reserved vars in
|
||||
if not (VarSet.is_empty inter) then
|
||||
let clash = VarSet.choose inter in
|
||||
raise (Error (Reserved_name clash))
|
||||
else vars
|
||||
|
||||
let check_reserved_name var =
|
||||
if SSet.mem var.value reserved then
|
||||
raise (Error (Reserved_name var))
|
||||
|
||||
(* Checking the linearity of patterns *)
|
||||
|
||||
open! AST
|
||||
|
||||
let rec vars_of_pattern env = function
|
||||
PConstr p -> vars_of_pconstr env p
|
||||
| PUnit _ | PFalse _ | PTrue _
|
||||
| PInt _ | PNat _ | PBytes _
|
||||
| PString _ | PWild _ -> env
|
||||
| PVar var ->
|
||||
if VarSet.mem var env then
|
||||
raise (Error (Non_linear_pattern var))
|
||||
else VarSet.add var env
|
||||
| PList l -> vars_of_plist env l
|
||||
| PTuple t -> Utils.nsepseq_foldl vars_of_pattern env t.value
|
||||
| PPar p -> vars_of_pattern env p.value.inside
|
||||
| PRecord p -> vars_of_fields env p.value.ne_elements
|
||||
| PTyped p -> vars_of_pattern env p.value.pattern
|
||||
|
||||
and vars_of_fields env fields =
|
||||
Utils.nsepseq_foldl vars_of_field_pattern env fields
|
||||
|
||||
and vars_of_field_pattern env field =
|
||||
let var = field.value.field_name in
|
||||
if VarSet.mem var env then
|
||||
raise (Error (Non_linear_pattern var))
|
||||
else
|
||||
let p = field.value.pattern
|
||||
in vars_of_pattern (VarSet.add var env) p
|
||||
|
||||
and vars_of_pconstr env = function
|
||||
PNone _ -> env
|
||||
| PSomeApp {value=_, pattern; _} ->
|
||||
vars_of_pattern env pattern
|
||||
| PConstrApp {value=_, Some pattern; _} ->
|
||||
vars_of_pattern env pattern
|
||||
| PConstrApp {value=_,None; _} -> env
|
||||
|
||||
and vars_of_plist env = function
|
||||
PListComp {value; _} ->
|
||||
Utils.sepseq_foldl vars_of_pattern env value.elements
|
||||
| PCons {value; _} ->
|
||||
let head, _, tail = value in
|
||||
List.fold_left vars_of_pattern env [head; tail]
|
||||
|
||||
let check_linearity = vars_of_pattern VarSet.empty
|
||||
|
||||
(* Checking patterns *)
|
||||
|
||||
let check_pattern p =
|
||||
check_linearity p |> check_reserved_names |> ignore
|
||||
|
||||
(* Checking variants for duplicates *)
|
||||
|
||||
let check_variants variants =
|
||||
let add acc {value; _} =
|
||||
if VarSet.mem value.constr acc then
|
||||
raise (Error (Duplicate_variant value.constr))
|
||||
else VarSet.add value.constr acc in
|
||||
let variants =
|
||||
List.fold_left add VarSet.empty variants
|
||||
in ignore variants
|
||||
|
||||
(* Checking record fields *)
|
||||
|
||||
let check_fields fields =
|
||||
let add acc {value; _} =
|
||||
if VarSet.mem (value: field_decl).field_name acc then
|
||||
raise (Error (Duplicate_field value.field_name))
|
||||
else VarSet.add value.field_name acc in
|
||||
let fields =
|
||||
List.fold_left add VarSet.empty fields
|
||||
in ignore fields
|
16
src/passes/1-parser/cameligo/Scoping.mli
Normal file
16
src/passes/1-parser/cameligo/Scoping.mli
Normal file
@ -0,0 +1,16 @@
|
||||
(* This module exports checks on scoping, called from the parser. *)
|
||||
|
||||
type t =
|
||||
Reserved_name of AST.variable
|
||||
| Duplicate_variant of AST.variable
|
||||
| Non_linear_pattern of AST.variable
|
||||
| Duplicate_field of AST.variable
|
||||
|
||||
type error = t
|
||||
|
||||
exception Error of t
|
||||
|
||||
val check_reserved_name : AST.variable -> unit
|
||||
val check_pattern : AST.pattern -> unit
|
||||
val check_variants : AST.variant Region.reg list -> unit
|
||||
val check_fields : AST.field_decl Region.reg list -> unit
|
@ -1,10 +0,0 @@
|
||||
#!/bin/sh
|
||||
|
||||
set -e
|
||||
|
||||
if test -d ../../.git; then
|
||||
echo true > dot_git_is_dir
|
||||
else
|
||||
echo false > dot_git_is_dir
|
||||
cat .git >> dot_git_is_dir
|
||||
fi
|
@ -1,14 +1,21 @@
|
||||
;; Build of the lexer
|
||||
|
||||
(ocamllex LexToken)
|
||||
|
||||
;; Build of the parser
|
||||
|
||||
(menhir
|
||||
(merge_into Parser)
|
||||
(modules ParToken Parser)
|
||||
(flags -la 1 --table --strict --explain --external-tokens LexToken))
|
||||
|
||||
;; Build of the parser as a library
|
||||
|
||||
(library
|
||||
(name parser_cameligo)
|
||||
(public_name ligo.parser.cameligo)
|
||||
(modules AST cameligo Parser ParserLog LexToken)
|
||||
(modules
|
||||
Scoping AST cameligo Parser ParserLog LexToken)
|
||||
(libraries
|
||||
menhirLib
|
||||
parser_shared
|
||||
@ -20,6 +27,18 @@
|
||||
(pps bisect_ppx --conditional))
|
||||
(flags (:standard -open Simple_utils -open Parser_shared)))
|
||||
|
||||
;; Build of the unlexer (for covering the
|
||||
;; error states of the LR automaton)
|
||||
|
||||
(executable
|
||||
(name Unlexer)
|
||||
(libraries str)
|
||||
(preprocess
|
||||
(pps bisect_ppx --conditional))
|
||||
(modules Unlexer))
|
||||
|
||||
;; Local build of a standalone lexer
|
||||
|
||||
(executable
|
||||
(name LexerMain)
|
||||
(libraries parser_cameligo)
|
||||
@ -28,6 +47,8 @@
|
||||
(pps bisect_ppx --conditional))
|
||||
(flags (:standard -open Parser_shared -open Parser_cameligo)))
|
||||
|
||||
;; Local build of a standalone parser
|
||||
|
||||
(executable
|
||||
(name ParserMain)
|
||||
(libraries parser_cameligo)
|
||||
@ -37,19 +58,16 @@
|
||||
(pps bisect_ppx --conditional))
|
||||
(flags (:standard -open Simple_utils -open Parser_shared -open Parser_cameligo)))
|
||||
|
||||
(executable
|
||||
(name Unlexer)
|
||||
(libraries str)
|
||||
(preprocess
|
||||
(pps bisect_ppx --conditional))
|
||||
(modules Unlexer))
|
||||
;; Build of the covering of error states in the LR automaton
|
||||
|
||||
(rule
|
||||
(targets Parser.msg)
|
||||
(deps (:script_messages ../../../../vendors/ligo-utils/simple-utils/messages.sh) Parser.mly LexToken.mli ParToken.mly)
|
||||
(action (run %{script_messages} --lex-tokens=LexToken.mli --par-tokens=ParToken.mly Parser.mly)))
|
||||
|
||||
;; Build of all the LIGO source file that cover all error states
|
||||
|
||||
(rule
|
||||
(targets all.ligo)
|
||||
(targets all.mligo)
|
||||
(deps (:script_cover ../../../../vendors/ligo-utils/simple-utils/cover.sh) Parser.mly LexToken.mli ParToken.mly Parser.msg Unlexer.exe)
|
||||
(action (run %{script_cover} --lex-tokens=LexToken.mli --par-tokens=ParToken.mly --ext=ligo --unlexer=./Unlexer.exe --messages=Parser.msg --dir=. --concatenate Parser.mly )))
|
||||
(action (run %{script_cover} --lex-tokens=LexToken.mli --par-tokens=ParToken.mly --ext=mligo --unlexer=./Unlexer.exe --messages=Parser.msg --dir=. --concatenate Parser.mly )))
|
@ -7,10 +7,7 @@
|
||||
parser_shared
|
||||
parser_pascaligo
|
||||
parser_cameligo
|
||||
parser_reasonligo
|
||||
)
|
||||
parser_reasonligo)
|
||||
(preprocess
|
||||
(pps ppx_let bisect_ppx --conditional)
|
||||
)
|
||||
(flags (:standard -w +1..62-4-9-44-40-42-48-30@39@33 -open Simple_utils -open Parser_shared))
|
||||
)
|
||||
(pps ppx_let bisect_ppx --conditional))
|
||||
(flags (:standard -w +1..62-4-9-44-40-42-48-30@39@33 -open Simple_utils -open Parser_shared)))
|
||||
|
@ -1,129 +1,103 @@
|
||||
open Trace
|
||||
|
||||
module Parser = Parser_pascaligo.Parser
|
||||
module AST = Parser_pascaligo.AST
|
||||
module ParserLog = Parser_pascaligo.ParserLog
|
||||
module LexToken = Parser_pascaligo.LexToken
|
||||
module Lexer = Lexer.Make(LexToken)
|
||||
module SyntaxError = Parser_pascaligo.SyntaxError
|
||||
module Scoping = Parser_pascaligo.Scoping
|
||||
module Parser = Parser_pascaligo.Parser
|
||||
|
||||
module Errors = struct
|
||||
module Errors =
|
||||
struct
|
||||
let reserved_name Region.{value; region} =
|
||||
let title () = Printf.sprintf "reserved name \"%s\"" value in
|
||||
let message () = "" in
|
||||
let data = [
|
||||
("location",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ region)]
|
||||
in error ~data title message
|
||||
|
||||
let non_linear_pattern Region.{value; region} =
|
||||
let title () =
|
||||
Printf.sprintf "repeated variable \"%s\" in this pattern" value in
|
||||
let message () = "" in
|
||||
let data = [
|
||||
("location",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ region)]
|
||||
in error ~data title message
|
||||
|
||||
let duplicate_parameter Region.{value; region} =
|
||||
let title () =
|
||||
Printf.sprintf "duplicate parameter \"%s\"" value in
|
||||
let message () = "" in
|
||||
let data = [
|
||||
("location",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ region)]
|
||||
in error ~data title message
|
||||
|
||||
let duplicate_variant Region.{value; region} =
|
||||
let title () =
|
||||
Printf.sprintf "duplicate variant \"%s\" in this\
|
||||
type declaration" value in
|
||||
let message () = "" in
|
||||
let data = [
|
||||
("location",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ region)]
|
||||
in error ~data title message
|
||||
|
||||
let unrecognized_error source (start: Lexing.position)
|
||||
(stop: Lexing.position) lexbuf =
|
||||
let title () = "unrecognized error" in
|
||||
let file =
|
||||
if source = "" then ""
|
||||
else
|
||||
Format.sprintf "In file \"%s|%s\"" start.pos_fname source in
|
||||
let message () =
|
||||
Format.sprintf
|
||||
"Parse error at \"%s\" from (%d, %d) to (%d, %d). %s\n"
|
||||
(Lexing.lexeme lexbuf)
|
||||
start.pos_lnum (start.pos_cnum - start.pos_bol)
|
||||
stop.pos_lnum (stop.pos_cnum - stop.pos_bol)
|
||||
file in
|
||||
let loc = Region.make ~start:(Pos.from_byte start)
|
||||
~stop:(Pos.from_byte stop) in
|
||||
let data = [
|
||||
("unrecognized_loc",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ loc)]
|
||||
in error ~data title message
|
||||
|
||||
let parser_error source (start: Lexing.position)
|
||||
(stop: Lexing.position) lexbuf =
|
||||
let title () = "parser error" in
|
||||
let file =
|
||||
if source = "" then ""
|
||||
else
|
||||
Format.sprintf "In file \"%s|%s\"" start.pos_fname source in
|
||||
let message () =
|
||||
Format.sprintf
|
||||
"Parse error at \"%s\" from (%d, %d) to (%d, %d). %s\n"
|
||||
(Lexing.lexeme lexbuf)
|
||||
start.pos_lnum (start.pos_cnum - start.pos_bol)
|
||||
stop.pos_lnum (stop.pos_cnum - stop.pos_bol)
|
||||
file in
|
||||
let loc =
|
||||
if start.pos_cnum = -1 then
|
||||
Region.make
|
||||
~start:(Pos.min ~file:source) ~stop:(Pos.from_byte stop)
|
||||
else
|
||||
Region.make ~start:(Pos.from_byte start)
|
||||
~stop:(Pos.from_byte stop) in
|
||||
let data =
|
||||
[("parser_loc",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ loc)] in
|
||||
error ~data title message
|
||||
|
||||
let lexer_error (e: Lexer.error AST.reg) =
|
||||
let title () = "lexer error" in
|
||||
let message () = Lexer.error_to_string e.value in
|
||||
let data = [
|
||||
("parser_loc",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ e.region
|
||||
)
|
||||
] in
|
||||
error ~data title message
|
||||
|
||||
let reserved_name Region.{value; region} =
|
||||
let title () = Printf.sprintf "reserved name \"%s\"" value in
|
||||
let message () = "" in
|
||||
let data = [
|
||||
("location",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ region)
|
||||
] in
|
||||
error ~data title message
|
||||
|
||||
let non_linear_pattern Region.{value; region} =
|
||||
let title () = Printf.sprintf "repeated variable \"%s\" in this pattern" value in
|
||||
let message () = "" in
|
||||
let data = [
|
||||
("location",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ region)
|
||||
] in
|
||||
error ~data title message
|
||||
|
||||
let duplicate_parameter Region.{value; region} =
|
||||
let title () = Printf.sprintf "duplicate parameter \"%s\"" value in
|
||||
let message () = "" in
|
||||
let data = [
|
||||
("location",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ region)
|
||||
] in
|
||||
error ~data title message
|
||||
|
||||
let duplicate_variant Region.{value; region} =
|
||||
let title () = Printf.sprintf "duplicate variant \"%s\" in this\
|
||||
type declaration" value in
|
||||
let message () = "" in
|
||||
let data = [
|
||||
("location",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ region)
|
||||
] in
|
||||
error ~data title message
|
||||
|
||||
let detached_attributes (attrs: AST.attributes) =
|
||||
let title () = "detached attributes" in
|
||||
let message () = "" in
|
||||
let data = [
|
||||
("location",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ attrs.region)
|
||||
] in
|
||||
error ~data title message
|
||||
|
||||
let parser_error source (start: Lexing.position) (end_: Lexing.position) lexbuf =
|
||||
let title () = "parser error" in
|
||||
let file = if source = "" then
|
||||
""
|
||||
else
|
||||
Format.sprintf "In file \"%s|%s\"" start.pos_fname source
|
||||
in
|
||||
let str = Format.sprintf
|
||||
"Parse error at \"%s\" from (%d, %d) to (%d, %d). %s\n"
|
||||
(Lexing.lexeme lexbuf)
|
||||
start.pos_lnum (start.pos_cnum - start.pos_bol)
|
||||
end_.pos_lnum (end_.pos_cnum - end_.pos_bol)
|
||||
file
|
||||
in
|
||||
let message () = str in
|
||||
let loc = if start.pos_cnum = -1 then
|
||||
Region.make
|
||||
~start: Pos.min
|
||||
~stop:(Pos.from_byte end_)
|
||||
else
|
||||
Region.make
|
||||
~start:(Pos.from_byte start)
|
||||
~stop:(Pos.from_byte end_)
|
||||
in
|
||||
let data =
|
||||
[
|
||||
("parser_loc",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ loc
|
||||
)
|
||||
]
|
||||
in
|
||||
error ~data title message
|
||||
|
||||
let unrecognized_error source (start: Lexing.position) (end_: Lexing.position) lexbuf =
|
||||
let title () = "unrecognized error" in
|
||||
let file = if source = "" then
|
||||
""
|
||||
else
|
||||
Format.sprintf "In file \"%s|%s\"" start.pos_fname source
|
||||
in
|
||||
let str = Format.sprintf
|
||||
"Parse error at \"%s\" from (%d, %d) to (%d, %d). %s\n"
|
||||
(Lexing.lexeme lexbuf)
|
||||
start.pos_lnum (start.pos_cnum - start.pos_bol)
|
||||
end_.pos_lnum (end_.pos_cnum - end_.pos_bol)
|
||||
file
|
||||
in
|
||||
let message () = str in
|
||||
let loc = Region.make
|
||||
~start:(Pos.from_byte start)
|
||||
~stop:(Pos.from_byte end_)
|
||||
in
|
||||
let data = [
|
||||
("unrecognized_loc",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ loc
|
||||
)
|
||||
] in
|
||||
error ~data title message
|
||||
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ e.region)]
|
||||
in error ~data title message
|
||||
end
|
||||
|
||||
open Errors
|
||||
@ -133,33 +107,27 @@ type 'a parser = (Lexing.lexbuf -> LexToken.token) -> Lexing.lexbuf -> 'a
|
||||
let parse (parser: 'a parser) source lexbuf =
|
||||
let Lexer.{read; close; _} = Lexer.open_token_stream None in
|
||||
let result =
|
||||
try
|
||||
ok (parser read lexbuf)
|
||||
with
|
||||
SyntaxError.Error (Non_linear_pattern var) ->
|
||||
fail @@ (non_linear_pattern var)
|
||||
| SyntaxError.Error (Duplicate_parameter name) ->
|
||||
fail @@ (duplicate_parameter name)
|
||||
| SyntaxError.Error (Duplicate_variant name) ->
|
||||
fail @@ (duplicate_variant name)
|
||||
| SyntaxError.Error (Reserved_name name) ->
|
||||
fail @@ (reserved_name name)
|
||||
| SyntaxError.Error (Detached_attributes attrs) ->
|
||||
fail @@ (detached_attributes attrs)
|
||||
try ok (parser read lexbuf) with
|
||||
Lexer.Error e ->
|
||||
fail @@ lexer_error e
|
||||
| Parser.Error ->
|
||||
let start = Lexing.lexeme_start_p lexbuf in
|
||||
let end_ = Lexing.lexeme_end_p lexbuf in
|
||||
fail @@ (parser_error source start end_ lexbuf)
|
||||
| Lexer.Error e ->
|
||||
fail @@ (lexer_error e)
|
||||
let stop = Lexing.lexeme_end_p lexbuf in
|
||||
fail @@ parser_error source start stop lexbuf
|
||||
| Scoping.Error (Scoping.Non_linear_pattern var) ->
|
||||
fail @@ non_linear_pattern var
|
||||
| Scoping.Error (Duplicate_parameter name) ->
|
||||
fail @@ duplicate_parameter name
|
||||
| Scoping.Error (Duplicate_variant name) ->
|
||||
fail @@ duplicate_variant name
|
||||
| Scoping.Error (Reserved_name name) ->
|
||||
fail @@ reserved_name name
|
||||
| _ ->
|
||||
let _ = Printexc.print_backtrace Pervasives.stdout in
|
||||
let () = Printexc.print_backtrace Pervasives.stdout in
|
||||
let start = Lexing.lexeme_start_p lexbuf in
|
||||
let end_ = Lexing.lexeme_end_p lexbuf in
|
||||
fail @@ (unrecognized_error source start end_ lexbuf)
|
||||
in
|
||||
close ();
|
||||
result
|
||||
let stop = Lexing.lexeme_end_p lexbuf in
|
||||
fail @@ unrecognized_error source start stop lexbuf
|
||||
in close (); result
|
||||
|
||||
let parse_file (source: string) : AST.t result =
|
||||
let pp_input =
|
||||
|
@ -1,21 +1,18 @@
|
||||
(* This file provides an interface to the PascaLIGO parser. *)
|
||||
(** This file provides an interface to the PascaLIGO parser. *)
|
||||
|
||||
open Trace
|
||||
|
||||
module Parser = Parser_pascaligo.Parser
|
||||
module AST = Parser_pascaligo.AST
|
||||
module ParserLog = Parser_pascaligo.ParserLog
|
||||
module LexToken = Parser_pascaligo.LexToken
|
||||
|
||||
|
||||
(** Open a PascaLIGO filename given by string and convert into an abstract syntax tree. *)
|
||||
val parse_file : string -> (AST.t result)
|
||||
(** Open a PascaLIGO filename given by string and convert into an
|
||||
abstract syntax tree. *)
|
||||
val parse_file : string -> AST.t Trace.result
|
||||
|
||||
(** Convert a given string into a PascaLIGO abstract syntax tree *)
|
||||
val parse_string : string -> AST.t result
|
||||
val parse_string : string -> AST.t Trace.result
|
||||
|
||||
(** Parse a given string as a PascaLIGO expression and return an expression AST.
|
||||
(** Parse a given string as a PascaLIGO expression and return an
|
||||
expression AST.
|
||||
|
||||
This is intended to be used for interactive interpreters, or other scenarios
|
||||
where you would want to parse a PascaLIGO expression outside of a contract. *)
|
||||
val parse_expression : string -> AST.expr result
|
||||
This is intended to be used for interactive interpreters, or other
|
||||
scenarios where you would want to parse a PascaLIGO expression
|
||||
outside of a contract. *)
|
||||
val parse_expression : string -> AST.expr Trace.result
|
||||
|
@ -1,5 +1,4 @@
|
||||
$HOME/git/OCaml-build/Makefile
|
||||
$HOME/git/OCaml-build/Makefile.cfg
|
||||
$HOME/git/ligo/vendors/ligo-utils/simple-utils/pos.mli
|
||||
$HOME/git/ligo/vendors/ligo-utils/simple-utils/pos.ml
|
||||
$HOME/git/ligo/vendors/ligo-utils/simple-utils/region.mli
|
||||
@ -19,5 +18,8 @@ $HOME/git/ligo/vendors/ligo-utils/simple-utils/region.ml
|
||||
../shared/ParserAPI.mli
|
||||
../shared/ParserAPI.ml
|
||||
../shared/LexerUnit.ml
|
||||
../shared/ParserUnit.mli
|
||||
../shared/ParserUnit.ml
|
||||
../shared/Memo.mli
|
||||
../shared/Memo.ml
|
||||
Stubs/Simple_utils.ml
|
||||
|
@ -25,6 +25,7 @@ type 'a reg = 'a Region.reg
|
||||
|
||||
type keyword = Region.t
|
||||
type kwd_and = Region.t
|
||||
type kwd_attributes = Region.t
|
||||
type kwd_begin = Region.t
|
||||
type kwd_block = Region.t
|
||||
type kwd_case = Region.t
|
||||
@ -144,12 +145,13 @@ type t = {
|
||||
|
||||
and ast = t
|
||||
|
||||
and attributes = attribute list reg
|
||||
|
||||
and declaration =
|
||||
TypeDecl of type_decl reg
|
||||
| ConstDecl of const_decl reg
|
||||
| FunDecl of fun_decl reg
|
||||
| AttrDecl of attr_decl
|
||||
|
||||
and attr_decl = string reg ne_injection reg
|
||||
|
||||
and const_decl = {
|
||||
kwd_const : kwd_const;
|
||||
@ -159,7 +161,7 @@ and const_decl = {
|
||||
equal : equal;
|
||||
init : expr;
|
||||
terminator : semi option;
|
||||
attributes : attributes;
|
||||
attributes : attr_decl option
|
||||
}
|
||||
|
||||
(* Type declarations *)
|
||||
@ -217,7 +219,7 @@ and fun_decl = {
|
||||
block_with : (block reg * kwd_with) option;
|
||||
return : expr;
|
||||
terminator : semi option;
|
||||
attributes : attributes;
|
||||
attributes : attr_decl option
|
||||
}
|
||||
|
||||
and parameters = (param_decl, semi) nsepseq par reg
|
||||
@ -260,6 +262,7 @@ and statements = (statement, semi) nsepseq
|
||||
and statement =
|
||||
Instr of instruction
|
||||
| Data of data_decl
|
||||
| Attr of attr_decl
|
||||
|
||||
and data_decl =
|
||||
LocalConst of const_decl reg
|
||||
@ -562,6 +565,7 @@ and field_assign = {
|
||||
equal : equal;
|
||||
field_expr : expr
|
||||
}
|
||||
|
||||
and record = field_assign reg ne_injection
|
||||
|
||||
and projection = {
|
||||
|
@ -28,6 +28,11 @@ type lexeme = string
|
||||
|
||||
(* TOKENS *)
|
||||
|
||||
type attribute = {
|
||||
header : string;
|
||||
string : lexeme Region.reg
|
||||
}
|
||||
|
||||
type t =
|
||||
(* Literals *)
|
||||
|
||||
@ -151,8 +156,7 @@ val mk_kwd : lexeme -> Region.t -> (token, kwd_err) result
|
||||
val mk_string : lexeme -> Region.t -> token
|
||||
val mk_bytes : lexeme -> Region.t -> token
|
||||
val mk_constr : lexeme -> Region.t -> token
|
||||
val mk_attr : lexeme -> Region.t -> (token, attr_err) result
|
||||
val mk_attr2 : lexeme -> Region.t -> (token, attr_err) result
|
||||
val mk_attr : string -> lexeme -> Region.t -> (token, attr_err) result
|
||||
val eof : Region.t -> token
|
||||
|
||||
(* Predicates *)
|
||||
|
@ -26,6 +26,11 @@ let rollback buffer =
|
||||
|
||||
(* TOKENS *)
|
||||
|
||||
type attribute = {
|
||||
header : string;
|
||||
string : lexeme Region.reg
|
||||
}
|
||||
|
||||
type t =
|
||||
(* Literals *)
|
||||
|
||||
@ -144,6 +149,11 @@ let proj_token = function
|
||||
| Constr Region.{region; value} ->
|
||||
region, sprintf "Constr \"%s\"" value
|
||||
|
||||
(*
|
||||
| Attr {header; string={region; value}} ->
|
||||
region, sprintf "Attr (\"%s\",\"%s\")" header value
|
||||
*)
|
||||
|
||||
(* Symbols *)
|
||||
|
||||
| SEMI region -> region, "SEMI"
|
||||
@ -312,6 +322,7 @@ let to_lexeme = function
|
||||
|
||||
| EOF _ -> ""
|
||||
|
||||
(* CONVERSIONS *)
|
||||
|
||||
let to_string token ?(offsets=true) mode =
|
||||
let region, val_str = proj_token token in
|
||||
@ -365,7 +376,7 @@ let keywords = [
|
||||
|
||||
let reserved =
|
||||
let open SSet in
|
||||
empty |> add "args"
|
||||
empty |> add "arguments"
|
||||
|
||||
let constructors = [
|
||||
(fun reg -> False reg);
|
||||
@ -489,8 +500,6 @@ let eof region = EOF region
|
||||
|
||||
type sym_err = Invalid_symbol
|
||||
|
||||
type attr_err = Invalid_attribute
|
||||
|
||||
let mk_sym lexeme region =
|
||||
match lexeme with
|
||||
(* Lexemes in common with all concrete syntaxes *)
|
||||
@ -539,10 +548,9 @@ let mk_constr lexeme region =
|
||||
|
||||
(* Attributes *)
|
||||
|
||||
let mk_attr _lexeme _region =
|
||||
Error Invalid_attribute
|
||||
type attr_err = Invalid_attribute
|
||||
|
||||
let mk_attr2 _lexeme _region =
|
||||
let mk_attr _header _string _region =
|
||||
Error Invalid_attribute
|
||||
|
||||
(* Predicates *)
|
||||
|
@ -7,3 +7,8 @@ module IO =
|
||||
end
|
||||
|
||||
module M = LexerUnit.Make (IO) (Lexer.Make (LexToken))
|
||||
|
||||
let () =
|
||||
match M.trace () with
|
||||
Stdlib.Ok _ -> ()
|
||||
| Error msg -> Printf.eprintf "\027[31m%s\027[0m%!" msg
|
||||
|
5
src/passes/1-parser/pascaligo/Makefile.cfg
Normal file
5
src/passes/1-parser/pascaligo/Makefile.cfg
Normal file
@ -0,0 +1,5 @@
|
||||
SHELL := dash
|
||||
BFLAGS := -strict-sequence -w +A-48-4 -g
|
||||
|
||||
clean::
|
||||
> rm -f Parser.msg Parser.msg.map Parser.msg.states Version.ml
|
39
src/passes/1-parser/pascaligo/Misc/pascaligo.ml
Normal file
39
src/passes/1-parser/pascaligo/Misc/pascaligo.ml
Normal file
@ -0,0 +1,39 @@
|
||||
|
||||
module ParserLog = Parser_pascaligo.ParserLog
|
||||
module ParErr = Parser_pascaligo.ParErr
|
||||
module SSet = Utils.String.Set
|
||||
|
||||
(* Mock options. TODO: Plug in cmdliner. *)
|
||||
|
||||
let pre_options =
|
||||
EvalOpt.make
|
||||
~libs:[]
|
||||
~verbose:SSet.empty
|
||||
~offsets:true
|
||||
~mode:`Point
|
||||
~cmd:EvalOpt.Quiet
|
||||
~mono:true (* Monolithic API of Menhir for now *)
|
||||
(* ~input:None *)
|
||||
(* ~expr:true *)
|
||||
|
||||
module Parser =
|
||||
struct
|
||||
type ast = AST.t
|
||||
type expr = AST.expr
|
||||
include Parser_pascaligo.Parser
|
||||
end
|
||||
|
||||
module ParserLog =
|
||||
struct
|
||||
type ast = AST.t
|
||||
type expr = AST.expr
|
||||
include Parser_pascaligo.ParserLog
|
||||
end
|
||||
|
||||
module PreUnit = ParserUnit.Make (Lexer)(AST)(Parser)(ParErr)(ParserLog)
|
||||
module Front = ParserAPI.Make (Lexer)(Parser)(ParErr)
|
||||
|
||||
let issue_error point =
|
||||
let error = Front.format_error ~offsets:true (* TODO: CLI *)
|
||||
`Point (* TODO: CLI *) point
|
||||
in Stdlib.Error error
|
@ -58,13 +58,13 @@ let message =
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 64 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 517 ->
|
||||
| 543 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 29 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 32 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 515 ->
|
||||
| 541 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 35 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
@ -78,23 +78,9 @@ let message =
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 67 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 70 ->
|
||||
| 68 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 71 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 72 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 73 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 80 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 81 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 76 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 77 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 78 ->
|
||||
| 84 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 85 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
@ -102,241 +88,225 @@ let message =
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 87 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 88 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 512 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 358 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 359 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 499 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 362 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 360 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 361 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 363 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 364 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 365 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 366 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 367 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 475 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 476 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 477 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 478 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 496 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 503 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 502 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 371 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 372 ->
|
||||
| 514 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 373 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 374 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 507 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 377 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 375 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 376 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 378 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 379 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 380 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 381 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 382 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 383 ->
|
||||
| 484 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 485 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 486 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 487 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 504 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 511 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 510 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 386 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 387 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 384 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 385 ->
|
||||
| 388 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 389 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 390 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 391 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 393 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 395 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 399 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 396 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 397 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 375 ->
|
||||
| 398 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 381 ->
|
||||
| 402 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 399 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 400 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 404 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 408 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 405 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 406 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 492 ->
|
||||
| 390 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 493 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 494 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 407 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 488 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 408 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 452 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 447 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 453 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 409 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 410 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 416 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 420 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 421 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 411 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 424 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 425 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 426 ->
|
||||
| 396 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 413 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 414 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 415 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 435 ->
|
||||
| 500 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 436 ->
|
||||
| 501 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 437 ->
|
||||
| 502 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 440 ->
|
||||
| 416 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 441 ->
|
||||
| 496 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 469 ->
|
||||
| 417 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 470 ->
|
||||
| 461 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 473 ->
|
||||
| 456 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 472 ->
|
||||
| 462 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 438 ->
|
||||
| 418 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 467 ->
|
||||
| 419 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 439 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 69 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 428 ->
|
||||
| 425 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 429 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 430 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 431 ->
|
||||
| 420 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 432 ->
|
||||
| 433 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 508 ->
|
||||
| 434 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 521 ->
|
||||
| 435 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 159 ->
|
||||
| 422 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 523 ->
|
||||
| 424 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 137 ->
|
||||
| 444 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 150 ->
|
||||
| 445 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 166 ->
|
||||
| 446 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 167 ->
|
||||
| 449 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 158 ->
|
||||
| 450 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 173 ->
|
||||
| 478 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 152 ->
|
||||
| 479 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 168 ->
|
||||
| 482 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 169 ->
|
||||
| 481 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 175 ->
|
||||
| 447 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 177 ->
|
||||
| 476 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 179 ->
|
||||
| 448 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 437 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 438 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 439 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 440 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 441 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 536 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 515 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 516 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 517 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 518 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 519 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 520 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 529 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 532 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 524 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 525 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 547 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 181 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 183 ->
|
||||
| 549 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 160 ->
|
||||
| 159 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 170 ->
|
||||
| 172 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 157 ->
|
||||
| 188 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 163 ->
|
||||
| 189 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 187 ->
|
||||
| 180 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 92 ->
|
||||
| 195 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 318 ->
|
||||
| 174 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 319 ->
|
||||
| 190 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 322 ->
|
||||
| 191 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 323 ->
|
||||
| 197 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 356 ->
|
||||
| 199 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 351 ->
|
||||
| 201 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 353 ->
|
||||
| 203 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 93 ->
|
||||
| 205 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 94 ->
|
||||
| 182 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 338 ->
|
||||
| 192 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 95 ->
|
||||
| 179 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 96 ->
|
||||
| 185 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 209 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 91 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 342 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
@ -346,169 +316,231 @@ let message =
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 347 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 349 ->
|
||||
| 371 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 97 ->
|
||||
| 366 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 136 ->
|
||||
| 368 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 92 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 93 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 362 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 94 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 95 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 144 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 145 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 148 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 149 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 364 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 96 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 158 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 100 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 217 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 218 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 220 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 221 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 224 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 225 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 358 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 353 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 355 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 101 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 195 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 196 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 198 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 199 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 202 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 203 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 334 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 329 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 331 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 102 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 350 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 336 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 338 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 103 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 326 ->
|
||||
| 332 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 312 ->
|
||||
| 330 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 314 ->
|
||||
| 333 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 104 ->
|
||||
| 334 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 308 ->
|
||||
| 328 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 306 ->
|
||||
| 156 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 105 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 320 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 321 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 322 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 323 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 324 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 137 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 138 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 139 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 140 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 151 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 106 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 107 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 309 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 310 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 304 ->
|
||||
| 154 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 134 ->
|
||||
| 177 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 106 ->
|
||||
| 312 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 296 ->
|
||||
| 315 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 297 ->
|
||||
| 316 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 298 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 299 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 300 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 107 ->
|
||||
| 133 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 108 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 285 ->
|
||||
| 69 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 286 ->
|
||||
| 70 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 132 ->
|
||||
| 71 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 155 ->
|
||||
| 72 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 79 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 80 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 75 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 76 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 77 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 109 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 110 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 111 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 112 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 114 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 117 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 230 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 231 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 269 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 293 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 270 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 272 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 273 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 294 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 300 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 299 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 303 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 302 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 240 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 283 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 284 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 287 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 288 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 291 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 292 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 128 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 110 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 113 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 208 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 209 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 247 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 271 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 248 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 250 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 251 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 272 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 278 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 277 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 281 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 280 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 218 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 261 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 262 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 265 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 266 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 269 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 255 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 257 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 219 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 244 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 245 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 253 ->
|
||||
| 279 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 241 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 210 ->
|
||||
| 266 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 267 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 275 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 211 ->
|
||||
| 263 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 223 ->
|
||||
| 232 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 224 ->
|
||||
| 297 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 240 ->
|
||||
| 233 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 225 ->
|
||||
| 245 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 226 ->
|
||||
| 246 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 234 ->
|
||||
| 262 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 114 ->
|
||||
| 247 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 248 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 256 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 118 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 206 ->
|
||||
| 122 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 119 ->
|
||||
| 228 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 125 ->
|
||||
| 123 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 130 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| _ ->
|
||||
raise Not_found
|
||||
|
@ -6,39 +6,6 @@
|
||||
open Region
|
||||
open AST
|
||||
|
||||
type statement_attributes_mixed =
|
||||
PInstr of instruction
|
||||
| PData of data_decl
|
||||
| PAttributes of attributes
|
||||
|
||||
let attributes_to_statement (statement, statements) =
|
||||
if (List.length statements = 0) then
|
||||
match statement with
|
||||
| PInstr i -> Instr i, []
|
||||
| PData d -> Data d, []
|
||||
| PAttributes a ->
|
||||
let open! SyntaxError in
|
||||
raise (Error (Detached_attributes a))
|
||||
else (
|
||||
let statements = (Region.ghost, statement) :: statements in
|
||||
let rec inner result = function
|
||||
| (t, PData (LocalConst const)) :: (_, PAttributes a) :: rest ->
|
||||
inner (result @ [(t, Data (LocalConst {const with value = {const.value with attributes = a}}))]) rest
|
||||
| (t, PData (LocalFun func)) :: (_, PAttributes a) :: rest ->
|
||||
inner (result @ [(t, Data (LocalFun {func with value = {func.value with attributes = a}}))]) rest
|
||||
| (t, PData d) :: rest ->
|
||||
inner (result @ [(t, Data d)]) rest
|
||||
| (t, PInstr i) :: rest ->
|
||||
inner (result @ [(t, Instr i)]) rest
|
||||
| (_, PAttributes _) :: rest ->
|
||||
inner result rest
|
||||
| [] ->
|
||||
result
|
||||
in
|
||||
let result = inner [] statements in
|
||||
(snd (List.hd result), List.tl result)
|
||||
)
|
||||
|
||||
(* END HEADER *)
|
||||
%}
|
||||
|
||||
@ -146,12 +113,21 @@ declaration:
|
||||
type_decl { TypeDecl $1 }
|
||||
| const_decl { ConstDecl $1 }
|
||||
| fun_decl { FunDecl $1 }
|
||||
| attr_decl { AttrDecl $1 }
|
||||
|
||||
(* Attribute declarations *)
|
||||
|
||||
attr_decl:
|
||||
open_attr_decl ";"? { $1 }
|
||||
|
||||
open_attr_decl:
|
||||
ne_injection("attributes","<string>") { $1 }
|
||||
|
||||
(* Type declarations *)
|
||||
|
||||
type_decl:
|
||||
"type" type_name "is" type_expr ";"? {
|
||||
ignore (SyntaxError.check_reserved_name $2);
|
||||
Scoping.check_reserved_name $2;
|
||||
let stop =
|
||||
match $5 with
|
||||
Some region -> region
|
||||
@ -219,7 +195,7 @@ type_tuple:
|
||||
|
||||
sum_type:
|
||||
"|"? nsepseq(variant,"|") {
|
||||
SyntaxError.check_variants (Utils.nsepseq_to_list $2);
|
||||
Scoping.check_variants (Utils.nsepseq_to_list $2);
|
||||
let region = nsepseq_to_region (fun x -> x.region) $2
|
||||
in TSum {region; value=$2} }
|
||||
|
||||
@ -234,7 +210,7 @@ record_type:
|
||||
"record" sep_or_term_list(field_decl,";") "end" {
|
||||
let ne_elements, terminator = $2 in
|
||||
let () = Utils.nsepseq_to_list ne_elements
|
||||
|> SyntaxError.check_fields in
|
||||
|> Scoping.check_fields in
|
||||
let region = cover $1 $3
|
||||
and value = {opening = Kwd $1;
|
||||
ne_elements;
|
||||
@ -268,21 +244,19 @@ fun_expr:
|
||||
colon = $3;
|
||||
ret_type = $4;
|
||||
kwd_is = $5;
|
||||
return = $6
|
||||
}
|
||||
return = $6}
|
||||
in {region; value} }
|
||||
|
||||
(* Function declarations *)
|
||||
|
||||
open_fun_decl:
|
||||
"function" fun_name parameters ":" type_expr "is"
|
||||
block
|
||||
"with" expr {
|
||||
let fun_name = SyntaxError.check_reserved_name $2 in
|
||||
block "with" expr {
|
||||
Scoping.check_reserved_name $2;
|
||||
let stop = expr_to_region $9 in
|
||||
let region = cover $1 stop
|
||||
and value = {kwd_function = $1;
|
||||
fun_name;
|
||||
fun_name = $2;
|
||||
param = $3;
|
||||
colon = $4;
|
||||
ret_type = $5;
|
||||
@ -290,14 +264,15 @@ open_fun_decl:
|
||||
block_with = Some ($7, $8);
|
||||
return = $9;
|
||||
terminator = None;
|
||||
attributes = {value = []; region = Region.ghost}}
|
||||
in {region; value} }
|
||||
attributes = None}
|
||||
in {region; value}
|
||||
}
|
||||
| "function" fun_name parameters ":" type_expr "is" expr {
|
||||
let fun_name = SyntaxError.check_reserved_name $2 in
|
||||
Scoping.check_reserved_name $2;
|
||||
let stop = expr_to_region $7 in
|
||||
let region = cover $1 stop
|
||||
and value = {kwd_function = $1;
|
||||
fun_name;
|
||||
fun_name = $2;
|
||||
param = $3;
|
||||
colon = $4;
|
||||
ret_type = $5;
|
||||
@ -305,39 +280,36 @@ open_fun_decl:
|
||||
block_with = None;
|
||||
return = $7;
|
||||
terminator = None;
|
||||
attributes = {value = []; region = Region.ghost}}
|
||||
attributes = None}
|
||||
in {region; value} }
|
||||
|
||||
fun_decl:
|
||||
open_fun_decl semi_attributes {
|
||||
let attributes, terminator = $2 in
|
||||
{$1 with value = {$1.value with terminator = terminator; attributes = attributes}}
|
||||
}
|
||||
open_fun_decl ";"? {
|
||||
{$1 with value = {$1.value with terminator=$2}} }
|
||||
|
||||
parameters:
|
||||
par(nsepseq(param_decl,";")) {
|
||||
let params =
|
||||
Utils.nsepseq_to_list ($1.value: _ par).inside
|
||||
in SyntaxError.check_parameters params;
|
||||
$1 }
|
||||
in Scoping.check_parameters params; $1 }
|
||||
|
||||
param_decl:
|
||||
"var" var ":" param_type {
|
||||
let var = SyntaxError.check_reserved_name $2 in
|
||||
Scoping.check_reserved_name $2;
|
||||
let stop = type_expr_to_region $4 in
|
||||
let region = cover $1 stop
|
||||
and value = {kwd_var = $1;
|
||||
var;
|
||||
var = $2;
|
||||
colon = $3;
|
||||
param_type = $4}
|
||||
in ParamVar {region; value}
|
||||
}
|
||||
| "const" var ":" param_type {
|
||||
let var = SyntaxError.check_reserved_name $2 in
|
||||
Scoping.check_reserved_name $2;
|
||||
let stop = type_expr_to_region $4 in
|
||||
let region = cover $1 stop
|
||||
and value = {kwd_const = $1;
|
||||
var;
|
||||
var = $2;
|
||||
colon = $3;
|
||||
param_type = $4}
|
||||
in ParamConst {region; value} }
|
||||
@ -350,7 +322,7 @@ block:
|
||||
let statements, terminator = $2 in
|
||||
let region = cover $1 $3
|
||||
and value = {opening = Begin $1;
|
||||
statements = attributes_to_statement statements;
|
||||
statements;
|
||||
terminator;
|
||||
closing = End $3}
|
||||
in {region; value}
|
||||
@ -359,15 +331,15 @@ block:
|
||||
let statements, terminator = $3 in
|
||||
let region = cover $1 $4
|
||||
and value = {opening = Block ($1,$2);
|
||||
statements = attributes_to_statement statements;
|
||||
statements;
|
||||
terminator;
|
||||
closing = Block $4}
|
||||
in {region; value} }
|
||||
|
||||
statement:
|
||||
instruction { PInstr $1 }
|
||||
| open_data_decl { PData $1 }
|
||||
| attributes { PAttributes $1 }
|
||||
instruction { Instr $1 }
|
||||
| open_data_decl { Data $1 }
|
||||
| open_attr_decl { Attr $1 }
|
||||
|
||||
open_data_decl:
|
||||
open_const_decl { LocalConst $1 }
|
||||
@ -385,10 +357,9 @@ open_const_decl:
|
||||
equal;
|
||||
init;
|
||||
terminator = None;
|
||||
attributes = {value = []; region = Region.ghost}}
|
||||
attributes = None}
|
||||
in {region; value} }
|
||||
|
||||
|
||||
open_var_decl:
|
||||
"var" unqualified_decl(":=") {
|
||||
let name, colon, var_type, assign, init, stop = $2 in
|
||||
@ -399,33 +370,18 @@ open_var_decl:
|
||||
var_type;
|
||||
assign;
|
||||
init;
|
||||
terminator = None;
|
||||
}
|
||||
terminator=None}
|
||||
in {region; value} }
|
||||
|
||||
unqualified_decl(OP):
|
||||
var ":" type_expr OP expr {
|
||||
let var = SyntaxError.check_reserved_name $1 in
|
||||
Scoping.check_reserved_name $1;
|
||||
let region = expr_to_region $5
|
||||
in var, $2, $3, $4, $5, region }
|
||||
|
||||
attributes:
|
||||
"attributes" "[" nsepseq(String,";") "]" {
|
||||
let region = cover $1 $4 in
|
||||
let value = (Utils.nsepseq_to_list $3) in
|
||||
{region; value}
|
||||
}
|
||||
|
||||
semi_attributes:
|
||||
/* empty */ { {value = []; region = Region.ghost}, None }
|
||||
| ";" { {value = []; region = Region.ghost}, Some $1 }
|
||||
| ";" attributes ";" { $2, Some $1 }
|
||||
in $1, $2, $3, $4, $5, region }
|
||||
|
||||
const_decl:
|
||||
open_const_decl semi_attributes {
|
||||
let attributes, terminator = $2 in
|
||||
{$1 with value = {$1.value with terminator = terminator; attributes = attributes }}
|
||||
}
|
||||
open_const_decl ";"? {
|
||||
{$1 with value = {$1.value with terminator=$2}} }
|
||||
|
||||
instruction:
|
||||
conditional { Cond $1 }
|
||||
@ -589,7 +545,7 @@ clause_block:
|
||||
let statements, terminator = $2 in
|
||||
let region = cover $1 $3 in
|
||||
let value = {lbrace = $1;
|
||||
inside = attributes_to_statement statements, terminator;
|
||||
inside = statements, terminator;
|
||||
rbrace = $3} in
|
||||
ShortBlock {value; region} }
|
||||
|
||||
@ -629,7 +585,7 @@ cases(rhs):
|
||||
|
||||
case_clause(rhs):
|
||||
pattern "->" rhs {
|
||||
SyntaxError.check_pattern $1;
|
||||
Scoping.check_pattern $1;
|
||||
fun rhs_to_region ->
|
||||
let start = pattern_to_region $1 in
|
||||
let region = cover start (rhs_to_region $3)
|
||||
@ -671,10 +627,10 @@ for_loop:
|
||||
in For (ForInt {region; value})
|
||||
}
|
||||
| "for" var arrow_clause? "in" collection expr block {
|
||||
let var = SyntaxError.check_reserved_name $2 in
|
||||
Scoping.check_reserved_name $2;
|
||||
let region = cover $1 $7.region in
|
||||
let value = {kwd_for = $1;
|
||||
var;
|
||||
var = $2;
|
||||
bind_to = $3;
|
||||
kwd_in = $4;
|
||||
collection = $5;
|
||||
@ -689,13 +645,13 @@ collection:
|
||||
|
||||
var_assign:
|
||||
var ":=" expr {
|
||||
let name = SyntaxError.check_reserved_name $1 in
|
||||
let region = cover name.region (expr_to_region $3)
|
||||
and value = {name; assign=$2; expr=$3}
|
||||
Scoping.check_reserved_name $1;
|
||||
let region = cover $1.region (expr_to_region $3)
|
||||
and value = {name=$1; assign=$2; expr=$3}
|
||||
in {region; value} }
|
||||
|
||||
arrow_clause:
|
||||
"->" var { $1, SyntaxError.check_reserved_name $2 }
|
||||
"->" var { Scoping.check_reserved_name $2; ($1,$2) }
|
||||
|
||||
(* Expressions *)
|
||||
|
||||
|
@ -114,29 +114,25 @@ let rec print_tokens state ast =
|
||||
Utils.nseq_iter (print_decl state) decl;
|
||||
print_token state eof "EOF"
|
||||
|
||||
and print_attributes state attributes =
|
||||
let attributes = List.fold_left (fun all a -> all ^ a.value ^ ";") "" attributes.value in
|
||||
let line =
|
||||
sprintf "attributes[%s]"
|
||||
attributes
|
||||
in Buffer.add_string state#buffer line
|
||||
and print_attr_decl state =
|
||||
print_ne_injection state "attributes" print_string
|
||||
|
||||
and print_decl state = function
|
||||
TypeDecl decl -> print_type_decl state decl
|
||||
| ConstDecl decl -> print_const_decl state decl
|
||||
| FunDecl decl -> print_fun_decl state decl
|
||||
| AttrDecl decl -> print_attr_decl state decl
|
||||
|
||||
and print_const_decl state {value; _} =
|
||||
let {kwd_const; name; colon; const_type;
|
||||
equal; init; terminator; attributes} = value in
|
||||
equal; init; terminator; _} = value in
|
||||
print_token state kwd_const "const";
|
||||
print_var state name;
|
||||
print_token state colon ":";
|
||||
print_type_expr state const_type;
|
||||
print_token state equal "=";
|
||||
print_expr state init;
|
||||
print_terminator state terminator;
|
||||
print_attributes state attributes
|
||||
print_terminator state terminator
|
||||
|
||||
and print_type_decl state {value; _} =
|
||||
let {kwd_type; name; kwd_is;
|
||||
@ -206,7 +202,7 @@ and print_type_tuple state {value; _} =
|
||||
and print_fun_decl state {value; _} =
|
||||
let {kwd_function; fun_name; param; colon;
|
||||
ret_type; kwd_is; block_with;
|
||||
return; terminator; attributes } = value in
|
||||
return; terminator; _} = value in
|
||||
print_token state kwd_function "function";
|
||||
print_var state fun_name;
|
||||
print_parameters state param;
|
||||
@ -220,7 +216,6 @@ and print_fun_decl state {value; _} =
|
||||
print_token state kwd_with "with");
|
||||
print_expr state return;
|
||||
print_terminator state terminator;
|
||||
print_attributes state attributes
|
||||
|
||||
and print_fun_expr state {value; _} =
|
||||
let {kwd_function; param; colon;
|
||||
@ -296,6 +291,7 @@ and print_statements state sequence =
|
||||
and print_statement state = function
|
||||
Instr instr -> print_instruction state instr
|
||||
| Data data -> print_data_decl state data
|
||||
| Attr attr -> print_attr_decl state attr
|
||||
|
||||
and print_instruction state = function
|
||||
Cond {value; _} -> print_conditional state value
|
||||
@ -688,10 +684,10 @@ and print_opening state lexeme = function
|
||||
print_token state kwd lexeme
|
||||
| KwdBracket (kwd, lbracket) ->
|
||||
print_token state kwd lexeme;
|
||||
print_token state lbracket "{"
|
||||
print_token state lbracket "["
|
||||
|
||||
and print_closing state = function
|
||||
RBracket rbracket -> print_token state rbracket "}"
|
||||
RBracket rbracket -> print_token state rbracket "]"
|
||||
| End kwd_end -> print_token state kwd_end "end"
|
||||
|
||||
and print_binding state {value; _} =
|
||||
@ -848,21 +844,27 @@ and pp_declaration state = function
|
||||
| FunDecl {value; region} ->
|
||||
pp_loc_node state "FunDecl" region;
|
||||
pp_fun_decl state value
|
||||
| AttrDecl {value; region} ->
|
||||
pp_loc_node state "AttrDecl" region;
|
||||
pp_attr_decl state value
|
||||
|
||||
and pp_attr_decl state = pp_ne_injection pp_string state
|
||||
|
||||
and pp_fun_decl state decl =
|
||||
let arity = 5 in
|
||||
let () =
|
||||
let state = state#pad 5 0 in
|
||||
let state = state#pad arity 0 in
|
||||
pp_ident state decl.fun_name in
|
||||
let () =
|
||||
let state = state#pad 5 1 in
|
||||
let state = state#pad arity 1 in
|
||||
pp_node state "<parameters>";
|
||||
pp_parameters state decl.param in
|
||||
let () =
|
||||
let state = state#pad 5 2 in
|
||||
let state = state#pad arity 2 in
|
||||
pp_node state "<return type>";
|
||||
pp_type_expr (state#pad 1 0) decl.ret_type in
|
||||
let () =
|
||||
let state = state#pad 5 3 in
|
||||
let state = state#pad arity 3 in
|
||||
pp_node state "<body>";
|
||||
let statements =
|
||||
match decl.block_with with
|
||||
@ -870,15 +872,16 @@ and pp_fun_decl state decl =
|
||||
| None -> Instr (Skip Region.ghost), [] in
|
||||
pp_statements state statements in
|
||||
let () =
|
||||
let state = state#pad 5 4 in
|
||||
let state = state#pad arity 4 in
|
||||
pp_node state "<return>";
|
||||
pp_expr (state#pad 1 0) decl.return
|
||||
in ()
|
||||
|
||||
and pp_const_decl state decl =
|
||||
pp_ident (state#pad 3 0) decl.name;
|
||||
pp_type_expr (state#pad 3 1) decl.const_type;
|
||||
pp_expr (state#pad 3 2) decl.init
|
||||
let arity = 3 in
|
||||
pp_ident (state#pad arity 0) decl.name;
|
||||
pp_type_expr (state#pad arity 1) decl.const_type;
|
||||
pp_expr (state#pad arity 2) decl.init
|
||||
|
||||
and pp_type_expr state = function
|
||||
TProd cartesian ->
|
||||
@ -979,6 +982,9 @@ and pp_statement state = function
|
||||
| Data data_decl ->
|
||||
pp_node state "Data";
|
||||
pp_data_decl (state#pad 1 0) data_decl
|
||||
| Attr attr_decl ->
|
||||
pp_node state "Attr";
|
||||
pp_attr_decl state attr_decl.value
|
||||
|
||||
and pp_instruction state = function
|
||||
Cond {value; region} ->
|
||||
|
@ -18,6 +18,7 @@ val print_tokens : state -> AST.t -> unit
|
||||
val print_path : state -> AST.path -> unit
|
||||
val print_pattern : state -> AST.pattern -> unit
|
||||
val print_instruction : state -> AST.instruction -> unit
|
||||
val print_expr : state -> AST.expr -> unit
|
||||
|
||||
(** {1 Printing tokens from the AST in a string} *)
|
||||
|
||||
@ -30,6 +31,7 @@ val pattern_to_string :
|
||||
val instruction_to_string :
|
||||
offsets:bool -> mode:[`Point|`Byte] -> AST.instruction -> string
|
||||
|
||||
(** {1 Pretty-printing of the AST} *)
|
||||
(** {1 Pretty-printing of AST nodes} *)
|
||||
|
||||
val pp_ast : state -> AST.t -> unit
|
||||
val pp_expr : state -> AST.expr -> unit
|
||||
|
@ -6,100 +6,97 @@ module IO =
|
||||
let options = EvalOpt.read "PascaLIGO" ext
|
||||
end
|
||||
|
||||
module ExtParser =
|
||||
module Parser =
|
||||
struct
|
||||
type ast = AST.t
|
||||
type expr = AST.expr
|
||||
include Parser
|
||||
end
|
||||
|
||||
module ExtParserLog =
|
||||
module ParserLog =
|
||||
struct
|
||||
type ast = AST.t
|
||||
type expr = AST.expr
|
||||
include ParserLog
|
||||
end
|
||||
|
||||
module MyLexer = Lexer.Make (LexToken)
|
||||
module Lexer = Lexer.Make (LexToken)
|
||||
|
||||
module Unit =
|
||||
ParserUnit.Make (IO)(MyLexer)(AST)(ExtParser)(ParErr)(ExtParserLog)
|
||||
ParserUnit.Make (Lexer)(AST)(Parser)(ParErr)(ParserLog)(IO)
|
||||
|
||||
open! SyntaxError
|
||||
(* Main *)
|
||||
|
||||
let () =
|
||||
try Unit.run () with
|
||||
(* Ad hoc errors from the parser *)
|
||||
let issue_error point =
|
||||
let error = Unit.format_error ~offsets:IO.options#offsets
|
||||
IO.options#mode point
|
||||
in Stdlib.Error error
|
||||
|
||||
Error (Reserved_name name) ->
|
||||
let () = Unit.close_all () in
|
||||
let parse parser : ('a,string) Stdlib.result =
|
||||
try parser () with
|
||||
(* Scoping errors *)
|
||||
|
||||
| Scoping.Error (Scoping.Duplicate_parameter name) ->
|
||||
let token =
|
||||
MyLexer.Token.mk_ident name.Region.value name.Region.region in
|
||||
Lexer.Token.mk_ident name.Region.value name.Region.region in
|
||||
(match token with
|
||||
Stdlib.Error _ ->
|
||||
assert false (* Should not fail if [name] is valid. *)
|
||||
(* Cannot fail because [name] is a not a
|
||||
reserved name for the lexer. *)
|
||||
Stdlib.Error _ -> assert false
|
||||
| Ok invalid ->
|
||||
let point = "Reserved name.\nHint: Change the name.\n",
|
||||
None, invalid in
|
||||
let error =
|
||||
Unit.format_error ~offsets:IO.options#offsets
|
||||
IO.options#mode point
|
||||
in Printf.eprintf "\027[31m%s\027[0m%!" error)
|
||||
issue_error ("Duplicate parameter.\nHint: Change the name.\n",
|
||||
None, invalid))
|
||||
|
||||
| Error (Duplicate_parameter name) ->
|
||||
let () = Unit.close_all () in
|
||||
| Scoping.Error (Scoping.Reserved_name name) ->
|
||||
let token =
|
||||
MyLexer.Token.mk_ident name.Region.value name.Region.region in
|
||||
Lexer.Token.mk_ident name.Region.value name.Region.region in
|
||||
(match token with
|
||||
Stdlib.Error _ ->
|
||||
assert false (* Should not fail if [name] is valid. *)
|
||||
(* Cannot fail because [name] is a not a
|
||||
reserved name for the lexer. *)
|
||||
Stdlib.Error _ -> assert false
|
||||
| Ok invalid ->
|
||||
let point = "Duplicate parameter.\nHint: Change the name.\n",
|
||||
None, invalid in
|
||||
let error =
|
||||
Unit.format_error ~offsets:IO.options#offsets
|
||||
IO.options#mode point
|
||||
in Printf.eprintf "\027[31m%s\027[0m%!" error)
|
||||
issue_error
|
||||
("Reserved name.\nHint: Change the name.\n", None, invalid))
|
||||
|
||||
| Error (Duplicate_variant name) ->
|
||||
let () = Unit.close_all () in
|
||||
| Scoping.Error (Scoping.Duplicate_variant name) ->
|
||||
let token =
|
||||
MyLexer.Token.mk_constr name.Region.value name.Region.region in
|
||||
let point = "Duplicate variant in this sum type declaration.\n\
|
||||
Hint: Change the name.\n",
|
||||
None, token in
|
||||
let error =
|
||||
Unit.format_error ~offsets:IO.options#offsets
|
||||
IO.options#mode point
|
||||
in Printf.eprintf "\027[31m%s\027[0m%!" error
|
||||
Lexer.Token.mk_constr name.Region.value name.Region.region in
|
||||
let point = "Duplicate constructor in this sum type declaration.\n\
|
||||
Hint: Change the constructor.\n",
|
||||
None, token
|
||||
in issue_error point
|
||||
|
||||
| Error (Non_linear_pattern var) ->
|
||||
let () = Unit.close_all () in
|
||||
| Scoping.Error (Scoping.Non_linear_pattern var) ->
|
||||
let token =
|
||||
MyLexer.Token.mk_ident var.Region.value var.Region.region in
|
||||
Lexer.Token.mk_ident var.Region.value var.Region.region in
|
||||
(match token with
|
||||
Stdlib.Error _ ->
|
||||
assert false (* Should not fail if [name] is valid. *)
|
||||
(* Cannot fail because [var] is a not a
|
||||
reserved name for the lexer. *)
|
||||
Stdlib.Error _ -> assert false
|
||||
| Ok invalid ->
|
||||
let point = "Repeated variable in this pattern.\n\
|
||||
Hint: Change the name.\n",
|
||||
None, invalid in
|
||||
let error =
|
||||
Unit.format_error ~offsets:IO.options#offsets
|
||||
IO.options#mode point
|
||||
in Printf.eprintf "\027[31m%s\027[0m%!" error)
|
||||
None, invalid
|
||||
in issue_error point)
|
||||
|
||||
| Error (Duplicate_field name) ->
|
||||
let () = Unit.close_all () in
|
||||
| Scoping.Error (Scoping.Duplicate_field name) ->
|
||||
let token =
|
||||
MyLexer.Token.mk_ident name.Region.value name.Region.region in
|
||||
Lexer.Token.mk_ident name.Region.value name.Region.region in
|
||||
(match token with
|
||||
Stdlib.Error _ ->
|
||||
assert false (* Should not fail if [name] is valid. *)
|
||||
(* Cannot fail because [name] is a not a
|
||||
reserved name for the lexer. *)
|
||||
Stdlib.Error _ -> assert false
|
||||
| Ok invalid ->
|
||||
let point = "Duplicate field name in this record declaration.\n\
|
||||
Hint: Change the name.\n",
|
||||
None, invalid in
|
||||
let error =
|
||||
Unit.format_error ~offsets:IO.options#offsets
|
||||
IO.options#mode point
|
||||
in Printf.eprintf "\027[31m%s\027[0m%!" error)
|
||||
None, invalid
|
||||
in issue_error point)
|
||||
|
||||
let () =
|
||||
if IO.options#expr
|
||||
then match parse (fun () -> Unit.parse Unit.parse_expr) with
|
||||
Stdlib.Ok _ -> ()
|
||||
| Error msg -> Printf.eprintf "\027[31m%s\027[0m%!" msg
|
||||
else match parse (fun () -> Unit.parse Unit.parse_contract) with
|
||||
Stdlib.Ok _ -> ()
|
||||
| Error msg -> Printf.eprintf "\027[31m%s\027[0m%!" msg
|
||||
|
@ -1,12 +1,12 @@
|
||||
[@@@warning "-42"]
|
||||
|
||||
|
||||
type t =
|
||||
Reserved_name of AST.variable
|
||||
| Duplicate_parameter of AST.variable
|
||||
| Duplicate_variant of AST.variable
|
||||
| Non_linear_pattern of AST.variable
|
||||
| Duplicate_field of AST.variable
|
||||
| Detached_attributes of AST.attributes
|
||||
|
||||
type error = t
|
||||
|
||||
@ -95,11 +95,6 @@ let check_reserved_names vars =
|
||||
let check_reserved_name var =
|
||||
if SSet.mem var.value reserved then
|
||||
raise (Error (Reserved_name var))
|
||||
else var
|
||||
|
||||
let check_reserved_name_opt = function
|
||||
Some var -> ignore (check_reserved_name var)
|
||||
| None -> ()
|
||||
|
||||
(* Checking the linearity of patterns *)
|
||||
|
18
src/passes/1-parser/pascaligo/Scoping.mli
Normal file
18
src/passes/1-parser/pascaligo/Scoping.mli
Normal file
@ -0,0 +1,18 @@
|
||||
(* This module exports checks on scoping, called from the parser. *)
|
||||
|
||||
type t =
|
||||
Reserved_name of AST.variable
|
||||
| Duplicate_parameter of AST.variable
|
||||
| Duplicate_variant of AST.variable
|
||||
| Non_linear_pattern of AST.variable
|
||||
| Duplicate_field of AST.variable
|
||||
|
||||
type error = t
|
||||
|
||||
exception Error of t
|
||||
|
||||
val check_reserved_name : AST.variable -> unit
|
||||
val check_pattern : AST.pattern -> unit
|
||||
val check_variants : AST.variant Region.reg list -> unit
|
||||
val check_parameters : AST.param_decl list -> unit
|
||||
val check_fields : AST.field_decl Region.reg list -> unit
|
@ -1,27 +0,0 @@
|
||||
type t =
|
||||
Reserved_name of AST.variable
|
||||
| Duplicate_parameter of AST.variable
|
||||
| Duplicate_variant of AST.variable
|
||||
| Non_linear_pattern of AST.variable
|
||||
| Duplicate_field of AST.variable
|
||||
| Detached_attributes of AST.attributes
|
||||
|
||||
type error = t
|
||||
|
||||
exception Error of t
|
||||
|
||||
module Ord :
|
||||
sig
|
||||
type t = AST.variable
|
||||
val compare : t -> t -> int
|
||||
end
|
||||
|
||||
module VarSet : Set.S with type elt = Ord.t
|
||||
|
||||
val check_reserved_name : AST.variable -> AST.variable
|
||||
val check_reserved_name_opt : AST.variable option -> unit
|
||||
val check_reserved_names : VarSet.t -> VarSet.t
|
||||
val check_pattern : AST.pattern -> unit
|
||||
val check_variants : AST.variant Region.reg list -> unit
|
||||
val check_parameters : AST.param_decl list -> unit
|
||||
val check_fields : AST.field_decl Region.reg list -> unit
|
@ -63,12 +63,12 @@ function claim (var store : store) : list (operation) * store is
|
||||
case store.backers[sender] of
|
||||
None ->
|
||||
failwith ("Not a backer.")
|
||||
| Some (amount) ->
|
||||
| Some (quantity) ->
|
||||
if balance >= store.goal or store.funded then
|
||||
failwith ("Goal reached: no refund.")
|
||||
else
|
||||
begin
|
||||
operations.0.foo := list [transaction (unit, sender, amount)];
|
||||
operations.0.foo := list [transaction (unit, sender, quantity)];
|
||||
remove sender from map store.backers
|
||||
end
|
||||
end
|
||||
|
@ -1,10 +0,0 @@
|
||||
#!/bin/sh
|
||||
|
||||
set -e
|
||||
|
||||
if test -d ../../.git; then
|
||||
echo true > dot_git_is_dir
|
||||
else
|
||||
echo false > dot_git_is_dir
|
||||
cat .git >> dot_git_is_dir
|
||||
fi
|
@ -15,7 +15,7 @@
|
||||
(name parser_pascaligo)
|
||||
(public_name ligo.parser.pascaligo)
|
||||
(modules
|
||||
SyntaxError AST pascaligo Parser ParserLog LexToken)
|
||||
Scoping AST pascaligo Parser ParserLog LexToken ParErr)
|
||||
(libraries
|
||||
menhirLib
|
||||
parser_shared
|
||||
@ -53,31 +53,20 @@
|
||||
(name ParserMain)
|
||||
(libraries parser_pascaligo)
|
||||
(modules
|
||||
ParErr ParserMain)
|
||||
ParserMain)
|
||||
(preprocess
|
||||
(pps bisect_ppx --conditional))
|
||||
(flags (:standard -open Simple_utils -open Parser_shared -open Parser_pascaligo)))
|
||||
|
||||
;; Les deux directives (rule) qui suivent sont pour le dev local.
|
||||
;; Il suffit de faire "dune build Parser.exe" pour avoir un Parser.exe dans le dossier.
|
||||
;; Pour le purger, il faut faire "dune clean".
|
||||
;(rule
|
||||
; (targets Parser.exe)
|
||||
; (deps ParserMain.exe)
|
||||
; (action (copy ParserMain.exe Parser.exe))
|
||||
; (mode promote-until-clean))
|
||||
|
||||
;(rule
|
||||
; (targets Lexer.exe)
|
||||
; (deps LexerMain.exe)
|
||||
; (action (copy LexerMain.exe Lexer.exe))
|
||||
; (mode promote-until-clean))
|
||||
;; Build of the covering of error states in the LR automaton
|
||||
|
||||
(rule
|
||||
(targets Parser.msg)
|
||||
(deps (:script_messages ../../../../vendors/ligo-utils/simple-utils/messages.sh) Parser.mly LexToken.mli ParToken.mly)
|
||||
(action (run %{script_messages} --lex-tokens=LexToken.mli --par-tokens=ParToken.mly Parser.mly)))
|
||||
|
||||
;; Build of all the LIGO source file that cover all error states
|
||||
|
||||
(rule
|
||||
(targets all.ligo)
|
||||
(deps (:script_cover ../../../../vendors/ligo-utils/simple-utils/cover.sh) Parser.mly LexToken.mli ParToken.mly Parser.msg Unlexer.exe)
|
||||
|
@ -1,5 +1,5 @@
|
||||
module Parser = Parser
|
||||
module AST = AST
|
||||
module Lexer = Lexer
|
||||
module LexToken = LexToken
|
||||
module AST = AST
|
||||
module Parser = Parser
|
||||
module ParserLog = ParserLog
|
||||
|
@ -6,18 +6,17 @@ module ParserLog = Parser_cameligo.ParserLog
|
||||
module LexToken = Parser_reasonligo.LexToken
|
||||
module Lexer = Lexer.Make(LexToken)
|
||||
module SyntaxError = Parser_reasonligo.SyntaxError
|
||||
module Scoping = Parser_cameligo.Scoping
|
||||
|
||||
module Errors = struct
|
||||
|
||||
module Errors =
|
||||
struct
|
||||
let lexer_error (e: Lexer.error AST.reg) =
|
||||
let title () = "lexer error" in
|
||||
let message () = Lexer.error_to_string e.value in
|
||||
let data = [
|
||||
("parser_loc",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ e.region
|
||||
)
|
||||
] in
|
||||
error ~data title message
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ e.region)]
|
||||
in error ~data title message
|
||||
|
||||
let wrong_function_arguments expr =
|
||||
let title () = "wrong function arguments" in
|
||||
@ -25,68 +24,58 @@ module Errors = struct
|
||||
let expression_loc = AST.expr_to_region expr in
|
||||
let data = [
|
||||
("location",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ expression_loc)
|
||||
] in
|
||||
error ~data title message
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ expression_loc)]
|
||||
in error ~data title message
|
||||
|
||||
let parser_error source (start: Lexing.position) (end_: Lexing.position) lexbuf =
|
||||
let parser_error source (start: Lexing.position)
|
||||
(end_: Lexing.position) lexbuf =
|
||||
let title () = "parser error" in
|
||||
let file = if source = "" then
|
||||
""
|
||||
let file =
|
||||
if source = "" then ""
|
||||
else
|
||||
Format.sprintf "In file \"%s|%s\"" start.pos_fname source
|
||||
in
|
||||
let str = Format.sprintf
|
||||
Format.sprintf "In file \"%s|%s\"" start.pos_fname source in
|
||||
let str =
|
||||
Format.sprintf
|
||||
"Parse error at \"%s\" from (%d, %d) to (%d, %d). %s\n"
|
||||
(Lexing.lexeme lexbuf)
|
||||
start.pos_lnum (start.pos_cnum - start.pos_bol)
|
||||
end_.pos_lnum (end_.pos_cnum - end_.pos_bol)
|
||||
file
|
||||
in
|
||||
file in
|
||||
let message () = str in
|
||||
let loc = if start.pos_cnum = -1 then
|
||||
Region.make
|
||||
~start: Pos.min
|
||||
let loc =
|
||||
if start.pos_cnum = -1
|
||||
then Region.make
|
||||
~start:(Pos.min ~file:source)
|
||||
~stop:(Pos.from_byte end_)
|
||||
else
|
||||
Region.make
|
||||
else Region.make
|
||||
~start:(Pos.from_byte start)
|
||||
~stop:(Pos.from_byte end_)
|
||||
in
|
||||
~stop:(Pos.from_byte end_) in
|
||||
let data =
|
||||
[
|
||||
("parser_loc",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ loc
|
||||
)
|
||||
]
|
||||
in
|
||||
error ~data title message
|
||||
[("parser_loc",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ loc)]
|
||||
in error ~data title message
|
||||
|
||||
let unrecognized_error source (start: Lexing.position) (end_: Lexing.position) lexbuf =
|
||||
let title () = "unrecognized error" in
|
||||
let file = if source = "" then
|
||||
""
|
||||
let file =
|
||||
if source = "" then ""
|
||||
else
|
||||
Format.sprintf "In file \"%s|%s\"" start.pos_fname source
|
||||
in
|
||||
let str = Format.sprintf
|
||||
Format.sprintf "In file \"%s|%s\"" start.pos_fname source in
|
||||
let str =
|
||||
Format.sprintf
|
||||
"Parse error at \"%s\" from (%d, %d) to (%d, %d). %s\n"
|
||||
(Lexing.lexeme lexbuf)
|
||||
start.pos_lnum (start.pos_cnum - start.pos_bol)
|
||||
end_.pos_lnum (end_.pos_cnum - end_.pos_bol)
|
||||
file
|
||||
in
|
||||
file in
|
||||
let message () = str in
|
||||
let loc = Region.make
|
||||
~start:(Pos.from_byte start)
|
||||
~stop:(Pos.from_byte end_)
|
||||
in
|
||||
~stop:(Pos.from_byte end_) in
|
||||
let data = [
|
||||
("location",
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ loc
|
||||
)
|
||||
] in
|
||||
error ~data title message
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ loc)]
|
||||
in error ~data title message
|
||||
|
||||
end
|
||||
|
||||
|
@ -1,5 +1,4 @@
|
||||
$HOME/git/OCaml-build/Makefile
|
||||
$HOME/git/OCaml-build/Makefile.cfg
|
||||
$HOME/git/ligo/vendors/ligo-utils/simple-utils/pos.mli
|
||||
$HOME/git/ligo/vendors/ligo-utils/simple-utils/pos.ml
|
||||
$HOME/git/ligo/vendors/ligo-utils/simple-utils/region.mli
|
||||
@ -22,7 +21,8 @@ $HOME/git/ligo/vendors/ligo-utils/simple-utils/region.ml
|
||||
../shared/ParserUnit.ml
|
||||
Stubs/Simple_utils.ml
|
||||
Stubs/Parser_cameligo.ml
|
||||
../cameligo/AST.mli
|
||||
../cameligo/AST.ml
|
||||
../cameligo/ParserLog.mli
|
||||
../cameligo/ParserLog.ml
|
||||
../cameligo/Scoping.mli
|
||||
../cameligo/Scoping.ml
|
@ -143,8 +143,7 @@ val mk_int : lexeme -> Region.t -> (token, int_err) result
|
||||
val mk_nat : lexeme -> Region.t -> (token, nat_err) result
|
||||
val mk_mutez : lexeme -> Region.t -> (token, int_err) result
|
||||
val mk_ident : lexeme -> Region.t -> (token, ident_err) result
|
||||
val mk_attr : lexeme -> Region.t -> (token, attr_err) result
|
||||
val mk_attr2 : lexeme -> Region.t -> (token, attr_err) result
|
||||
val mk_attr : string -> lexeme -> Region.t -> (token, attr_err) result
|
||||
val mk_sym : lexeme -> Region.t -> (token, sym_err) result
|
||||
val mk_kwd : lexeme -> Region.t -> (token, kwd_err) result
|
||||
val mk_string : lexeme -> Region.t -> token
|
||||
|
@ -453,11 +453,10 @@ let mk_constr lexeme region = mk_constr' lexeme region lexicon
|
||||
|
||||
(* Attributes *)
|
||||
|
||||
let mk_attr lexeme region =
|
||||
Ok (Attr { value = lexeme; region })
|
||||
|
||||
let mk_attr2 _lexeme _region =
|
||||
Error Invalid_attribute
|
||||
let mk_attr header lexeme region =
|
||||
if header = "[@" then
|
||||
Ok (Attr Region.{value=lexeme; region})
|
||||
else Error Invalid_attribute
|
||||
|
||||
(* Predicates *)
|
||||
|
||||
|
5
src/passes/1-parser/reasonligo/Makefile.cfg
Normal file
5
src/passes/1-parser/reasonligo/Makefile.cfg
Normal file
@ -0,0 +1,5 @@
|
||||
SHELL := dash
|
||||
BFLAGS := -strict-sequence -w +A-48-4 -g
|
||||
|
||||
clean::
|
||||
> rm -f Parser.msg Parser.msg.map Parser.msg.states Version.ml
|
65
src/passes/1-parser/reasonligo/Misc/Misc.ml
Normal file
65
src/passes/1-parser/reasonligo/Misc/Misc.ml
Normal file
@ -0,0 +1,65 @@
|
||||
type error =
|
||||
IntErr of LexToken.int_err
|
||||
| IdentErr of LexToken.ident_err
|
||||
| NatErr of LexToken.nat_err
|
||||
| SymErr of LexToken.sym_err
|
||||
| KwdErr of LexToken.kwd_err
|
||||
|
||||
let rec first_of_expr = function
|
||||
ECase {value; _} ->
|
||||
(match LexToken.mk_kwd "switch" value.kwd_match with
|
||||
Error e -> Error (KwdErr e)
|
||||
| Ok token -> Ok token)
|
||||
| ECond {value; _} ->
|
||||
(match LexToken.mk_kwd "if" value.kwd_if with
|
||||
Error e -> Error (KwdErr e)
|
||||
| Ok token -> Ok token)
|
||||
| EPar {value; _} ->
|
||||
(match LexToken.mk_sym "(" value.lpar with
|
||||
Error e -> Error (SymErr e)
|
||||
| Ok token -> Ok token)
|
||||
| EAnnot {value; _} ->
|
||||
(match LexToken.mk_sym "(" value.lpar with
|
||||
Error e -> Error (SymErr e)
|
||||
| Ok token -> Ok token)
|
||||
| EUnit {value=opening, _; _} ->
|
||||
(match LexToken.mk_sym "(" opening with
|
||||
Error e -> Error (SymErr e)
|
||||
| Ok token -> Ok token)
|
||||
| EBytes b ->
|
||||
Ok (LexToken.mk_bytes (fst b.value) b.region)
|
||||
| EVar v ->
|
||||
(match LexToken.mk_ident v.value v.region with
|
||||
Error e -> Error (IdentErr e)
|
||||
| Ok token -> Ok token)
|
||||
| ESeq {value; _} ->
|
||||
let opening =
|
||||
match value.compound with
|
||||
BeginEnd (opening, _)
|
||||
| Braces (opening, _)
|
||||
| Brackets (opening, _) -> opening
|
||||
in (match LexToken.mk_sym "{" opening with
|
||||
Error e -> Error (SymErr e)
|
||||
| Ok token -> Ok token)
|
||||
| EProj {value; _} ->
|
||||
let structure = value.struct_name in
|
||||
(match LexToken.mk_ident structure.value structure.region with
|
||||
Error e -> Error (IdentErr e)
|
||||
| Ok token -> Ok token)
|
||||
| EFun {value; _} ->
|
||||
(match LexToken.mk_kwd "fun" value.kwd_fun with
|
||||
Error e -> Error (KwdErr e)
|
||||
| Ok token -> Ok token)
|
||||
| _ -> failwith "TODO"
|
||||
(*
|
||||
|
||||
| ELogic expr -> first_of_logic_expr expr
|
||||
| EArith expr -> first_of_arith_expr expr
|
||||
| EString expr -> first_of_string_expr expr
|
||||
| EList expr -> first_of_list_expr expr
|
||||
| EConstr expr -> first_of_constr_expr expr
|
||||
| ECall {value=expr,_; _} -> first_of_expr expr
|
||||
| ERecord {value; _} -> (*field_assign reg ne_injection *)
|
||||
| ETuple {value; _} -> (* (expr, comma) nsepseq *)
|
||||
| ELetIn {value; _} -> first_of_let_in value
|
||||
*)
|
@ -46,9 +46,9 @@ let message =
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 11 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 509 ->
|
||||
| 528 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 503 ->
|
||||
| 61 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 48 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
@ -68,335 +68,387 @@ let message =
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 14 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 60 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 65 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 505 ->
|
||||
| 70 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 145 ->
|
||||
| 524 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 146 ->
|
||||
| 185 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 144 ->
|
||||
| 186 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 329 ->
|
||||
| 184 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 331 ->
|
||||
| 302 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 330 ->
|
||||
| 304 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 61 ->
|
||||
| 303 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 66 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 69 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 64 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 59 ->
|
||||
| 183 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 143 ->
|
||||
| 311 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 338 ->
|
||||
| 313 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 340 ->
|
||||
| 312 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 339 ->
|
||||
| 191 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 151 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 152 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 78 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 325 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 327 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 326 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 92 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 155 ->
|
||||
| 192 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 118 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 125 ->
|
||||
| 298 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 87 ->
|
||||
| 300 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 105 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 107 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 108 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 106 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 88 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 93 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 80 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 81 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 82 ->
|
||||
| 299 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 132 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 334 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 336 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 335 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 133 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 136 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 137 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 157 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 159 ->
|
||||
| 195 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 158 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 512 ->
|
||||
| 165 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 218 ->
|
||||
| 127 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 514 ->
|
||||
| 145 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 216 ->
|
||||
| 147 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 250 ->
|
||||
| 148 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 248 ->
|
||||
| 146 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 249 ->
|
||||
| 128 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 230 ->
|
||||
| 133 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 235 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 252 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 254 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 255 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 258 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 219 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 226 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 227 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 260 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 262 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 264 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 266 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 194 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 195 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 206 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 215 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 199 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 207 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 208 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 196 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 197 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 198 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 256 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 257 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 277 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 233 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 279 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 67 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 463 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 464 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 387 ->
|
||||
| 120 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 121 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 122 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 120 ->
|
||||
| 172 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 466 ->
|
||||
| 307 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 467 ->
|
||||
| 309 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 308 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 173 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 176 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 177 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 197 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 199 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 198 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 59 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 531 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 225 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 533 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 223 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 257 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 255 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 256 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 237 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 242 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 259 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 261 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 262 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 265 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 226 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 233 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 234 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 267 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 269 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 271 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 273 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 201 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 202 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 213 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 222 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 206 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 214 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 215 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 203 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 204 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 205 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 263 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 284 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 240 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 286 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 72 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 483 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 492 ->
|
||||
| 484 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 469 ->
|
||||
| 423 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 470 ->
|
||||
| 161 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 468 ->
|
||||
| 162 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 471 ->
|
||||
| 160 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 472 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 473 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 475 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 476 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 477 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 478 ->
|
||||
| 486 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 487 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 488 ->
|
||||
| 504 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 474 ->
|
||||
| 513 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 498 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 499 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 497 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 465 ->
|
||||
| 488 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 321 ->
|
||||
| 489 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 490 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 492 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 493 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 494 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 495 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 509 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 510 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 491 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 520 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 518 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 485 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 372 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 366 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 367 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 369 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 368 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 365 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 76 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 446 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 326 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 332 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 333 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 336 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 337 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 328 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 339 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 100 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 78 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 80 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 315 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 316 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 318 ->
|
||||
| 117 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 317 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 314 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 71 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 410 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 298 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 304 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 305 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 308 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 309 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 300 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 178 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 73 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 75 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 419 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 420 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 77 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 160 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 412 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 413 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 415 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 416 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 193 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 229 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 74 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 447 ->
|
||||
| 82 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 448 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 456 ->
|
||||
| 449 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 457 ->
|
||||
| 451 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 459 ->
|
||||
| 452 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 200 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 236 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 79 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 467 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 468 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 476 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 477 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 479 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 480 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 469 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 470 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 81 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 460 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 449 ->
|
||||
| 461 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 450 ->
|
||||
| 455 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 76 ->
|
||||
| 454 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 458 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 348 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 356 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 360 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 359 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 355 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 349 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 457 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 340 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 341 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 346 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 347 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 342 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 343 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 344 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 84 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 85 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 318 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 323 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 324 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 389 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 436 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 437 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 438 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 439 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 440 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 441 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 425 ->
|
||||
| 435 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 422 ->
|
||||
| 325 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 428 ->
|
||||
| 362 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 429 ->
|
||||
| 363 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 434 ->
|
||||
| 373 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 438 ->
|
||||
| 374 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 437 ->
|
||||
| 413 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 433 ->
|
||||
| 420 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 423 ->
|
||||
| 408 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 427 ->
|
||||
| 409 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 162 ->
|
||||
| 407 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 163 ->
|
||||
| 375 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 290 ->
|
||||
| 376 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 295 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 296 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 357 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 400 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 401 ->
|
||||
| 377 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 402 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
@ -406,105 +458,69 @@ let message =
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 405 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 399 ->
|
||||
| 417 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 297 ->
|
||||
| 418 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 311 ->
|
||||
| 401 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 312 ->
|
||||
| 429 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 322 ->
|
||||
| 427 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 323 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 377 ->
|
||||
| 364 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 384 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 342 ->
|
||||
| 385 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 343 ->
|
||||
| 383 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 324 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 344 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 345 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 346 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 370 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 371 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 372 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 373 ->
|
||||
| 378 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 379 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 380 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 369 ->
|
||||
| 394 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 395 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 396 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 397 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 399 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 398 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 393 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 391 ->
|
||||
| 320 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 313 ->
|
||||
| 321 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 348 ->
|
||||
| 86 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 349 ->
|
||||
| 87 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 347 ->
|
||||
| 88 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 350 ->
|
||||
| 89 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 351 ->
|
||||
| 90 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 352 ->
|
||||
| 91 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 359 ->
|
||||
| 96 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 360 ->
|
||||
| 97 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 361 ->
|
||||
| 98 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 362 ->
|
||||
| 111 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 364 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 363 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 358 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 292 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 293 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 164 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 165 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 166 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 167 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 168 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 169 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 174 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 175 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 176 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 188 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| 237 ->
|
||||
| 244 ->
|
||||
"<YOUR SYNTAX ERROR MESSAGE HERE>\n"
|
||||
| _ ->
|
||||
raise Not_found
|
||||
|
@ -148,6 +148,7 @@ declaration:
|
||||
|
||||
type_decl:
|
||||
"type" type_name "=" type_expr {
|
||||
Scoping.check_reserved_name $2;
|
||||
let region = cover $1 (type_expr_to_region $4)
|
||||
and value = {kwd_type = $1;
|
||||
name = $2;
|
||||
@ -192,6 +193,7 @@ core_type:
|
||||
|
||||
sum_type:
|
||||
"|" nsepseq(variant,"|") {
|
||||
Scoping.check_variants (Utils.nsepseq_to_list $2);
|
||||
let region = nsepseq_to_region (fun x -> x.region) $2
|
||||
in TSum {region; value=$2} }
|
||||
|
||||
@ -205,6 +207,8 @@ variant:
|
||||
record_type:
|
||||
"{" sep_or_term_list(field_decl,",") "}" {
|
||||
let ne_elements, terminator = $2 in
|
||||
let () = Utils.nsepseq_to_list ne_elements
|
||||
|> Scoping.check_fields in
|
||||
let region = cover $1 $3
|
||||
and value = {compound = Braces ($1,$3); ne_elements; terminator}
|
||||
in TRecord {region; value} }
|
||||
@ -240,6 +244,7 @@ es6_func:
|
||||
|
||||
let_binding:
|
||||
"<ident>" type_annotation? "=" expr {
|
||||
Scoping.check_reserved_name $1;
|
||||
{binders = PVar $1, []; lhs_type=$2; eq=$3; let_rhs=$4}
|
||||
}
|
||||
| "_" type_annotation? "=" expr {
|
||||
@ -249,12 +254,15 @@ let_binding:
|
||||
{binders = PUnit $1, []; lhs_type=$2; eq=$3; let_rhs=$4}
|
||||
}
|
||||
| record_pattern type_annotation? "=" expr {
|
||||
Scoping.check_pattern (PRecord $1);
|
||||
{binders = PRecord $1, []; lhs_type=$2; eq=$3; let_rhs=$4}
|
||||
}
|
||||
| par(closed_irrefutable) type_annotation? "=" expr {
|
||||
Scoping.check_pattern $1.value.inside;
|
||||
{binders = PPar $1, []; lhs_type=$2; eq=$3; let_rhs=$4}
|
||||
}
|
||||
| tuple(sub_irrefutable) type_annotation? "=" expr {
|
||||
Utils.nsepseq_iter Scoping.check_pattern $1;
|
||||
let hd, tl = $1 in
|
||||
let start = pattern_to_region hd in
|
||||
let stop = last fst tl in
|
||||
@ -419,8 +427,11 @@ fun_expr:
|
||||
let region = cover start stop in
|
||||
|
||||
let rec arg_to_pattern = function
|
||||
EVar v -> PVar v
|
||||
EVar v ->
|
||||
Scoping.check_reserved_name v;
|
||||
PVar v
|
||||
| EAnnot {region; value = {inside = EVar v, colon, typ; _}} ->
|
||||
Scoping.check_reserved_name v;
|
||||
let value = {pattern = PVar v; colon; type_expr = typ}
|
||||
in PTyped {region; value}
|
||||
| EPar p ->
|
||||
@ -468,8 +479,9 @@ fun_expr:
|
||||
arg_to_pattern (EAnnot e), []
|
||||
| ETuple {value = fun_args; _} ->
|
||||
let bindings =
|
||||
List.map (arg_to_pattern <@ snd) (snd fun_args)
|
||||
in arg_to_pattern (fst fun_args), bindings
|
||||
List.map (arg_to_pattern <@ snd) (snd fun_args) in
|
||||
List.iter Scoping.check_pattern bindings;
|
||||
arg_to_pattern (fst fun_args), bindings
|
||||
| EUnit e ->
|
||||
arg_to_pattern (EUnit e), []
|
||||
| e -> let open! SyntaxError
|
||||
@ -555,6 +567,7 @@ cases(right_expr):
|
||||
|
||||
case_clause(right_expr):
|
||||
"|" pattern "=>" right_expr ";"? {
|
||||
Scoping.check_pattern $2;
|
||||
let start = pattern_to_region $2
|
||||
and stop = expr_to_region $4 in
|
||||
let region = cover start stop
|
||||
|
@ -6,32 +6,37 @@ module IO =
|
||||
let options = EvalOpt.read "ReasonLIGO" ext
|
||||
end
|
||||
|
||||
module ExtParser =
|
||||
module Parser =
|
||||
struct
|
||||
type ast = AST.t
|
||||
type expr = AST.expr
|
||||
include Parser
|
||||
end
|
||||
|
||||
module ExtParserLog =
|
||||
module ParserLog =
|
||||
struct
|
||||
type ast = AST.t
|
||||
type expr = AST.expr
|
||||
include ParserLog
|
||||
end
|
||||
|
||||
module MyLexer = Lexer.Make (LexToken)
|
||||
module Lexer = Lexer.Make (LexToken)
|
||||
|
||||
module Unit =
|
||||
ParserUnit.Make (IO)(MyLexer)(AST)(ExtParser)(ParErr)(ExtParserLog)
|
||||
ParserUnit.Make (Lexer)(AST)(Parser)(ParErr)(ParserLog)(IO)
|
||||
|
||||
(* Main *)
|
||||
|
||||
let () =
|
||||
try Unit.run () with
|
||||
(* Ad hoc errors from the parsers *)
|
||||
let issue_error point =
|
||||
let error = Unit.format_error ~offsets:IO.options#offsets
|
||||
IO.options#mode point
|
||||
in Stdlib.Error error
|
||||
|
||||
let parse parser : ('a,string) Stdlib.result =
|
||||
try parser () with
|
||||
(* Ad hoc errors from the parser *)
|
||||
|
||||
SyntaxError.Error (SyntaxError.WrongFunctionArguments expr) ->
|
||||
let () = Unit.close_all () in
|
||||
let msg = "It looks like you are defining a function, \
|
||||
however we do not\n\
|
||||
understand the parameters declaration.\n\
|
||||
@ -41,4 +46,61 @@ let () =
|
||||
and reg = AST.expr_to_region expr in
|
||||
let error = Unit.short_error ~offsets:IO.options#offsets
|
||||
IO.options#mode msg reg
|
||||
in Printf.eprintf "\027[31m%s\027[0m%!" error
|
||||
in Stdlib.Error error
|
||||
|
||||
(* Scoping errors *)
|
||||
|
||||
| Scoping.Error (Scoping.Reserved_name name) ->
|
||||
let token =
|
||||
Lexer.Token.mk_ident name.Region.value name.Region.region in
|
||||
(match token with
|
||||
(* Cannot fail because [name] is a not a
|
||||
reserved name for the lexer. *)
|
||||
Stdlib.Error _ -> assert false
|
||||
| Ok invalid ->
|
||||
issue_error
|
||||
("Reserved name.\nHint: Change the name.\n", None, invalid))
|
||||
|
||||
| Scoping.Error (Scoping.Duplicate_variant name) ->
|
||||
let token =
|
||||
Lexer.Token.mk_constr name.Region.value name.Region.region in
|
||||
let point = "Duplicate constructor in this sum type declaration.\n\
|
||||
Hint: Change the constructor.\n",
|
||||
None, token
|
||||
in issue_error point
|
||||
|
||||
| Scoping.Error (Scoping.Non_linear_pattern var) ->
|
||||
let token =
|
||||
Lexer.Token.mk_ident var.Region.value var.Region.region in
|
||||
(match token with
|
||||
(* Cannot fail because [var] is a not a
|
||||
reserved name for the lexer. *)
|
||||
Stdlib.Error _ -> assert false
|
||||
| Ok invalid ->
|
||||
let point = "Repeated variable in this pattern.\n\
|
||||
Hint: Change the name.\n",
|
||||
None, invalid
|
||||
in issue_error point)
|
||||
|
||||
| Scoping.Error (Scoping.Duplicate_field name) ->
|
||||
let token =
|
||||
Lexer.Token.mk_ident name.Region.value name.Region.region in
|
||||
(match token with
|
||||
(* Cannot fail because [name] is a not a
|
||||
reserved name for the lexer. *)
|
||||
Stdlib.Error _ -> assert false
|
||||
| Ok invalid ->
|
||||
let point =
|
||||
"Duplicate field name in this record declaration.\n\
|
||||
Hint: Change the name.\n",
|
||||
None, invalid
|
||||
in issue_error point)
|
||||
|
||||
let () =
|
||||
if IO.options#expr
|
||||
then match parse (fun () -> Unit.parse Unit.parse_expr) with
|
||||
Stdlib.Ok _ -> ()
|
||||
| Error msg -> Printf.eprintf "\027[31m%s\027[0m%!" msg
|
||||
else match parse (fun () -> Unit.parse Unit.parse_contract) with
|
||||
Stdlib.Ok _ -> ()
|
||||
| Error msg -> Printf.eprintf "\027[31m%s\027[0m%!" msg
|
||||
|
@ -1,10 +1,16 @@
|
||||
;; Build of the lexer
|
||||
|
||||
(ocamllex LexToken)
|
||||
|
||||
;; Build of the parser
|
||||
|
||||
(menhir
|
||||
(merge_into Parser)
|
||||
(modules ParToken Parser)
|
||||
(flags -la 1 --table --explain --strict --external-tokens LexToken))
|
||||
|
||||
;; Build of the parser as a library
|
||||
|
||||
(library
|
||||
(name parser_reasonligo)
|
||||
(public_name ligo.parser.reasonligo)
|
||||
@ -22,6 +28,18 @@
|
||||
(pps bisect_ppx --conditional))
|
||||
(flags (:standard -open Simple_utils -open Parser_shared -open Parser_cameligo)))
|
||||
|
||||
;; Build of the unlexer (for covering the
|
||||
;; error states of the LR automaton)
|
||||
|
||||
(executable
|
||||
(name Unlexer)
|
||||
(libraries str)
|
||||
(preprocess
|
||||
(pps bisect_ppx --conditional))
|
||||
(modules Unlexer))
|
||||
|
||||
;; Local build of a standalone lexer
|
||||
|
||||
(executable
|
||||
(name LexerMain)
|
||||
(libraries parser_reasonligo)
|
||||
@ -30,6 +48,8 @@
|
||||
(pps bisect_ppx --conditional))
|
||||
(flags (:standard -open Parser_shared -open Parser_reasonligo)))
|
||||
|
||||
;; Local build of a standalone parser
|
||||
|
||||
(executable
|
||||
(name ParserMain)
|
||||
(libraries
|
||||
@ -41,19 +61,16 @@
|
||||
(pps bisect_ppx --conditional))
|
||||
(flags (:standard -open Simple_utils -open Parser_cameligo -open Parser_shared -open Parser_reasonligo)))
|
||||
|
||||
(executable
|
||||
(name Unlexer)
|
||||
(libraries str)
|
||||
(preprocess
|
||||
(pps bisect_ppx --conditional))
|
||||
(modules Unlexer))
|
||||
;; Build of the covering of error states in the LR automaton
|
||||
|
||||
(rule
|
||||
(targets Parser.msg)
|
||||
(deps (:script_messages ../../../../vendors/ligo-utils/simple-utils/messages.sh) Parser.mly LexToken.mli ParToken.mly)
|
||||
(action (run %{script_messages} --lex-tokens=LexToken.mli --par-tokens=ParToken.mly Parser.mly )))
|
||||
|
||||
;; Build of all the LIGO source file that cover all error states
|
||||
|
||||
(rule
|
||||
(targets all.ligo)
|
||||
(targets all.religo)
|
||||
(deps (:script_cover ../../../../vendors/ligo-utils/simple-utils/cover.sh) Parser.mly LexToken.mli ParToken.mly Parser.msg Unlexer.exe)
|
||||
(action (run %{script_cover} --lex-tokens=LexToken.mli --par-tokens=ParToken.mly --ext=ligo --unlexer=./Unlexer.exe --messages=Parser.msg --dir=. --concatenate Parser.mly )))
|
||||
(action (run %{script_cover} --lex-tokens=LexToken.mli --par-tokens=ParToken.mly --ext=religo --unlexer=./Unlexer.exe --messages=Parser.msg --dir=. --concatenate Parser.mly )))
|
||||
|
@ -14,10 +14,11 @@ type options = <
|
||||
offsets : bool;
|
||||
mode : [`Byte | `Point];
|
||||
cmd : command;
|
||||
mono : bool
|
||||
mono : bool;
|
||||
expr : bool
|
||||
>
|
||||
|
||||
let make ~input ~libs ~verbose ~offsets ~mode ~cmd ~mono =
|
||||
let make ~input ~libs ~verbose ~offsets ~mode ~cmd ~mono ~expr =
|
||||
object
|
||||
method input = input
|
||||
method libs = libs
|
||||
@ -26,6 +27,7 @@ let make ~input ~libs ~verbose ~offsets ~mode ~cmd ~mono =
|
||||
method mode = mode
|
||||
method cmd = cmd
|
||||
method mono = mono
|
||||
method expr = expr
|
||||
end
|
||||
|
||||
(** {1 Auxiliary functions} *)
|
||||
@ -42,17 +44,18 @@ let abort msg =
|
||||
let help language extension () =
|
||||
let file = Filename.basename Sys.argv.(0) in
|
||||
printf "Usage: %s [<option> ...] [<input>%s | \"-\"]\n" file extension;
|
||||
printf "where <input>%s is the %s source file (default: stdin)," extension language;
|
||||
printf "where <input>%s is the %s source file (default: stdin),\n" extension language;
|
||||
print "and each <option> (if any) is one of the following:";
|
||||
print " -I <paths> Library paths (colon-separated)";
|
||||
print " -c, --copy Print lexemes of tokens and markup (lexer)";
|
||||
print " -t, --tokens Print tokens (lexer)";
|
||||
print " -u, --units Print tokens and markup (lexer)";
|
||||
print " -t, --tokens Print tokens";
|
||||
print " -u, --units Print lexical units";
|
||||
print " -c, --copy Print lexemes and markup";
|
||||
print " -q, --quiet No output, except errors (default)";
|
||||
print " --columns Columns for source locations";
|
||||
print " --bytes Bytes for source locations";
|
||||
print " --mono Use Menhir monolithic API";
|
||||
print " --verbose=<stages> cmdline, cpp, ast-tokens, ast (colon-separated)";
|
||||
print " --expr Parse an expression";
|
||||
print " --verbose=<stages> cli, cpp, ast-tokens, ast (colon-separated)";
|
||||
print " --version Commit hash on stdout";
|
||||
print " -h, --help This help";
|
||||
exit 0
|
||||
@ -74,6 +77,7 @@ and input = ref None
|
||||
and libs = ref []
|
||||
and verb_str = ref ""
|
||||
and mono = ref false
|
||||
and expr = ref false
|
||||
|
||||
let split_at_colon = Str.(split (regexp ":"))
|
||||
|
||||
@ -94,6 +98,7 @@ let specs language extension =
|
||||
noshort, "columns", set columns true, None;
|
||||
noshort, "bytes", set bytes true, None;
|
||||
noshort, "mono", set mono true, None;
|
||||
noshort, "expr", set expr true, None;
|
||||
noshort, "verbose", None, Some add_verbose;
|
||||
'h', "help", Some (help language extension), None;
|
||||
noshort, "version", Some version, None
|
||||
@ -129,7 +134,8 @@ let print_opt () =
|
||||
printf "quiet = %b\n" !quiet;
|
||||
printf "columns = %b\n" !columns;
|
||||
printf "bytes = %b\n" !bytes;
|
||||
printf "mono = %b\b" !mono;
|
||||
printf "mono = %b\n" !mono;
|
||||
printf "expr = %b\n" !expr;
|
||||
printf "verbose = %s\n" !verb_str;
|
||||
printf "input = %s\n" (string_of quote !input);
|
||||
printf "libs = %s\n" (string_of_path !libs)
|
||||
@ -137,7 +143,7 @@ let print_opt () =
|
||||
|
||||
let check extension =
|
||||
let () =
|
||||
if Utils.String.Set.mem "cmdline" !verbose then print_opt () in
|
||||
if Utils.String.Set.mem "cli" !verbose then print_opt () in
|
||||
|
||||
let input =
|
||||
match !input with
|
||||
@ -158,11 +164,12 @@ let check extension =
|
||||
and offsets = not !columns
|
||||
and mode = if !bytes then `Byte else `Point
|
||||
and mono = !mono
|
||||
and expr = !expr
|
||||
and verbose = !verbose
|
||||
and libs = !libs in
|
||||
|
||||
let () =
|
||||
if Utils.String.Set.mem "cmdline" verbose then
|
||||
if Utils.String.Set.mem "cli" verbose then
|
||||
begin
|
||||
printf "\nEXPORTED COMMAND LINE\n";
|
||||
printf "copy = %b\n" copy;
|
||||
@ -172,6 +179,7 @@ let check extension =
|
||||
printf "offsets = %b\n" offsets;
|
||||
printf "mode = %s\n" (if mode = `Byte then "`Byte" else "`Point");
|
||||
printf "mono = %b\n" mono;
|
||||
printf "expr = %b\n" expr;
|
||||
printf "verbose = %s\n" !verb_str;
|
||||
printf "input = %s\n" (string_of quote input);
|
||||
printf "libs = %s\n" (string_of_path libs)
|
||||
@ -186,7 +194,7 @@ let check extension =
|
||||
| false, false, false, true -> Tokens
|
||||
| _ -> abort "Choose one of -q, -c, -u, -t."
|
||||
|
||||
in make ~input ~libs ~verbose ~offsets ~mode ~cmd ~mono
|
||||
in make ~input ~libs ~verbose ~offsets ~mode ~cmd ~mono ~expr
|
||||
|
||||
(** {1 Parsing the command-line options} *)
|
||||
|
||||
@ -195,7 +203,7 @@ let read language extension =
|
||||
Getopt.parse_cmdline (specs language extension) anonymous;
|
||||
(verb_str :=
|
||||
let apply e a =
|
||||
if a <> "" then Printf.sprintf "%s, %s" e a else e
|
||||
if a = "" then e else Printf.sprintf "%s, %s" e a
|
||||
in Utils.String.Set.fold apply !verbose "");
|
||||
check extension
|
||||
with Getopt.Error msg -> abort msg
|
||||
|
@ -1,4 +1,4 @@
|
||||
(** Parsing the command-line options of PascaLIGO *)
|
||||
(** Parsing the command-line options of LIGO *)
|
||||
|
||||
(** The type [command] denotes some possible behaviours of the
|
||||
compiler. The constructors are
|
||||
@ -23,12 +23,11 @@ type command = Quiet | Copy | Units | Tokens
|
||||
(** The type [options] gathers the command-line options.
|
||||
{ul
|
||||
|
||||
{li If the field [input] is [Some src], the name of the
|
||||
PascaLIGO source file, with the extension ".ligo", is
|
||||
[src]. If [input] is [Some "-"] or [None], the source file
|
||||
is read from standard input.}
|
||||
{li If the field [input] is [Some src], the name of the LIGO
|
||||
source file is [src]. If [input] is [Some "-"] or [None],
|
||||
the source file is read from standard input.}
|
||||
|
||||
{li The field [libs] is the paths where to find PascaLIGO files
|
||||
{li The field [libs] is the paths where to find LIGO files
|
||||
for inclusion (#include).}
|
||||
|
||||
{li The field [verbose] is a set of stages of the compiler
|
||||
@ -41,8 +40,14 @@ type command = Quiet | Copy | Units | Tokens
|
||||
{li If the value [mode] is [`Byte], then the unit in which
|
||||
source positions and regions are expressed in messages is
|
||||
the byte. If [`Point], the unit is unicode points.}
|
||||
}
|
||||
*)
|
||||
|
||||
{li If the field [mono] is [true], then the monolithic API of
|
||||
Menhir is called, otherwise the incremental API is.}
|
||||
|
||||
{li If the field [expr] is [true], then the parser for
|
||||
expressions is used, otherwise a full-fledged contract is
|
||||
expected.}
|
||||
} *)
|
||||
type options = <
|
||||
input : string option;
|
||||
libs : string list;
|
||||
@ -50,7 +55,8 @@ type options = <
|
||||
offsets : bool;
|
||||
mode : [`Byte | `Point];
|
||||
cmd : command;
|
||||
mono : bool
|
||||
mono : bool;
|
||||
expr : bool
|
||||
>
|
||||
|
||||
val make :
|
||||
@ -61,6 +67,7 @@ val make :
|
||||
mode:[`Byte | `Point] ->
|
||||
cmd:command ->
|
||||
mono:bool ->
|
||||
expr:bool ->
|
||||
options
|
||||
|
||||
(** Parsing the command-line options on stdin. The first parameter is
|
||||
|
@ -77,8 +77,7 @@ module type TOKEN =
|
||||
val mk_string : lexeme -> Region.t -> token
|
||||
val mk_bytes : lexeme -> Region.t -> token
|
||||
val mk_constr : lexeme -> Region.t -> token
|
||||
val mk_attr : lexeme -> Region.t -> (token, attr_err) result
|
||||
val mk_attr2 : lexeme -> Region.t -> (token, attr_err) result
|
||||
val mk_attr : string -> lexeme -> Region.t -> (token, attr_err) result
|
||||
val eof : Region.t -> token
|
||||
|
||||
(* Predicates *)
|
||||
@ -108,6 +107,8 @@ module type TOKEN =
|
||||
* a function [get_pos] that returns the current position, and
|
||||
* a function [get_last] that returns the region of the last
|
||||
recognised token.
|
||||
* a function [get_file] that returns the name of the file being scanned
|
||||
(empty string if [stdin]).
|
||||
|
||||
Note that a module [Token] is exported too, because the signature
|
||||
of the exported functions depend on it.
|
||||
@ -140,6 +141,7 @@ module type S =
|
||||
get_win : unit -> window;
|
||||
get_pos : unit -> Pos.t;
|
||||
get_last : unit -> Region.t;
|
||||
get_file : unit -> file_path;
|
||||
close : unit -> unit
|
||||
}
|
||||
|
||||
|
@ -119,8 +119,7 @@ module type TOKEN =
|
||||
val mk_string : lexeme -> Region.t -> token
|
||||
val mk_bytes : lexeme -> Region.t -> token
|
||||
val mk_constr : lexeme -> Region.t -> token
|
||||
val mk_attr : lexeme -> Region.t -> (token, attr_err) result
|
||||
val mk_attr2 : lexeme -> Region.t -> (token, attr_err) result
|
||||
val mk_attr : string -> lexeme -> Region.t -> (token, attr_err) result
|
||||
val eof : Region.t -> token
|
||||
|
||||
(* Predicates *)
|
||||
@ -164,6 +163,7 @@ module type S =
|
||||
get_win : unit -> window;
|
||||
get_pos : unit -> Pos.t;
|
||||
get_last : unit -> Region.t;
|
||||
get_file : unit -> file_path;
|
||||
close : unit -> unit
|
||||
}
|
||||
|
||||
@ -177,7 +177,8 @@ module type S =
|
||||
|
||||
exception Error of error Region.reg
|
||||
|
||||
val format_error : ?offsets:bool -> [`Byte | `Point] ->
|
||||
val format_error :
|
||||
?offsets:bool -> [`Byte | `Point] ->
|
||||
error Region.reg -> file:bool -> string
|
||||
end
|
||||
|
||||
@ -441,9 +442,9 @@ module Make (Token: TOKEN) : (S with module Token = Token) =
|
||||
exception Error of error Region.reg
|
||||
|
||||
let format_error ?(offsets=true) mode Region.{region; value} ~file =
|
||||
let msg = error_to_string value in
|
||||
let reg = region#to_string ~file ~offsets mode in
|
||||
sprintf "\027[31mLexical error %s:\n%s\027[0m%!" reg msg
|
||||
let msg = error_to_string value
|
||||
and reg = region#to_string ~file ~offsets mode
|
||||
in sprintf "Lexical error %s:\n%s" reg msg
|
||||
|
||||
let fail region value = raise (Error Region.{region; value})
|
||||
|
||||
@ -530,17 +531,9 @@ module Make (Token: TOKEN) : (S with module Token = Token) =
|
||||
Ok token -> token, state
|
||||
| Error Token.Reserved_name -> fail region (Reserved_name lexeme)
|
||||
|
||||
let mk_attr state buffer attr =
|
||||
let mk_attr header attr state buffer =
|
||||
let region, _, state = sync state buffer in
|
||||
match Token.mk_attr attr region with
|
||||
Ok token ->
|
||||
token, state
|
||||
| Error Token.Invalid_attribute ->
|
||||
fail region Invalid_attribute
|
||||
|
||||
let mk_attr2 state buffer attr =
|
||||
let region, _, state = sync state buffer in
|
||||
match Token.mk_attr2 attr region with
|
||||
match Token.mk_attr header attr region with
|
||||
Ok token ->
|
||||
token, state
|
||||
| Error Token.Invalid_attribute ->
|
||||
@ -579,6 +572,7 @@ let capital = ['A'-'Z']
|
||||
let letter = small | capital
|
||||
let ident = small (letter | '_' | digit)*
|
||||
let constr = capital (letter | '_' | digit)*
|
||||
let attr = ident | constr
|
||||
let hexa_digit = digit | ['A'-'F']
|
||||
let byte = hexa_digit hexa_digit
|
||||
let byte_seq = byte | byte (byte | '_')* byte
|
||||
@ -586,8 +580,8 @@ let bytes = "0x" (byte_seq? as seq)
|
||||
let esc = "\\n" | "\\\"" | "\\\\" | "\\b"
|
||||
| "\\r" | "\\t" | "\\x" byte
|
||||
let pascaligo_sym = "=/=" | '#' | ":="
|
||||
let cameligo_sym = "<>" | "::" | "||" | "&&" | "[@"
|
||||
let reasonligo_sym = '!' | "=>" | "!=" | "==" | "++" | "..." | "||" | "&&" | "[@"
|
||||
let cameligo_sym = "<>" | "::" | "||" | "&&"
|
||||
let reasonligo_sym = '!' | "=>" | "!=" | "==" | "++" | "..." | "||" | "&&"
|
||||
|
||||
let symbol =
|
||||
';' | ',' | '(' | ')'| '[' | ']' | '{' | '}'
|
||||
@ -618,16 +612,19 @@ and scan state = parse
|
||||
| '\t'+ { scan (push_tabs state lexbuf) lexbuf }
|
||||
| ident { mk_ident state lexbuf |> enqueue }
|
||||
| constr { mk_constr state lexbuf |> enqueue }
|
||||
| bytes { (mk_bytes seq) state lexbuf |> enqueue }
|
||||
| bytes { mk_bytes seq state lexbuf |> enqueue }
|
||||
| natural 'n' { mk_nat state lexbuf |> enqueue }
|
||||
| natural "mutez" { mk_mutez state lexbuf |> enqueue }
|
||||
| natural "tz" { mk_tz state lexbuf |> enqueue }
|
||||
| decimal "tz" { mk_tz_decimal state lexbuf |> enqueue }
|
||||
| natural "tz"
|
||||
| natural "tez" { mk_tz state lexbuf |> enqueue }
|
||||
| decimal "tz"
|
||||
| decimal "tez" { mk_tz_decimal state lexbuf |> enqueue }
|
||||
| natural { mk_int state lexbuf |> enqueue }
|
||||
| symbol { mk_sym state lexbuf |> enqueue }
|
||||
| eof { mk_eof state lexbuf |> enqueue }
|
||||
| "[@" (ident|constr as attr) "]" { mk_attr state lexbuf attr |> enqueue }
|
||||
| "[@@" (ident|constr as attr) "]" { mk_attr2 state lexbuf attr |> enqueue }
|
||||
| "[@" (attr as a) "]" { mk_attr "[@" a state lexbuf |> enqueue }
|
||||
| "[@@" (attr as a) "]" { mk_attr "[@@" a state lexbuf |> enqueue }
|
||||
|
||||
| '"' { let opening, _, state = sync state lexbuf in
|
||||
let thread = {opening; len=1; acc=['"']} in
|
||||
scan_string thread state lexbuf |> mk_string |> enqueue }
|
||||
@ -676,8 +673,7 @@ and scan state = parse
|
||||
and file = Filename.basename file in
|
||||
let pos = state.pos#set ~file ~line ~offset:0 in
|
||||
let state = {state with pos} in
|
||||
scan state lexbuf
|
||||
}
|
||||
scan state lexbuf }
|
||||
|
||||
(* Some special errors
|
||||
|
||||
@ -864,6 +860,7 @@ type instance = {
|
||||
get_win : unit -> window;
|
||||
get_pos : unit -> Pos.t;
|
||||
get_last : unit -> Region.t;
|
||||
get_file : unit -> file_path;
|
||||
close : unit -> unit
|
||||
}
|
||||
|
||||
@ -871,7 +868,7 @@ let open_token_stream file_path_opt =
|
||||
let file_path = match file_path_opt with
|
||||
None | Some "-" -> ""
|
||||
| Some file_path -> file_path in
|
||||
let pos = Pos.min#set_file file_path in
|
||||
let pos = Pos.min ~file:file_path in
|
||||
let buf_reg = ref (pos#byte, pos#byte)
|
||||
and first_call = ref true
|
||||
and decoder = Uutf.decoder ~encoding:`UTF_8 `Manual in
|
||||
@ -886,7 +883,8 @@ let open_token_stream file_path_opt =
|
||||
|
||||
let get_pos () = !state.pos
|
||||
and get_last () = !state.last
|
||||
and get_win () = !state.window in
|
||||
and get_win () = !state.window
|
||||
and get_file () = file_path in
|
||||
|
||||
let patch_buffer (start, stop) buffer =
|
||||
let open Lexing in
|
||||
@ -958,7 +956,7 @@ let open_token_stream file_path_opt =
|
||||
None | Some "-" -> ()
|
||||
| Some file_path -> reset ~file:file_path buffer
|
||||
and close () = close_in cin in
|
||||
{read = read_token; buffer; get_win; get_pos; get_last; close}
|
||||
{read = read_token; buffer; get_win; get_pos; get_last; get_file; close}
|
||||
|
||||
end (* of functor [Make] in HEADER *)
|
||||
(* END TRAILER *)
|
||||
|
@ -1,7 +1,5 @@
|
||||
(** Embedding the LIGO lexer in a debug module *)
|
||||
|
||||
let sprintf = Printf.sprintf
|
||||
|
||||
module type S =
|
||||
sig
|
||||
module Lexer : Lexer.S
|
||||
@ -15,12 +13,12 @@ module type S =
|
||||
|
||||
val trace :
|
||||
?offsets:bool -> [`Byte | `Point] ->
|
||||
file_path option -> EvalOpt.command -> unit
|
||||
file_path option -> EvalOpt.command ->
|
||||
(unit, string) Stdlib.result
|
||||
end
|
||||
|
||||
module Make (Lexer: Lexer.S) : (S with module Lexer = Lexer) =
|
||||
struct
|
||||
|
||||
module Lexer = Lexer
|
||||
module Token = Lexer.Token
|
||||
|
||||
@ -49,28 +47,29 @@ module Make (Lexer: Lexer.S) : (S with module Lexer = Lexer) =
|
||||
|
||||
type file_path = string
|
||||
|
||||
let trace ?(offsets=true) mode file_path_opt command : unit =
|
||||
let trace ?(offsets=true) mode file_path_opt command :
|
||||
(unit, string) Stdlib.result =
|
||||
try
|
||||
let Lexer.{read; buffer; close; _} =
|
||||
Lexer.open_token_stream file_path_opt
|
||||
and cout = stdout in
|
||||
let log = output_token ~offsets mode command cout
|
||||
and close_all () = close (); close_out cout in
|
||||
Lexer.open_token_stream file_path_opt in
|
||||
let log = output_token ~offsets mode command stdout
|
||||
and close_all () = close (); close_out stdout in
|
||||
let rec iter () =
|
||||
match read ~log buffer with
|
||||
token ->
|
||||
if Token.is_eof token then close_all ()
|
||||
if Token.is_eof token
|
||||
then Stdlib.Ok ()
|
||||
else iter ()
|
||||
| exception Lexer.Error e ->
|
||||
| exception Lexer.Error error ->
|
||||
let file =
|
||||
match file_path_opt with
|
||||
None | Some "-" -> false
|
||||
| Some _ -> true in
|
||||
let msg =
|
||||
Lexer.format_error ~offsets mode e ~file
|
||||
in prerr_string msg;
|
||||
close_all ()
|
||||
in iter ()
|
||||
with Sys_error msg -> Utils.highlight (sprintf "%s\n" msg)
|
||||
Lexer.format_error ~offsets mode ~file error
|
||||
in Stdlib.Error msg in
|
||||
let result = iter ()
|
||||
in (close_all (); result)
|
||||
with Sys_error msg -> Stdlib.Error msg
|
||||
|
||||
end
|
||||
|
@ -11,7 +11,8 @@ module type S =
|
||||
|
||||
val trace :
|
||||
?offsets:bool -> [`Byte | `Point] ->
|
||||
file_path option -> EvalOpt.command -> unit
|
||||
file_path option -> EvalOpt.command ->
|
||||
(unit, string) Stdlib.result
|
||||
end
|
||||
|
||||
module Make (Lexer: Lexer.S) : S with module Lexer = Lexer
|
||||
|
@ -1,21 +1,20 @@
|
||||
(* Functor to build a standalone LIGO lexer *)
|
||||
|
||||
module type S =
|
||||
module type IO =
|
||||
sig
|
||||
val ext : string (* LIGO file extension *)
|
||||
val options : EvalOpt.options (* CLI options *)
|
||||
end
|
||||
|
||||
module Make (IO: S) (Lexer: Lexer.S) =
|
||||
module Make (IO: IO) (Lexer: Lexer.S) =
|
||||
struct
|
||||
open Printf
|
||||
module SSet = Utils.String.Set
|
||||
|
||||
(* Error printing and exception tracing *)
|
||||
|
||||
let () = Printexc.record_backtrace true
|
||||
|
||||
let external_ text =
|
||||
Utils.highlight (sprintf "External error: %s" text); exit 1
|
||||
|
||||
(* Preprocessing the input source and opening the input channels *)
|
||||
|
||||
(* Path for CPP inclusions (#include) *)
|
||||
@ -48,18 +47,62 @@ module Make (IO: S) (Lexer: Lexer.S) =
|
||||
sprintf "cpp -traditional-cpp%s %s > %s"
|
||||
lib_path file pp_input
|
||||
|
||||
let () =
|
||||
if Utils.String.Set.mem "cpp" IO.options#verbose
|
||||
then eprintf "%s\n%!" cpp_cmd;
|
||||
if Sys.command cpp_cmd <> 0 then
|
||||
external_ (sprintf "the command \"%s\" failed." cpp_cmd)
|
||||
|
||||
(* Running the lexer on the input file *)
|
||||
|
||||
let scan () : (Lexer.token list, string) Stdlib.result =
|
||||
(* Preprocessing the input *)
|
||||
|
||||
if SSet.mem "cpp" IO.options#verbose
|
||||
then eprintf "%s\n%!" cpp_cmd
|
||||
else ();
|
||||
|
||||
if Sys.command cpp_cmd <> 0 then
|
||||
let msg =
|
||||
sprintf "External error: the command \"%s\" failed." cpp_cmd
|
||||
in Stdlib.Error msg
|
||||
else
|
||||
try
|
||||
let Lexer.{read; buffer; close; _} =
|
||||
Lexer.open_token_stream (Some pp_input) in
|
||||
let close_all () = close (); close_out stdout in
|
||||
let rec read_tokens tokens =
|
||||
match read ~log:(fun _ _ -> ()) buffer with
|
||||
token ->
|
||||
if Lexer.Token.is_eof token
|
||||
then Stdlib.Ok (List.rev tokens)
|
||||
else read_tokens (token::tokens)
|
||||
| exception Lexer.Error error ->
|
||||
let file =
|
||||
match IO.options#input with
|
||||
None | Some "-" -> false
|
||||
| Some _ -> true in
|
||||
let msg =
|
||||
Lexer.format_error ~offsets:IO.options#offsets
|
||||
IO.options#mode ~file error
|
||||
in Stdlib.Error msg in
|
||||
let result = read_tokens []
|
||||
in close_all (); result
|
||||
with Sys_error msg -> close_out stdout; Stdlib.Error msg
|
||||
|
||||
(* Tracing the lexing (effectful) *)
|
||||
|
||||
module Log = LexerLog.Make (Lexer)
|
||||
|
||||
let () = Log.trace ~offsets:IO.options#offsets
|
||||
IO.options#mode (Some pp_input)
|
||||
let trace () : (unit, string) Stdlib.result =
|
||||
(* Preprocessing the input *)
|
||||
|
||||
if SSet.mem "cpp" IO.options#verbose
|
||||
then eprintf "%s\n%!" cpp_cmd
|
||||
else ();
|
||||
|
||||
if Sys.command cpp_cmd <> 0 then
|
||||
let msg =
|
||||
sprintf "External error: the command \"%s\" failed." cpp_cmd
|
||||
in Stdlib.Error msg
|
||||
else
|
||||
Log.trace ~offsets:IO.options#offsets
|
||||
IO.options#mode
|
||||
(Some pp_input)
|
||||
IO.options#cmd
|
||||
|
||||
end
|
||||
|
13
src/passes/1-parser/shared/LexerUnit.mli
Normal file
13
src/passes/1-parser/shared/LexerUnit.mli
Normal file
@ -0,0 +1,13 @@
|
||||
(* Functor to build a standalone LIGO lexer *)
|
||||
|
||||
module type IO =
|
||||
sig
|
||||
val ext : string (* LIGO file extension *)
|
||||
val options : EvalOpt.options (* CLI options *)
|
||||
end
|
||||
|
||||
module Make (IO: IO) (Lexer: Lexer.S) :
|
||||
sig
|
||||
val scan : unit -> (Lexer.token list, string) Stdlib.result
|
||||
val trace : unit -> (unit, string) Stdlib.result
|
||||
end
|
163
src/passes/1-parser/shared/Memo.ml
Normal file
163
src/passes/1-parser/shared/Memo.ml
Normal file
@ -0,0 +1,163 @@
|
||||
module Region = Simple_utils.Region
|
||||
|
||||
type macro = {
|
||||
origin : Region.t; (* Not ghost *)
|
||||
current : Region.t (* Maybe ghost *)
|
||||
}
|
||||
|
||||
type location =
|
||||
Loc of Region.t (* Not ghost *)
|
||||
| Link of macro
|
||||
|
||||
(* Regions must not be ghosts and strings must not be empty. *)
|
||||
|
||||
type valid_lexeme = string Region.reg (* Not ghost, not empty. *)
|
||||
type invalid_lexeme = string Region.reg (* Not ghost, empty if EOF. *)
|
||||
|
||||
type phase =
|
||||
Lexer
|
||||
| Parser of valid_lexeme option * invalid_lexeme
|
||||
| Scoping
|
||||
|
||||
type error = <
|
||||
location : location;
|
||||
message : string; (* Sentence ending with a period *)
|
||||
hint : string; (* Suggestion to solve the issue *)
|
||||
help : string (* Off-program help *)
|
||||
>
|
||||
|
||||
type invalid_error = Ghost_region
|
||||
|
||||
let check_loc = function
|
||||
Loc reg ->
|
||||
if reg#is_ghost then
|
||||
Stdlib.Error Ghost_region
|
||||
else Ok ()
|
||||
| Link {origin; _} ->
|
||||
if origin#is_ghost then
|
||||
Stdlib.Error Ghost_region
|
||||
else Ok ()
|
||||
|
||||
let make_error ~location ~message ~hint ~help =
|
||||
match check_loc location with
|
||||
Stdlib.Ok () ->
|
||||
Ok (object
|
||||
method location = location
|
||||
method message = message
|
||||
method hint = hint
|
||||
method help = help
|
||||
end)
|
||||
| Error _ as e -> e
|
||||
|
||||
type warning = <
|
||||
location : location;
|
||||
message : string; (* Sentence ending with a period *)
|
||||
hint : string; (* Idem *)
|
||||
>
|
||||
|
||||
type invalid_warning = invalid_error
|
||||
|
||||
let make_warning ~location ~message ~hint =
|
||||
match check_loc location with
|
||||
Stdlib.Ok () ->
|
||||
Ok (object
|
||||
method location = location
|
||||
method message = message
|
||||
method hint = hint
|
||||
method help = help
|
||||
end)
|
||||
| Error _ as e -> e
|
||||
|
||||
type kind =
|
||||
Error of error (* Failure of an external invariant *)
|
||||
| Internal of string (* Failure of an internal invariant *)
|
||||
| External of string (* Failure of an external process *)
|
||||
| Warning of warning
|
||||
| Info of (unit -> string) (* Log *)
|
||||
|
||||
type entry = <
|
||||
phase : phase;
|
||||
kind : kind
|
||||
>
|
||||
|
||||
type invalid_entry =
|
||||
Ghost_lexeme
|
||||
| Empty_lexeme
|
||||
|
||||
let check_phase = function
|
||||
Parser (Some valid_lexeme, invalid_lexeme) ->
|
||||
let open Region in
|
||||
if valid_lexeme.region#is_ghost
|
||||
|| invalid_lexeme.region#is_ghost
|
||||
then Stdlib.Error Ghost_lexeme
|
||||
else if valid_lexeme.value = ""
|
||||
then Stdlib.Error Empty_lexeme
|
||||
else Ok ()
|
||||
| Parser (None, invalid_lexeme) ->
|
||||
if invalid_lexeme.region#is_ghost
|
||||
then Stdlib.Error Ghost_lexeme
|
||||
else Ok ()
|
||||
| Lexer
|
||||
| Scoping -> Ok ()
|
||||
|
||||
let make_entry ~phase ~kind =
|
||||
match check_phase phase with
|
||||
Stdlib.Error _ as e -> e
|
||||
| Ok () -> Ok (object
|
||||
method phase = phase
|
||||
method kind = kind
|
||||
end)
|
||||
|
||||
type memo = <
|
||||
mode : [`Byte | `Point]; (* Bytes vs UTF-8 *)
|
||||
offsets : bool; (* [true] for horizontal offsets *)
|
||||
log : entry FQueue.t
|
||||
>
|
||||
|
||||
type t = memo
|
||||
|
||||
let empty_memo ~mode ~offsets : memo =
|
||||
object
|
||||
method mode = mode
|
||||
method offsets = offsets
|
||||
method log = FQueue.empty
|
||||
method enqueue entry = {< log = FQueue.enq entry log >}
|
||||
method dequeue =
|
||||
match FQueue.deq log with
|
||||
None -> None
|
||||
| Some (log, entry) -> Some ({< log=log >}, entry)
|
||||
end
|
||||
|
||||
let sprintf = Printf.sprintf
|
||||
|
||||
let string_of_entry ~(file:bool) entry : string =
|
||||
let reg = entry#region#to_string
|
||||
~file
|
||||
~offsets:entry#offsets
|
||||
error#mode in
|
||||
let string =
|
||||
match error#phase with
|
||||
Parser (None, invalid_lexeme) ->
|
||||
(match invalid_lexeme.Region.value with
|
||||
"" -> sprintf "Parse error %s" reg (* EOF *)
|
||||
| lexeme -> sprintf "Parse error %s, before \"%s\""
|
||||
reg lexeme)
|
||||
| Parser (Some valid_lexeme, invalid_lexeme) ->
|
||||
let string =
|
||||
sprintf "Parse error %s, after \"%s\""
|
||||
reg valid_lexeme.Region.value in
|
||||
(match invalid_lexeme.Region.value with
|
||||
"" -> string (* EOF *)
|
||||
| lexeme -> sprintf "%s and before \"%s\"" string lexeme)
|
||||
| Lexer ->
|
||||
sprintf "Lexical error %s" reg
|
||||
| Scoping ->
|
||||
sprintf "Scoping error %s" reg in
|
||||
let string =
|
||||
string
|
||||
^ (if error#message = "" then "."
|
||||
else ":\n" ^ error#message) ^ "\n" in
|
||||
let string =
|
||||
string ^ (if error#hint = "" then ""
|
||||
else sprintf "Hint: %s\n" error#hint)
|
||||
in string
|
120
src/passes/1-parser/shared/Memo.mli
Normal file
120
src/passes/1-parser/shared/Memo.mli
Normal file
@ -0,0 +1,120 @@
|
||||
(* This module defines compilation memos. *)
|
||||
|
||||
(* Locations *)
|
||||
|
||||
module Region = Simple_utils.Region
|
||||
|
||||
type macro = private <
|
||||
origin : Region.t; (* Not ghost *)
|
||||
current : Region.t (* Maybe ghost *)
|
||||
>
|
||||
|
||||
type location = private
|
||||
Loc of Region.t (* Not ghost *)
|
||||
| Link of macro
|
||||
|
||||
type invalid_loc = Ghost_region
|
||||
|
||||
val make_loc :
|
||||
Region.t -> (location, invalid_loc) Stdlib.result
|
||||
|
||||
val make_link :
|
||||
origin:Region.t ->
|
||||
current:Region.t ->
|
||||
(location, invalid_loc) Stdlib.result
|
||||
|
||||
type 'a located = <
|
||||
value : 'a;
|
||||
location : location
|
||||
>
|
||||
|
||||
val make_located : value:'a -> location:location -> 'a located
|
||||
|
||||
(* Lexemes *)
|
||||
|
||||
type lexeme = string location (* Not ghost, empty => EOF *)
|
||||
|
||||
type window = <
|
||||
valid_lexeme : lexeme option;
|
||||
invalid_lexeme : lexeme
|
||||
>
|
||||
|
||||
val make_window : ?valid:lexeme -> invalid:lexeme -> window
|
||||
|
||||
(* Compilation phases *)
|
||||
|
||||
type phase =
|
||||
Lexer
|
||||
| Parser of window
|
||||
| Scoping
|
||||
|
||||
(* Messages *)
|
||||
|
||||
type message = private string
|
||||
|
||||
type invalid_message = Empty_message
|
||||
|
||||
val make_message : string -> (message, invalid_error) Stdlib.result
|
||||
val string_of_message : message -> string
|
||||
|
||||
(* Errors *)
|
||||
|
||||
type error = <
|
||||
location : location;
|
||||
message : message; (* Non-empty string (ending with a period) *)
|
||||
hint : string; (* Suggestion to solve the issue (may be empty) *)
|
||||
help : string (* Off-program help (may be empty) *)
|
||||
>
|
||||
|
||||
val make_error :
|
||||
location:location ->
|
||||
message:message ->
|
||||
hint:string ->
|
||||
help:string ->
|
||||
error
|
||||
|
||||
(* Warnings *)
|
||||
|
||||
type warning = <
|
||||
location : location;
|
||||
message : message; (* Non-empty string (ending with a period) *)
|
||||
hint : string; (* May empty *)
|
||||
>
|
||||
|
||||
val make_warning :
|
||||
location:location ->
|
||||
message:message ->
|
||||
hint:string ->
|
||||
warning
|
||||
|
||||
(* Kinds of entries *)
|
||||
|
||||
type kind =
|
||||
Error of error (* Failure of an external invariant *)
|
||||
| Internal of message (* Failure of an internal invariant (non-empty) *)
|
||||
| External of message (* Failure of an external process (non-empty) *)
|
||||
| Warning of warning
|
||||
| Info of (unit -> message) (* Log (not-empty) *)
|
||||
|
||||
type entry = private <
|
||||
phase : phase;
|
||||
kind : kind
|
||||
>
|
||||
|
||||
val make_entry : phase:phase -> kind:kind -> entry
|
||||
|
||||
val string_of_entry : file:bool -> entry -> string
|
||||
|
||||
(* Memos *)
|
||||
|
||||
type memo = <
|
||||
mode : [`Byte | `Point]; (* Bytes vs UTF-8 *)
|
||||
offsets : bool; (* [true] for horizontal offsets *)
|
||||
log : entry FQueue.t;
|
||||
enqueue : entry -> memo;
|
||||
dequeue : (memo * entry) option
|
||||
>
|
||||
|
||||
type t = memo
|
||||
|
||||
val empty_memo : mode:[`Byte | `Point] -> offsets:bool -> memo
|
@ -77,28 +77,6 @@ module Make (Lexer: Lexer.S)
|
||||
|
||||
exception Point of error
|
||||
|
||||
let failure get_win checkpoint =
|
||||
let message = ParErr.message (state checkpoint) in
|
||||
match get_win () with
|
||||
Lexer.Nil -> assert false
|
||||
| Lexer.One invalid ->
|
||||
raise (Point (message, None, invalid))
|
||||
| Lexer.Two (invalid, valid) ->
|
||||
raise (Point (message, Some valid, invalid))
|
||||
|
||||
(* The two Menhir APIs are called from the following two functions. *)
|
||||
|
||||
let incr_contract Lexer.{read; buffer; get_win; close; _} : Parser.ast =
|
||||
let supplier = I.lexer_lexbuf_to_supplier read buffer
|
||||
and failure = failure get_win in
|
||||
let parser = Parser.Incremental.contract buffer.Lexing.lex_curr_p in
|
||||
let ast = I.loop_handle success failure supplier parser
|
||||
in close (); ast
|
||||
|
||||
let mono_contract = Parser.contract
|
||||
|
||||
(* Errors *)
|
||||
|
||||
let format_error ?(offsets=true) mode (msg, valid_opt, invalid) =
|
||||
let invalid_region = Lexer.Token.to_region invalid in
|
||||
let header =
|
||||
@ -119,9 +97,37 @@ module Make (Lexer: Lexer.S)
|
||||
let header = header ^ trailer in
|
||||
header ^ (if msg = "" then ".\n" else ":\n" ^ msg)
|
||||
|
||||
let short_error ?(offsets=true) mode msg (invalid_region: Region.t) =
|
||||
let () = assert (not (invalid_region#is_ghost)) in
|
||||
let header =
|
||||
"Parse error " ^ invalid_region#to_string ~offsets mode in
|
||||
header ^ (if msg = "" then ".\n" else ":\n" ^ msg)
|
||||
let failure get_win checkpoint =
|
||||
let message = ParErr.message (state checkpoint) in
|
||||
match get_win () with
|
||||
Lexer.Nil -> assert false
|
||||
| Lexer.One invalid ->
|
||||
raise (Point (message, None, invalid))
|
||||
| Lexer.Two (invalid, valid) ->
|
||||
raise (Point (message, Some valid, invalid))
|
||||
|
||||
(* The monolithic API of Menhir *)
|
||||
|
||||
let mono_contract = Parser.contract
|
||||
|
||||
let mono_expr = Parser.interactive_expr
|
||||
|
||||
(* Incremental API of Menhir *)
|
||||
|
||||
module Incr = Parser.Incremental
|
||||
|
||||
let incr_contract Lexer.{read; buffer; get_win; close; _} =
|
||||
let supplier = I.lexer_lexbuf_to_supplier read buffer
|
||||
and failure = failure get_win in
|
||||
let parser = Incr.contract buffer.Lexing.lex_curr_p in
|
||||
let ast = I.loop_handle success failure supplier parser
|
||||
in close (); ast
|
||||
|
||||
let incr_expr Lexer.{read; buffer; get_win; close; _} =
|
||||
let supplier = I.lexer_lexbuf_to_supplier read buffer
|
||||
and failure = failure get_win in
|
||||
let parser = Incr.interactive_expr buffer.Lexing.lex_curr_p in
|
||||
let expr = I.loop_handle success failure supplier parser
|
||||
in close (); expr
|
||||
|
||||
end
|
||||
|
@ -2,6 +2,9 @@
|
||||
|
||||
module Region = Simple_utils.Region
|
||||
|
||||
(* The signature generated by Menhir with additional type definitions
|
||||
for [ast] and [expr]. *)
|
||||
|
||||
module type PARSER =
|
||||
sig
|
||||
(* The type of tokens. *)
|
||||
@ -16,8 +19,10 @@ module type PARSER =
|
||||
|
||||
(* The monolithic API. *)
|
||||
|
||||
val interactive_expr : (Lexing.lexbuf -> token) -> Lexing.lexbuf -> expr
|
||||
val contract : (Lexing.lexbuf -> token) -> Lexing.lexbuf -> ast
|
||||
val interactive_expr :
|
||||
(Lexing.lexbuf -> token) -> Lexing.lexbuf -> expr
|
||||
val contract :
|
||||
(Lexing.lexbuf -> token) -> Lexing.lexbuf -> ast
|
||||
|
||||
(* The incremental API. *)
|
||||
|
||||
@ -42,14 +47,15 @@ module Make (Lexer: Lexer.S)
|
||||
(Parser: PARSER with type token = Lexer.Token.token)
|
||||
(ParErr: sig val message : int -> string end) :
|
||||
sig
|
||||
(* Monolithic and incremental APIs of Menhir for parsing *)
|
||||
(* The monolithic API of Menhir *)
|
||||
|
||||
val mono_contract :
|
||||
(Lexing.lexbuf -> Lexer.token) -> Lexing.lexbuf -> Parser.ast
|
||||
val incr_contract :
|
||||
Lexer.instance -> Parser.ast
|
||||
|
||||
(* Error handling *)
|
||||
val mono_expr :
|
||||
(Lexing.lexbuf -> Lexer.token) -> Lexing.lexbuf -> Parser.expr
|
||||
|
||||
(* Incremental API of Menhir *)
|
||||
|
||||
type message = string
|
||||
type valid = Parser.token
|
||||
@ -58,9 +64,8 @@ module Make (Lexer: Lexer.S)
|
||||
|
||||
exception Point of error
|
||||
|
||||
val format_error :
|
||||
?offsets:bool -> [`Byte | `Point] -> error -> string
|
||||
val incr_contract : Lexer.instance -> Parser.ast
|
||||
val incr_expr : Lexer.instance -> Parser.expr
|
||||
|
||||
val short_error :
|
||||
?offsets:bool -> [`Byte | `Point] -> message -> Region.t -> string
|
||||
val format_error : ?offsets:bool -> [`Point | `Byte] -> error -> string
|
||||
end
|
||||
|
@ -1,6 +1,8 @@
|
||||
(* Functor to build a standalone LIGO parser *)
|
||||
|
||||
module type S =
|
||||
module Region = Simple_utils.Region
|
||||
|
||||
module type IO =
|
||||
sig
|
||||
val ext : string (* LIGO file extension *)
|
||||
val options : EvalOpt.options (* CLI options *)
|
||||
@ -10,40 +12,35 @@ module type Pretty =
|
||||
sig
|
||||
type state
|
||||
type ast
|
||||
val pp_ast :
|
||||
state -> ast -> unit
|
||||
type expr
|
||||
|
||||
val mk_state :
|
||||
offsets:bool -> mode:[`Point|`Byte] -> buffer:Buffer.t -> state
|
||||
val print_tokens :
|
||||
state -> ast -> unit
|
||||
|
||||
val pp_ast : state -> ast -> unit
|
||||
val pp_expr : state -> expr -> unit
|
||||
val print_tokens : state -> ast -> unit
|
||||
val print_expr : state -> expr -> unit
|
||||
end
|
||||
|
||||
module Make (IO: S)
|
||||
(Lexer: Lexer.S)
|
||||
module Make (Lexer: Lexer.S)
|
||||
(AST: sig type t type expr end)
|
||||
(Parser: ParserAPI.PARSER
|
||||
with type ast = AST.t
|
||||
and type expr = AST.expr
|
||||
and type token = Lexer.token)
|
||||
(ParErr: sig val message : int -> string end)
|
||||
(ParserLog: Pretty with type ast = AST.t) =
|
||||
(ParserLog: Pretty with type ast = AST.t
|
||||
and type expr = AST.expr)
|
||||
(IO: IO) =
|
||||
struct
|
||||
open Printf
|
||||
module SSet = Utils.String.Set
|
||||
|
||||
(* Error printing and exception tracing *)
|
||||
|
||||
let () = Printexc.record_backtrace true
|
||||
|
||||
let external_ text =
|
||||
Utils.highlight (sprintf "External error: %s" text); exit 1
|
||||
|
||||
(* Extracting the input file *)
|
||||
|
||||
let file =
|
||||
match IO.options#input with
|
||||
None | Some "-" -> false
|
||||
| Some _ -> true
|
||||
|
||||
(* Preprocessing the input source and opening the input channels *)
|
||||
|
||||
(* Path for CPP inclusions (#include) *)
|
||||
@ -62,9 +59,10 @@ module Make (IO: S)
|
||||
let suffix = ".pp" ^ IO.ext
|
||||
|
||||
let pp_input =
|
||||
if Utils.String.Set.mem "cpp" IO.options#verbose
|
||||
if SSet.mem "cpp" IO.options#verbose
|
||||
then prefix ^ suffix
|
||||
else let pp_input, pp_out = Filename.open_temp_file prefix suffix
|
||||
else let pp_input, pp_out =
|
||||
Filename.open_temp_file prefix suffix
|
||||
in close_out pp_out; pp_input
|
||||
|
||||
let cpp_cmd =
|
||||
@ -76,100 +74,161 @@ module Make (IO: S)
|
||||
sprintf "cpp -traditional-cpp%s %s > %s"
|
||||
lib_path file pp_input
|
||||
|
||||
let () =
|
||||
if Utils.String.Set.mem "cpp" IO.options#verbose
|
||||
then eprintf "%s\n%!" cpp_cmd;
|
||||
if Sys.command cpp_cmd <> 0 then
|
||||
external_ (sprintf "the command \"%s\" failed." cpp_cmd)
|
||||
(* Error handling (reexported from [ParserAPI]) *)
|
||||
|
||||
(* Instanciating the lexer *)
|
||||
type message = string
|
||||
type valid = Parser.token
|
||||
type invalid = Parser.token
|
||||
type error = message * valid option * invalid
|
||||
|
||||
module ParserFront = ParserAPI.Make (Lexer) (Parser) (ParErr)
|
||||
exception Point of error
|
||||
|
||||
let format_error = ParserFront.format_error
|
||||
let short_error = ParserFront.short_error
|
||||
(* Instantiating the parser *)
|
||||
|
||||
let lexer_inst = Lexer.open_token_stream (Some pp_input)
|
||||
let Lexer.{read; buffer; get_win; get_pos; get_last; close} = lexer_inst
|
||||
module Front = ParserAPI.Make (Lexer)(Parser)(ParErr)
|
||||
|
||||
and cout = stdout
|
||||
let format_error = Front.format_error
|
||||
|
||||
let close_all () = close (); close_out cout
|
||||
let short_error ?(offsets=true) mode msg (reg: Region.t) =
|
||||
sprintf "Parse error %s:\n%s" (reg#to_string ~offsets mode) msg
|
||||
|
||||
(* Tokeniser *)
|
||||
(* Parsing an expression *)
|
||||
|
||||
module Log = LexerLog.Make (Lexer)
|
||||
|
||||
let log = Log.output_token ~offsets:IO.options#offsets
|
||||
IO.options#mode IO.options#cmd cout
|
||||
|
||||
let tokeniser = read ~log
|
||||
|
||||
(* Main *)
|
||||
|
||||
let run () =
|
||||
let parse_expr lexer_inst tokeniser output state :
|
||||
(AST.expr, string) Stdlib.result =
|
||||
let close_all () =
|
||||
lexer_inst.Lexer.close (); close_out stdout in
|
||||
let lexbuf = lexer_inst.Lexer.buffer in
|
||||
let expr =
|
||||
try
|
||||
if IO.options#mono then
|
||||
Front.mono_expr tokeniser lexbuf
|
||||
else
|
||||
Front.incr_expr lexer_inst
|
||||
with exn -> close_all (); raise exn in
|
||||
let () =
|
||||
if SSet.mem "ast-tokens" IO.options#verbose then
|
||||
begin
|
||||
Buffer.clear output;
|
||||
ParserLog.print_expr state expr;
|
||||
Buffer.output_buffer stdout output
|
||||
end in
|
||||
let () =
|
||||
if SSet.mem "ast" IO.options#verbose then
|
||||
begin
|
||||
Buffer.clear output;
|
||||
ParserLog.pp_expr state expr;
|
||||
Buffer.output_buffer stdout output
|
||||
end
|
||||
in close_all (); Ok expr
|
||||
|
||||
(* Parsing a contract *)
|
||||
|
||||
let parse_contract lexer_inst tokeniser output state
|
||||
: (AST.t, string) Stdlib.result =
|
||||
let close_all () =
|
||||
lexer_inst.Lexer.close (); close_out stdout in
|
||||
let lexbuf = lexer_inst.Lexer.buffer in
|
||||
let ast =
|
||||
if IO.options#mono
|
||||
then ParserFront.mono_contract tokeniser buffer
|
||||
else ParserFront.incr_contract lexer_inst in
|
||||
if Utils.String.Set.mem "ast" IO.options#verbose
|
||||
then let buffer = Buffer.create 131 in
|
||||
let state = ParserLog.mk_state
|
||||
~offsets:IO.options#offsets
|
||||
~mode:IO.options#mode
|
||||
~buffer in
|
||||
begin
|
||||
ParserLog.pp_ast state ast;
|
||||
Buffer.output_buffer stdout buffer
|
||||
end
|
||||
else if Utils.String.Set.mem "ast-tokens" IO.options#verbose
|
||||
then let buffer = Buffer.create 131 in
|
||||
let state = ParserLog.mk_state
|
||||
~offsets:IO.options#offsets
|
||||
~mode:IO.options#mode
|
||||
~buffer in
|
||||
try
|
||||
if IO.options#mono then
|
||||
Front.mono_contract tokeniser lexbuf
|
||||
else
|
||||
Front.incr_contract lexer_inst
|
||||
with exn -> close_all (); raise exn in
|
||||
let () =
|
||||
if SSet.mem "ast-tokens" IO.options#verbose then
|
||||
begin
|
||||
Buffer.clear output;
|
||||
ParserLog.print_tokens state ast;
|
||||
Buffer.output_buffer stdout buffer
|
||||
Buffer.output_buffer stdout output
|
||||
end in
|
||||
let () =
|
||||
if SSet.mem "ast" IO.options#verbose then
|
||||
begin
|
||||
Buffer.clear output;
|
||||
ParserLog.pp_ast state ast;
|
||||
Buffer.output_buffer stdout output
|
||||
end
|
||||
with
|
||||
in close_all (); Ok ast
|
||||
|
||||
(* Wrapper for the parsers above *)
|
||||
|
||||
let parse parser =
|
||||
(* Preprocessing the input *)
|
||||
|
||||
if SSet.mem "cpp" IO.options#verbose
|
||||
then eprintf "%s\n%!" cpp_cmd
|
||||
else ();
|
||||
|
||||
if Sys.command cpp_cmd <> 0 then
|
||||
let msg =
|
||||
sprintf "External error: \"%s\" failed." cpp_cmd
|
||||
in Stdlib.Error msg
|
||||
else
|
||||
(* Instantiating the lexer *)
|
||||
|
||||
let lexer_inst = Lexer.open_token_stream (Some pp_input) in
|
||||
|
||||
(* Making the tokeniser *)
|
||||
|
||||
let module Log = LexerLog.Make (Lexer) in
|
||||
|
||||
let log =
|
||||
Log.output_token ~offsets:IO.options#offsets
|
||||
IO.options#mode IO.options#cmd stdout in
|
||||
|
||||
let tokeniser = lexer_inst.Lexer.read ~log in
|
||||
|
||||
let output = Buffer.create 131 in
|
||||
let state = ParserLog.mk_state
|
||||
~offsets:IO.options#offsets
|
||||
~mode:IO.options#mode
|
||||
~buffer:output in
|
||||
|
||||
(* Calling the specific parser (that is, the parameter) *)
|
||||
|
||||
match parser lexer_inst tokeniser output state with
|
||||
Stdlib.Error _ as error -> error
|
||||
| Stdlib.Ok _ as node -> node
|
||||
|
||||
(* Lexing errors *)
|
||||
|
||||
Lexer.Error err ->
|
||||
close_all ();
|
||||
let msg =
|
||||
| exception Lexer.Error err ->
|
||||
let file =
|
||||
match IO.options#input with
|
||||
None | Some "-" -> false
|
||||
| Some _ -> true in
|
||||
let error =
|
||||
Lexer.format_error ~offsets:IO.options#offsets
|
||||
IO.options#mode err ~file
|
||||
in prerr_string msg
|
||||
in Stdlib.Error error
|
||||
|
||||
(* Incremental API of Menhir *)
|
||||
|
||||
| ParserFront.Point point ->
|
||||
let () = close_all () in
|
||||
| exception Front.Point point ->
|
||||
let error =
|
||||
ParserFront.format_error ~offsets:IO.options#offsets
|
||||
Front.format_error ~offsets:IO.options#offsets
|
||||
IO.options#mode point
|
||||
in eprintf "\027[31m%s\027[0m%!" error
|
||||
in Stdlib.Error error
|
||||
|
||||
(* Monolithic API of Menhir *)
|
||||
|
||||
| Parser.Error ->
|
||||
let () = close_all () in
|
||||
| exception Parser.Error ->
|
||||
let invalid, valid_opt =
|
||||
match get_win () with
|
||||
match lexer_inst.Lexer.get_win () with
|
||||
Lexer.Nil ->
|
||||
assert false (* Safe: There is always at least EOF. *)
|
||||
| Lexer.One invalid -> invalid, None
|
||||
| Lexer.Two (invalid, valid) -> invalid, Some valid in
|
||||
let point = "", valid_opt, invalid in
|
||||
let error =
|
||||
ParserFront.format_error ~offsets:IO.options#offsets
|
||||
Front.format_error ~offsets:IO.options#offsets
|
||||
IO.options#mode point
|
||||
in eprintf "\027[31m%s\027[0m%!" error
|
||||
in Stdlib.Error error
|
||||
|
||||
(* I/O errors *)
|
||||
|
||||
| Sys_error msg -> Utils.highlight msg
|
||||
| exception Sys_error error -> Stdlib.Error error
|
||||
|
||||
end
|
||||
|
71
src/passes/1-parser/shared/ParserUnit.mli
Normal file
71
src/passes/1-parser/shared/ParserUnit.mli
Normal file
@ -0,0 +1,71 @@
|
||||
(* Functor to build a standalone LIGO parser *)
|
||||
|
||||
module Region = Simple_utils.Region
|
||||
|
||||
module type IO =
|
||||
sig
|
||||
val ext : string (* LIGO file extension *)
|
||||
val options : EvalOpt.options (* CLI options *)
|
||||
end
|
||||
|
||||
module type Pretty =
|
||||
sig
|
||||
type state
|
||||
type ast
|
||||
type expr
|
||||
|
||||
val mk_state :
|
||||
offsets:bool -> mode:[`Point|`Byte] -> buffer:Buffer.t -> state
|
||||
|
||||
val pp_ast : state -> ast -> unit
|
||||
val pp_expr : state -> expr -> unit
|
||||
val print_tokens : state -> ast -> unit
|
||||
val print_expr : state -> expr -> unit
|
||||
end
|
||||
|
||||
module Make (Lexer: Lexer.S)
|
||||
(AST: sig type t type expr end)
|
||||
(Parser: ParserAPI.PARSER
|
||||
with type ast = AST.t
|
||||
and type expr = AST.expr
|
||||
and type token = Lexer.token)
|
||||
(ParErr: sig val message : int -> string end)
|
||||
(ParserLog: Pretty with type ast = AST.t
|
||||
and type expr = AST.expr)
|
||||
(IO: IO) :
|
||||
sig
|
||||
(* Error handling (reexported from [ParserAPI]) *)
|
||||
|
||||
type message = string
|
||||
type valid = Parser.token
|
||||
type invalid = Parser.token
|
||||
type error = message * valid option * invalid
|
||||
|
||||
exception Point of error
|
||||
|
||||
val format_error :
|
||||
?offsets:bool -> [`Byte | `Point] -> error -> string
|
||||
|
||||
val short_error :
|
||||
?offsets:bool -> [`Point | `Byte] -> string -> Region.t -> string
|
||||
|
||||
(* Parsers *)
|
||||
|
||||
val parse :
|
||||
(Lexer.instance ->
|
||||
(Lexing.lexbuf -> Lexer.token) ->
|
||||
Buffer.t -> ParserLog.state -> ('a, string) result) ->
|
||||
('a, string) result
|
||||
|
||||
val parse_contract :
|
||||
Lexer.instance ->
|
||||
(Lexing.lexbuf -> Lexer.token) ->
|
||||
Buffer.t -> ParserLog.state ->
|
||||
(AST.t, string) Stdlib.result
|
||||
|
||||
val parse_expr :
|
||||
Lexer.instance ->
|
||||
(Lexing.lexbuf -> Lexer.token) ->
|
||||
Buffer.t -> ParserLog.state -> (AST.expr, string) Stdlib.result
|
||||
|
||||
end
|
@ -12,8 +12,5 @@
|
||||
(preprocess
|
||||
(pps
|
||||
ppx_let
|
||||
bisect_ppx --conditional
|
||||
)
|
||||
)
|
||||
(flags (:standard -w +1..62-4-9-44-40-42-48-30@39@33 -open Simple_utils ))
|
||||
)
|
||||
bisect_ppx --conditional))
|
||||
(flags (:standard -w +1..62-4-9-44-40-42-48-30@39@33 -open Simple_utils)))
|
||||
|
@ -4,13 +4,14 @@ open Ast_simplified
|
||||
module Raw = Parser.Pascaligo.AST
|
||||
module SMap = Map.String
|
||||
module SSet = Set.Make (String)
|
||||
module ParserLog = Parser_pascaligo.ParserLog
|
||||
|
||||
open Combinators
|
||||
|
||||
let nseq_to_list (hd, tl) = hd :: tl
|
||||
let npseq_to_list (hd, tl) = hd :: (List.map snd tl)
|
||||
let pseq_to_list = function
|
||||
| None -> []
|
||||
None -> []
|
||||
| Some lst -> npseq_to_list lst
|
||||
let get_value : 'a Raw.reg -> 'a = fun x -> x.value
|
||||
let is_compiler_generated name = String.contains (Var.to_name name) '#'
|
||||
@ -132,7 +133,7 @@ module Errors = struct
|
||||
fun () -> Format.asprintf "%a" Location.pp_lift @@ pattern_loc) ;
|
||||
(** TODO: The labelled arguments should be flowing from the CLI. *)
|
||||
("pattern",
|
||||
fun () -> Parser.Pascaligo.ParserLog.pattern_to_string
|
||||
fun () -> ParserLog.pattern_to_string
|
||||
~offsets:true ~mode:`Point p)
|
||||
] in
|
||||
error ~data title message
|
||||
@ -168,7 +169,7 @@ module Errors = struct
|
||||
(** TODO: The labelled arguments should be flowing from the CLI. *)
|
||||
let data = [
|
||||
("instruction",
|
||||
fun () -> Parser.Pascaligo.ParserLog.instruction_to_string
|
||||
fun () -> ParserLog.instruction_to_string
|
||||
~offsets:true ~mode:`Point t)
|
||||
] in
|
||||
error ~data title message
|
||||
@ -562,10 +563,11 @@ and simpl_tuple_expression ?loc (lst:Raw.expr list) : expression result =
|
||||
| [] -> return @@ e_literal Literal_unit
|
||||
| [hd] -> simpl_expression hd
|
||||
| lst ->
|
||||
let%bind lst = bind_list @@ List.map simpl_expression lst in
|
||||
return @@ e_tuple ?loc lst
|
||||
let%bind lst = bind_list @@ List.map simpl_expression lst
|
||||
in return @@ e_tuple ?loc lst
|
||||
|
||||
and simpl_data_declaration : Raw.data_decl -> _ result = fun t ->
|
||||
and simpl_data_declaration : Raw.data_decl -> _ result =
|
||||
fun t ->
|
||||
match t with
|
||||
| LocalVar x ->
|
||||
let (x , loc) = r_split x in
|
||||
@ -578,15 +580,26 @@ and simpl_data_declaration : Raw.data_decl -> _ result = fun t ->
|
||||
let name = x.name.value in
|
||||
let%bind t = simpl_type_expression x.const_type in
|
||||
let%bind expression = simpl_expression x.init in
|
||||
let inline = List.exists (fun (f: Raw.attribute) -> f.value = "\"inline\"") x.attributes.value in
|
||||
return_let_in ~loc (Var.of_name name , Some t) inline expression
|
||||
let inline =
|
||||
match x.attributes with
|
||||
None -> false
|
||||
| Some {value; _} ->
|
||||
npseq_to_list value.ne_elements
|
||||
|> List.exists (fun Region.{value; _} -> value = "\"inline\"")
|
||||
in return_let_in ~loc (Var.of_name name, Some t) inline expression
|
||||
| LocalFun f ->
|
||||
let (f , loc) = r_split f in
|
||||
let%bind (binder, expr) = simpl_fun_decl ~loc f in
|
||||
let inline = List.exists (fun (f: Raw.attribute) -> f.value = "\"inline\"") f.attributes.value in
|
||||
return_let_in ~loc binder inline expr
|
||||
let inline =
|
||||
match f.attributes with
|
||||
None -> false
|
||||
| Some {value; _} ->
|
||||
npseq_to_list value.ne_elements
|
||||
|> List.exists (fun Region.{value; _} -> value = "\"inline\"")
|
||||
in return_let_in ~loc binder inline expr
|
||||
|
||||
and simpl_param : Raw.param_decl -> (expression_variable * type_expression) result =
|
||||
and simpl_param :
|
||||
Raw.param_decl -> (expression_variable * type_expression) result =
|
||||
fun t ->
|
||||
match t with
|
||||
| ParamConst c ->
|
||||
@ -601,11 +614,18 @@ and simpl_param : Raw.param_decl -> (expression_variable * type_expression) resu
|
||||
ok (type_name , type_expression)
|
||||
|
||||
and simpl_fun_decl :
|
||||
loc:_ -> Raw.fun_decl -> ((expression_variable * type_expression option) * expression) result =
|
||||
loc:_ -> Raw.fun_decl ->
|
||||
((expression_variable * type_expression option) * expression) result =
|
||||
fun ~loc x ->
|
||||
let open! Raw in
|
||||
let {fun_name;param;ret_type;block_with;return; attributes} : fun_decl = x in
|
||||
let inline = List.exists (fun (a: Raw.attribute) -> a.value = "\"inline\"") attributes.value in
|
||||
let {fun_name; param; ret_type; block_with;
|
||||
return; attributes} : fun_decl = x in
|
||||
let inline =
|
||||
match attributes with
|
||||
None -> false
|
||||
| Some {value; _} ->
|
||||
npseq_to_list value.ne_elements
|
||||
|> List.exists (fun Region.{value; _} -> value = "\"inline\"") in
|
||||
let statements =
|
||||
match block_with with
|
||||
| Some (block,_) -> npseq_to_list block.value.statements
|
||||
@ -615,9 +635,7 @@ and simpl_fun_decl :
|
||||
a, [] -> (
|
||||
let%bind input = simpl_param a in
|
||||
let (binder , input_type) = input in
|
||||
let%bind instructions = bind_list
|
||||
@@ List.map simpl_statement
|
||||
@@ statements in
|
||||
let%bind instructions = simpl_statement_list statements in
|
||||
let%bind result = simpl_expression return in
|
||||
let%bind output_type = simpl_type_expression ret_type in
|
||||
let body = instructions in
|
||||
@ -647,9 +665,7 @@ and simpl_fun_decl :
|
||||
ass
|
||||
in
|
||||
bind_list @@ List.mapi aux params in
|
||||
let%bind instructions = bind_list
|
||||
@@ List.map simpl_statement
|
||||
@@ statements in
|
||||
let%bind instructions = simpl_statement_list statements in
|
||||
let%bind result = simpl_expression return in
|
||||
let%bind output_type = simpl_type_expression ret_type in
|
||||
let body = tpl_declarations @ instructions in
|
||||
@ -673,9 +689,7 @@ and simpl_fun_expression :
|
||||
a, [] -> (
|
||||
let%bind input = simpl_param a in
|
||||
let (binder , input_type) = input in
|
||||
let%bind instructions = bind_list
|
||||
@@ List.map simpl_statement
|
||||
@@ statements in
|
||||
let%bind instructions = simpl_statement_list statements in
|
||||
let%bind result = simpl_expression return in
|
||||
let%bind output_type = simpl_type_expression ret_type in
|
||||
let body = instructions in
|
||||
@ -705,9 +719,7 @@ and simpl_fun_expression :
|
||||
ass
|
||||
in
|
||||
bind_list @@ List.mapi aux params in
|
||||
let%bind instructions = bind_list
|
||||
@@ List.map simpl_statement
|
||||
@@ statements in
|
||||
let%bind instructions = simpl_statement_list statements in
|
||||
let%bind result = simpl_expression return in
|
||||
let%bind output_type = simpl_type_expression ret_type in
|
||||
let body = tpl_declarations @ instructions in
|
||||
@ -721,44 +733,39 @@ and simpl_fun_expression :
|
||||
)
|
||||
)
|
||||
|
||||
and simpl_declaration : Raw.declaration -> declaration Location.wrap result =
|
||||
fun t ->
|
||||
let open! Raw in
|
||||
match t with
|
||||
| TypeDecl x ->
|
||||
let decl, loc = r_split x in
|
||||
let {name;type_expr} : Raw.type_decl = decl in
|
||||
let%bind type_expression = simpl_type_expression type_expr in
|
||||
ok @@ Location.wrap ~loc (Declaration_type
|
||||
(Var.of_name name.value, type_expression))
|
||||
|
||||
| ConstDecl x ->
|
||||
let simpl_const_decl = fun {name;const_type; init; attributes} ->
|
||||
let%bind expression = simpl_expression init in
|
||||
let%bind t = simpl_type_expression const_type in
|
||||
let type_annotation = Some t in
|
||||
let inline = List.exists (fun (a: Raw.attribute) -> a.value = "\"inline\"") attributes.value in
|
||||
ok @@ Declaration_constant
|
||||
(Var.of_name name.value, type_annotation, inline, expression)
|
||||
in bind_map_location simpl_const_decl (Location.lift_region x)
|
||||
| FunDecl x ->
|
||||
let decl, loc = r_split x in
|
||||
let%bind ((name, ty_opt), expr) = simpl_fun_decl ~loc decl in
|
||||
let inline = List.exists (fun (a: Raw.attribute) -> a.value = "\"inline\"") x.value.attributes.value in
|
||||
ok @@ Location.wrap ~loc (Declaration_constant (name, ty_opt, inline, expr))
|
||||
|
||||
and simpl_statement : Raw.statement -> (_ -> expression result) result =
|
||||
fun s ->
|
||||
match s with
|
||||
| Instr i -> simpl_instruction i
|
||||
| Data d -> simpl_data_declaration d
|
||||
and simpl_statement_list statements =
|
||||
let open Raw in
|
||||
let rec hook acc = function
|
||||
[] -> acc
|
||||
| [Attr _] ->
|
||||
(* Detached attributes are erased. TODO: Warning. *)
|
||||
acc
|
||||
| Attr _ :: (Attr _ :: _ as statements) ->
|
||||
(* Detached attributes are erased. TODO: Warning. *)
|
||||
hook acc statements
|
||||
| Attr decl :: Data (LocalConst {value; region}) :: statements ->
|
||||
let new_const =
|
||||
Data (LocalConst {value = {value with attributes = Some decl}; region})
|
||||
in hook acc (new_const :: statements)
|
||||
| Attr decl :: Data (LocalFun {value; region}) :: statements ->
|
||||
let new_fun =
|
||||
Data (LocalFun {value = {value with attributes = Some decl}; region})
|
||||
in hook acc (new_fun :: statements)
|
||||
| Attr _ :: statements ->
|
||||
(* Detached attributes are erased. TODO: Warning. *)
|
||||
hook acc statements
|
||||
| Instr i :: statements ->
|
||||
hook (simpl_instruction i :: acc) statements
|
||||
| Data d :: statements ->
|
||||
hook (simpl_data_declaration d :: acc) statements
|
||||
in bind_list @@ hook [] (List.rev statements)
|
||||
|
||||
and simpl_single_instruction : Raw.instruction -> (_ -> expression result) result =
|
||||
fun t ->
|
||||
match t with
|
||||
| ProcCall x -> (
|
||||
let ((f, args) , loc) = r_split x in
|
||||
let (args , args_loc) = r_split args in
|
||||
let (f, args) , loc = r_split x in
|
||||
let args, args_loc = r_split args in
|
||||
let args' = npseq_to_list args.inside in
|
||||
match f with
|
||||
| EVar name -> (
|
||||
@ -1057,10 +1064,10 @@ and simpl_cases : type a . (Raw.pattern * a) list -> (a, unit) matching result =
|
||||
let aux (x , y) =
|
||||
let error =
|
||||
let title () = "Pattern" in
|
||||
(** TODO: The labelled arguments should be flowing from the CLI. *)
|
||||
(* TODO: The labelled arguments should be flowing from the CLI. *)
|
||||
let content () =
|
||||
Printf.sprintf "Pattern : %s"
|
||||
(Parser.Pascaligo.ParserLog.pattern_to_string
|
||||
(ParserLog.pattern_to_string
|
||||
~offsets:true ~mode:`Point x) in
|
||||
error title content in
|
||||
let%bind x' =
|
||||
@ -1071,23 +1078,22 @@ and simpl_cases : type a . (Raw.pattern * a) list -> (a, unit) matching result =
|
||||
ok @@ ez_match_variant constrs
|
||||
|
||||
and simpl_instruction : Raw.instruction -> (_ -> expression result) result =
|
||||
fun t ->
|
||||
trace (simplifying_instruction t) @@ simpl_single_instruction t
|
||||
fun t -> trace (simplifying_instruction t) @@ simpl_single_instruction t
|
||||
|
||||
and simpl_statements : Raw.statements -> (_ -> expression result) result =
|
||||
fun ss ->
|
||||
let lst = npseq_to_list ss in
|
||||
let%bind fs = bind_map_list simpl_statement lst in
|
||||
fun statements ->
|
||||
let lst = npseq_to_list statements in
|
||||
let%bind fs = simpl_statement_list lst in
|
||||
let aux : _ -> (expression option -> expression result) -> _ =
|
||||
fun prec cur ->
|
||||
let%bind res = cur prec in
|
||||
ok @@ Some res in
|
||||
let%bind res = cur prec
|
||||
in ok @@ Some res in
|
||||
ok @@ fun (expr' : _ option) ->
|
||||
let%bind ret = bind_fold_right_list aux expr' fs in
|
||||
ok @@ Option.unopt_exn ret
|
||||
|
||||
and simpl_block : Raw.block -> (_ -> expression result) result = fun t ->
|
||||
simpl_statements t.statements
|
||||
and simpl_block : Raw.block -> (_ -> expression result) result =
|
||||
fun t -> simpl_statements t.statements
|
||||
|
||||
and simpl_for_int : Raw.for_int -> (_ -> expression result) result = fun fi ->
|
||||
(* cond part *)
|
||||
@ -1263,11 +1269,13 @@ and simpl_for_collect : Raw.for_collect -> (_ -> expression result) result = fun
|
||||
(* STEP 5 *)
|
||||
let rec add_return (expr : expression) = match expr.expression with
|
||||
| E_sequence (a,b) -> e_sequence a (add_return b)
|
||||
| _ -> e_sequence expr (e_variable (Var.of_name "#COMPILER#acc")) in (* TODO fresh *)
|
||||
| _ -> (* TODO fresh *)
|
||||
e_sequence expr (e_variable (Var.of_name "#COMPILER#acc")) in
|
||||
let for_body = add_return for_body in
|
||||
(* STEP 6 *)
|
||||
let for_body =
|
||||
let ( arg_access: Types.access_path -> expression ) = e_accessor (e_variable (Var.of_name "arguments")) in (* TODO fresh *)
|
||||
let ( arg_access: Types.access_path -> expression ) =
|
||||
e_accessor (e_variable (Var.of_name "arguments")) in (* TODO fresh *)
|
||||
( match fc.collection with
|
||||
| Map _ ->
|
||||
let acc = arg_access [Access_tuple 0 ] in
|
||||
@ -1290,7 +1298,8 @@ and simpl_for_collect : Raw.for_collect -> (_ -> expression result) result = fun
|
||||
let fold = e_constant op_name [lambda; collect ; init_record] in
|
||||
(* STEP 8 *)
|
||||
let assign_back (prev : expression option) (captured_varname : string) : expression option =
|
||||
let access = e_accessor (e_variable (Var.of_name "#COMPILER#folded_record")) (* TODO fresh *)
|
||||
let access = (* TODO fresh *)
|
||||
e_accessor (e_variable (Var.of_name "#COMPILER#folded_record"))
|
||||
[Access_record captured_varname] in
|
||||
let assign = e_assign captured_varname [] access in
|
||||
match prev with
|
||||
@ -1303,6 +1312,74 @@ and simpl_for_collect : Raw.for_collect -> (_ -> expression result) result = fun
|
||||
| None -> e_skip ()
|
||||
| Some seq -> e_let_in (Var.of_name "#COMPILER#folded_record", None) false fold seq in (* TODO fresh *)
|
||||
return_statement @@ final_sequence
|
||||
(*
|
||||
and simpl_declaration : Raw.declaration -> declaration Location.wrap result =
|
||||
*)
|
||||
|
||||
let simpl_program : Raw.ast -> program result = fun t ->
|
||||
bind_list @@ List.map simpl_declaration @@ nseq_to_list t.decl
|
||||
and simpl_declaration_list declarations :
|
||||
Ast_simplified.declaration Location.wrap list result =
|
||||
let open Raw in
|
||||
let rec hook acc = function
|
||||
[] -> acc
|
||||
| [AttrDecl _] ->
|
||||
(* Detached attributes are erased. TODO: Warning. *)
|
||||
acc
|
||||
| AttrDecl _ :: (AttrDecl _ :: _ as declarations) ->
|
||||
(* Detached attributes are erased. TODO: Warning. *)
|
||||
hook acc declarations
|
||||
| AttrDecl decl :: ConstDecl {value; region} :: declarations ->
|
||||
let new_const =
|
||||
ConstDecl {value = {value with attributes = Some decl}; region}
|
||||
in hook acc (new_const :: declarations)
|
||||
| AttrDecl decl :: FunDecl {value; region} :: declarations ->
|
||||
let new_fun =
|
||||
FunDecl {value = {value with attributes = Some decl}; region}
|
||||
in hook acc (new_fun :: declarations)
|
||||
| AttrDecl _ :: declarations ->
|
||||
(* Detached attributes are erased. TODO: Warning. *)
|
||||
hook acc declarations
|
||||
| TypeDecl decl :: declarations ->
|
||||
let decl, loc = r_split decl in
|
||||
let {name; type_expr} : Raw.type_decl = decl in
|
||||
let%bind type_expression = simpl_type_expression type_expr in
|
||||
let new_decl =
|
||||
Declaration_type (Var.of_name name.value, type_expression) in
|
||||
let res = Location.wrap ~loc new_decl in
|
||||
hook (bind_list_cons res acc) declarations
|
||||
| ConstDecl decl :: declarations ->
|
||||
let simpl_const_decl =
|
||||
fun {name;const_type; init; attributes} ->
|
||||
let%bind expression = simpl_expression init in
|
||||
let%bind t = simpl_type_expression const_type in
|
||||
let type_annotation = Some t in
|
||||
let inline =
|
||||
match attributes with
|
||||
None -> false
|
||||
| Some {value; _} ->
|
||||
npseq_to_list value.ne_elements
|
||||
|> List.exists (fun Region.{value; _} -> value = "\"inline\"") in
|
||||
let new_decl =
|
||||
Declaration_constant
|
||||
(Var.of_name name.value, type_annotation, inline, expression)
|
||||
in ok new_decl in
|
||||
let%bind res =
|
||||
bind_map_location simpl_const_decl (Location.lift_region decl)
|
||||
in hook (bind_list_cons res acc) declarations
|
||||
| FunDecl fun_decl :: declarations ->
|
||||
let decl, loc = r_split fun_decl in
|
||||
let%bind ((name, ty_opt), expr) = simpl_fun_decl ~loc decl in
|
||||
let inline =
|
||||
match fun_decl.value.attributes with
|
||||
None -> false
|
||||
| Some {value; _} ->
|
||||
npseq_to_list value.ne_elements
|
||||
|> List.exists (fun Region.{value; _} -> value = "\"inline\"") in
|
||||
let new_decl =
|
||||
Declaration_constant (name, ty_opt, inline, expr) in
|
||||
let res = Location.wrap ~loc new_decl in
|
||||
hook (bind_list_cons res acc) declarations
|
||||
in
|
||||
hook (ok @@ []) (List.rev declarations)
|
||||
|
||||
let simpl_program : Raw.ast -> program result =
|
||||
fun t -> simpl_declaration_list @@ nseq_to_list t.decl
|
||||
|
@ -1,17 +1,19 @@
|
||||
const x: int = 1; attributes ["inline"];
|
||||
const x : int = 1; attributes ["inline"]
|
||||
|
||||
function foo (const a : int) : int is
|
||||
block {
|
||||
const test: int = 2 + a; attributes ["inline"];
|
||||
} with test;
|
||||
begin
|
||||
const test : int = 2 + a;
|
||||
attributes ["inline"];
|
||||
end with test;
|
||||
attributes ["inline"];
|
||||
|
||||
const y: int = 1; attributes ["inline"; "other"];
|
||||
const y : int = 1; attributes ["inline"; "other"]
|
||||
|
||||
function bar (const b : int) : int is
|
||||
block {
|
||||
function test (const z : int) : int is begin
|
||||
begin
|
||||
function test (const z : int) : int is
|
||||
begin
|
||||
const r : int = 2 + b + z
|
||||
end with r;
|
||||
attributes ["inline"; "foo"; "bar"];
|
||||
} with test(b);
|
||||
attributes ["inline"; "foo"; "bar"]
|
||||
end with test(b)
|
||||
|
@ -1,13 +1,10 @@
|
||||
let x = 1 [@@inline]
|
||||
|
||||
let foo (a: int): int = (
|
||||
let test = 2 + a [@@inline] in
|
||||
test
|
||||
) [@@inline]
|
||||
let foo (a: int): int =
|
||||
(let test = 2 + a [@@inline] in test) [@@inline]
|
||||
|
||||
let y = 1 [@@inline][@@other]
|
||||
|
||||
let bar (b: int): int = (
|
||||
let test = fun (z: int) -> 2 + b + z [@@inline][@@foo][@@bar] in
|
||||
test b
|
||||
)
|
||||
let bar (b: int): int =
|
||||
let test = fun (z: int) -> 2 + b + z [@@inline][@@foo][@@bar]
|
||||
in test b
|
||||
|
@ -25,8 +25,8 @@
|
||||
(action (run ./parser_negative_tests.exe))
|
||||
(deps
|
||||
../passes/1-parser/pascaligo/all.ligo
|
||||
../passes/1-parser/cameligo/all.ligo
|
||||
../passes/1-parser/reasonligo/all.ligo
|
||||
../passes/1-parser/cameligo/all.mligo
|
||||
../passes/1-parser/reasonligo/all.religo
|
||||
))
|
||||
|
||||
(alias
|
||||
|
@ -1 +1 @@
|
||||
let args = 1;
|
||||
let arguments = 1;
|
@ -6,10 +6,10 @@ let pascaligo_sdata = {
|
||||
erroneous_source_file = "../passes/1-parser/pascaligo/all.ligo" ;
|
||||
parser = Parser.Pascaligo.parse_expression }
|
||||
let cameligo_sdata = {
|
||||
erroneous_source_file = "../passes/1-parser/cameligo/all.ligo" ;
|
||||
erroneous_source_file = "../passes/1-parser/cameligo/all.mligo" ;
|
||||
parser = Parser.Cameligo.parse_expression }
|
||||
let reasonligo_sdata = {
|
||||
erroneous_source_file = "../passes/1-parser/reasonligo/all.ligo" ;
|
||||
erroneous_source_file = "../passes/1-parser/reasonligo/all.religo" ;
|
||||
parser = Parser.Reasonligo.parse_expression }
|
||||
|
||||
let get_exp_as_string filename =
|
||||
|
49
vendors/ligo-utils/simple-utils/pos.ml
vendored
49
vendors/ligo-utils/simple-utils/pos.ml
vendored
@ -23,9 +23,10 @@ type t = <
|
||||
|
||||
is_ghost : bool;
|
||||
|
||||
to_string : ?offsets:bool -> [`Byte | `Point] -> string;
|
||||
compact : ?offsets:bool -> [`Byte | `Point] -> string;
|
||||
anonymous : ?offsets:bool -> [`Byte | `Point] -> string
|
||||
to_string :
|
||||
?file:bool -> ?offsets:bool -> [`Byte | `Point] -> string;
|
||||
compact :
|
||||
?file:bool -> ?offsets:bool -> [`Byte | `Point] -> string;
|
||||
>
|
||||
|
||||
type pos = t
|
||||
@ -61,8 +62,6 @@ let make ~byte ~point_num ~point_bol =
|
||||
let pos = pos#set_offset offset
|
||||
in pos
|
||||
|
||||
(* The string must not contain '\n'. See [new_line]. *)
|
||||
|
||||
method shift_bytes len =
|
||||
{< byte = Lexing.{byte with pos_cnum = byte.pos_cnum + len};
|
||||
point_num = point_num + len >}
|
||||
@ -77,11 +76,13 @@ let make ~byte ~point_num ~point_bol =
|
||||
pos_bol = byte.pos_cnum};
|
||||
point_bol = point_num >}
|
||||
|
||||
(* The string must not contain '\n'. See [add_line]. *)
|
||||
|
||||
method new_line string =
|
||||
let len = String.length string
|
||||
in (self#shift_bytes len)#add_nl
|
||||
|
||||
method is_ghost = byte = Lexing.dummy_pos
|
||||
method is_ghost = (byte = Lexing.dummy_pos)
|
||||
|
||||
method file = byte.Lexing.pos_fname
|
||||
|
||||
@ -99,24 +100,30 @@ let make ~byte ~point_num ~point_bol =
|
||||
|
||||
method byte_offset = byte.Lexing.pos_cnum
|
||||
|
||||
method to_string ?(offsets=true) mode =
|
||||
let offset = self#offset mode in
|
||||
let horizontal, value =
|
||||
if offsets then "character", offset else "column", offset + 1
|
||||
in sprintf "File \"%s\", line %i, %s %i"
|
||||
self#file self#line horizontal value
|
||||
|
||||
method compact ?(offsets=true) mode =
|
||||
method to_string ?(file=true) ?(offsets=true) mode =
|
||||
if self#is_ghost then "ghost"
|
||||
else
|
||||
let offset = self#offset mode in
|
||||
sprintf "%s:%i:%i"
|
||||
self#file self#line (if offsets then offset else offset + 1)
|
||||
let horizontal, value =
|
||||
if offsets then
|
||||
"character", offset
|
||||
else "column", offset + 1 in
|
||||
if file && self#file <> "" then
|
||||
sprintf "File \"%s\", line %i, %s %i"
|
||||
self#file self#line horizontal value
|
||||
else sprintf "Line %i, %s %i"
|
||||
self#line horizontal value
|
||||
|
||||
method anonymous ?(offsets=true) mode =
|
||||
method compact ?(file=true) ?(offsets=true) mode =
|
||||
if self#is_ghost then "ghost"
|
||||
else sprintf "%i:%i" self#line
|
||||
(if offsets then self#offset mode else self#column mode)
|
||||
else
|
||||
let horizontal =
|
||||
if offsets then self#offset mode
|
||||
else self#column mode in
|
||||
if file && self#file <> "" then
|
||||
sprintf "%s:%i:%i" self#file self#line horizontal
|
||||
else
|
||||
sprintf "%i:%i" self#line horizontal
|
||||
end
|
||||
|
||||
let from_byte byte =
|
||||
@ -126,7 +133,9 @@ let from_byte byte =
|
||||
|
||||
let ghost = make ~byte:Lexing.dummy_pos ~point_num:(-1) ~point_bol:(-1)
|
||||
|
||||
let min = make ~byte:Lexing.dummy_pos ~point_num:0 ~point_bol:0
|
||||
let min ~file =
|
||||
let pos = make ~byte:Lexing.dummy_pos ~point_num:0 ~point_bol:0
|
||||
in pos#set_file file
|
||||
|
||||
(* Comparisons *)
|
||||
|
||||
|
28
vendors/ligo-utils/simple-utils/pos.mli
vendored
28
vendors/ligo-utils/simple-utils/pos.mli
vendored
@ -62,19 +62,20 @@ type t = <
|
||||
(* Payload *)
|
||||
|
||||
byte : Lexing.position;
|
||||
point_num : int;
|
||||
point_bol : int;
|
||||
file : string;
|
||||
line : int;
|
||||
point_num : int; (* point_num >= point_bol *)
|
||||
point_bol : int; (* point_bol >= 0 *)
|
||||
file : string; (* May be empty *)
|
||||
line : int; (* line > 0 *)
|
||||
|
||||
(* Setters *)
|
||||
|
||||
set_file : string -> t;
|
||||
set_line : int -> t;
|
||||
set_offset : int -> t;
|
||||
|
||||
set : file:string -> line:int -> offset:int -> t;
|
||||
|
||||
new_line : string -> t;
|
||||
new_line : string -> t; (* String must be "\n" or "\c\r" *)
|
||||
add_nl : t;
|
||||
|
||||
shift_bytes : int -> t;
|
||||
@ -93,9 +94,10 @@ type t = <
|
||||
|
||||
(* Conversions to [string] *)
|
||||
|
||||
to_string : ?offsets:bool -> [`Byte | `Point] -> string;
|
||||
compact : ?offsets:bool -> [`Byte | `Point] -> string;
|
||||
anonymous : ?offsets:bool -> [`Byte | `Point] -> string
|
||||
to_string :
|
||||
?file:bool -> ?offsets:bool -> [`Byte | `Point] -> string;
|
||||
compact :
|
||||
?file:bool -> ?offsets:bool -> [`Byte | `Point] -> string;
|
||||
>
|
||||
|
||||
(** A shorthand after an [open Pos].
|
||||
@ -104,18 +106,20 @@ type pos = t
|
||||
|
||||
(** {1 Constructors} *)
|
||||
|
||||
val make : byte:Lexing.position -> point_num:int -> point_bol:int -> t
|
||||
val make :
|
||||
byte:Lexing.position -> point_num:int -> point_bol:int -> t
|
||||
|
||||
val from_byte : Lexing.position -> t
|
||||
|
||||
(** {1 Special positions} *)
|
||||
|
||||
(** The value [ghost] is the same as {! Lexing.dummy_pos}.
|
||||
(** The value [ghost] based on the same as {! Lexing.dummy_pos}.
|
||||
*)
|
||||
val ghost : t
|
||||
|
||||
(** Lexing convention: line [1], offsets to [0] and file to [""].
|
||||
(** Lexing convention: line [1], offset to [0].
|
||||
*)
|
||||
val min : t
|
||||
val min : file:string -> t
|
||||
|
||||
(** {1 Comparisons} *)
|
||||
|
||||
|
6
vendors/ligo-utils/simple-utils/region.ml
vendored
6
vendors/ligo-utils/simple-utils/region.ml
vendored
@ -101,8 +101,8 @@ let make ~(start: Pos.t) ~(stop: Pos.t) =
|
||||
if start#is_ghost || stop#is_ghost then "ghost"
|
||||
else
|
||||
let prefix = if file then start#file ^ ":" else ""
|
||||
and start_str = start#anonymous ~offsets mode
|
||||
and stop_str = stop#anonymous ~offsets mode in
|
||||
and start_str = start#compact ~file:false ~offsets mode
|
||||
and stop_str = stop#compact ~file:false ~offsets mode in
|
||||
if start#file = stop#file then
|
||||
if start#line = stop#line then
|
||||
sprintf "%s%s-%i" prefix start_str
|
||||
@ -120,7 +120,7 @@ let ghost = make ~start:Pos.ghost ~stop:Pos.ghost
|
||||
|
||||
let wrap_ghost value = {value ; region = ghost}
|
||||
|
||||
let min = make ~start:Pos.min ~stop:Pos.min
|
||||
let min ~file = make ~start:(Pos.min ~file) ~stop:(Pos.min ~file)
|
||||
|
||||
(* Comparisons *)
|
||||
|
||||
|
6
vendors/ligo-utils/simple-utils/region.mli
vendored
6
vendors/ligo-utils/simple-utils/region.mli
vendored
@ -54,7 +54,7 @@
|
||||
|
||||
{li The method [compact] has the same signature as and calling
|
||||
convention as [to_string], except that the resulting string
|
||||
is more compact.}}
|
||||
is shorter (usually for debugging or tracing).}}
|
||||
*)
|
||||
type t = <
|
||||
start : Pos.t;
|
||||
@ -86,7 +86,7 @@ type t = <
|
||||
*)
|
||||
type region = t
|
||||
|
||||
(** The type ['a reg] enables the concept of something of type ['a] to
|
||||
(** The type ['a reg] enables the concept of some value of type ['a] to
|
||||
be related to a region in a source file.
|
||||
*)
|
||||
type 'a reg = {region: t; value: 'a}
|
||||
@ -119,7 +119,7 @@ val wrap_ghost : 'a -> 'a reg
|
||||
(** Occasionnally, we may need a minimum region. It is here made of
|
||||
two minimal positions.
|
||||
*)
|
||||
val min : t
|
||||
val min : file:string -> t
|
||||
|
||||
(** {1 Comparisons} *)
|
||||
|
||||
|
268
vendors/ligo-utils/simple-utils/trace.ml
vendored
268
vendors/ligo-utils/simple-utils/trace.ml
vendored
@ -1,4 +1,4 @@
|
||||
(** Trace tutorial
|
||||
(* Trace tutorial
|
||||
|
||||
This module guides the reader through the writing of a simplified
|
||||
version of the trace monad [result], and the definition of a few
|
||||
@ -6,13 +6,13 @@
|
||||
*)
|
||||
|
||||
module Trace_tutorial = struct
|
||||
(** The trace monad is fairly similar to the predefined [option]
|
||||
(* The trace monad is fairly similar to the predefined [option]
|
||||
type. It is an instance of the predefined [result] type. *)
|
||||
|
||||
type annotation = string
|
||||
type error = string
|
||||
|
||||
(** The type ['a result] is used by the trace monad to both model an
|
||||
(* The type ['a result] is used by the trace monad to both model an
|
||||
expected value of type ['a] or the failure to obtain it, instead
|
||||
of working directly with ['a] values and handling separately
|
||||
errors, for example by means of exceptions. (See the type [('a,'b)
|
||||
@ -23,36 +23,36 @@ module Trace_tutorial = struct
|
||||
list of annotations (information about past successful
|
||||
computations), or it is a list of errors accumulated so far.
|
||||
The former case is denoted by the data constructor [Ok], and the
|
||||
second by [Error].
|
||||
*)
|
||||
second by [Error]. *)
|
||||
|
||||
type nonrec 'a result = ('a * annotation list, error list) result
|
||||
(*
|
||||
= Ok of 'a * annotation list
|
||||
| Error of error list
|
||||
*)
|
||||
|
||||
(** The function [divide_trace] shows the basic use of the trace
|
||||
monad.
|
||||
*)
|
||||
(* The function [divide_trace] shows the basic use of the trace
|
||||
monad. *)
|
||||
|
||||
let divide_trace a b =
|
||||
if b = 0
|
||||
then Error [Printf.sprintf "division by zero: %d/%d" a b]
|
||||
else Ok (a/b, [])
|
||||
|
||||
(** The function [divide_three] shows that when composing two
|
||||
(* The function [divide_three] shows that when composing two
|
||||
functions, if the first call fails, the error is passed along
|
||||
and the second call is not evaluated. (A pattern called
|
||||
"error-passing style").
|
||||
*)
|
||||
"error-passing style"). *)
|
||||
|
||||
let divide_three a b c =
|
||||
match divide_trace a b with
|
||||
Ok (a_div_b , _) -> divide_trace a_div_b c
|
||||
| errors -> errors
|
||||
|
||||
(** The function [divide_three_annot] shows that when composing two
|
||||
(* The function [divide_three_annot] shows that when composing two
|
||||
functions, if both calls are successful, the lists of
|
||||
annotations are joined.
|
||||
*)
|
||||
annotations are joined. *)
|
||||
|
||||
let divide_three_annot a b c =
|
||||
match divide_trace a b with
|
||||
Ok (a_div_b, annot1) -> (
|
||||
@ -62,21 +62,19 @@ module Trace_tutorial = struct
|
||||
| errors -> errors)
|
||||
| errors -> errors
|
||||
|
||||
(** The systematic matching of the result of each call in a function
|
||||
(* The systematic matching of the result of each call in a function
|
||||
composition is bulky, so we define a [bind] function which takes
|
||||
a function [f: 'a -> 'b result] and applies it to a current ['a
|
||||
result] (not ['a]).
|
||||
{ul
|
||||
{li If the current result is an error, then [bind]
|
||||
returns that same error without calling [f];}
|
||||
|
||||
{li otherwise [bind] unwraps the [Ok] of the current result
|
||||
* If the current result is an error, then [bind]
|
||||
returns that same error without calling [f];
|
||||
|
||||
* otherwise [bind] unwraps the [Ok] of the current result
|
||||
and calls [f] on it:
|
||||
{ul
|
||||
{li That call itself may return an error;}
|
||||
{li if not, [bind] combines the annotations and returns the last
|
||||
result.}}}}
|
||||
*)
|
||||
* That call itself may return an error;}
|
||||
* if not, [bind] combines the annotations and returns the
|
||||
last result. *)
|
||||
let bind (f: 'a -> 'b result) : 'a result -> 'b result =
|
||||
function
|
||||
Ok (x, annot) -> (
|
||||
@ -85,36 +83,36 @@ module Trace_tutorial = struct
|
||||
| errors -> ignore annot; errors)
|
||||
| Error _ as e -> e
|
||||
|
||||
(** The function [divide_three_bind] is equivalent to the verbose
|
||||
[divide_three] above, but makes use of [bind].
|
||||
*)
|
||||
(* The function [divide_three_bind] is equivalent to the verbose
|
||||
[divide_three] above, but makes use of [bind]. *)
|
||||
|
||||
let divide_three_bind a b c =
|
||||
let maybe_a_div_b = divide_trace a b in
|
||||
let continuation a_div_b = divide_trace a_div_b c
|
||||
in bind continuation maybe_a_div_b
|
||||
|
||||
(** The operator [(>>?)] is a redefinition of [bind] that makes the
|
||||
(* The operator [(>>?)] is a redefinition of [bind] that makes the
|
||||
program shorter, at the cost of a slightly
|
||||
awkward reading because the two parameters are swapped.
|
||||
*)
|
||||
awkward reading because the two parameters are swapped. *)
|
||||
|
||||
let (>>?) x f = bind f x
|
||||
|
||||
(** The function [divide_three_bind_symbol] is equivalent to
|
||||
[divide_three_bind], but makes use of the operator [(>>?)].
|
||||
*)
|
||||
(* The function [divide_three_bind_symbol] is equivalent to
|
||||
[divide_three_bind], but makes use of the operator [(>>?)]. *)
|
||||
|
||||
let divide_three_bind_symbol a b c =
|
||||
let maybe_a_div_b = divide_trace a b in
|
||||
let continuation a_div_b = divide_trace a_div_b c in
|
||||
maybe_a_div_b >>? continuation
|
||||
|
||||
(** The function [divide_three_bind_symbol'] is equivalent to
|
||||
(* The function [divide_three_bind_symbol'] is equivalent to
|
||||
[divide_three_bind_symbol], where the two temporary [let]
|
||||
definitions are inlined for a more compact reading.
|
||||
*)
|
||||
definitions are inlined for a more compact reading. *)
|
||||
|
||||
let divide_three_bind_symbol' a b c =
|
||||
divide_trace a b >>? (fun a_div_b -> divide_trace a_div_b c)
|
||||
|
||||
(** This is now fairly legible, but chaining many such functions is
|
||||
(* This is now fairly legible, but chaining many such functions is
|
||||
not the usual way of writing code. We use the PPX extension to
|
||||
the OCaml compiler [ppx_let] to add some syntactic sugar.
|
||||
The extension framework PPX is enabled by adding the following
|
||||
@ -123,26 +121,26 @@ module Trace_tutorial = struct
|
||||
[(preprocess
|
||||
(pps simple-utils.ppx_let_generalized))]
|
||||
The extension [ppx_let] requires the module [Let_syntax] to be
|
||||
defined.
|
||||
*)
|
||||
defined. *)
|
||||
|
||||
module Let_syntax = struct
|
||||
let bind m ~f = m >>? f
|
||||
module Open_on_rhs_bind = struct end
|
||||
end
|
||||
|
||||
(** The function [divide_three_bind_ppx_let] is equivalent to the
|
||||
(* The function [divide_three_bind_ppx_let] is equivalent to the
|
||||
function [divide_three_bind_symbol']. The only difference is
|
||||
that the module [Open_on_rhs_bind] is implicitly opened around
|
||||
the expression on the righ-hand side of the [=] sign, namely
|
||||
[divide_trace a b].
|
||||
*)
|
||||
[divide_trace a b]. *)
|
||||
|
||||
let divide_three_bind_ppx_let a b c =
|
||||
let%bind a_div_b = divide_trace a b
|
||||
in divide_trace a_div_b c
|
||||
|
||||
(** The function [divide_many_bind_ppx_let] shows how well this
|
||||
notation composes.
|
||||
*)
|
||||
notation composes. *)
|
||||
|
||||
let divide_many_bind_ppx_let a b c d e f =
|
||||
let x = a in
|
||||
let%bind x = divide_trace x b in
|
||||
@ -153,34 +151,35 @@ module Trace_tutorial = struct
|
||||
in Ok (x, [])
|
||||
|
||||
(** The function [ok] is a shorthand for an [Ok] without
|
||||
annotations.
|
||||
*)
|
||||
annotations. *)
|
||||
|
||||
let ok x = Ok (x, [])
|
||||
|
||||
(** The function [map] lifts a regular ['a -> 'b] function on values to
|
||||
a function on results, of type ['a result -> 'b result].
|
||||
*)
|
||||
(* The function [map] lifts a regular ['a -> 'b] function on values to
|
||||
a function on results, of type ['a result -> 'b result]. *)
|
||||
|
||||
let map f = function
|
||||
Ok (x, annotations) -> Ok (f x, annotations)
|
||||
| e -> e
|
||||
|
||||
(** The function [bind_list] turns a list of results of type [('a
|
||||
(* The function [bind_list] turns a list of results of type [('a
|
||||
result) list] into a result of list, of type [('a list) result],
|
||||
as follows.
|
||||
{ul
|
||||
{li If the list only contains [Ok] values, it strips the [Ok]
|
||||
* If the list only contains [Ok] values, it strips the [Ok]
|
||||
of each element and returns that list wrapped with [Ok].}
|
||||
{li Otherwise, one or more of the elements of the input list
|
||||
|
||||
* Otherwise, one or more of the elements of the input list
|
||||
is [Error], then [bind_list] returns the first error in the
|
||||
list.}}
|
||||
list.
|
||||
*)
|
||||
|
||||
let rec bind_list = function
|
||||
[] -> ok []
|
||||
| hd::tl ->
|
||||
hd >>? fun hd ->
|
||||
bind_list tl >>? fun tl ->
|
||||
ok @@ hd::tl
|
||||
(** A major feature of [Trace] is that it enables having a stack of
|
||||
(* A major feature of [Trace] is that it enables having a stack of
|
||||
errors (that should act as a simplified stack frame), rather
|
||||
than a unique error. It is done by using the function
|
||||
[trace]. For instance, let's say that you have a function that
|
||||
@ -198,17 +197,17 @@ module Trace_tutorial = struct
|
||||
trace (simple_error "error getting key") @@
|
||||
get key map
|
||||
in ...]
|
||||
And this will pass along the error triggered by [get key map].
|
||||
*)
|
||||
And this will pass along the error triggered by [get key map]. *)
|
||||
|
||||
let trace err = function
|
||||
Error e -> Error (err::e)
|
||||
| ok -> ok
|
||||
|
||||
(** The real trace monad is very similar to the one that we have
|
||||
(* The real trace monad is very similar to the one that we have
|
||||
defined above. The main difference is that the errors and
|
||||
annotations are structured data (instead of plain strings) and
|
||||
are generated lazily.
|
||||
*)
|
||||
are generated lazily. *)
|
||||
|
||||
let the_end = "End of the tutorial."
|
||||
|
||||
end (* end Trace_tutorial. *)
|
||||
@ -239,8 +238,7 @@ module JSON_string_utils = struct
|
||||
match assoc j with
|
||||
None -> j
|
||||
| Some assoc -> `Assoc (
|
||||
List.map (fun (k', v') -> (k', if k = k' then v else v')) assoc
|
||||
)
|
||||
List.map (fun (k', v') -> (k', if k = k' then v else v')) assoc)
|
||||
|
||||
let swap f l r = f r l
|
||||
|
||||
@ -264,38 +262,39 @@ module JSON_string_utils = struct
|
||||
let (||) l r = l |> default r
|
||||
|
||||
let (|^) = bind2 (^)
|
||||
|
||||
end
|
||||
|
||||
type 'a thunk = unit -> 'a
|
||||
|
||||
(** Errors are encoded in JSON. This is because different libraries
|
||||
(* Errors are encoded in JSON. This is because different libraries
|
||||
will implement their own helpers, and we do not want to hardcode
|
||||
in their type how they are supposed to interact.
|
||||
*)
|
||||
in their type how they are supposed to interact. *)
|
||||
|
||||
type error = J.t
|
||||
|
||||
(** Thunks are used because computing some errors can be costly, and
|
||||
(* Thunks are used because computing some errors can be costly, and
|
||||
we do not want to spend most of our time building errors. Instead,
|
||||
their computation is deferred.
|
||||
*)
|
||||
their computation is deferred.*)
|
||||
|
||||
type error_thunk = error thunk
|
||||
|
||||
(** Annotations should be used in debug mode to aggregate information
|
||||
(* Annotations should be used in debug mode to aggregate information
|
||||
about some value history. Where it was produced, when it was
|
||||
modified, etc. It is currently not being used. *)
|
||||
|
||||
type annotation = J.t
|
||||
|
||||
(** Even in debug mode, building annotations can be quite
|
||||
(* Even in debug mode, building annotations can be quite
|
||||
resource-intensive. Instead, a thunk is passed, that is computed
|
||||
only when debug information is queried (typically before a print).
|
||||
*)
|
||||
only when debug information is queried (typically before a print).*)
|
||||
|
||||
type annotation_thunk = annotation thunk
|
||||
|
||||
(** Types of traced elements. It might be good to rename it [trace] at
|
||||
some point.
|
||||
*)
|
||||
(* Types of traced elements. It might be good to rename it [trace] at
|
||||
some point. *)
|
||||
|
||||
type nonrec 'a result = ('a * annotation_thunk list, error_thunk) result
|
||||
|
||||
(*
|
||||
= Ok of 'a * annotation_thunk list
|
||||
| Error of error_thunk
|
||||
@ -308,29 +307,28 @@ let ok x = Ok (x, [])
|
||||
|
||||
let fail err = Error err
|
||||
|
||||
(** {1 Monadic operators} *)
|
||||
(* Monadic operators *)
|
||||
|
||||
let bind f = function
|
||||
Ok (x, ann) -> (
|
||||
Error _ as e -> e
|
||||
| Ok (x, ann) ->
|
||||
match f x with
|
||||
Ok (x', ann') -> Ok (x', ann' @ ann)
|
||||
| Error _ as e' -> ignore ann; e')
|
||||
| Error _ as e -> e
|
||||
| Error _ as e' -> ignore ann; e'
|
||||
|
||||
let map f = function
|
||||
Ok (x, annotations) -> Ok (f x, annotations)
|
||||
| Error _ as e -> e
|
||||
|
||||
(** The lexical convention usually adopted for the bind function is
|
||||
(* The lexical convention usually adopted for the bind function is
|
||||
[>>=], but ours comes from the Tezos code base, where the [result]
|
||||
bind is [>>?], and [Lwt]'s (threading library) is [>>=], and the
|
||||
combination of both is [>>=?].
|
||||
*)
|
||||
combination of both is [>>=?]. *)
|
||||
|
||||
let (>>?) x f = bind f x
|
||||
let (>>|?) x f = map f x
|
||||
|
||||
(**
|
||||
Used by PPX_let, an OCaml preprocessor.
|
||||
(* Used by PPX_let, an OCaml preprocessor.
|
||||
What it does is that, when you only care about the case where a result isn't
|
||||
an error, instead of writing:
|
||||
[
|
||||
@ -344,21 +342,20 @@ let (>>|?) x f = map f x
|
||||
]
|
||||
This is much more typical of OCaml. This makes the code more
|
||||
readable, easy to write and refactor. It is used pervasively in
|
||||
LIGO.
|
||||
*)
|
||||
LIGO. *)
|
||||
|
||||
module Let_syntax = struct
|
||||
let bind m ~f = m >>? f
|
||||
module Open_on_rhs_bind = struct end
|
||||
end
|
||||
|
||||
(* Build a thunk from a constant. *)
|
||||
|
||||
(** Build a thunk from a constant.
|
||||
*)
|
||||
let thunk x () = x
|
||||
|
||||
(** Build a standard error, with a title, a message, an error code and
|
||||
some data.
|
||||
*)
|
||||
(* Build a standard error, with a title, a message, an error code and
|
||||
some data. *)
|
||||
|
||||
let mk_error
|
||||
?(error_code : int thunk option) ?(message : string thunk option)
|
||||
?(data : (string * string thunk) list option)
|
||||
@ -373,9 +370,11 @@ let mk_error
|
||||
let type' = Some ("type" , `String "error") in
|
||||
let children' = Some ("children" , `List children) in
|
||||
let infos' = Some ("infos" , `List infos) in
|
||||
`Assoc (X_option.collapse_list [ error_code' ; title' ; message' ; data' ; type' ; children' ; infos' ])
|
||||
`Assoc (X_option.collapse_list
|
||||
[error_code'; title'; message'; data'; type'; children'; infos'])
|
||||
|
||||
let error ?data ?error_code ?children ?infos title message () = mk_error ?data ?error_code ?children ?infos ~title:(title) ~message:(message) ()
|
||||
let error ?data ?error_code ?children ?infos title message () =
|
||||
mk_error ?data ?error_code ?children ?infos ~title ~message ()
|
||||
|
||||
let prepend_child = fun child err ->
|
||||
let open JSON_string_utils in
|
||||
@ -389,9 +388,8 @@ let patch_children = fun children err ->
|
||||
let open JSON_string_utils in
|
||||
patch err "children" (`List (List.map (fun f -> f ()) children))
|
||||
|
||||
(**
|
||||
Build a standard info, with a title, a message, an info code and some data.
|
||||
*)
|
||||
(* Build a standard info, with a title, a message, an info code and some data. *)
|
||||
|
||||
let mk_info
|
||||
?(info_code : int thunk option) ?(message : string thunk option)
|
||||
?(data : (string * string thunk) list option)
|
||||
@ -405,7 +403,8 @@ let mk_info
|
||||
let type' = Some ("type" , `String "info") in
|
||||
`Assoc (X_option.collapse_list [ error_code' ; title' ; message' ; data' ; type' ])
|
||||
|
||||
let info ?data ?info_code title message () = mk_info ?data ?info_code ~title:(title) ~message:(message) ()
|
||||
let info ?data ?info_code title message () =
|
||||
mk_info ?data ?info_code ~title ~message ()
|
||||
|
||||
let prepend_info = fun info err ->
|
||||
let open JSON_string_utils in
|
||||
@ -416,32 +415,31 @@ let prepend_info = fun info err ->
|
||||
patch err "infos" (`List infos)
|
||||
|
||||
|
||||
(** Helpers that ideally should not be used in production.
|
||||
*)
|
||||
(* Helpers that ideally should not be used in production. *)
|
||||
|
||||
let simple_error str () = mk_error ~title:(thunk str) ()
|
||||
let simple_info str () = mk_info ~title:(thunk str) ()
|
||||
let simple_fail str = fail @@ simple_error str
|
||||
let internal_assertion_failure str = simple_error ("assertion failed: " ^ str)
|
||||
|
||||
(** To be used when you only want to signal an error. It can be useful
|
||||
when followed by [trace_strong].
|
||||
*)
|
||||
(* To be used when you only want to signal an error. It can be useful
|
||||
when followed by [trace_strong]. *)
|
||||
|
||||
let dummy_fail = simple_fail "dummy"
|
||||
|
||||
let trace info = function
|
||||
Ok _ as o -> o
|
||||
| Error err -> Error (fun () -> prepend_info (info ()) (err ()))
|
||||
|
||||
(** Erase the current error stack, and replace it by the given
|
||||
(* Erase the current error stack, and replace it by the given
|
||||
error. It's useful when using [Assert] and you want to discard its
|
||||
autogenerated message.
|
||||
*)
|
||||
autogenerated message. *)
|
||||
|
||||
let trace_strong err = function
|
||||
Ok _ as o -> o
|
||||
| Error _ -> Error err
|
||||
|
||||
(**
|
||||
Sometimes, when you have a list of potentially erroneous elements, you need
|
||||
(* Sometimes, when you have a list of potentially erroneous elements, you need
|
||||
to retrieve all the errors, instead of just the first one. In that case, do:
|
||||
[let type_list lst =
|
||||
let%bind lst' =
|
||||
@ -451,8 +449,7 @@ let trace_strong err = function
|
||||
Where before you would have written:
|
||||
[let type_list lst =
|
||||
let%bind lst' = bind_map_list type_element lst in
|
||||
...]
|
||||
*)
|
||||
...] *)
|
||||
let trace_list err lst =
|
||||
let oks =
|
||||
let aux = function
|
||||
@ -468,30 +465,24 @@ let trace_list err lst =
|
||||
| [] -> ok oks
|
||||
| errs -> fail (fun () -> patch_children errs err)
|
||||
|
||||
(**
|
||||
Trace, but with an error which generation may itself fail.
|
||||
*)
|
||||
(* Trace, but with an error which generation may itself fail. *)
|
||||
|
||||
let trace_r err_thunk_may_fail = function
|
||||
| Ok _ as o -> o
|
||||
| Error _ -> (
|
||||
Ok _ as o -> o
|
||||
| Error _ ->
|
||||
match err_thunk_may_fail () with
|
||||
| Ok (err, annotations) -> ignore annotations; Error (err)
|
||||
Ok (err, annotations) -> ignore annotations; Error (err)
|
||||
| Error errors_while_generating_error ->
|
||||
(* TODO: the complexity could be O(n*n) in the worst case,
|
||||
this should use some catenable lists. *)
|
||||
Error (errors_while_generating_error)
|
||||
)
|
||||
|
||||
(**
|
||||
`trace_f f error` yields a function that acts the same as `f`, but with an
|
||||
error frame that has one more error.
|
||||
*)
|
||||
let trace_f f error x =
|
||||
trace error @@ f x
|
||||
(* [trace_f f error] yields a function that acts the same as `f`, but with an
|
||||
error frame that has one more error. *)
|
||||
|
||||
let trace_f f error x = trace error @@ f x
|
||||
(* Same, but for functions with 2 parameters. *)
|
||||
|
||||
(**
|
||||
Same, but for functions with 2 parameters.
|
||||
*)
|
||||
let trace_f_2 f error x y =
|
||||
trace error @@ f x y
|
||||
|
||||
@ -520,7 +511,7 @@ let to_option = function
|
||||
Convert an option to a result, with a given error if the parameter is None.
|
||||
*)
|
||||
let trace_option error = function
|
||||
| None -> fail error
|
||||
None -> fail error
|
||||
| Some s -> ok s
|
||||
|
||||
(** Utilities to interact with other data-structure. [bind_t] takes
|
||||
@ -535,22 +526,14 @@ let trace_option error = function
|
||||
*)
|
||||
|
||||
let bind_map_option f = function
|
||||
| None -> ok None
|
||||
None -> ok None
|
||||
| Some s -> f s >>? fun x -> ok (Some x)
|
||||
|
||||
let rec bind_list = function
|
||||
| [] -> ok []
|
||||
| hd :: tl -> (
|
||||
hd >>? fun hd ->
|
||||
bind_list tl >>? fun tl ->
|
||||
ok @@ hd :: tl
|
||||
)
|
||||
|
||||
let bind_ne_list = fun (hd , tl) ->
|
||||
hd >>? fun hd ->
|
||||
bind_list tl >>? fun tl ->
|
||||
ok @@ (hd , tl)
|
||||
|
||||
[] -> ok []
|
||||
| hd::tl -> hd >>? fun hd -> bind_list tl >>? fun tl -> ok @@ hd :: tl
|
||||
let bind_ne_list (hd, tl) =
|
||||
hd >>? fun hd -> bind_list tl >>? fun tl -> ok @@ (hd, tl)
|
||||
let bind_smap (s:_ X_map.String.t) =
|
||||
let open X_map.String in
|
||||
let aux k v prev =
|
||||
@ -718,6 +701,10 @@ let bind_fold_map_pair f acc (a, b) =
|
||||
let bind_map_triple f (a, b, c) =
|
||||
bind_and3 (f a, f b, f c)
|
||||
|
||||
let bind_list_cons v lst =
|
||||
lst >>? fun lst ->
|
||||
ok (v::lst)
|
||||
|
||||
let rec bind_chain : ('a -> 'a result) list -> 'a -> 'a result = fun fs x ->
|
||||
match fs with
|
||||
| [] -> ok x
|
||||
@ -726,7 +713,6 @@ let rec bind_chain : ('a -> 'a result) list -> 'a -> 'a result = fun fs x ->
|
||||
bind aux (ok x)
|
||||
)
|
||||
|
||||
|
||||
(**
|
||||
Wraps a call that might trigger an exception in a result.
|
||||
*)
|
||||
|
Loading…
Reference in New Issue
Block a user