summaryrefslogtreecommitdiffstats
path: root/parser.y
diff options
context:
space:
mode:
authorKaz Kylheku <kaz@kylheku.com>2021-05-27 06:51:23 -0700
committerKaz Kylheku <kaz@kylheku.com>2021-05-27 06:51:23 -0700
commit2f8bbbb5e50b0297bd531da3388589716a8cf0f2 (patch)
tree13161ca135e5cdcfa37a5d50004251c0dc4e00bd /parser.y
parent77bc5d4cb9e364a56835fa46226e50faa422fe45 (diff)
downloadtxr-2f8bbbb5e50b0297bd531da3388589716a8cf0f2.tar.gz
txr-2f8bbbb5e50b0297bd531da3388589716a8cf0f2.tar.bz2
txr-2f8bbbb5e50b0297bd531da3388589716a8cf0f2.zip
json: support quasiquoting.
* parser.h (end_of_json_unquote): Declared. * parser.l (JPUNC, NJPUNC): Add ~ and * characters to set of JSON punctuators. (grammar): Allow closing brace character in NESTED, SPECIAL and QSPECIAL statues to be a token. This is because it occurs as a lookahead character in this situation #J{"foo":~expr}. The lexer switches from the JSON to the NESTED start state when it scans the ~ token, so that expr is treated as Lisp. But then } is consumed as a lookahead token by the parser in that same mode; when we pop back to JSON mode, the } token has already been scanned in NESTED mode. We add two new rules in JSON mode to the lexer to recognize the ~ unquote and ~* splicing unquote. Both have to push the NESTED start condition. (end_of_json_unquote): New function. * parser.y (JSPLICE): New token. (json_val): Logic for unquoting. The array and hash rules must now be prepared to deal with json_vals and json_pairs now producing a list object instead of a hash or vector. That is the signal that the data contains active quasiquotes and must be translated to the special literal syntax for quasiquoted vectors and hashes. Here we also add the rules for ~ and ~* unquoting syntax, including managing the lexer's transition back to the JSON start condition. (json_vals, json_pairs): We add the logic here to recognize unquotes in quasiquoting state. This is more clever than the way it is done in the Lisp areas of the grammar. If no quasiquotes occur, we construct a vector or hash, respectively, and add to it. If unquotes occur and if we are nested in a quasiquote, we switch the object to a list, and continue it that way. (yybadtoken): Handle JSPLICE. * lex.yy.c.shipped, y.tab.c.shipped, y.tab.h.shipped: Updated.
Diffstat (limited to 'parser.y')
-rw-r--r--parser.y57
1 files changed, 47 insertions, 10 deletions
diff --git a/parser.y b/parser.y
index 7725111a..ddd78bd1 100644
--- a/parser.y
+++ b/parser.y
@@ -126,7 +126,7 @@ INLINE val expand_form_ver(val form, int ver)
%token <val> NUMBER METANUM JSKW
%token <val> HASH_N_EQUALS HASH_N_HASH
-%token <chr> REGCHAR REGTOKEN LITCHAR SPLICE OLD_AT
+%token <chr> REGCHAR REGTOKEN LITCHAR SPLICE JSPLICE OLD_AT
%token <chr> CONSDOT LAMBDOT UREFDOT OREFDOT UOREFDOT
%type <val> spec hash_semi_or_n_expr hash_semi_or_i_expr
@@ -949,7 +949,20 @@ json_val : NUMBER { $$ = $1; }
| '[' ']' { $$ = vector(0, nil); }
| '[' json_vals ']' { $$ = $2; }
| '{' '}' { $$ = make_hash(nil, nil, t); }
- | '{' json_pairs '}' { $$ = $2; }
+ | '{' json_pairs '}' { $$ = if3(hashp($2),
+ $2,
+ rl(cons(hash_lit_s,
+ cons(nil, $2)), $2)); }
+ | '~' { parser->quasi_level--; }
+ n_dot_expr { parser->quasi_level++;
+ end_of_json_unquote(scnr);
+ $$ = rl(rlc(list(sys_unquote_s, $3, nao), $3),
+ num(parser->lineno)); }
+ | JSPLICE { parser->quasi_level--; }
+ n_dot_expr { parser->quasi_level++;
+ end_of_json_unquote(scnr);
+ $$ = rl(rlc(list(sys_splice_s, $3, nao), $3),
+ num(parser->lineno)); }
| HASH_N_EQUALS { parser_circ_def(parser, $1, unique_s); }
json_val { parser_circ_def(parser, $1, $3);
$$ = $3; }
@@ -962,9 +975,19 @@ json_val : NUMBER { $$ = $1; }
yybadtok(yychar, lit("JSON hash")); }
;
-json_vals : json_val { $$ = vector(one, $1); }
- | json_vals ',' json_val { vec_push($1, $3);
- $$ = $1; }
+json_vals : json_val { $$ = if3(parser->quasi_level > 0 &&
+ unquotes_occur($1, 0),
+ cons($1, nil),
+ vector(one, $1)); }
+ | json_vals ',' json_val { if (consp($1))
+ { $$ = cons($3, $1); }
+ else if (parser->quasi_level > 0 &&
+ unquotes_occur($3, 0))
+ { val li = list_vec($1);
+ $$ = cons($3, li); }
+ else
+ { vec_push($1, $3);
+ $$ = $1; } }
| json_vals json_val { yyerr("missing comma in JSON array");
$$ = $1; }
| json_vals error { yyerr("bad element in JSON array");
@@ -972,14 +995,27 @@ json_vals : json_val { $$ = vector(one, $1); }
;
json_pairs : json_val ':' json_val { if (!stringp($1))
- yyerr("non-string key in JSON hash");
- $$ = make_hash(nil, nil, t);
- sethash($$, $1, $3); }
+ yyerr("non-string key in JSON hash");
+ if (parser->quasi_level > 0 &&
+ (unquotes_occur($1, 0) ||
+ unquotes_occur($3, 0)))
+ { $$ = cons(list($1, $3, nao), nil); }
+ else
+ { $$ = make_hash(nil, nil, t);
+ sethash($$, $1, $3); } }
| json_pairs ','
json_val ':' json_val { if (!stringp($3))
yyerr("non-string key in JSON hash");
- sethash($1, $3, $5);
- $$ = $1; }
+ if (consp($1))
+ { $$ = cons(list($3, $5, nao), $1); }
+ else if (parser->quasi_level > 0 &&
+ ((unquotes_occur($3, 0)) ||
+ unquotes_occur($5, 0)))
+ { val pa = hash_pairs($1);
+ $$ = cons(list($3, $5, nao), pa); }
+ else
+ { sethash($1, $3, $5);
+ $$ = $1; } }
| json_val json_val { yyerr("missing colon in JSON hash"); }
| json_val ':' json_val
error { yyerr("missing comma in JSON hash"); }
@@ -2057,6 +2093,7 @@ void yybadtoken(parser_t *parser, int tok, val context)
case REGTOKEN: problem = lit("regular expression token"); break;
case LITCHAR: problem = lit("string literal character"); break;
case SPLICE: problem = lit("*"); break;
+ case JSPLICE: problem = lit("~*"); break;
case CONSDOT:
case LAMBDOT: problem = lit("consing dot"); break;
case DOTDOT: problem = lit(".."); break;