X-Git-Url: https://git.saurik.com/bison.git/blobdiff_plain/776209d6c747eaf2efd87d818464f82fb14448b1..7431029172f190eb3c5c63a8828e723276551ca6:/tests/regression.at diff --git a/tests/regression.at b/tests/regression.at index 2f10d066..3620200a 100644 --- a/tests/regression.at +++ b/tests/regression.at @@ -1,5 +1,5 @@ # Bison Regressions. -*- Autotest -*- -# Copyright 2001 Free Software Foundation, Inc. +# Copyright (C) 2001, 2002 Free Software Foundation, Inc. # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by @@ -18,247 +18,102 @@ AT_BANNER([[Regression tests.]]) -## ------------------ ## -## Duplicate string. ## -## ------------------ ## +## ------------------------- ## +## Early token definitions. ## +## ------------------------- ## -AT_SETUP([Duplicate string]) -AT_DATA([duplicate.y], -[[/* `Bison -v' used to dump core when two tokens are defined with the same - string, as LE and GE below. */ +AT_SETUP([Early token definitions]) -%token NUM -%token LE "<=" -%token GE "<=" +# Found in GCJ: they expect the tokens to be defined before the user +# prologue, so that they can use the token definitions in it. + +AT_DATA([input.y], +[[%{ +void yyerror (const char *s); +int yylex (void); +%} +%union +{ + int val; +}; +%{ +#ifndef MY_TOKEN +# error "MY_TOKEN not defined." +#endif +%} +%token MY_TOKEN %% -exp: '(' exp ')' | NUM ; +exp: MY_TOKEN; %% ]]) -AT_CHECK([bison -v duplicate.y -o duplicate.c], 0, ignore, ignore) +AT_CHECK([bison input.y -o input.c]) +AT_CHECK([$CC $CFLAGS $CPPFLAGS input.c -c]) AT_CLEANUP -## ------------------------- ## -## Unresolved SR Conflicts. ## -## ------------------------- ## - -AT_SETUP([Unresolved SR Conflicts]) - -AT_DATA([input.y], -[[%token NUM OP -%% -exp: exp OP exp | NUM; -]]) - -AT_CHECK([bison input.y -o input.c -v], 0, [], -[input.y contains 1 shift/reduce conflict. -]) - -# Check the contents of the report. -AT_CHECK([cat input.output], [], -[[State 4 contains 1 shift/reduce conflict. - - -Grammar - - Number, Line, Rule - 1 3 exp -> exp OP exp - 2 3 exp -> NUM - - -Terminals, with rules where they appear - -$ (-1) -error (256) -NUM (257) 2 -OP (258) 1 - - -Nonterminals, with rules where they appear - -exp (5) - on left: 1 2, on right: 1 - - -state 0 - - NUM shift, and go to state 1 - - exp go to state 2 - - - -state 1 - - exp -> NUM . (rule 2) - - $default reduce using rule 2 (exp) - - - -state 2 - - exp -> exp . OP exp (rule 1) - - $ shift, and go to state 5 - OP shift, and go to state 3 - - - -state 3 - exp -> exp OP . exp (rule 1) +## ---------------- ## +## Braces parsing. ## +## ---------------- ## - NUM shift, and go to state 1 - exp go to state 4 - - - -state 4 - - exp -> exp . OP exp (rule 1) - exp -> exp OP exp . (rule 1) - - OP shift, and go to state 3 - - OP [reduce using rule 1 (exp)] - $default reduce using rule 1 (exp) - - - -state 5 - - $ shift, and go to state 6 - - - -state 6 - - $default accept - - -]]) - -AT_CLEANUP - - -## --------------------- ## -## Solved SR Conflicts. ## -## --------------------- ## - -AT_SETUP([Solved SR Conflicts]) +AT_SETUP([Braces parsing]) AT_DATA([input.y], -[[%token NUM OP -%right OP +[[/* Bison used to swallow the character after `}'. */ + +%% +exp: { tests = {{{{{{{{{{}}}}}}}}}}; }; %% -exp: exp OP exp | NUM; ]]) -AT_CHECK([bison input.y -o input.c -v], 0, [], []) - -# Check the contents of the report. -AT_CHECK([cat input.output], [], -[[Conflict in state 4 between rule 1 and token OP resolved as shift. - - -Grammar - - Number, Line, Rule - 1 4 exp -> exp OP exp - 2 4 exp -> NUM - - -Terminals, with rules where they appear - -$ (-1) -error (256) -NUM (257) 2 -OP (258) 1 - - -Nonterminals, with rules where they appear - -exp (5) - on left: 1 2, on right: 1 - - -state 0 - - NUM shift, and go to state 1 - - exp go to state 2 - - - -state 1 - - exp -> NUM . (rule 2) - - $default reduce using rule 2 (exp) - - - -state 2 - - exp -> exp . OP exp (rule 1) - - $ shift, and go to state 5 - OP shift, and go to state 3 - - - -state 3 - - exp -> exp OP . exp (rule 1) - - NUM shift, and go to state 1 - - exp go to state 4 - - - -state 4 - - exp -> exp . OP exp (rule 1) - exp -> exp OP exp . (rule 1) - - OP shift, and go to state 3 - - $default reduce using rule 1 (exp) +AT_CHECK([bison -v input.y -o input.c]) +AT_CHECK([fgrep 'tests = {{{{{{{{{{}}}}}}}}}};' input.c], 0, [ignore]) +AT_CLEANUP -state 5 - $ shift, and go to state 6 +## ------------------ ## +## Duplicate string. ## +## ------------------ ## +AT_SETUP([Duplicate string]) -state 6 +AT_DATA([input.y], +[[/* `Bison -v' used to dump core when two tokens are defined with the same + string, as LE and GE below. */ - $default accept +%token NUM +%token LE "<=" +%token GE "<=" +%% +exp: '(' exp ')' | NUM ; +%% +]]) +AT_CHECK([bison -v input.y -o input.c], 0, [], +[[input.y:6: warning: symbol `"<="' used more than once as a literal string ]]) AT_CLEANUP - - ## ------------------- ## ## Rule Line Numbers. ## ## ------------------- ## AT_SETUP([Rule Line Numbers]) +AT_KEYWORDS([report]) + AT_DATA([input.y], [[%% expr: @@ -286,25 +141,29 @@ expr: { -} +}; ]]) -AT_CHECK([bison input.y -o input.c -v], 0, [], []) +AT_CHECK([bison input.y -o input.c -v]) # Check the contents of the report. AT_CHECK([cat input.output], [], [[Grammar - Number, Line, Rule - 1 2 @1 -> /* empty */ - 2 2 expr -> 'a' @1 'b' - 3 15 @2 -> /* empty */ - 4 15 expr -> @2 'c' + 0 $axiom: expr $ + + 1 @1: /* empty */ + + 2 expr: 'a' @1 'b' + + 3 @2: /* empty */ + + 4 expr: @2 'c' Terminals, with rules where they appear -$ (-1) +$ (0) 0 'a' (97) 2 'b' (98) 2 'c' (99) 4 @@ -313,22 +172,26 @@ error (256) Nonterminals, with rules where they appear -expr (6) - on left: 2 4 -@1 (7) +$axiom (6) + on left: 0 +expr (7) + on left: 2 4, on right: 0 +@1 (8) on left: 1, on right: 2 -@2 (8) +@2 (9) on left: 3, on right: 4 state 0 + $axiom -> . expr $ (rule 0) + 'a' shift, and go to state 1 $default reduce using rule 3 (@2) - expr go to state 6 - @2 go to state 2 + expr go to state 2 + @2 go to state 3 @@ -338,115 +201,62 @@ state 1 $default reduce using rule 1 (@1) - @1 go to state 3 + @1 go to state 4 state 2 - expr -> @2 . 'c' (rule 4) + $axiom -> expr . $ (rule 0) - 'c' shift, and go to state 4 + $ shift, and go to state 5 state 3 - expr -> 'a' @1 . 'b' (rule 2) + expr -> @2 . 'c' (rule 4) - 'b' shift, and go to state 5 + 'c' shift, and go to state 6 state 4 - expr -> @2 'c' . (rule 4) + expr -> 'a' @1 . 'b' (rule 2) - $default reduce using rule 4 (expr) + 'b' shift, and go to state 7 state 5 - expr -> 'a' @1 'b' . (rule 2) - - $default reduce using rule 2 (expr) + $axiom -> expr $ . (rule 0) + $default accept state 6 - $ shift, and go to state 7 - - - -state 7 + expr -> @2 'c' . (rule 4) - $default accept + $default reduce using rule 4 (expr) -]]) -AT_CLEANUP +state 7 + expr -> 'a' @1 'b' . (rule 2) + $default reduce using rule 2 (expr) -## -------------------- ## -## %expect not enough. ## -## -------------------- ## -AT_SETUP([%expect not enough]) -AT_DATA([input.y], -[[%token NUM OP -%expect 0 -%% -exp: exp OP exp | NUM; ]]) -AT_CHECK([bison input.y -o input.c], 1, [], -[input.y contains 1 shift/reduce conflict. -expected 0 shift/reduce conflicts -]) AT_CLEANUP -## --------------- ## -## %expect right. ## -## --------------- ## - -AT_SETUP([%expect right]) - -AT_DATA([input.y], -[[%token NUM OP -%expect 1 -%% -exp: exp OP exp | NUM; -]]) - -AT_CHECK([bison input.y -o input.c], 0) -AT_CLEANUP - - -## ------------------ ## -## %expect too much. ## -## ------------------ ## - -AT_SETUP([%expect too much]) - -AT_DATA([input.y], -[[%token NUM OP -%expect 2 -%% -exp: exp OP exp | NUM; -]]) - -AT_CHECK([bison input.y -o input.c], 1, [], -[input.y contains 1 shift/reduce conflict. -expected 2 shift/reduce conflicts -]) -AT_CLEANUP - ## ---------------------- ## ## Mixing %token styles. ## @@ -465,159 +275,86 @@ exp: ; %% ]]) -AT_CHECK([bison -v input.y -o input.c], 0, ignore, ignore) +AT_CHECK([bison -v input.y -o input.c]) AT_CLEANUP -## ---------------------- ## -## %union and --defines. ## -## ---------------------- ## - - -AT_SETUP([%union and --defines]) - -AT_DATA([union.y], -[%union -{ - int integer; - char *string ; -} -%% -exp: {}; -]) - -AT_CHECK([bison --defines union.y]) - -AT_CLEANUP +## ---------------- ## +## Invalid inputs. ## +## ---------------- ## -## --------------------------------------- ## -## Duplicate '/' in C comments in %union ## -## --------------------------------------- ## - - -AT_SETUP([%union and C comments]) - -AT_DATA([union-comment.y], -[%union -{ - /* The int. */ int integer; - /* The string. */ char *string ; -} -%% -exp: {}; -]) - -AT_CHECK([bison union-comment.y]) -AT_CHECK([fgrep '//*' union-comment.tab.c], [1], []) - -AT_CLEANUP - - -## ----------------- ## -## Invalid input 1. ## -## ----------------- ## - - -AT_SETUP([Invalid input: 1]) +AT_SETUP([Invalid inputs]) AT_DATA([input.y], [[%% ? -]]) - -AT_CHECK([bison input.y], [1], [], -[[input.y:2: invalid input: `?' -input.y:3: fatal error: no rules in the input grammar -]]) - -AT_CLEANUP - - -## ----------------- ## -## Invalid input 2. ## -## ----------------- ## - - -AT_SETUP([Invalid input: 2]) - -AT_DATA([input.y], -[[%% default: 'a' } +%& +%a +%- +%{ ]]) AT_CHECK([bison input.y], [1], [], -[[input.y:2: invalid input: `}' +[[input.y:2.1: invalid character: `?' +input.y:3.14: invalid character: `}' +input.y:4.1: invalid character: `%' +input.y:4.2: invalid character: `&' +input.y:5.1: invalid character: `%' +input.y:6.1: invalid character: `%' +input.y:6.2: invalid character: `-' +input.y:7.1-8.0: unexpected end of file in a prologue +input.y:7.1-8.0: parse error, unexpected PROLOGUE, expecting ";" or "|" +input.y:5.2: symbol a is used, but is not defined as a token and has no rules ]]) AT_CLEANUP -## -------------------- ## -## Invalid %directive. ## -## -------------------- ## +## ------------------- ## +## Token definitions. ## +## ------------------- ## -AT_SETUP([Invalid %directive]) +AT_SETUP([Token definitions]) +# Bison managed, when fed with `%token 'f' "f"' to #define 'f'! AT_DATA([input.y], -[[%invalid -]]) - -AT_CHECK([bison input.y], [1], [], -[[input.y:1: unrecognized: %invalid -input.y:1: Skipping to next % -input.y:2: fatal error: no input grammar +[%{ +void yyerror (const char *s); +int yylex (void); +%} +[%token YYEOF 0 "end of file" +%token 'a' "a" +%token b "b" +%token c 'c' +%token 'd' d +%% +exp: "a"; ]]) +AT_CHECK([bison input.y -o input.c]) +AT_CHECK([$CC $CFLAGS $CPPFLAGS input.c -c]) AT_CLEANUP -## --------------------- ## -## Invalid CPP headers. ## -## --------------------- ## - -# AT_TEST_CPP_GUARD_H([INPUT-FILE-BASE) -# ------------------------------------- -m4_define([AT_TEST_CPP_GUARD_H], -[AT_SETUP([Invalid CPP guards: $1]) - -# Possibly create inner directories. -dirname=`AS_DIRNAME([$1])` -AS_MKDIR_P([$dirname]) - -AT_DATA([$1.y], -[%% -dummy: -]) - -AT_CHECK([bison --defines=$1.h $1.y]) - -# CPP should be happy with it. -AT_CHECK([$CC -E $1.h], 0, [ignore]) - -AT_CLEANUP -]) - -AT_TEST_CPP_GUARD_H([input/input]) -AT_TEST_CPP_GUARD_H([9foo]) - - - -## ------- ## -## Web2c. ## -## ------- ## +## -------------- ## +## Web2c Report. ## +## -------------- ## # The generation of the reduction was once wrong in Bison, and made it # miss some reductions. In the following test case, the reduction on # `undef_id_tok' in state 1 was missing. This is stripped down from # the actual web2c.y. -AT_SETUP([Web2c]) +AT_SETUP([Web2c Report]) + +AT_KEYWORDS([report]) AT_DATA([input.y], [[%token undef_id_tok const_id_tok @@ -645,66 +382,207 @@ AT_CHECK([bison -v input.y]) AT_CHECK([sed -n 's/ *$//;/^$/!p' input.output], 0, [[Grammar - Number, Line, Rule - 1 6 CONST_DEC_PART -> CONST_DEC_LIST - 2 10 CONST_DEC_LIST -> CONST_DEC - 3 12 CONST_DEC_LIST -> CONST_DEC_LIST CONST_DEC - 4 15 @1 -> /* empty */ - 5 15 CONST_DEC -> @1 undef_id_tok '=' const_id_tok ';' + 0 $axiom: CONST_DEC_PART $ + 1 CONST_DEC_PART: CONST_DEC_LIST + 2 CONST_DEC_LIST: CONST_DEC + 3 | CONST_DEC_LIST CONST_DEC + 4 @1: /* empty */ + 5 CONST_DEC: @1 undef_id_tok '=' const_id_tok ';' Terminals, with rules where they appear -$ (-1) +$ (0) 0 ';' (59) 5 '=' (61) 5 error (256) -undef_id_tok (257) 5 -const_id_tok (258) 5 +undef_id_tok (258) 5 +const_id_tok (259) 5 Nonterminals, with rules where they appear -CONST_DEC_PART (7) - on left: 1 -CONST_DEC_LIST (8) +$axiom (7) + on left: 0 +CONST_DEC_PART (8) + on left: 1, on right: 0 +CONST_DEC_LIST (9) on left: 2 3, on right: 1 3 -CONST_DEC (9) +CONST_DEC (10) on left: 5, on right: 2 3 -@1 (10) +@1 (11) on left: 4, on right: 5 state 0 + $axiom -> . CONST_DEC_PART $ (rule 0) $default reduce using rule 4 (@1) - CONST_DEC_PART go to state 9 - CONST_DEC_LIST go to state 1 - CONST_DEC go to state 2 - @1 go to state 3 + CONST_DEC_PART go to state 1 + CONST_DEC_LIST go to state 2 + CONST_DEC go to state 3 + @1 go to state 4 state 1 + $axiom -> CONST_DEC_PART . $ (rule 0) + $ shift, and go to state 5 +state 2 CONST_DEC_PART -> CONST_DEC_LIST . (rule 1) CONST_DEC_LIST -> CONST_DEC_LIST . CONST_DEC (rule 3) undef_id_tok reduce using rule 4 (@1) $default reduce using rule 1 (CONST_DEC_PART) - CONST_DEC go to state 4 - @1 go to state 3 -state 2 + CONST_DEC go to state 6 + @1 go to state 4 +state 3 CONST_DEC_LIST -> CONST_DEC . (rule 2) $default reduce using rule 2 (CONST_DEC_LIST) -state 3 - CONST_DEC -> @1 . undef_id_tok '=' const_id_tok ';' (rule 5) - undef_id_tok shift, and go to state 5 state 4 + CONST_DEC -> @1 . undef_id_tok '=' const_id_tok ';' (rule 5) + undef_id_tok shift, and go to state 7 +state 5 + $axiom -> CONST_DEC_PART $ . (rule 0) + $default accept +state 6 CONST_DEC_LIST -> CONST_DEC_LIST CONST_DEC . (rule 3) $default reduce using rule 3 (CONST_DEC_LIST) -state 5 +state 7 CONST_DEC -> @1 undef_id_tok . '=' const_id_tok ';' (rule 5) - '=' shift, and go to state 6 -state 6 + '=' shift, and go to state 8 +state 8 CONST_DEC -> @1 undef_id_tok '=' . const_id_tok ';' (rule 5) - const_id_tok shift, and go to state 7 -state 7 + const_id_tok shift, and go to state 9 +state 9 CONST_DEC -> @1 undef_id_tok '=' const_id_tok . ';' (rule 5) - ';' shift, and go to state 8 -state 8 + ';' shift, and go to state 10 +state 10 CONST_DEC -> @1 undef_id_tok '=' const_id_tok ';' . (rule 5) $default reduce using rule 5 (CONST_DEC) -state 9 - $ shift, and go to state 10 -state 10 - $default accept +]]) + +AT_CLEANUP + + +## --------------- ## +## Web2c Actions. ## +## --------------- ## + +# The generation of the mapping `state -> action' was once wrong in +# extremely specific situations. web2c.y exhibits this situation. +# Below is a stripped version of the grammar. It looks like one can +# simplify it further, but just don't: it is tuned to exhibit a bug, +# which disapears when applying sane grammar transformations. +# +# It used to be wrong on yydefact only: +# +# static const short yydefact[] = +# { +# - 2, 0, 1, 0, 0, 2, 3, 2, 5, 4, +# + 2, 0, 1, 0, 0, 0, 3, 2, 5, 4, +# 0, 0 +# }; +# +# but let's check all the tables. + + +AT_SETUP([Web2c Actions]) + +AT_KEYWORDS([report]) + +AT_DATA([input.y], +[[%% +statement: struct_stat; +struct_stat: /* empty. */ | if else; +if: "if" "const" "then" statement; +else: "else" statement; +%% +]]) + +AT_CHECK([bison -v input.y -o input.c]) + +# Check only the tables. We don't use --no-parser, because it is +# still to be implemented in the experimental branch of Bison. +AT_CHECK([[sed -n 's/ *$//;/^static const.*\[\] =/,/^}/p' input.c]], 0, +[[static const unsigned char yytranslate[] = +{ + 0, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, + 2, 2, 2, 2, 2, 2, 1, 2, 3, 4, + 5, 6 +}; +static const unsigned char yyprhs[] = +{ + 0, 0, 3, 5, 6, 9, 14 +}; +static const signed char yyrhs[] = +{ + 8, 0, -1, 9, -1, -1, 10, 11, -1, 3, + 4, 5, 8, -1, 6, 8, -1 +}; +static const unsigned char yyrline[] = +{ + 0, 2, 2, 3, 3, 4, 5 +}; +static const char *const yytname[] = +{ + "$", "error", "$undefined.", "\"if\"", "\"const\"", "\"then\"", + "\"else\"", "$axiom", "statement", "struct_stat", "if", "else", 0 +}; +static const short yytoknum[] = +{ + 0, 256, 257, 258, 259, 260, 261, -1 +}; +static const unsigned char yyr1[] = +{ + 0, 7, 8, 9, 9, 10, 11 +}; +static const unsigned char yyr2[] = +{ + 0, 2, 1, 0, 2, 4, 2 +}; +static const short yydefact[] = +{ + 3, 0, 0, 2, 0, 0, 0, 3, 4, 3, + 6, 5 +}; +static const short yydefgoto[] = +{ + -1, 2, 3, 4, 8 +}; +static const short yypact[] = +{ + -2, -1, 4,-32768, 0, 2,-32768, -2,-32768, -2, + -32768,-32768 +}; +static const short yypgoto[] = +{ + -32768, -7,-32768,-32768,-32768 +}; +static const short yytable[] = +{ + 10, 1, 11, 5, 6, 0, 7, 9 +}; +static const short yycheck[] = +{ + 7, 3, 9, 4, 0, -1, 6, 5 +}; +static const unsigned char yystos[] = +{ + 0, 3, 8, 9, 10, 4, 0, 6, 11, 5, + 8, 8 +}; ]]) AT_CLEANUP