Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 0 additions & 1 deletion Grammar/python.gram
Original file line number Diff line number Diff line change
Expand Up @@ -477,7 +477,6 @@ atom[expr_ty]:
| 'True' { _Py_Constant(Py_True, NULL, EXTRA) }
| 'False' { _Py_Constant(Py_False, NULL, EXTRA) }
| 'None' { _Py_Constant(Py_None, NULL, EXTRA) }
| '__new_parser__' { RAISE_SYNTAX_ERROR("You found it!") }
| &STRING strings
| NUMBER
| &'(' (tuple | group | genexp)
Expand Down
1 change: 0 additions & 1 deletion Lib/keyword.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,6 @@
'False',
'None',
'True',
'__new_parser__',
'and',
'as',
'assert',
Expand Down
1 change: 0 additions & 1 deletion Lib/pydoc.py
Original file line number Diff line number Diff line change
Expand Up @@ -1817,7 +1817,6 @@ class Helper:
'False': '',
'None': '',
'True': '',
'__new_parser__': '',
'and': 'BOOLEAN',
'as': 'with',
'assert': ('assert', ''),
Expand Down
54 changes: 10 additions & 44 deletions Parser/pegen/parse.c
Original file line number Diff line number Diff line change
Expand Up @@ -7,16 +7,16 @@ extern int Py_DebugFlag;
#else
#define D(x)
#endif
static const int n_keyword_lists = 15;
static const int n_keyword_lists = 9;
static KeywordToken *reserved_keywords[] = {
NULL,
NULL,
(KeywordToken[]) {
{"if", 510},
{"in", 518},
{"is", 526},
{"as", 531},
{"or", 532},
{"as", 530},
{"or", 531},
{NULL, -1},
},
(KeywordToken[]) {
Expand All @@ -25,7 +25,7 @@ static KeywordToken *reserved_keywords[] = {
{"for", 517},
{"def", 522},
{"not", 525},
{"and", 533},
{"and", 532},
{NULL, -1},
},
(KeywordToken[]) {
Expand Down Expand Up @@ -65,15 +65,6 @@ static KeywordToken *reserved_keywords[] = {
{"nonlocal", 509},
{NULL, -1},
},
NULL,
NULL,
NULL,
NULL,
NULL,
(KeywordToken[]) {
{"__new_parser__", 530},
{NULL, -1},
},
};
#define file_type 1000
#define interactive_type 1001
Expand Down Expand Up @@ -10567,7 +10558,6 @@ slice_rule(Parser *p)
// | 'True'
// | 'False'
// | 'None'
// | '__new_parser__'
// | &STRING strings
// | NUMBER
// | &'(' (tuple | group | genexp)
Expand Down Expand Up @@ -10711,30 +10701,6 @@ atom_rule(Parser *p)
D(fprintf(stderr, "%*c%s atom[%d-%d]: %s failed!\n", p->level, ' ',
p->error_indicator ? "ERROR!" : "-", _mark, p->mark, "'None'"));
}
{ // '__new_parser__'
if (p->error_indicator) {
D(p->level--);
return NULL;
}
D(fprintf(stderr, "%*c> atom[%d-%d]: %s\n", p->level, ' ', _mark, p->mark, "'__new_parser__'"));
Token * _keyword;
if (
(_keyword = _PyPegen_expect_token(p, 530)) // token='__new_parser__'
)
{
D(fprintf(stderr, "%*c+ atom[%d-%d]: %s succeeded!\n", p->level, ' ', _mark, p->mark, "'__new_parser__'"));
_res = RAISE_SYNTAX_ERROR ( "You found it!" );
if (_res == NULL && PyErr_Occurred()) {
p->error_indicator = 1;
D(p->level--);
return NULL;
}
goto done;
}
p->mark = _mark;
D(fprintf(stderr, "%*c%s atom[%d-%d]: %s failed!\n", p->level, ' ',
p->error_indicator ? "ERROR!" : "-", _mark, p->mark, "'__new_parser__'"));
}
{ // &STRING strings
if (p->error_indicator) {
D(p->level--);
Expand Down Expand Up @@ -17313,7 +17279,7 @@ _tmp_34_rule(Parser *p)
Token * _keyword;
expr_ty z;
if (
(_keyword = _PyPegen_expect_token(p, 531)) // token='as'
(_keyword = _PyPegen_expect_token(p, 530)) // token='as'
&&
(z = _PyPegen_name_token(p)) // NAME
)
Expand Down Expand Up @@ -17471,7 +17437,7 @@ _tmp_37_rule(Parser *p)
Token * _keyword;
expr_ty z;
if (
(_keyword = _PyPegen_expect_token(p, 531)) // token='as'
(_keyword = _PyPegen_expect_token(p, 530)) // token='as'
&&
(z = _PyPegen_name_token(p)) // NAME
)
Expand Down Expand Up @@ -17971,7 +17937,7 @@ _tmp_46_rule(Parser *p)
Token * _keyword;
expr_ty t;
if (
(_keyword = _PyPegen_expect_token(p, 531)) // token='as'
(_keyword = _PyPegen_expect_token(p, 530)) // token='as'
&&
(t = target_rule(p)) // target
)
Expand Down Expand Up @@ -18086,7 +18052,7 @@ _tmp_48_rule(Parser *p)
Token * _keyword;
expr_ty z;
if (
(_keyword = _PyPegen_expect_token(p, 531)) // token='as'
(_keyword = _PyPegen_expect_token(p, 530)) // token='as'
&&
(z = _PyPegen_name_token(p)) // NAME
)
Expand Down Expand Up @@ -23892,7 +23858,7 @@ _tmp_144_rule(Parser *p)
Token * _keyword;
expr_ty c;
if (
(_keyword = _PyPegen_expect_token(p, 532)) // token='or'
(_keyword = _PyPegen_expect_token(p, 531)) // token='or'
&&
(c = conjunction_rule(p)) // conjunction
)
Expand Down Expand Up @@ -23936,7 +23902,7 @@ _tmp_145_rule(Parser *p)
Token * _keyword;
expr_ty c;
if (
(_keyword = _PyPegen_expect_token(p, 533)) // token='and'
(_keyword = _PyPegen_expect_token(p, 532)) // token='and'
&&
(c = inversion_rule(p)) // inversion
)
Expand Down