/** @file | |
Parser-tokenizer link implementation. | |
Copyright (c) 2014, Intel Corporation. All rights reserved.<BR> | |
This program and the accompanying materials are licensed and made available under | |
the terms and conditions of the BSD License that accompanies this distribution. | |
The full text of the license may be found at | |
http://opensource.org/licenses/bsd-license.php. | |
THE PROGRAM IS DISTRIBUTED UNDER THE BSD LICENSE ON AN "AS IS" BASIS, | |
WITHOUT WARRANTIES OR REPRESENTATIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED. | |
**/ | |
#include "pgenheaders.h" | |
#include "tokenizer.h" | |
#include "node.h" | |
#include "grammar.h" | |
#include "parser.h" | |
#include "parsetok.h" | |
#include "errcode.h" | |
#include "graminit.h" | |
int Py_TabcheckFlag; | |
/* Forward */ | |
static node *parsetok(struct tok_state *, grammar *, int, perrdetail *, int *); | |
static void initerr(perrdetail *err_ret, const char* filename); | |
/* Parse input coming from a string. Return error code, print some errors. */ | |
node * | |
PyParser_ParseString(const char *s, grammar *g, int start, perrdetail *err_ret) | |
{ | |
return PyParser_ParseStringFlagsFilename(s, NULL, g, start, err_ret, 0); | |
} | |
node * | |
PyParser_ParseStringFlags(const char *s, grammar *g, int start, | |
perrdetail *err_ret, int flags) | |
{ | |
return PyParser_ParseStringFlagsFilename(s, NULL, | |
g, start, err_ret, flags); | |
} | |
node * | |
PyParser_ParseStringFlagsFilename(const char *s, const char *filename, | |
grammar *g, int start, | |
perrdetail *err_ret, int flags) | |
{ | |
int iflags = flags; | |
return PyParser_ParseStringFlagsFilenameEx(s, filename, g, start, | |
err_ret, &iflags); | |
} | |
node * | |
PyParser_ParseStringFlagsFilenameEx(const char *s, const char *filename, | |
grammar *g, int start, | |
perrdetail *err_ret, int *flags) | |
{ | |
struct tok_state *tok; | |
initerr(err_ret, filename); | |
if ((tok = PyTokenizer_FromString(s, start == file_input)) == NULL) { | |
err_ret->error = PyErr_Occurred() ? E_DECODE : E_NOMEM; | |
return NULL; | |
} | |
tok->filename = filename ? filename : "<string>"; | |
if (Py_TabcheckFlag || Py_VerboseFlag) { | |
tok->altwarning = (tok->filename != NULL); | |
if (Py_TabcheckFlag >= 2) | |
tok->alterror++; | |
} | |
return parsetok(tok, g, start, err_ret, flags); | |
} | |
/* Parse input coming from a file. Return error code, print some errors. */ | |
node * | |
PyParser_ParseFile(FILE *fp, const char *filename, grammar *g, int start, | |
char *ps1, char *ps2, perrdetail *err_ret) | |
{ | |
return PyParser_ParseFileFlags(fp, filename, g, start, ps1, ps2, | |
err_ret, 0); | |
} | |
node * | |
PyParser_ParseFileFlags(FILE *fp, const char *filename, grammar *g, int start, | |
char *ps1, char *ps2, perrdetail *err_ret, int flags) | |
{ | |
int iflags = flags; | |
return PyParser_ParseFileFlagsEx(fp, filename, g, start, ps1, ps2, err_ret, &iflags); | |
} | |
node * | |
PyParser_ParseFileFlagsEx(FILE *fp, const char *filename, grammar *g, int start, | |
char *ps1, char *ps2, perrdetail *err_ret, int *flags) | |
{ | |
struct tok_state *tok; | |
initerr(err_ret, filename); | |
if ((tok = PyTokenizer_FromFile(fp, ps1, ps2)) == NULL) { | |
err_ret->error = E_NOMEM; | |
return NULL; | |
} | |
tok->filename = filename; | |
if (Py_TabcheckFlag || Py_VerboseFlag) { | |
tok->altwarning = (filename != NULL); | |
if (Py_TabcheckFlag >= 2) | |
tok->alterror++; | |
} | |
return parsetok(tok, g, start, err_ret, flags); | |
} | |
#if 0 | |
static char with_msg[] = | |
"%s:%d: Warning: 'with' will become a reserved keyword in Python 2.6\n"; | |
static char as_msg[] = | |
"%s:%d: Warning: 'as' will become a reserved keyword in Python 2.6\n"; | |
static void | |
warn(const char *msg, const char *filename, int lineno) | |
{ | |
if (filename == NULL) | |
filename = "<string>"; | |
PySys_WriteStderr(msg, filename, lineno); | |
} | |
#endif | |
/* Parse input coming from the given tokenizer structure. | |
Return error code. */ | |
static node * | |
parsetok(struct tok_state *tok, grammar *g, int start, perrdetail *err_ret, | |
int *flags) | |
{ | |
parser_state *ps; | |
node *n; | |
int started = 0; | |
//int handling_import = 0; | |
//int handling_with = 0; | |
if ((ps = PyParser_New(g, start)) == NULL) { | |
fprintf(stderr, "no mem for new parser\n"); | |
err_ret->error = E_NOMEM; | |
PyTokenizer_Free(tok); | |
return NULL; | |
} | |
#ifdef PY_PARSER_REQUIRES_FUTURE_KEYWORD | |
if (*flags & PyPARSE_PRINT_IS_FUNCTION) { | |
ps->p_flags |= CO_FUTURE_PRINT_FUNCTION; | |
} | |
if (*flags & PyPARSE_UNICODE_LITERALS) { | |
ps->p_flags |= CO_FUTURE_UNICODE_LITERALS; | |
} | |
#endif | |
for (;;) { | |
char *a, *b; | |
int type; | |
size_t len; | |
char *str; | |
int col_offset; | |
type = PyTokenizer_Get(tok, &a, &b); | |
if (type == ERRORTOKEN) { | |
err_ret->error = tok->done; | |
break; | |
} | |
if (type == ENDMARKER && started) { | |
type = NEWLINE; /* Add an extra newline */ | |
//handling_with = 0; | |
//handling_import = 0; | |
started = 0; | |
/* Add the right number of dedent tokens, | |
except if a certain flag is given -- | |
codeop.py uses this. */ | |
if (tok->indent && | |
!(*flags & PyPARSE_DONT_IMPLY_DEDENT)) | |
{ | |
tok->pendin = -tok->indent; | |
tok->indent = 0; | |
} | |
} | |
else | |
started = 1; | |
len = b - a; /* XXX this may compute NULL - NULL */ | |
str = (char *) PyObject_MALLOC(len + 1); | |
if (str == NULL) { | |
fprintf(stderr, "no mem for next token\n"); | |
err_ret->error = E_NOMEM; | |
break; | |
} | |
if (len > 0) | |
strncpy(str, a, len); | |
str[len] = '\0'; | |
#ifdef PY_PARSER_REQUIRES_FUTURE_KEYWORD | |
#endif | |
if (a >= tok->line_start) | |
col_offset = a - tok->line_start; | |
else | |
col_offset = -1; | |
if ((err_ret->error = | |
PyParser_AddToken(ps, (int)type, str, tok->lineno, col_offset, | |
&(err_ret->expected))) != E_OK) { | |
if (err_ret->error != E_DONE) { | |
PyObject_FREE(str); | |
err_ret->token = type; | |
} | |
break; | |
} | |
} | |
if (err_ret->error == E_DONE) { | |
n = ps->p_tree; | |
ps->p_tree = NULL; | |
} | |
else | |
n = NULL; | |
#ifdef PY_PARSER_REQUIRES_FUTURE_KEYWORD | |
*flags = ps->p_flags; | |
#endif | |
PyParser_Delete(ps); | |
if (n == NULL) { | |
if (tok->lineno <= 1 && tok->done == E_EOF) | |
err_ret->error = E_EOF; | |
err_ret->lineno = tok->lineno; | |
if (tok->buf != NULL) { | |
char *text = NULL; | |
size_t len; | |
assert(tok->cur - tok->buf < INT_MAX); | |
err_ret->offset = (int)(tok->cur - tok->buf); | |
len = tok->inp - tok->buf; | |
#ifdef Py_USING_UNICODE | |
text = PyTokenizer_RestoreEncoding(tok, len, &err_ret->offset); | |
#endif | |
if (text == NULL) { | |
text = (char *) PyObject_MALLOC(len + 1); | |
if (text != NULL) { | |
if (len > 0) | |
strncpy(text, tok->buf, len); | |
text[len] = '\0'; | |
} | |
} | |
err_ret->text = text; | |
} | |
} else if (tok->encoding != NULL) { | |
/* 'nodes->n_str' uses PyObject_*, while 'tok->encoding' was | |
* allocated using PyMem_ | |
*/ | |
node* r = PyNode_New(encoding_decl); | |
if (r) | |
r->n_str = PyObject_MALLOC(strlen(tok->encoding)+1); | |
if (!r || !r->n_str) { | |
err_ret->error = E_NOMEM; | |
if (r) | |
PyObject_FREE(r); | |
n = NULL; | |
goto done; | |
} | |
strcpy(r->n_str, tok->encoding); | |
PyMem_FREE(tok->encoding); | |
tok->encoding = NULL; | |
r->n_nchildren = 1; | |
r->n_child = n; | |
n = r; | |
} | |
done: | |
PyTokenizer_Free(tok); | |
return n; | |
} | |
static void | |
initerr(perrdetail *err_ret, const char *filename) | |
{ | |
err_ret->error = E_OK; | |
err_ret->filename = filename; | |
err_ret->lineno = 0; | |
err_ret->offset = 0; | |
err_ret->text = NULL; | |
err_ret->token = -1; | |
err_ret->expected = -1; | |
} |