| /* | 
 |  * JSON lexer | 
 |  * | 
 |  * Copyright IBM, Corp. 2009 | 
 |  * | 
 |  * Authors: | 
 |  *  Anthony Liguori   <aliguori@us.ibm.com> | 
 |  * | 
 |  * This work is licensed under the terms of the GNU LGPL, version 2.1 or later. | 
 |  * See the COPYING.LIB file in the top-level directory. | 
 |  * | 
 |  */ | 
 |  | 
 | #include "qemu/osdep.h" | 
 | #include "json-parser-int.h" | 
 |  | 
 | #define MAX_TOKEN_SIZE (64ULL << 20) | 
 |  | 
 | /* | 
 |  * From RFC 8259 "The JavaScript Object Notation (JSON) Data | 
 |  * Interchange Format", with [comments in brackets]: | 
 |  * | 
 |  * The set of tokens includes six structural characters, strings, | 
 |  * numbers, and three literal names. | 
 |  * | 
 |  * These are the six structural characters: | 
 |  * | 
 |  *    begin-array     = ws %x5B ws  ; [ left square bracket | 
 |  *    begin-object    = ws %x7B ws  ; { left curly bracket | 
 |  *    end-array       = ws %x5D ws  ; ] right square bracket | 
 |  *    end-object      = ws %x7D ws  ; } right curly bracket | 
 |  *    name-separator  = ws %x3A ws  ; : colon | 
 |  *    value-separator = ws %x2C ws  ; , comma | 
 |  * | 
 |  * Insignificant whitespace is allowed before or after any of the six | 
 |  * structural characters. | 
 |  * [This lexer accepts it before or after any token, which is actually | 
 |  * the same, as the grammar always has structural characters between | 
 |  * other tokens.] | 
 |  * | 
 |  *    ws = *( | 
 |  *           %x20 /              ; Space | 
 |  *           %x09 /              ; Horizontal tab | 
 |  *           %x0A /              ; Line feed or New line | 
 |  *           %x0D )              ; Carriage return | 
 |  * | 
 |  * [...] three literal names: | 
 |  *    false null true | 
 |  *  [This lexer accepts [a-z]+, and leaves rejecting unknown literal | 
 |  *  names to the parser.] | 
 |  * | 
 |  * [Numbers:] | 
 |  * | 
 |  *    number = [ minus ] int [ frac ] [ exp ] | 
 |  *    decimal-point = %x2E       ; . | 
 |  *    digit1-9 = %x31-39         ; 1-9 | 
 |  *    e = %x65 / %x45            ; e E | 
 |  *    exp = e [ minus / plus ] 1*DIGIT | 
 |  *    frac = decimal-point 1*DIGIT | 
 |  *    int = zero / ( digit1-9 *DIGIT ) | 
 |  *    minus = %x2D               ; - | 
 |  *    plus = %x2B                ; + | 
 |  *    zero = %x30                ; 0 | 
 |  * | 
 |  * [Strings:] | 
 |  *    string = quotation-mark *char quotation-mark | 
 |  * | 
 |  *    char = unescaped / | 
 |  *        escape ( | 
 |  *            %x22 /          ; "    quotation mark  U+0022 | 
 |  *            %x5C /          ; \    reverse solidus U+005C | 
 |  *            %x2F /          ; /    solidus         U+002F | 
 |  *            %x62 /          ; b    backspace       U+0008 | 
 |  *            %x66 /          ; f    form feed       U+000C | 
 |  *            %x6E /          ; n    line feed       U+000A | 
 |  *            %x72 /          ; r    carriage return U+000D | 
 |  *            %x74 /          ; t    tab             U+0009 | 
 |  *            %x75 4HEXDIG )  ; uXXXX                U+XXXX | 
 |  *    escape = %x5C              ; \ | 
 |  *    quotation-mark = %x22      ; " | 
 |  *    unescaped = %x20-21 / %x23-5B / %x5D-10FFFF | 
 |  *    [This lexer accepts any non-control character after escape, and | 
 |  *    leaves rejecting invalid ones to the parser.] | 
 |  * | 
 |  * | 
 |  * Extensions over RFC 8259: | 
 |  * - Extra escape sequence in strings: | 
 |  *   0x27 (apostrophe) is recognized after escape, too | 
 |  * - Single-quoted strings: | 
 |  *   Like double-quoted strings, except they're delimited by %x27 | 
 |  *   (apostrophe) instead of %x22 (quotation mark), and can't contain | 
 |  *   unescaped apostrophe, but can contain unescaped quotation mark. | 
 |  * - Interpolation, if enabled: | 
 |  *   The lexer accepts %[A-Za-z0-9]*, and leaves rejecting invalid | 
 |  *   ones to the parser. | 
 |  * | 
 |  * Note: | 
 |  * - Input must be encoded in modified UTF-8. | 
 |  * - Decoding and validating is left to the parser. | 
 |  */ | 
 |  | 
 | enum json_lexer_state { | 
 |     IN_RECOVERY = 1, | 
 |     IN_DQ_STRING_ESCAPE, | 
 |     IN_DQ_STRING, | 
 |     IN_SQ_STRING_ESCAPE, | 
 |     IN_SQ_STRING, | 
 |     IN_ZERO, | 
 |     IN_EXP_DIGITS, | 
 |     IN_EXP_SIGN, | 
 |     IN_EXP_E, | 
 |     IN_MANTISSA, | 
 |     IN_MANTISSA_DIGITS, | 
 |     IN_DIGITS, | 
 |     IN_SIGN, | 
 |     IN_KEYWORD, | 
 |     IN_INTERP, | 
 |     IN_START, | 
 |     IN_START_INTERP,            /* must be IN_START + 1 */ | 
 | }; | 
 |  | 
 | QEMU_BUILD_BUG_ON(JSON_ERROR != 0); | 
 | QEMU_BUILD_BUG_ON(IN_RECOVERY != JSON_ERROR + 1); | 
 | QEMU_BUILD_BUG_ON((int)JSON_MIN <= (int)IN_START_INTERP); | 
 | QEMU_BUILD_BUG_ON(JSON_MAX >= 0x80); | 
 | QEMU_BUILD_BUG_ON(IN_START_INTERP != IN_START + 1); | 
 |  | 
 | #define LOOKAHEAD 0x80 | 
 | #define TERMINAL(state) [0 ... 0xFF] = ((state) | LOOKAHEAD) | 
 |  | 
 | static const uint8_t json_lexer[][256] =  { | 
 |     /* Relies on default initialization to IN_ERROR! */ | 
 |  | 
 |     /* error recovery */ | 
 |     [IN_RECOVERY] = { | 
 |         /* | 
 |          * Skip characters until a structural character, an ASCII | 
 |          * control character other than '\t', or impossible UTF-8 | 
 |          * bytes '\xFE', '\xFF'.  Structural characters and line | 
 |          * endings are promising resynchronization points.  Clients | 
 |          * may use the others to force the JSON parser into known-good | 
 |          * state; see docs/interop/qmp-spec.rst. | 
 |          */ | 
 |         [0 ... 0x1F] = IN_START | LOOKAHEAD, | 
 |         [0x20 ... 0xFD] = IN_RECOVERY, | 
 |         [0xFE ... 0xFF] = IN_START | LOOKAHEAD, | 
 |         ['\t'] = IN_RECOVERY, | 
 |         ['['] = IN_START | LOOKAHEAD, | 
 |         [']'] = IN_START | LOOKAHEAD, | 
 |         ['{'] = IN_START | LOOKAHEAD, | 
 |         ['}'] = IN_START | LOOKAHEAD, | 
 |         [':'] = IN_START | LOOKAHEAD, | 
 |         [','] = IN_START | LOOKAHEAD, | 
 |     }, | 
 |  | 
 |     /* double quote string */ | 
 |     [IN_DQ_STRING_ESCAPE] = { | 
 |         [0x20 ... 0xFD] = IN_DQ_STRING, | 
 |     }, | 
 |     [IN_DQ_STRING] = { | 
 |         [0x20 ... 0xFD] = IN_DQ_STRING, | 
 |         ['\\'] = IN_DQ_STRING_ESCAPE, | 
 |         ['"'] = JSON_STRING, | 
 |     }, | 
 |  | 
 |     /* single quote string */ | 
 |     [IN_SQ_STRING_ESCAPE] = { | 
 |         [0x20 ... 0xFD] = IN_SQ_STRING, | 
 |     }, | 
 |     [IN_SQ_STRING] = { | 
 |         [0x20 ... 0xFD] = IN_SQ_STRING, | 
 |         ['\\'] = IN_SQ_STRING_ESCAPE, | 
 |         ['\''] = JSON_STRING, | 
 |     }, | 
 |  | 
 |     /* Zero */ | 
 |     [IN_ZERO] = { | 
 |         TERMINAL(JSON_INTEGER), | 
 |         ['0' ... '9'] = JSON_ERROR, | 
 |         ['.'] = IN_MANTISSA, | 
 |     }, | 
 |  | 
 |     /* Float */ | 
 |     [IN_EXP_DIGITS] = { | 
 |         TERMINAL(JSON_FLOAT), | 
 |         ['0' ... '9'] = IN_EXP_DIGITS, | 
 |     }, | 
 |  | 
 |     [IN_EXP_SIGN] = { | 
 |         ['0' ... '9'] = IN_EXP_DIGITS, | 
 |     }, | 
 |  | 
 |     [IN_EXP_E] = { | 
 |         ['-'] = IN_EXP_SIGN, | 
 |         ['+'] = IN_EXP_SIGN, | 
 |         ['0' ... '9'] = IN_EXP_DIGITS, | 
 |     }, | 
 |  | 
 |     [IN_MANTISSA_DIGITS] = { | 
 |         TERMINAL(JSON_FLOAT), | 
 |         ['0' ... '9'] = IN_MANTISSA_DIGITS, | 
 |         ['e'] = IN_EXP_E, | 
 |         ['E'] = IN_EXP_E, | 
 |     }, | 
 |  | 
 |     [IN_MANTISSA] = { | 
 |         ['0' ... '9'] = IN_MANTISSA_DIGITS, | 
 |     }, | 
 |  | 
 |     /* Number */ | 
 |     [IN_DIGITS] = { | 
 |         TERMINAL(JSON_INTEGER), | 
 |         ['0' ... '9'] = IN_DIGITS, | 
 |         ['e'] = IN_EXP_E, | 
 |         ['E'] = IN_EXP_E, | 
 |         ['.'] = IN_MANTISSA, | 
 |     }, | 
 |  | 
 |     [IN_SIGN] = { | 
 |         ['0'] = IN_ZERO, | 
 |         ['1' ... '9'] = IN_DIGITS, | 
 |     }, | 
 |  | 
 |     /* keywords */ | 
 |     [IN_KEYWORD] = { | 
 |         TERMINAL(JSON_KEYWORD), | 
 |         ['a' ... 'z'] = IN_KEYWORD, | 
 |     }, | 
 |  | 
 |     /* interpolation */ | 
 |     [IN_INTERP] = { | 
 |         TERMINAL(JSON_INTERP), | 
 |         ['A' ... 'Z'] = IN_INTERP, | 
 |         ['a' ... 'z'] = IN_INTERP, | 
 |         ['0' ... '9'] = IN_INTERP, | 
 |     }, | 
 |  | 
 |     /* | 
 |      * Two start states: | 
 |      * - IN_START recognizes JSON tokens with our string extensions | 
 |      * - IN_START_INTERP additionally recognizes interpolation. | 
 |      */ | 
 |     [IN_START ... IN_START_INTERP] = { | 
 |         ['"'] = IN_DQ_STRING, | 
 |         ['\''] = IN_SQ_STRING, | 
 |         ['0'] = IN_ZERO, | 
 |         ['1' ... '9'] = IN_DIGITS, | 
 |         ['-'] = IN_SIGN, | 
 |         ['{'] = JSON_LCURLY, | 
 |         ['}'] = JSON_RCURLY, | 
 |         ['['] = JSON_LSQUARE, | 
 |         [']'] = JSON_RSQUARE, | 
 |         [','] = JSON_COMMA, | 
 |         [':'] = JSON_COLON, | 
 |         ['a' ... 'z'] = IN_KEYWORD, | 
 |         [' '] = IN_START, | 
 |         ['\t'] = IN_START, | 
 |         ['\r'] = IN_START, | 
 |         ['\n'] = IN_START, | 
 |     }, | 
 |     [IN_START_INTERP]['%'] = IN_INTERP, | 
 | }; | 
 |  | 
 | static inline uint8_t next_state(JSONLexer *lexer, char ch, bool flush, | 
 |                                  bool *char_consumed) | 
 | { | 
 |     uint8_t next; | 
 |  | 
 |     assert(lexer->state < ARRAY_SIZE(json_lexer)); | 
 |     next = json_lexer[lexer->state][(uint8_t)ch]; | 
 |     *char_consumed = !flush && !(next & LOOKAHEAD); | 
 |     return next & ~LOOKAHEAD; | 
 | } | 
 |  | 
 | void json_lexer_init(JSONLexer *lexer, bool enable_interpolation) | 
 | { | 
 |     lexer->start_state = lexer->state = enable_interpolation | 
 |         ? IN_START_INTERP : IN_START; | 
 |     lexer->token = g_string_sized_new(3); | 
 |     lexer->x = lexer->y = 0; | 
 | } | 
 |  | 
 | static void json_lexer_feed_char(JSONLexer *lexer, char ch, bool flush) | 
 | { | 
 |     int new_state; | 
 |     bool char_consumed = false; | 
 |  | 
 |     lexer->x++; | 
 |     if (ch == '\n') { | 
 |         lexer->x = 0; | 
 |         lexer->y++; | 
 |     } | 
 |  | 
 |     while (flush ? lexer->state != lexer->start_state : !char_consumed) { | 
 |         new_state = next_state(lexer, ch, flush, &char_consumed); | 
 |         if (char_consumed) { | 
 |             assert(!flush); | 
 |             g_string_append_c(lexer->token, ch); | 
 |         } | 
 |  | 
 |         switch (new_state) { | 
 |         case JSON_LCURLY: | 
 |         case JSON_RCURLY: | 
 |         case JSON_LSQUARE: | 
 |         case JSON_RSQUARE: | 
 |         case JSON_COLON: | 
 |         case JSON_COMMA: | 
 |         case JSON_INTERP: | 
 |         case JSON_INTEGER: | 
 |         case JSON_FLOAT: | 
 |         case JSON_KEYWORD: | 
 |         case JSON_STRING: | 
 |             json_message_process_token(lexer, lexer->token, new_state, | 
 |                                        lexer->x, lexer->y); | 
 |             /* fall through */ | 
 |         case IN_START: | 
 |             g_string_truncate(lexer->token, 0); | 
 |             new_state = lexer->start_state; | 
 |             break; | 
 |         case JSON_ERROR: | 
 |             json_message_process_token(lexer, lexer->token, JSON_ERROR, | 
 |                                        lexer->x, lexer->y); | 
 |             new_state = IN_RECOVERY; | 
 |             /* fall through */ | 
 |         case IN_RECOVERY: | 
 |             g_string_truncate(lexer->token, 0); | 
 |             break; | 
 |         default: | 
 |             break; | 
 |         } | 
 |         lexer->state = new_state; | 
 |     } | 
 |  | 
 |     /* Do not let a single token grow to an arbitrarily large size, | 
 |      * this is a security consideration. | 
 |      */ | 
 |     if (lexer->token->len > MAX_TOKEN_SIZE) { | 
 |         json_message_process_token(lexer, lexer->token, lexer->state, | 
 |                                    lexer->x, lexer->y); | 
 |         g_string_truncate(lexer->token, 0); | 
 |         lexer->state = lexer->start_state; | 
 |     } | 
 | } | 
 |  | 
 | void json_lexer_feed(JSONLexer *lexer, const char *buffer, size_t size) | 
 | { | 
 |     size_t i; | 
 |  | 
 |     for (i = 0; i < size; i++) { | 
 |         json_lexer_feed_char(lexer, buffer[i], false); | 
 |     } | 
 | } | 
 |  | 
 | void json_lexer_flush(JSONLexer *lexer) | 
 | { | 
 |     json_lexer_feed_char(lexer, 0, true); | 
 |     assert(lexer->state == lexer->start_state); | 
 |     json_message_process_token(lexer, lexer->token, JSON_END_OF_INPUT, | 
 |                                lexer->x, lexer->y); | 
 | } | 
 |  | 
 | void json_lexer_destroy(JSONLexer *lexer) | 
 | { | 
 |     g_string_free(lexer->token, true); | 
 | } |