case '-': return T_DASH;
case '[': return T_LBRACK;
case ']': return T_RBRACK;
+ case '{': return T_LCURLY;
+ case '}': return T_RCURLY;
case '&': return T_AND;
case '|': return T_OR;
case '+': return T_PLUS;
case '<': return T_LT;
case '>': return T_GT;
case '~': return T_NOT;
+ case ';': return T_SEMICOLON;
+ case ':': return T_COLON;
default: return T_MACRO_PUNCT;
}
else
{
tokenize_error__ (token, error);
+ ss_dealloc (&out);
return TOKENIZE_ERROR;
}
}
case SEG_INLINE_DATA:
case SEG_DOCUMENT:
case SEG_MACRO_BODY:
- *token = (struct token) { .type = T_STRING };
- ss_alloc_substring (&token->string, s);
+ case SEG_MACRO_NAME:
+ *token = (struct token) { .type = T_STRING, .string = ss_clone (s) };
return TOKENIZE_TOKEN;
case SEG_RESERVED_WORD:
return TOKENIZE_TOKEN;
case SEG_IDENTIFIER:
- *token = (struct token) { .type = T_ID };
- ss_alloc_substring (&token->string, s);
+ *token = (struct token) { .type = T_ID, .string = ss_clone (s) };
return TOKENIZE_TOKEN;
case SEG_MACRO_ID:
- *token = (struct token) { .type = T_MACRO_ID };
- ss_alloc_substring (&token->string, s);
+ *token = (struct token) { .type = T_MACRO_ID, .string = ss_clone (s)};
return TOKENIZE_TOKEN;
case SEG_PUNCT:
*token = (struct token) { .type = scan_punct__ (s) };
if (token->type == T_MACRO_PUNCT)
- ss_alloc_substring (&token->string, s);
+ token->string = ss_clone (s);
return TOKENIZE_TOKEN;
case SEG_SHBANG:
return TOKENIZE_EMPTY;
case SEG_START_DOCUMENT:
- *token = (struct token) { .type = T_ID };
- ss_alloc_substring (&token->string, ss_cstr ("DOCUMENT"));
+ *token = (struct token) {
+ .type = T_ID,
+ .string = ss_clone (ss_cstr ("DOCUMENT"))
+ };
return TOKENIZE_TOKEN;
case SEG_START_COMMAND: