#include "parser_pos.cpp" // #undef at the bottom of this file #define ALLOW_NEWLINE build_context.strict_style Token token_end_of_line(AstFile *f, Token tok) { u8 const *start = f->tokenizer.start + tok.pos.offset; u8 const *s = start; while (*s && *s != '\n' && s < f->tokenizer.end) { s += 1; } tok.pos.column += cast(i32)(s - start) - 1; return tok; } gbString get_file_line_as_string(TokenPos const &pos, i32 *offset_) { AstFile *file = thread_safe_get_ast_file_from_id(pos.file_id); if (file == nullptr) { return nullptr; } isize offset = pos.offset; u8 *start = file->tokenizer.start; u8 *end = file->tokenizer.end; isize len = end-start; if (len < offset) { return nullptr; } u8 *pos_offset = start+offset; u8 *line_start = pos_offset; u8 *line_end = pos_offset; while (line_start >= start) { if (*line_start == '\n') { line_start += 1; break; } line_start -= 1; } while (line_end < end) { if (*line_end == '\n') { line_end -= 1; break; } line_end += 1; } String the_line = make_string(line_start, line_end-line_start); the_line = string_trim_whitespace(the_line); if (offset_) *offset_ = cast(i32)(pos_offset - the_line.text); return gb_string_make_length(heap_allocator(), the_line.text, the_line.len); } isize ast_node_size(AstKind kind) { return align_formula_isize(gb_size_of(AstCommonStuff) + ast_variant_sizes[kind], gb_align_of(void *)); } gb_global std::atomic global_total_node_memory_allocated; // NOTE(bill): And this below is why is I/we need a new language! Discriminated unions are a pain in C/C++ Ast *alloc_ast_node(AstFile *f, AstKind kind) { gbAllocator a = ast_allocator(f); isize size = ast_node_size(kind); Ast *node = cast(Ast *)gb_alloc(a, size); node->kind = kind; node->file_id = f ? f->id : 0; global_total_node_memory_allocated += size; return node; } Ast *clone_ast(Ast *node); Array clone_ast_array(Array const &array) { Array result = {}; if (array.count > 0) { result = array_make(ast_allocator(nullptr), array.count); for_array(i, array) { result[i] = clone_ast(array[i]); } } return result; } Slice clone_ast_array(Slice const &array) { Slice result = {}; if (array.count > 0) { result = slice_clone(permanent_allocator(), array); for_array(i, array) { result[i] = clone_ast(array[i]); } } return result; } Ast *clone_ast(Ast *node) { if (node == nullptr) { return nullptr; } AstFile *f = node->thread_safe_file(); Ast *n = alloc_ast_node(f, node->kind); gb_memmove(n, node, ast_node_size(node->kind)); switch (n->kind) { default: GB_PANIC("Unhandled Ast %.*s", LIT(ast_strings[n->kind])); break; case Ast_Invalid: break; case Ast_Ident: n->Ident.entity = nullptr; break; case Ast_Implicit: break; case Ast_Undef: break; case Ast_BasicLit: break; case Ast_BasicDirective: break; case Ast_PolyType: n->PolyType.type = clone_ast(n->PolyType.type); n->PolyType.specialization = clone_ast(n->PolyType.specialization); break; case Ast_Ellipsis: n->Ellipsis.expr = clone_ast(n->Ellipsis.expr); break; case Ast_ProcGroup: n->ProcGroup.args = clone_ast_array(n->ProcGroup.args); break; case Ast_ProcLit: n->ProcLit.type = clone_ast(n->ProcLit.type); n->ProcLit.body = clone_ast(n->ProcLit.body); n->ProcLit.where_clauses = clone_ast_array(n->ProcLit.where_clauses); break; case Ast_CompoundLit: n->CompoundLit.type = clone_ast(n->CompoundLit.type); n->CompoundLit.elems = clone_ast_array(n->CompoundLit.elems); break; case Ast_BadExpr: break; case Ast_TagExpr: n->TagExpr.expr = clone_ast(n->TagExpr.expr); break; case Ast_UnaryExpr: n->UnaryExpr.expr = clone_ast(n->UnaryExpr.expr); break; case Ast_BinaryExpr: n->BinaryExpr.left = clone_ast(n->BinaryExpr.left); n->BinaryExpr.right = clone_ast(n->BinaryExpr.right); break; case Ast_ParenExpr: n->ParenExpr.expr = clone_ast(n->ParenExpr.expr); break; case Ast_SelectorExpr: n->SelectorExpr.expr = clone_ast(n->SelectorExpr.expr); n->SelectorExpr.selector = clone_ast(n->SelectorExpr.selector); break; case Ast_ImplicitSelectorExpr: n->ImplicitSelectorExpr.selector = clone_ast(n->ImplicitSelectorExpr.selector); break; case Ast_SelectorCallExpr: n->SelectorCallExpr.expr = clone_ast(n->SelectorCallExpr.expr); n->SelectorCallExpr.call = clone_ast(n->SelectorCallExpr.call); break; case Ast_IndexExpr: n->IndexExpr.expr = clone_ast(n->IndexExpr.expr); n->IndexExpr.index = clone_ast(n->IndexExpr.index); break; case Ast_MatrixIndexExpr: n->MatrixIndexExpr.expr = clone_ast(n->MatrixIndexExpr.expr); n->MatrixIndexExpr.row_index = clone_ast(n->MatrixIndexExpr.row_index); n->MatrixIndexExpr.column_index = clone_ast(n->MatrixIndexExpr.column_index); break; case Ast_DerefExpr: n->DerefExpr.expr = clone_ast(n->DerefExpr.expr); break; case Ast_SliceExpr: n->SliceExpr.expr = clone_ast(n->SliceExpr.expr); n->SliceExpr.low = clone_ast(n->SliceExpr.low); n->SliceExpr.high = clone_ast(n->SliceExpr.high); break; case Ast_CallExpr: n->CallExpr.proc = clone_ast(n->CallExpr.proc); n->CallExpr.args = clone_ast_array(n->CallExpr.args); break; case Ast_FieldValue: n->FieldValue.field = clone_ast(n->FieldValue.field); n->FieldValue.value = clone_ast(n->FieldValue.value); break; case Ast_EnumFieldValue: n->EnumFieldValue.name = clone_ast(n->EnumFieldValue.name); n->EnumFieldValue.value = clone_ast(n->EnumFieldValue.value); break; case Ast_TernaryIfExpr: n->TernaryIfExpr.x = clone_ast(n->TernaryIfExpr.x); n->TernaryIfExpr.cond = clone_ast(n->TernaryIfExpr.cond); n->TernaryIfExpr.y = clone_ast(n->TernaryIfExpr.y); break; case Ast_TernaryWhenExpr: n->TernaryWhenExpr.x = clone_ast(n->TernaryWhenExpr.x); n->TernaryWhenExpr.cond = clone_ast(n->TernaryWhenExpr.cond); n->TernaryWhenExpr.y = clone_ast(n->TernaryWhenExpr.y); break; case Ast_OrElseExpr: n->OrElseExpr.x = clone_ast(n->OrElseExpr.x); n->OrElseExpr.y = clone_ast(n->OrElseExpr.y); break; case Ast_OrReturnExpr: n->OrReturnExpr.expr = clone_ast(n->OrReturnExpr.expr); break; case Ast_TypeAssertion: n->TypeAssertion.expr = clone_ast(n->TypeAssertion.expr); n->TypeAssertion.type = clone_ast(n->TypeAssertion.type); break; case Ast_TypeCast: n->TypeCast.type = clone_ast(n->TypeCast.type); n->TypeCast.expr = clone_ast(n->TypeCast.expr); break; case Ast_AutoCast: n->AutoCast.expr = clone_ast(n->AutoCast.expr); break; case Ast_InlineAsmExpr: n->InlineAsmExpr.param_types = clone_ast_array(n->InlineAsmExpr.param_types); n->InlineAsmExpr.return_type = clone_ast(n->InlineAsmExpr.return_type); n->InlineAsmExpr.asm_string = clone_ast(n->InlineAsmExpr.asm_string); n->InlineAsmExpr.constraints_string = clone_ast(n->InlineAsmExpr.constraints_string); break; case Ast_BadStmt: break; case Ast_EmptyStmt: break; case Ast_ExprStmt: n->ExprStmt.expr = clone_ast(n->ExprStmt.expr); break; case Ast_TagStmt: n->TagStmt.stmt = clone_ast(n->TagStmt.stmt); break; case Ast_AssignStmt: n->AssignStmt.lhs = clone_ast_array(n->AssignStmt.lhs); n->AssignStmt.rhs = clone_ast_array(n->AssignStmt.rhs); break; case Ast_BlockStmt: n->BlockStmt.label = clone_ast(n->BlockStmt.label); n->BlockStmt.stmts = clone_ast_array(n->BlockStmt.stmts); break; case Ast_IfStmt: n->IfStmt.label = clone_ast(n->IfStmt.label); n->IfStmt.init = clone_ast(n->IfStmt.init); n->IfStmt.cond = clone_ast(n->IfStmt.cond); n->IfStmt.body = clone_ast(n->IfStmt.body); n->IfStmt.else_stmt = clone_ast(n->IfStmt.else_stmt); break; case Ast_WhenStmt: n->WhenStmt.cond = clone_ast(n->WhenStmt.cond); n->WhenStmt.body = clone_ast(n->WhenStmt.body); n->WhenStmt.else_stmt = clone_ast(n->WhenStmt.else_stmt); break; case Ast_ReturnStmt: n->ReturnStmt.results = clone_ast_array(n->ReturnStmt.results); break; case Ast_ForStmt: n->ForStmt.label = clone_ast(n->ForStmt.label); n->ForStmt.init = clone_ast(n->ForStmt.init); n->ForStmt.cond = clone_ast(n->ForStmt.cond); n->ForStmt.post = clone_ast(n->ForStmt.post); n->ForStmt.body = clone_ast(n->ForStmt.body); break; case Ast_RangeStmt: n->RangeStmt.label = clone_ast(n->RangeStmt.label); n->RangeStmt.vals = clone_ast_array(n->RangeStmt.vals); n->RangeStmt.expr = clone_ast(n->RangeStmt.expr); n->RangeStmt.body = clone_ast(n->RangeStmt.body); break; case Ast_UnrollRangeStmt: n->UnrollRangeStmt.val0 = clone_ast(n->UnrollRangeStmt.val0); n->UnrollRangeStmt.val1 = clone_ast(n->UnrollRangeStmt.val1); n->UnrollRangeStmt.expr = clone_ast(n->UnrollRangeStmt.expr); n->UnrollRangeStmt.body = clone_ast(n->UnrollRangeStmt.body); break; case Ast_CaseClause: n->CaseClause.list = clone_ast_array(n->CaseClause.list); n->CaseClause.stmts = clone_ast_array(n->CaseClause.stmts); n->CaseClause.implicit_entity = nullptr; break; case Ast_SwitchStmt: n->SwitchStmt.label = clone_ast(n->SwitchStmt.label); n->SwitchStmt.init = clone_ast(n->SwitchStmt.init); n->SwitchStmt.tag = clone_ast(n->SwitchStmt.tag); n->SwitchStmt.body = clone_ast(n->SwitchStmt.body); break; case Ast_TypeSwitchStmt: n->TypeSwitchStmt.label = clone_ast(n->TypeSwitchStmt.label); n->TypeSwitchStmt.tag = clone_ast(n->TypeSwitchStmt.tag); n->TypeSwitchStmt.body = clone_ast(n->TypeSwitchStmt.body); break; case Ast_DeferStmt: n->DeferStmt.stmt = clone_ast(n->DeferStmt.stmt); break; case Ast_BranchStmt: n->BranchStmt.label = clone_ast(n->BranchStmt.label); break; case Ast_UsingStmt: n->UsingStmt.list = clone_ast_array(n->UsingStmt.list); break; case Ast_BadDecl: break; case Ast_ForeignBlockDecl: n->ForeignBlockDecl.foreign_library = clone_ast(n->ForeignBlockDecl.foreign_library); n->ForeignBlockDecl.body = clone_ast(n->ForeignBlockDecl.body); n->ForeignBlockDecl.attributes = clone_ast_array(n->ForeignBlockDecl.attributes); break; case Ast_Label: n->Label.name = clone_ast(n->Label.name); break; case Ast_ValueDecl: n->ValueDecl.names = clone_ast_array(n->ValueDecl.names); n->ValueDecl.type = clone_ast(n->ValueDecl.type); n->ValueDecl.values = clone_ast_array(n->ValueDecl.values); n->ValueDecl.attributes = clone_ast_array(n->ValueDecl.attributes); break; case Ast_Attribute: n->Attribute.elems = clone_ast_array(n->Attribute.elems); break; case Ast_Field: n->Field.names = clone_ast_array(n->Field.names); n->Field.type = clone_ast(n->Field.type); break; case Ast_FieldList: n->FieldList.list = clone_ast_array(n->FieldList.list); break; case Ast_TypeidType: n->TypeidType.specialization = clone_ast(n->TypeidType.specialization); break; case Ast_HelperType: n->HelperType.type = clone_ast(n->HelperType.type); break; case Ast_DistinctType: n->DistinctType.type = clone_ast(n->DistinctType.type); break; case Ast_ProcType: n->ProcType.params = clone_ast(n->ProcType.params); n->ProcType.results = clone_ast(n->ProcType.results); break; case Ast_RelativeType: n->RelativeType.tag = clone_ast(n->RelativeType.tag); n->RelativeType.type = clone_ast(n->RelativeType.type); break; case Ast_PointerType: n->PointerType.type = clone_ast(n->PointerType.type); n->PointerType.tag = clone_ast(n->PointerType.tag); break; case Ast_MultiPointerType: n->MultiPointerType.type = clone_ast(n->MultiPointerType.type); break; case Ast_ArrayType: n->ArrayType.count = clone_ast(n->ArrayType.count); n->ArrayType.elem = clone_ast(n->ArrayType.elem); n->ArrayType.tag = clone_ast(n->ArrayType.tag); break; case Ast_DynamicArrayType: n->DynamicArrayType.elem = clone_ast(n->DynamicArrayType.elem); break; case Ast_StructType: n->StructType.fields = clone_ast_array(n->StructType.fields); n->StructType.polymorphic_params = clone_ast(n->StructType.polymorphic_params); n->StructType.align = clone_ast(n->StructType.align); n->StructType.where_clauses = clone_ast_array(n->StructType.where_clauses); break; case Ast_UnionType: n->UnionType.variants = clone_ast_array(n->UnionType.variants); n->UnionType.polymorphic_params = clone_ast(n->UnionType.polymorphic_params); n->UnionType.where_clauses = clone_ast_array(n->UnionType.where_clauses); break; case Ast_EnumType: n->EnumType.base_type = clone_ast(n->EnumType.base_type); n->EnumType.fields = clone_ast_array(n->EnumType.fields); break; case Ast_BitSetType: n->BitSetType.elem = clone_ast(n->BitSetType.elem); n->BitSetType.underlying = clone_ast(n->BitSetType.underlying); break; case Ast_MapType: n->MapType.count = clone_ast(n->MapType.count); n->MapType.key = clone_ast(n->MapType.key); n->MapType.value = clone_ast(n->MapType.value); break; case Ast_MatrixType: n->MatrixType.row_count = clone_ast(n->MatrixType.row_count); n->MatrixType.column_count = clone_ast(n->MatrixType.column_count); n->MatrixType.elem = clone_ast(n->MatrixType.elem); break; } return n; } void error(Ast *node, char const *fmt, ...) { Token token = {}; TokenPos end_pos = {}; if (node != nullptr) { token = ast_token(node); end_pos = ast_end_pos(node); } va_list va; va_start(va, fmt); error_va(token.pos, end_pos, fmt, va); va_end(va); if (node != nullptr && node->file_id != 0) { AstFile *f = node->thread_safe_file(); f->error_count += 1; } } void error_no_newline(Ast *node, char const *fmt, ...) { Token token = {}; if (node != nullptr) { token = ast_token(node); } va_list va; va_start(va, fmt); error_no_newline_va(token.pos, fmt, va); va_end(va); if (node != nullptr && node->file_id != 0) { AstFile *f = node->thread_safe_file(); f->error_count += 1; } } void warning(Ast *node, char const *fmt, ...) { Token token = {}; TokenPos end_pos = {}; if (node != nullptr) { token = ast_token(node); end_pos = ast_end_pos(node); } va_list va; va_start(va, fmt); warning_va(token.pos, end_pos, fmt, va); va_end(va); } void syntax_error(Ast *node, char const *fmt, ...) { Token token = {}; TokenPos end_pos = {}; if (node != nullptr) { token = ast_token(node); end_pos = ast_end_pos(node); } va_list va; va_start(va, fmt); syntax_error_va(token.pos, end_pos, fmt, va); va_end(va); if (node != nullptr && node->file_id != 0) { AstFile *f = node->thread_safe_file(); f->error_count += 1; } } bool ast_node_expect(Ast *node, AstKind kind) { if (node->kind != kind) { syntax_error(node, "Expected %.*s, got %.*s", LIT(ast_strings[kind]), LIT(ast_strings[node->kind])); return false; } return true; } bool ast_node_expect2(Ast *node, AstKind kind0, AstKind kind1) { if (node->kind != kind0 && node->kind != kind1) { syntax_error(node, "Expected %.*s or %.*s, got %.*s", LIT(ast_strings[kind0]), LIT(ast_strings[kind1]), LIT(ast_strings[node->kind])); return false; } return true; } Ast *ast_bad_expr(AstFile *f, Token begin, Token end) { Ast *result = alloc_ast_node(f, Ast_BadExpr); result->BadExpr.begin = begin; result->BadExpr.end = end; return result; } Ast *ast_tag_expr(AstFile *f, Token token, Token name, Ast *expr) { Ast *result = alloc_ast_node(f, Ast_TagExpr); result->TagExpr.token = token; result->TagExpr.name = name; result->TagExpr.expr = expr; return result; } Ast *ast_tag_stmt(AstFile *f, Token token, Token name, Ast *stmt) { Ast *result = alloc_ast_node(f, Ast_TagStmt); result->TagStmt.token = token; result->TagStmt.name = name; result->TagStmt.stmt = stmt; return result; } Ast *ast_unary_expr(AstFile *f, Token op, Ast *expr) { Ast *result = alloc_ast_node(f, Ast_UnaryExpr); result->UnaryExpr.op = op; result->UnaryExpr.expr = expr; return result; } Ast *ast_binary_expr(AstFile *f, Token op, Ast *left, Ast *right) { Ast *result = alloc_ast_node(f, Ast_BinaryExpr); if (left == nullptr) { syntax_error(op, "No lhs expression for binary expression '%.*s'", LIT(op.string)); left = ast_bad_expr(f, op, op); } if (right == nullptr) { syntax_error(op, "No rhs expression for binary expression '%.*s'", LIT(op.string)); right = ast_bad_expr(f, op, op); } result->BinaryExpr.op = op; result->BinaryExpr.left = left; result->BinaryExpr.right = right; return result; } Ast *ast_paren_expr(AstFile *f, Ast *expr, Token open, Token close) { Ast *result = alloc_ast_node(f, Ast_ParenExpr); result->ParenExpr.expr = expr; result->ParenExpr.open = open; result->ParenExpr.close = close; return result; } Ast *ast_call_expr(AstFile *f, Ast *proc, Array const &args, Token open, Token close, Token ellipsis) { Ast *result = alloc_ast_node(f, Ast_CallExpr); result->CallExpr.proc = proc; result->CallExpr.args = slice_from_array(args); result->CallExpr.open = open; result->CallExpr.close = close; result->CallExpr.ellipsis = ellipsis; return result; } Ast *ast_selector_expr(AstFile *f, Token token, Ast *expr, Ast *selector) { Ast *result = alloc_ast_node(f, Ast_SelectorExpr); result->SelectorExpr.token = token; result->SelectorExpr.expr = expr; result->SelectorExpr.selector = selector; return result; } Ast *ast_implicit_selector_expr(AstFile *f, Token token, Ast *selector) { Ast *result = alloc_ast_node(f, Ast_ImplicitSelectorExpr); result->ImplicitSelectorExpr.token = token; result->ImplicitSelectorExpr.selector = selector; return result; } Ast *ast_selector_call_expr(AstFile *f, Token token, Ast *expr, Ast *call) { Ast *result = alloc_ast_node(f, Ast_SelectorCallExpr); result->SelectorCallExpr.token = token; result->SelectorCallExpr.expr = expr; result->SelectorCallExpr.call = call; return result; } Ast *ast_index_expr(AstFile *f, Ast *expr, Ast *index, Token open, Token close) { Ast *result = alloc_ast_node(f, Ast_IndexExpr); result->IndexExpr.expr = expr; result->IndexExpr.index = index; result->IndexExpr.open = open; result->IndexExpr.close = close; return result; } Ast *ast_slice_expr(AstFile *f, Ast *expr, Token open, Token close, Token interval, Ast *low, Ast *high) { Ast *result = alloc_ast_node(f, Ast_SliceExpr); result->SliceExpr.expr = expr; result->SliceExpr.open = open; result->SliceExpr.close = close; result->SliceExpr.interval = interval; result->SliceExpr.low = low; result->SliceExpr.high = high; return result; } Ast *ast_deref_expr(AstFile *f, Ast *expr, Token op) { Ast *result = alloc_ast_node(f, Ast_DerefExpr); result->DerefExpr.expr = expr; result->DerefExpr.op = op; return result; } Ast *ast_matrix_index_expr(AstFile *f, Ast *expr, Token open, Token close, Token interval, Ast *row, Ast *column) { Ast *result = alloc_ast_node(f, Ast_MatrixIndexExpr); result->MatrixIndexExpr.expr = expr; result->MatrixIndexExpr.row_index = row; result->MatrixIndexExpr.column_index = column; result->MatrixIndexExpr.open = open; result->MatrixIndexExpr.close = close; return result; } Ast *ast_ident(AstFile *f, Token token) { Ast *result = alloc_ast_node(f, Ast_Ident); result->Ident.token = token; return result; } Ast *ast_implicit(AstFile *f, Token token) { Ast *result = alloc_ast_node(f, Ast_Implicit); result->Implicit = token; return result; } Ast *ast_undef(AstFile *f, Token token) { Ast *result = alloc_ast_node(f, Ast_Undef); result->Undef = token; return result; } ExactValue exact_value_from_token(AstFile *f, Token const &token) { String s = token.string; switch (token.kind) { case Token_Rune: if (!unquote_string(ast_allocator(f), &s, 0)) { syntax_error(token, "Invalid rune literal"); } break; case Token_String: if (!unquote_string(ast_allocator(f), &s, 0, s.text[0] == '`')) { syntax_error(token, "Invalid string literal"); } break; } return exact_value_from_basic_literal(token.kind, s); } String string_value_from_token(AstFile *f, Token const &token) { ExactValue value = exact_value_from_token(f, token); String str = {}; if (value.kind == ExactValue_String) { str = value.value_string; } return str; } Ast *ast_basic_lit(AstFile *f, Token basic_lit) { Ast *result = alloc_ast_node(f, Ast_BasicLit); result->BasicLit.token = basic_lit; result->tav.mode = Addressing_Constant; result->tav.value = exact_value_from_token(f, basic_lit); return result; } Ast *ast_basic_directive(AstFile *f, Token token, Token name) { Ast *result = alloc_ast_node(f, Ast_BasicDirective); result->BasicDirective.token = token; result->BasicDirective.name = name; return result; } Ast *ast_ellipsis(AstFile *f, Token token, Ast *expr) { Ast *result = alloc_ast_node(f, Ast_Ellipsis); result->Ellipsis.token = token; result->Ellipsis.expr = expr; return result; } Ast *ast_proc_group(AstFile *f, Token token, Token open, Token close, Array const &args) { Ast *result = alloc_ast_node(f, Ast_ProcGroup); result->ProcGroup.token = token; result->ProcGroup.open = open; result->ProcGroup.close = close; result->ProcGroup.args = slice_from_array(args); return result; } Ast *ast_proc_lit(AstFile *f, Ast *type, Ast *body, u64 tags, Token where_token, Array const &where_clauses) { Ast *result = alloc_ast_node(f, Ast_ProcLit); result->ProcLit.type = type; result->ProcLit.body = body; result->ProcLit.tags = tags; result->ProcLit.where_token = where_token; result->ProcLit.where_clauses = slice_from_array(where_clauses); return result; } Ast *ast_field_value(AstFile *f, Ast *field, Ast *value, Token eq) { Ast *result = alloc_ast_node(f, Ast_FieldValue); result->FieldValue.field = field; result->FieldValue.value = value; result->FieldValue.eq = eq; return result; } Ast *ast_enum_field_value(AstFile *f, Ast *name, Ast *value, CommentGroup *docs, CommentGroup *comment) { Ast *result = alloc_ast_node(f, Ast_EnumFieldValue); result->EnumFieldValue.name = name; result->EnumFieldValue.value = value; result->EnumFieldValue.docs = docs; result->EnumFieldValue.comment = comment; return result; } Ast *ast_compound_lit(AstFile *f, Ast *type, Array const &elems, Token open, Token close) { Ast *result = alloc_ast_node(f, Ast_CompoundLit); result->CompoundLit.type = type; result->CompoundLit.elems = slice_from_array(elems); result->CompoundLit.open = open; result->CompoundLit.close = close; return result; } Ast *ast_ternary_if_expr(AstFile *f, Ast *x, Ast *cond, Ast *y) { Ast *result = alloc_ast_node(f, Ast_TernaryIfExpr); result->TernaryIfExpr.x = x; result->TernaryIfExpr.cond = cond; result->TernaryIfExpr.y = y; return result; } Ast *ast_ternary_when_expr(AstFile *f, Ast *x, Ast *cond, Ast *y) { Ast *result = alloc_ast_node(f, Ast_TernaryWhenExpr); result->TernaryWhenExpr.x = x; result->TernaryWhenExpr.cond = cond; result->TernaryWhenExpr.y = y; return result; } Ast *ast_or_else_expr(AstFile *f, Ast *x, Token const &token, Ast *y) { Ast *result = alloc_ast_node(f, Ast_OrElseExpr); result->OrElseExpr.x = x; result->OrElseExpr.token = token; result->OrElseExpr.y = y; return result; } Ast *ast_or_return_expr(AstFile *f, Ast *expr, Token const &token) { Ast *result = alloc_ast_node(f, Ast_OrReturnExpr); result->OrReturnExpr.expr = expr; result->OrReturnExpr.token = token; return result; } Ast *ast_type_assertion(AstFile *f, Ast *expr, Token dot, Ast *type) { Ast *result = alloc_ast_node(f, Ast_TypeAssertion); result->TypeAssertion.expr = expr; result->TypeAssertion.dot = dot; result->TypeAssertion.type = type; return result; } Ast *ast_type_cast(AstFile *f, Token token, Ast *type, Ast *expr) { Ast *result = alloc_ast_node(f, Ast_TypeCast); result->TypeCast.token = token; result->TypeCast.type = type; result->TypeCast.expr = expr; return result; } Ast *ast_auto_cast(AstFile *f, Token token, Ast *expr) { Ast *result = alloc_ast_node(f, Ast_AutoCast); result->AutoCast.token = token; result->AutoCast.expr = expr; return result; } Ast *ast_inline_asm_expr(AstFile *f, Token token, Token open, Token close, Array const ¶m_types, Ast *return_type, Ast *asm_string, Ast *constraints_string, bool has_side_effects, bool is_align_stack, InlineAsmDialectKind dialect) { Ast *result = alloc_ast_node(f, Ast_InlineAsmExpr); result->InlineAsmExpr.token = token; result->InlineAsmExpr.open = open; result->InlineAsmExpr.close = close; result->InlineAsmExpr.param_types = slice_from_array(param_types); result->InlineAsmExpr.return_type = return_type; result->InlineAsmExpr.asm_string = asm_string; result->InlineAsmExpr.constraints_string = constraints_string; result->InlineAsmExpr.has_side_effects = has_side_effects; result->InlineAsmExpr.is_align_stack = is_align_stack; result->InlineAsmExpr.dialect = dialect; return result; } Ast *ast_bad_stmt(AstFile *f, Token begin, Token end) { Ast *result = alloc_ast_node(f, Ast_BadStmt); result->BadStmt.begin = begin; result->BadStmt.end = end; return result; } Ast *ast_empty_stmt(AstFile *f, Token token) { Ast *result = alloc_ast_node(f, Ast_EmptyStmt); result->EmptyStmt.token = token; return result; } Ast *ast_expr_stmt(AstFile *f, Ast *expr) { Ast *result = alloc_ast_node(f, Ast_ExprStmt); result->ExprStmt.expr = expr; return result; } Ast *ast_assign_stmt(AstFile *f, Token op, Array const &lhs, Array const &rhs) { Ast *result = alloc_ast_node(f, Ast_AssignStmt); result->AssignStmt.op = op; result->AssignStmt.lhs = slice_from_array(lhs); result->AssignStmt.rhs = slice_from_array(rhs); return result; } Ast *ast_block_stmt(AstFile *f, Array const &stmts, Token open, Token close) { Ast *result = alloc_ast_node(f, Ast_BlockStmt); result->BlockStmt.stmts = slice_from_array(stmts); result->BlockStmt.open = open; result->BlockStmt.close = close; return result; } Ast *ast_if_stmt(AstFile *f, Token token, Ast *init, Ast *cond, Ast *body, Ast *else_stmt) { Ast *result = alloc_ast_node(f, Ast_IfStmt); result->IfStmt.token = token; result->IfStmt.init = init; result->IfStmt.cond = cond; result->IfStmt.body = body; result->IfStmt.else_stmt = else_stmt; return result; } Ast *ast_when_stmt(AstFile *f, Token token, Ast *cond, Ast *body, Ast *else_stmt) { Ast *result = alloc_ast_node(f, Ast_WhenStmt); result->WhenStmt.token = token; result->WhenStmt.cond = cond; result->WhenStmt.body = body; result->WhenStmt.else_stmt = else_stmt; return result; } Ast *ast_return_stmt(AstFile *f, Token token, Array const &results) { Ast *result = alloc_ast_node(f, Ast_ReturnStmt); result->ReturnStmt.token = token; result->ReturnStmt.results = slice_from_array(results); return result; } Ast *ast_for_stmt(AstFile *f, Token token, Ast *init, Ast *cond, Ast *post, Ast *body) { Ast *result = alloc_ast_node(f, Ast_ForStmt); result->ForStmt.token = token; result->ForStmt.init = init; result->ForStmt.cond = cond; result->ForStmt.post = post; result->ForStmt.body = body; return result; } Ast *ast_range_stmt(AstFile *f, Token token, Slice vals, Token in_token, Ast *expr, Ast *body) { Ast *result = alloc_ast_node(f, Ast_RangeStmt); result->RangeStmt.token = token; result->RangeStmt.vals = vals; result->RangeStmt.in_token = in_token; result->RangeStmt.expr = expr; result->RangeStmt.body = body; return result; } Ast *ast_unroll_range_stmt(AstFile *f, Token unroll_token, Token for_token, Ast *val0, Ast *val1, Token in_token, Ast *expr, Ast *body) { Ast *result = alloc_ast_node(f, Ast_UnrollRangeStmt); result->UnrollRangeStmt.unroll_token = unroll_token; result->UnrollRangeStmt.for_token = for_token; result->UnrollRangeStmt.val0 = val0; result->UnrollRangeStmt.val1 = val1; result->UnrollRangeStmt.in_token = in_token; result->UnrollRangeStmt.expr = expr; result->UnrollRangeStmt.body = body; return result; } Ast *ast_switch_stmt(AstFile *f, Token token, Ast *init, Ast *tag, Ast *body) { Ast *result = alloc_ast_node(f, Ast_SwitchStmt); result->SwitchStmt.token = token; result->SwitchStmt.init = init; result->SwitchStmt.tag = tag; result->SwitchStmt.body = body; result->SwitchStmt.partial = false; return result; } Ast *ast_type_switch_stmt(AstFile *f, Token token, Ast *tag, Ast *body) { Ast *result = alloc_ast_node(f, Ast_TypeSwitchStmt); result->TypeSwitchStmt.token = token; result->TypeSwitchStmt.tag = tag; result->TypeSwitchStmt.body = body; result->TypeSwitchStmt.partial = false; return result; } Ast *ast_case_clause(AstFile *f, Token token, Array const &list, Array const &stmts) { Ast *result = alloc_ast_node(f, Ast_CaseClause); result->CaseClause.token = token; result->CaseClause.list = slice_from_array(list); result->CaseClause.stmts = slice_from_array(stmts); return result; } Ast *ast_defer_stmt(AstFile *f, Token token, Ast *stmt) { Ast *result = alloc_ast_node(f, Ast_DeferStmt); result->DeferStmt.token = token; result->DeferStmt.stmt = stmt; return result; } Ast *ast_branch_stmt(AstFile *f, Token token, Ast *label) { Ast *result = alloc_ast_node(f, Ast_BranchStmt); result->BranchStmt.token = token; result->BranchStmt.label = label; return result; } Ast *ast_using_stmt(AstFile *f, Token token, Array const &list) { Ast *result = alloc_ast_node(f, Ast_UsingStmt); result->UsingStmt.token = token; result->UsingStmt.list = slice_from_array(list); return result; } Ast *ast_bad_decl(AstFile *f, Token begin, Token end) { Ast *result = alloc_ast_node(f, Ast_BadDecl); result->BadDecl.begin = begin; result->BadDecl.end = end; return result; } Ast *ast_field(AstFile *f, Array const &names, Ast *type, Ast *default_value, u32 flags, Token tag, CommentGroup *docs, CommentGroup *comment) { Ast *result = alloc_ast_node(f, Ast_Field); result->Field.names = slice_from_array(names); result->Field.type = type; result->Field.default_value = default_value; result->Field.flags = flags; result->Field.tag = tag; result->Field.docs = docs; result->Field.comment = comment; return result; } Ast *ast_field_list(AstFile *f, Token token, Array const &list) { Ast *result = alloc_ast_node(f, Ast_FieldList); result->FieldList.token = token; result->FieldList.list = slice_from_array(list); return result; } Ast *ast_typeid_type(AstFile *f, Token token, Ast *specialization) { Ast *result = alloc_ast_node(f, Ast_TypeidType); result->TypeidType.token = token; result->TypeidType.specialization = specialization; return result; } Ast *ast_helper_type(AstFile *f, Token token, Ast *type) { Ast *result = alloc_ast_node(f, Ast_HelperType); result->HelperType.token = token; result->HelperType.type = type; return result; } Ast *ast_distinct_type(AstFile *f, Token token, Ast *type) { Ast *result = alloc_ast_node(f, Ast_DistinctType); result->DistinctType.token = token; result->DistinctType.type = type; return result; } Ast *ast_poly_type(AstFile *f, Token token, Ast *type, Ast *specialization) { Ast *result = alloc_ast_node(f, Ast_PolyType); result->PolyType.token = token; result->PolyType.type = type; result->PolyType.specialization = specialization; return result; } Ast *ast_proc_type(AstFile *f, Token token, Ast *params, Ast *results, u64 tags, ProcCallingConvention calling_convention, bool generic, bool diverging) { Ast *result = alloc_ast_node(f, Ast_ProcType); result->ProcType.token = token; result->ProcType.params = params; result->ProcType.results = results; result->ProcType.tags = tags; result->ProcType.calling_convention = calling_convention; result->ProcType.generic = generic; result->ProcType.diverging = diverging; return result; } Ast *ast_relative_type(AstFile *f, Ast *tag, Ast *type) { Ast *result = alloc_ast_node(f, Ast_RelativeType); result->RelativeType.tag = tag; result->RelativeType.type = type; return result; } Ast *ast_pointer_type(AstFile *f, Token token, Ast *type) { Ast *result = alloc_ast_node(f, Ast_PointerType); result->PointerType.token = token; result->PointerType.type = type; return result; } Ast *ast_multi_pointer_type(AstFile *f, Token token, Ast *type) { Ast *result = alloc_ast_node(f, Ast_MultiPointerType); result->MultiPointerType.token = token; result->MultiPointerType.type = type; return result; } Ast *ast_array_type(AstFile *f, Token token, Ast *count, Ast *elem) { Ast *result = alloc_ast_node(f, Ast_ArrayType); result->ArrayType.token = token; result->ArrayType.count = count; result->ArrayType.elem = elem; return result; } Ast *ast_dynamic_array_type(AstFile *f, Token token, Ast *elem) { Ast *result = alloc_ast_node(f, Ast_DynamicArrayType); result->DynamicArrayType.token = token; result->DynamicArrayType.elem = elem; return result; } Ast *ast_struct_type(AstFile *f, Token token, Slice fields, isize field_count, Ast *polymorphic_params, bool is_packed, bool is_raw_union, Ast *align, Token where_token, Array const &where_clauses) { Ast *result = alloc_ast_node(f, Ast_StructType); result->StructType.token = token; result->StructType.fields = fields; result->StructType.field_count = field_count; result->StructType.polymorphic_params = polymorphic_params; result->StructType.is_packed = is_packed; result->StructType.is_raw_union = is_raw_union; result->StructType.align = align; result->StructType.where_token = where_token; result->StructType.where_clauses = slice_from_array(where_clauses); return result; } Ast *ast_union_type(AstFile *f, Token token, Array const &variants, Ast *polymorphic_params, Ast *align, UnionTypeKind kind, Token where_token, Array const &where_clauses) { Ast *result = alloc_ast_node(f, Ast_UnionType); result->UnionType.token = token; result->UnionType.variants = slice_from_array(variants); result->UnionType.polymorphic_params = polymorphic_params; result->UnionType.align = align; result->UnionType.kind = kind; result->UnionType.where_token = where_token; result->UnionType.where_clauses = slice_from_array(where_clauses); return result; } Ast *ast_enum_type(AstFile *f, Token token, Ast *base_type, Array const &fields) { Ast *result = alloc_ast_node(f, Ast_EnumType); result->EnumType.token = token; result->EnumType.base_type = base_type; result->EnumType.fields = slice_from_array(fields); return result; } Ast *ast_bit_set_type(AstFile *f, Token token, Ast *elem, Ast *underlying) { Ast *result = alloc_ast_node(f, Ast_BitSetType); result->BitSetType.token = token; result->BitSetType.elem = elem; result->BitSetType.underlying = underlying; return result; } Ast *ast_map_type(AstFile *f, Token token, Ast *key, Ast *value) { Ast *result = alloc_ast_node(f, Ast_MapType); result->MapType.token = token; result->MapType.key = key; result->MapType.value = value; return result; } Ast *ast_matrix_type(AstFile *f, Token token, Ast *row_count, Ast *column_count, Ast *elem) { Ast *result = alloc_ast_node(f, Ast_MatrixType); result->MatrixType.token = token; result->MatrixType.row_count = row_count; result->MatrixType.column_count = column_count; result->MatrixType.elem = elem; return result; } Ast *ast_foreign_block_decl(AstFile *f, Token token, Ast *foreign_library, Ast *body, CommentGroup *docs) { Ast *result = alloc_ast_node(f, Ast_ForeignBlockDecl); result->ForeignBlockDecl.token = token; result->ForeignBlockDecl.foreign_library = foreign_library; result->ForeignBlockDecl.body = body; result->ForeignBlockDecl.docs = docs; result->ForeignBlockDecl.attributes.allocator = heap_allocator(); return result; } Ast *ast_label_decl(AstFile *f, Token token, Ast *name) { Ast *result = alloc_ast_node(f, Ast_Label); result->Label.token = token; result->Label.name = name; return result; } Ast *ast_value_decl(AstFile *f, Array const &names, Ast *type, Array const &values, bool is_mutable, CommentGroup *docs, CommentGroup *comment) { Ast *result = alloc_ast_node(f, Ast_ValueDecl); result->ValueDecl.names = slice_from_array(names); result->ValueDecl.type = type; result->ValueDecl.values = slice_from_array(values); result->ValueDecl.is_mutable = is_mutable; result->ValueDecl.docs = docs; result->ValueDecl.comment = comment; result->ValueDecl.attributes.allocator = heap_allocator(); return result; } Ast *ast_package_decl(AstFile *f, Token token, Token name, CommentGroup *docs, CommentGroup *comment) { Ast *result = alloc_ast_node(f, Ast_PackageDecl); result->PackageDecl.token = token; result->PackageDecl.name = name; result->PackageDecl.docs = docs; result->PackageDecl.comment = comment; return result; } Ast *ast_import_decl(AstFile *f, Token token, Token relpath, Token import_name, CommentGroup *docs, CommentGroup *comment) { Ast *result = alloc_ast_node(f, Ast_ImportDecl); result->ImportDecl.token = token; result->ImportDecl.relpath = relpath; result->ImportDecl.import_name = import_name; result->ImportDecl.docs = docs; result->ImportDecl.comment = comment; return result; } Ast *ast_foreign_import_decl(AstFile *f, Token token, Array filepaths, Token library_name, CommentGroup *docs, CommentGroup *comment) { Ast *result = alloc_ast_node(f, Ast_ForeignImportDecl); result->ForeignImportDecl.token = token; result->ForeignImportDecl.filepaths = slice_from_array(filepaths); result->ForeignImportDecl.library_name = library_name; result->ForeignImportDecl.docs = docs; result->ForeignImportDecl.comment = comment; result->ForeignImportDecl.attributes.allocator = heap_allocator(); return result; } Ast *ast_attribute(AstFile *f, Token token, Token open, Token close, Array const &elems) { Ast *result = alloc_ast_node(f, Ast_Attribute); result->Attribute.token = token; result->Attribute.open = open; result->Attribute.elems = slice_from_array(elems); result->Attribute.close = close; return result; } bool next_token0(AstFile *f) { if (f->curr_token_index+1 < f->tokens.count) { f->curr_token = f->tokens[++f->curr_token_index]; return true; } syntax_error(f->curr_token, "Token is EOF"); return false; } Token consume_comment(AstFile *f, isize *end_line_) { Token tok = f->curr_token; GB_ASSERT(tok.kind == Token_Comment); isize end_line = tok.pos.line; if (tok.string[1] == '*') { for (isize i = 2; i < tok.string.len; i++) { if (tok.string[i] == '\n') { end_line++; } } } if (end_line_) *end_line_ = end_line; next_token0(f); if (f->curr_token.pos.line > tok.pos.line || tok.kind == Token_EOF) { end_line++; } return tok; } CommentGroup *consume_comment_group(AstFile *f, isize n, isize *end_line_) { Array list = {}; list.allocator = heap_allocator(); isize end_line = f->curr_token.pos.line; if (f->curr_token_index == 1 && f->prev_token.kind == Token_Comment && f->prev_token.pos.line+1 == f->curr_token.pos.line) { // NOTE(bill): Special logic for the first comment in the file array_add(&list, f->prev_token); } while (f->curr_token.kind == Token_Comment && f->curr_token.pos.line <= end_line+n) { array_add(&list, consume_comment(f, &end_line)); } if (end_line_) *end_line_ = end_line; CommentGroup *comments = nullptr; if (list.count > 0) { comments = gb_alloc_item(permanent_allocator(), CommentGroup); comments->list = slice_from_array(list); array_add(&f->comments, comments); } return comments; } void consume_comment_groups(AstFile *f, Token prev) { if (f->curr_token.kind == Token_Comment) { CommentGroup *comment = nullptr; isize end_line = 0; if (f->curr_token.pos.line == prev.pos.line) { comment = consume_comment_group(f, 0, &end_line); if (f->curr_token.pos.line != end_line || f->curr_token.kind == Token_EOF) { f->line_comment = comment; } } end_line = -1; while (f->curr_token.kind == Token_Comment) { comment = consume_comment_group(f, 1, &end_line); } if (end_line+1 == f->curr_token.pos.line || end_line < 0) { f->lead_comment = comment; } GB_ASSERT(f->curr_token.kind != Token_Comment); } } gb_inline bool ignore_newlines(AstFile *f) { return f->expr_level > 0; } Token advance_token(AstFile *f) { f->lead_comment = nullptr; f->line_comment = nullptr; f->prev_token_index = f->curr_token_index; Token prev = f->prev_token = f->curr_token; bool ok = next_token0(f); if (ok) { switch (f->curr_token.kind) { case Token_Comment: consume_comment_groups(f, prev); break; case Token_Semicolon: if (ignore_newlines(f) && f->curr_token.string == "\n") { advance_token(f); } break; } } return prev; } bool peek_token_kind(AstFile *f, TokenKind kind) { for (isize i = f->curr_token_index+1; i < f->tokens.count; i++) { Token tok = f->tokens[i]; if (kind != Token_Comment && tok.kind == Token_Comment) { continue; } return tok.kind == kind; } return false; } Token peek_token(AstFile *f) { for (isize i = f->curr_token_index+1; i < f->tokens.count; i++) { Token tok = f->tokens[i]; if (tok.kind == Token_Comment) { continue; } return tok; } return {}; } bool skip_possible_newline(AstFile *f) { if (token_is_newline(f->curr_token)) { advance_token(f); return true; } return false; } bool skip_possible_newline_for_literal(AstFile *f) { Token curr = f->curr_token; if (token_is_newline(curr)) { Token next = peek_token(f); if (curr.pos.line+1 >= next.pos.line) { switch (next.kind) { case Token_OpenBrace: case Token_else: case Token_where: advance_token(f); return true; } } } return false; } String token_to_string(Token const &tok) { String p = token_strings[tok.kind]; if (token_is_newline(tok)) { p = str_lit("newline"); } return p; } Token expect_token(AstFile *f, TokenKind kind) { Token prev = f->curr_token; if (prev.kind != kind) { String c = token_strings[kind]; String p = token_to_string(prev); syntax_error(f->curr_token, "Expected '%.*s', got '%.*s'", LIT(c), LIT(p)); if (prev.kind == Token_EOF) { gb_exit(1); } } advance_token(f); return prev; } Token expect_token_after(AstFile *f, TokenKind kind, char const *msg) { Token prev = f->curr_token; if (prev.kind != kind) { String p = token_to_string(prev); syntax_error(f->curr_token, "Expected '%.*s' after %s, got '%.*s'", LIT(token_strings[kind]), msg, LIT(p)); } advance_token(f); return prev; } bool is_token_range(TokenKind kind) { switch (kind) { case Token_Ellipsis: case Token_RangeFull: case Token_RangeHalf: return true; } return false; } bool is_token_range(Token tok) { return is_token_range(tok.kind); } Token expect_operator(AstFile *f) { Token prev = f->curr_token; if ((prev.kind == Token_in || prev.kind == Token_not_in) && (f->expr_level >= 0 || f->allow_in_expr)) { // okay } else if (prev.kind == Token_if || prev.kind == Token_when) { // okay } else if (prev.kind == Token_or_else || prev.kind == Token_or_return) { // okay } else if (!gb_is_between(prev.kind, Token__OperatorBegin+1, Token__OperatorEnd-1)) { String p = token_to_string(prev); syntax_error(f->curr_token, "Expected an operator, got '%.*s'", LIT(p)); } else if (!f->allow_range && is_token_range(prev)) { String p = token_to_string(prev); syntax_error(f->curr_token, "Expected an non-range operator, got '%.*s'", LIT(p)); } if (f->curr_token.kind == Token_Ellipsis) { syntax_warning(f->curr_token, "'..' for ranges has now be deprecated, prefer '..='"); f->tokens[f->curr_token_index].flags |= TokenFlag_Replace; } advance_token(f); return prev; } Token expect_keyword(AstFile *f) { Token prev = f->curr_token; if (!gb_is_between(prev.kind, Token__KeywordBegin+1, Token__KeywordEnd-1)) { String p = token_to_string(prev); syntax_error(f->curr_token, "Expected a keyword, got '%.*s'", LIT(p)); } advance_token(f); return prev; } bool allow_token(AstFile *f, TokenKind kind) { Token prev = f->curr_token; if (prev.kind == kind) { advance_token(f); return true; } return false; } Token expect_closing_brace_of_field_list(AstFile *f) { Token token = f->curr_token; if (allow_token(f, Token_CloseBrace)) { return token; } bool ok = true; if (!f->allow_newline) { ok = !skip_possible_newline(f); } if (ok && allow_token(f, Token_Semicolon)) { String p = token_to_string(token); syntax_error(token_end_of_line(f, f->prev_token), "Expected a comma, got a %.*s", LIT(p)); } return expect_token(f, Token_CloseBrace); } bool is_blank_ident(String str) { if (str.len == 1) { return str[0] == '_'; } return false; } bool is_blank_ident(Token token) { if (token.kind == Token_Ident) { return is_blank_ident(token.string); } return false; } bool is_blank_ident(Ast *node) { if (node->kind == Ast_Ident) { ast_node(i, Ident, node); return is_blank_ident(i->token.string); } return false; } // NOTE(bill): Go to next statement to prevent numerous error messages popping up void fix_advance_to_next_stmt(AstFile *f) { for (;;) { Token t = f->curr_token; switch (t.kind) { case Token_EOF: case Token_Semicolon: return; case Token_package: case Token_foreign: case Token_import: case Token_if: case Token_for: case Token_when: case Token_return: case Token_switch: case Token_defer: case Token_using: case Token_break: case Token_continue: case Token_fallthrough: case Token_Hash: { if (t.pos == f->fix_prev_pos && f->fix_count < PARSER_MAX_FIX_COUNT) { f->fix_count++; return; } if (f->fix_prev_pos < t.pos) { f->fix_prev_pos = t.pos; f->fix_count = 0; // NOTE(bill): Reset return; } // NOTE(bill): Reaching here means there is a parsing bug } break; } advance_token(f); } } Token expect_closing(AstFile *f, TokenKind kind, String const &context) { if (f->curr_token.kind != kind && f->curr_token.kind == Token_Semicolon && (f->curr_token.string == "\n" || f->curr_token.kind == Token_EOF)) { if (f->allow_newline) { Token tok = f->prev_token; tok.pos.column += cast(i32)tok.string.len; syntax_error(tok, "Missing ',' before newline in %.*s", LIT(context)); } advance_token(f); } return expect_token(f, kind); } void assign_removal_flag_to_semicolon(AstFile *f) { // NOTE(bill): this is used for rewriting files to strip unneeded semicolons Token *prev_token = &f->tokens[f->prev_token_index]; Token *curr_token = &f->tokens[f->curr_token_index]; GB_ASSERT(prev_token->kind == Token_Semicolon); if (prev_token->string == ";") { bool ok = false; if (curr_token->pos.line > prev_token->pos.line) { ok = true; } else if (curr_token->pos.line == prev_token->pos.line) { switch (curr_token->kind) { case Token_CloseBrace: case Token_CloseParen: case Token_EOF: ok = true; break; } } if (ok) { if (build_context.strict_style) { syntax_error(*prev_token, "Found unneeded semicolon"); } else if (build_context.strict_style_init_only && f->pkg->kind == Package_Init) { syntax_error(*prev_token, "Found unneeded semicolon"); } prev_token->flags |= TokenFlag_Remove; } } } void expect_semicolon(AstFile *f) { Token prev_token = {}; if (allow_token(f, Token_Semicolon)) { assign_removal_flag_to_semicolon(f); return; } switch (f->curr_token.kind) { case Token_CloseBrace: case Token_CloseParen: if (f->curr_token.pos.line == f->prev_token.pos.line) { return; } break; } prev_token = f->prev_token; if (prev_token.kind == Token_Semicolon) { assign_removal_flag_to_semicolon(f); return; } if (f->curr_token.kind == Token_EOF) { return; } switch (f->curr_token.kind) { case Token_EOF: return; } if (f->curr_token.pos.line == f->prev_token.pos.line) { String p = token_to_string(f->curr_token); prev_token.pos = token_pos_end(prev_token); syntax_error(prev_token, "Expected ';', got %.*s", LIT(p)); fix_advance_to_next_stmt(f); } } Ast * parse_expr(AstFile *f, bool lhs); Ast * parse_proc_type(AstFile *f, Token proc_token); Array parse_stmt_list(AstFile *f); Ast * parse_stmt(AstFile *f); Ast * parse_body(AstFile *f); Ast * parse_do_body(AstFile *f, Token const &token, char const *msg); Ast * parse_block_stmt(AstFile *f, b32 is_when); Ast *parse_ident(AstFile *f, bool allow_poly_names=false) { Token token = f->curr_token; if (token.kind == Token_Ident) { advance_token(f); } else if (allow_poly_names && token.kind == Token_Dollar) { Token dollar = expect_token(f, Token_Dollar); Ast *name = ast_ident(f, expect_token(f, Token_Ident)); if (is_blank_ident(name)) { syntax_error(name, "Invalid polymorphic type definition with a blank identifier"); } return ast_poly_type(f, dollar, name, nullptr); } else { token.string = str_lit("_"); expect_token(f, Token_Ident); } return ast_ident(f, token); } Ast *parse_tag_expr(AstFile *f, Ast *expression) { Token token = expect_token(f, Token_Hash); Token name = expect_token(f, Token_Ident); return ast_tag_expr(f, token, name, expression); } Ast *unparen_expr(Ast *node) { for (;;) { if (node == nullptr) { return nullptr; } if (node->kind != Ast_ParenExpr) { return node; } node = node->ParenExpr.expr; } } Ast *unselector_expr(Ast *node) { node = unparen_expr(node); if (node == nullptr) { return nullptr; } while (node->kind == Ast_SelectorExpr) { node = node->SelectorExpr.selector; } return node; } Ast *strip_or_return_expr(Ast *node) { for (;;) { if (node == nullptr) { return node; } if (node->kind == Ast_OrReturnExpr) { node = node->OrReturnExpr.expr; } else if (node->kind == Ast_ParenExpr) { node = node->ParenExpr.expr; } else { return node; } } } Ast *parse_value(AstFile *f); Array parse_element_list(AstFile *f) { auto elems = array_make(heap_allocator()); while (f->curr_token.kind != Token_CloseBrace && f->curr_token.kind != Token_EOF) { Ast *elem = parse_value(f); if (f->curr_token.kind == Token_Eq) { Token eq = expect_token(f, Token_Eq); Ast *value = parse_value(f); elem = ast_field_value(f, elem, value, eq); } array_add(&elems, elem); if (!allow_field_separator(f)) { break; } } return elems; } CommentGroup *consume_line_comment(AstFile *f) { CommentGroup *comment = f->line_comment; if (f->line_comment == f->lead_comment) { f->lead_comment = nullptr; } f->line_comment = nullptr; return comment; } Array parse_enum_field_list(AstFile *f) { auto elems = array_make(heap_allocator()); while (f->curr_token.kind != Token_CloseBrace && f->curr_token.kind != Token_EOF) { CommentGroup *docs = f->lead_comment; CommentGroup *comment = nullptr; Ast *name = parse_value(f); Ast *value = nullptr; if (f->curr_token.kind == Token_Eq) { Token eq = expect_token(f, Token_Eq); value = parse_value(f); } comment = consume_line_comment(f); Ast *elem = ast_enum_field_value(f, name, value, docs, comment); array_add(&elems, elem); if (!allow_field_separator(f)) { break; } if (!elem->EnumFieldValue.comment) { elem->EnumFieldValue.comment = consume_line_comment(f); } } return elems; } Ast *parse_literal_value(AstFile *f, Ast *type) { Array elems = {}; Token open = expect_token(f, Token_OpenBrace); isize expr_level = f->expr_level; f->expr_level = 0; if (f->curr_token.kind != Token_CloseBrace) { elems = parse_element_list(f); } f->expr_level = expr_level; Token close = expect_closing(f, Token_CloseBrace, str_lit("compound literal")); return ast_compound_lit(f, type, elems, open, close); } Ast *parse_value(AstFile *f) { if (f->curr_token.kind == Token_OpenBrace) { return parse_literal_value(f, nullptr); } Ast *value; bool prev_allow_range = f->allow_range; f->allow_range = true; value = parse_expr(f, false); f->allow_range = prev_allow_range; return value; } Ast *parse_type_or_ident(AstFile *f); void check_proc_add_tag(AstFile *f, Ast *tag_expr, u64 *tags, ProcTag tag, String const &tag_name) { if (*tags & tag) { syntax_error(tag_expr, "Procedure tag already used: %.*s", LIT(tag_name)); } *tags |= tag; } bool is_foreign_name_valid(String const &name) { if (name.len == 0) { return false; } isize offset = 0; while (offset < name.len) { Rune rune; isize remaining = name.len - offset; isize width = utf8_decode(name.text+offset, remaining, &rune); if (rune == GB_RUNE_INVALID && width == 1) { return false; } else if (rune == GB_RUNE_BOM && remaining > 0) { return false; } if (offset == 0) { switch (rune) { case '-': case '$': case '.': case '_': break; default: if (!gb_char_is_alpha(cast(char)rune)) return false; break; } } else { switch (rune) { case '-': case '$': case '.': case '_': break; default: if (!gb_char_is_alphanumeric(cast(char)rune)) { return false; } break; } } offset += width; } return true; } void parse_proc_tags(AstFile *f, u64 *tags) { GB_ASSERT(tags != nullptr); while (f->curr_token.kind == Token_Hash) { Ast *tag_expr = parse_tag_expr(f, nullptr); ast_node(te, TagExpr, tag_expr); String tag_name = te->name.string; #define ELSE_IF_ADD_TAG(name) \ else if (tag_name == #name) { \ check_proc_add_tag(f, tag_expr, tags, ProcTag_##name, tag_name); \ } if (false) {} ELSE_IF_ADD_TAG(optional_ok) ELSE_IF_ADD_TAG(optional_second) ELSE_IF_ADD_TAG(require_results) ELSE_IF_ADD_TAG(bounds_check) ELSE_IF_ADD_TAG(no_bounds_check) ELSE_IF_ADD_TAG(type_assert) ELSE_IF_ADD_TAG(no_type_assert) else { syntax_error(tag_expr, "Unknown procedure type tag #%.*s", LIT(tag_name)); } #undef ELSE_IF_ADD_TAG } if ((*tags & ProcTag_bounds_check) && (*tags & ProcTag_no_bounds_check)) { syntax_error(f->curr_token, "You cannot apply both #bounds_check and #no_bounds_check to a procedure"); } if ((*tags & ProcTag_type_assert) && (*tags & ProcTag_no_type_assert)) { syntax_error(f->curr_token, "You cannot apply both #type_assert and #no_type_assert to a procedure"); } } Array parse_lhs_expr_list (AstFile *f); Array parse_rhs_expr_list (AstFile *f); Ast * parse_simple_stmt (AstFile *f, u32 flags); Ast * parse_type (AstFile *f); Ast * parse_call_expr (AstFile *f, Ast *operand); Ast * parse_struct_field_list(AstFile *f, isize *name_count_); Ast *parse_field_list(AstFile *f, isize *name_count_, u32 allowed_flags, TokenKind follow, bool allow_default_parameters, bool allow_typeid_token); Ast *parse_unary_expr(AstFile *f, bool lhs); Ast *convert_stmt_to_expr(AstFile *f, Ast *statement, String const &kind) { if (statement == nullptr) { return nullptr; } if (statement->kind == Ast_ExprStmt) { return statement->ExprStmt.expr; } syntax_error(f->curr_token, "Expected '%.*s', found a simple statement.", LIT(kind)); Token end = f->curr_token; if (f->tokens.count < f->curr_token_index) { end = f->tokens[f->curr_token_index+1]; } return ast_bad_expr(f, f->curr_token, end); } Ast *convert_stmt_to_body(AstFile *f, Ast *stmt) { if (stmt->kind == Ast_BlockStmt) { syntax_error(stmt, "Expected a normal statement rather than a block statement"); return stmt; } if (stmt->kind == Ast_EmptyStmt) { syntax_error(stmt, "Expected a non-empty statement"); } GB_ASSERT(is_ast_stmt(stmt) || is_ast_decl(stmt)); Token open = ast_token(stmt); Token close = ast_token(stmt); auto stmts = array_make(heap_allocator(), 0, 1); array_add(&stmts, stmt); return ast_block_stmt(f, stmts, open, close); } void check_polymorphic_params_for_type(AstFile *f, Ast *polymorphic_params, Token token) { if (polymorphic_params == nullptr) { return; } if (polymorphic_params->kind != Ast_FieldList) { return; } ast_node(fl, FieldList, polymorphic_params); for_array(fi, fl->list) { Ast *field = fl->list[fi]; if (field->kind != Ast_Field) { continue; } for_array(i, field->Field.names) { Ast *name = field->Field.names[i]; if (name->kind != field->Field.names[0]->kind) { syntax_error(name, "Mixture of polymorphic names using both $ and not for %.*s parameters", LIT(token.string)); return; } } } } bool ast_on_same_line(Token const &x, Ast *yp) { Token y = ast_token(yp); return x.pos.line == y.pos.line; } bool ast_on_same_line(Ast *x, Ast *y) { return ast_on_same_line(ast_token(x), y); } Ast *parse_force_inlining_operand(AstFile *f, Token token) { Ast *expr = parse_unary_expr(f, false); Ast *e = strip_or_return_expr(expr); if (e->kind != Ast_ProcLit && e->kind != Ast_CallExpr) { syntax_error(expr, "%.*s must be followed by a procedure literal or call, got %.*s", LIT(token.string), LIT(ast_strings[expr->kind])); return ast_bad_expr(f, token, f->curr_token); } ProcInlining pi = ProcInlining_none; if (token.kind == Token_Ident) { if (token.string == "force_inline") { pi = ProcInlining_inline; } else if (token.string == "force_no_inline") { pi = ProcInlining_no_inline; } } if (pi != ProcInlining_none) { if (e->kind == Ast_ProcLit) { if (expr->ProcLit.inlining != ProcInlining_none && expr->ProcLit.inlining != pi) { syntax_error(expr, "Cannot apply both '#force_inline' and '#force_no_inline' to a procedure literal"); } expr->ProcLit.inlining = pi; } else if (e->kind == Ast_CallExpr) { if (expr->CallExpr.inlining != ProcInlining_none && expr->CallExpr.inlining != pi) { syntax_error(expr, "Cannot apply both '#force_inline' and '#force_no_inline' to a procedure call"); } expr->CallExpr.inlining = pi; } } return expr; } Ast *parse_check_directive_for_statement(Ast *s, Token const &tag_token, u16 state_flag) { String name = tag_token.string; if (s == nullptr) { syntax_error(tag_token, "Invalid operand for #%.*s", LIT(name)); return nullptr; } if (s != nullptr && s->kind == Ast_EmptyStmt) { if (s->EmptyStmt.token.string == "\n") { syntax_error(tag_token, "#%.*s cannot be followed by a newline", LIT(name)); } else { syntax_error(tag_token, "#%.*s cannot be applied to an empty statement ';'", LIT(name)); } } if (s->state_flags & state_flag) { syntax_error(tag_token, "#%.*s has been applied multiple times", LIT(name)); } s->state_flags |= state_flag; switch (state_flag) { case StateFlag_bounds_check: if ((s->state_flags & StateFlag_no_bounds_check) != 0) { syntax_error(tag_token, "#bounds_check and #no_bounds_check cannot be applied together"); } break; case StateFlag_no_bounds_check: if ((s->state_flags & StateFlag_bounds_check) != 0) { syntax_error(tag_token, "#bounds_check and #no_bounds_check cannot be applied together"); } break; case StateFlag_type_assert: if ((s->state_flags & StateFlag_no_type_assert) != 0) { syntax_error(tag_token, "#type_assert and #no_type_assert cannot be applied together"); } break; case StateFlag_no_type_assert: if ((s->state_flags & StateFlag_type_assert) != 0) { syntax_error(tag_token, "#type_assert and #no_type_assert cannot be applied together"); } break; } switch (state_flag) { case StateFlag_bounds_check: case StateFlag_no_bounds_check: case StateFlag_type_assert: case StateFlag_no_type_assert: switch (s->kind) { case Ast_BlockStmt: case Ast_IfStmt: case Ast_WhenStmt: case Ast_ForStmt: case Ast_RangeStmt: case Ast_UnrollRangeStmt: case Ast_SwitchStmt: case Ast_TypeSwitchStmt: case Ast_ReturnStmt: case Ast_DeferStmt: case Ast_AssignStmt: // Okay break; case Ast_ValueDecl: if (!s->ValueDecl.is_mutable) { syntax_error(tag_token, "#%.*s may only be applied to a variable declaration, and not a constant value declaration", LIT(name)); } break; default: syntax_error(tag_token, "#%.*s may only be applied to the following statements: '{}', 'if', 'when', 'for', 'switch', 'return', 'defer', assignment, variable declaration", LIT(name)); break; } break; } return s; } Array parse_union_variant_list(AstFile *f) { auto variants = array_make(heap_allocator()); while (f->curr_token.kind != Token_CloseBrace && f->curr_token.kind != Token_EOF) { Ast *type = parse_type(f); if (type->kind != Ast_BadExpr) { array_add(&variants, type); } if (!allow_field_separator(f)) { break; } } return variants; } Ast *parse_operand(AstFile *f, bool lhs) { Ast *operand = nullptr; // Operand switch (f->curr_token.kind) { case Token_Ident: return parse_ident(f); case Token_Undef: return ast_undef(f, expect_token(f, Token_Undef)); case Token_context: return ast_implicit(f, expect_token(f, Token_context)); case Token_Integer: case Token_Float: case Token_Imag: case Token_Rune: return ast_basic_lit(f, advance_token(f)); case Token_String: return ast_basic_lit(f, advance_token(f)); case Token_OpenBrace: if (!lhs) return parse_literal_value(f, nullptr); break; case Token_OpenParen: { bool allow_newline; isize prev_expr_level; Token open, close; // NOTE(bill): Skip the Paren Expression open = expect_token(f, Token_OpenParen); if (f->prev_token.kind == Token_CloseParen) { close = expect_token(f, Token_CloseParen); syntax_error(open, "Invalid parentheses expression with no inside expression"); return ast_bad_expr(f, open, close); } prev_expr_level = f->expr_level; allow_newline = f->allow_newline; if (f->expr_level < 0) { f->allow_newline = false; } // NOTE(bill): enforce it to >0 f->expr_level = gb_max(f->expr_level, 0)+1; operand = parse_expr(f, false); f->allow_newline = allow_newline; f->expr_level = prev_expr_level; close = expect_token(f, Token_CloseParen); return ast_paren_expr(f, operand, open, close); } case Token_distinct: { Token token = expect_token(f, Token_distinct); Ast *type = parse_type(f); return ast_distinct_type(f, token, type); } case Token_Hash: { Token token = expect_token(f, Token_Hash); Token name = expect_token(f, Token_Ident); if (name.string == "type") { return ast_helper_type(f, token, parse_type(f)); } else if ( name.string == "simd") { Ast *tag = ast_basic_directive(f, token, name); Ast *original_type = parse_type(f); Ast *type = unparen_expr(original_type); switch (type->kind) { case Ast_ArrayType: type->ArrayType.tag = tag; break; default: syntax_error(type, "Expected a fixed array type after #%.*s, got %.*s", LIT(name.string), LIT(ast_strings[type->kind])); break; } return original_type; } else if (name.string == "soa") { Ast *tag = ast_basic_directive(f, token, name); Ast *original_type = parse_type(f); Ast *type = unparen_expr(original_type); switch (type->kind) { case Ast_ArrayType: type->ArrayType.tag = tag; break; case Ast_DynamicArrayType: type->DynamicArrayType.tag = tag; break; case Ast_PointerType: type->PointerType.tag = tag; break; default: syntax_error(type, "Expected an array or pointer type after #%.*s, got %.*s", LIT(name.string), LIT(ast_strings[type->kind])); break; } return original_type; } else if (name.string == "partial") { Ast *tag = ast_basic_directive(f, token, name); Ast *original_expr = parse_expr(f, lhs); Ast *expr = unparen_expr(original_expr); switch (expr->kind) { case Ast_ArrayType: syntax_error(expr, "#partial has been replaced with #sparse for non-contiguous enumerated array types"); break; case Ast_CompoundLit: expr->CompoundLit.tag = tag; break; default: syntax_error(expr, "Expected a compound literal after #%.*s, got %.*s", LIT(name.string), LIT(ast_strings[expr->kind])); break; } return original_expr; } else if (name.string == "sparse") { Ast *tag = ast_basic_directive(f, token, name); Ast *original_type = parse_type(f); Ast *type = unparen_expr(original_type); switch (type->kind) { case Ast_ArrayType: type->ArrayType.tag = tag; break; default: syntax_error(type, "Expected an enumerated array type after #%.*s, got %.*s", LIT(name.string), LIT(ast_strings[type->kind])); break; } return original_type; } else if (name.string == "bounds_check") { Ast *operand = parse_expr(f, lhs); return parse_check_directive_for_statement(operand, name, StateFlag_bounds_check); } else if (name.string == "no_bounds_check") { Ast *operand = parse_expr(f, lhs); return parse_check_directive_for_statement(operand, name, StateFlag_no_bounds_check); } else if (name.string == "type_assert") { Ast *operand = parse_expr(f, lhs); return parse_check_directive_for_statement(operand, name, StateFlag_type_assert); } else if (name.string == "no_type_assert") { Ast *operand = parse_expr(f, lhs); return parse_check_directive_for_statement(operand, name, StateFlag_no_type_assert); } else if (name.string == "relative") { Ast *tag = ast_basic_directive(f, token, name); tag = parse_call_expr(f, tag); Ast *type = parse_type(f); return ast_relative_type(f, tag, type); } else if (name.string == "force_inline" || name.string == "force_no_inline") { return parse_force_inlining_operand(f, name); } return ast_basic_directive(f, token, name); } // Parse Procedure Type or Literal or Group case Token_proc: { Token token = expect_token(f, Token_proc); if (f->curr_token.kind == Token_OpenBrace) { // ProcGroup Token open = expect_token(f, Token_OpenBrace); auto args = array_make(heap_allocator()); while (f->curr_token.kind != Token_CloseBrace && f->curr_token.kind != Token_EOF) { Ast *elem = parse_expr(f, false); array_add(&args, elem); if (!allow_field_separator(f)) { break; } } Token close = expect_token(f, Token_CloseBrace); if (args.count == 0) { syntax_error(token, "Expected a least 1 argument in a procedure group"); } return ast_proc_group(f, token, open, close, args); } Ast *type = parse_proc_type(f, token); Token where_token = {}; Array where_clauses = {}; u64 tags = 0; skip_possible_newline_for_literal(f); if (f->curr_token.kind == Token_where) { where_token = expect_token(f, Token_where); isize prev_level = f->expr_level; f->expr_level = -1; where_clauses = parse_rhs_expr_list(f); f->expr_level = prev_level; } parse_proc_tags(f, &tags); if ((tags & ProcTag_require_results) != 0) { syntax_error(f->curr_token, "#require_results has now been replaced as an attribute @(require_results) on the declaration"); tags &= ~ProcTag_require_results; } GB_ASSERT(type->kind == Ast_ProcType); type->ProcType.tags = tags; if (f->allow_type && f->expr_level < 0) { if (tags != 0) { syntax_error(token, "A procedure type cannot have suffix tags"); } if (where_token.kind != Token_Invalid) { syntax_error(where_token, "'where' clauses are not allowed on procedure types"); } return type; } skip_possible_newline_for_literal(f); if (allow_token(f, Token_Undef)) { if (where_token.kind != Token_Invalid) { syntax_error(where_token, "'where' clauses are not allowed on procedure literals without a defined body (replaced with ---)"); } return ast_proc_lit(f, type, nullptr, tags, where_token, where_clauses); } else if (f->curr_token.kind == Token_OpenBrace) { Ast *curr_proc = f->curr_proc; Ast *body = nullptr; f->curr_proc = type; body = parse_body(f); f->curr_proc = curr_proc; // Apply the tags directly to the body rather than the type if (tags & ProcTag_no_bounds_check) { body->state_flags |= StateFlag_no_bounds_check; } if (tags & ProcTag_bounds_check) { body->state_flags |= StateFlag_bounds_check; } if (tags & ProcTag_no_type_assert) { body->state_flags |= StateFlag_no_type_assert; } if (tags & ProcTag_type_assert) { body->state_flags |= StateFlag_type_assert; } return ast_proc_lit(f, type, body, tags, where_token, where_clauses); } else if (allow_token(f, Token_do)) { Ast *curr_proc = f->curr_proc; Ast *body = nullptr; f->curr_proc = type; body = convert_stmt_to_body(f, parse_stmt(f)); f->curr_proc = curr_proc; syntax_error(body, "'do' for procedure bodies is not allowed, prefer {}"); return ast_proc_lit(f, type, body, tags, where_token, where_clauses); } if (tags != 0) { syntax_error(token, "A procedure type cannot have suffix tags"); } if (where_token.kind != Token_Invalid) { syntax_error(where_token, "'where' clauses are not allowed on procedure types"); } return type; } // Check for Types case Token_Dollar: { Token token = expect_token(f, Token_Dollar); Ast *type = parse_ident(f); if (is_blank_ident(type)) { syntax_error(type, "Invalid polymorphic type definition with a blank identifier"); } Ast *specialization = nullptr; if (allow_token(f, Token_Quo)) { specialization = parse_type(f); } return ast_poly_type(f, token, type, specialization); } break; case Token_typeid: { Token token = expect_token(f, Token_typeid); return ast_typeid_type(f, token, nullptr); } break; case Token_Pointer: { Token token = expect_token(f, Token_Pointer); Ast *elem = parse_type(f); return ast_pointer_type(f, token, elem); } break; case Token_OpenBracket: { Token token = expect_token(f, Token_OpenBracket); Ast *count_expr = nullptr; if (f->curr_token.kind == Token_Pointer) { expect_token(f, Token_Pointer); expect_token(f, Token_CloseBracket); return ast_multi_pointer_type(f, token, parse_type(f)); } else if (f->curr_token.kind == Token_Question) { count_expr = ast_unary_expr(f, expect_token(f, Token_Question), nullptr); } else if (allow_token(f, Token_dynamic)) { expect_token(f, Token_CloseBracket); return ast_dynamic_array_type(f, token, parse_type(f)); } else if (f->curr_token.kind != Token_CloseBracket) { f->expr_level++; count_expr = parse_expr(f, false); f->expr_level--; } expect_token(f, Token_CloseBracket); return ast_array_type(f, token, count_expr, parse_type(f)); } break; case Token_map: { Token token = expect_token(f, Token_map); Ast *key = nullptr; Ast *value = nullptr; Token open, close; open = expect_token_after(f, Token_OpenBracket, "map"); key = parse_expr(f, true); close = expect_token(f, Token_CloseBracket); value = parse_type(f); return ast_map_type(f, token, key, value); } break; case Token_matrix: { Token token = expect_token(f, Token_matrix); Ast *row_count = nullptr; Ast *column_count = nullptr; Ast *type = nullptr; Token open, close; open = expect_token_after(f, Token_OpenBracket, "matrix"); row_count = parse_expr(f, true); expect_token(f, Token_Comma); column_count = parse_expr(f, true); close = expect_token(f, Token_CloseBracket); type = parse_type(f); return ast_matrix_type(f, token, row_count, column_count, type); } break; case Token_struct: { Token token = expect_token(f, Token_struct); Ast *polymorphic_params = nullptr; bool is_packed = false; bool is_raw_union = false; Ast *align = nullptr; if (allow_token(f, Token_OpenParen)) { isize param_count = 0; polymorphic_params = parse_field_list(f, ¶m_count, 0, Token_CloseParen, true, true); if (param_count == 0) { syntax_error(polymorphic_params, "Expected at least 1 polymorphic parameter"); polymorphic_params = nullptr; } expect_token_after(f, Token_CloseParen, "parameter list"); check_polymorphic_params_for_type(f, polymorphic_params, token); } isize prev_level; prev_level = f->expr_level; f->expr_level = -1; while (allow_token(f, Token_Hash)) { Token tag = expect_token_after(f, Token_Ident, "#"); if (tag.string == "packed") { if (is_packed) { syntax_error(tag, "Duplicate struct tag '#%.*s'", LIT(tag.string)); } is_packed = true; } else if (tag.string == "align") { if (align) { syntax_error(tag, "Duplicate struct tag '#%.*s'", LIT(tag.string)); } align = parse_expr(f, true); } else if (tag.string == "raw_union") { if (is_raw_union) { syntax_error(tag, "Duplicate struct tag '#%.*s'", LIT(tag.string)); } is_raw_union = true; } else { syntax_error(tag, "Invalid struct tag '#%.*s'", LIT(tag.string)); } } f->expr_level = prev_level; if (is_raw_union && is_packed) { is_packed = false; syntax_error(token, "'#raw_union' cannot also be '#packed'"); } Token where_token = {}; Array where_clauses = {}; skip_possible_newline_for_literal(f); if (f->curr_token.kind == Token_where) { where_token = expect_token(f, Token_where); prev_level = f->expr_level; f->expr_level = -1; where_clauses = parse_rhs_expr_list(f); f->expr_level = prev_level; } skip_possible_newline_for_literal(f); Token open = expect_token_after(f, Token_OpenBrace, "struct"); isize name_count = 0; Ast *fields = parse_struct_field_list(f, &name_count); Token close = expect_closing_brace_of_field_list(f); Slice decls = {}; if (fields != nullptr) { GB_ASSERT(fields->kind == Ast_FieldList); decls = fields->FieldList.list; } return ast_struct_type(f, token, decls, name_count, polymorphic_params, is_packed, is_raw_union, align, where_token, where_clauses); } break; case Token_union: { Token token = expect_token(f, Token_union); Ast *polymorphic_params = nullptr; Ast *align = nullptr; bool no_nil = false; bool maybe = false; bool shared_nil = false; UnionTypeKind union_kind = UnionType_Normal; Token start_token = f->curr_token; if (allow_token(f, Token_OpenParen)) { isize param_count = 0; polymorphic_params = parse_field_list(f, ¶m_count, 0, Token_CloseParen, true, true); if (param_count == 0) { syntax_error(polymorphic_params, "Expected at least 1 polymorphic parametric"); polymorphic_params = nullptr; } expect_token_after(f, Token_CloseParen, "parameter list"); check_polymorphic_params_for_type(f, polymorphic_params, token); } while (allow_token(f, Token_Hash)) { Token tag = expect_token_after(f, Token_Ident, "#"); if (tag.string == "align") { if (align) { syntax_error(tag, "Duplicate union tag '#%.*s'", LIT(tag.string)); } align = parse_expr(f, true); } else if (tag.string == "no_nil") { if (no_nil) { syntax_error(tag, "Duplicate union tag '#%.*s'", LIT(tag.string)); } no_nil = true; } else if (tag.string == "shared_nil") { if (shared_nil) { syntax_error(tag, "Duplicate union tag '#%.*s'", LIT(tag.string)); } shared_nil = true; } else if (tag.string == "maybe") { if (maybe) { syntax_error(tag, "Duplicate union tag '#%.*s'", LIT(tag.string)); } maybe = true; }else { syntax_error(tag, "Invalid union tag '#%.*s'", LIT(tag.string)); } } if (no_nil && shared_nil) { syntax_error(f->curr_token, "#shared_nil and #no_nil cannot be applied together"); } if (maybe) { syntax_error(f->curr_token, "#maybe functionality has now been merged with standard 'union' functionality"); } if (no_nil) { union_kind = UnionType_no_nil; } else if (shared_nil) { union_kind = UnionType_shared_nil; } skip_possible_newline_for_literal(f); Token where_token = {}; Array where_clauses = {}; if (f->curr_token.kind == Token_where) { where_token = expect_token(f, Token_where); isize prev_level = f->expr_level; f->expr_level = -1; where_clauses = parse_rhs_expr_list(f); f->expr_level = prev_level; } skip_possible_newline_for_literal(f); Token open = expect_token_after(f, Token_OpenBrace, "union"); auto variants = parse_union_variant_list(f); Token close = expect_closing_brace_of_field_list(f); return ast_union_type(f, token, variants, polymorphic_params, align, union_kind, where_token, where_clauses); } break; case Token_enum: { Token token = expect_token(f, Token_enum); Ast *base_type = nullptr; if (f->curr_token.kind != Token_OpenBrace) { base_type = parse_type(f); } skip_possible_newline_for_literal(f); Token open = expect_token(f, Token_OpenBrace); Array values = parse_enum_field_list(f); Token close = expect_closing_brace_of_field_list(f); return ast_enum_type(f, token, base_type, values); } break; case Token_bit_set: { Token token = expect_token(f, Token_bit_set); expect_token(f, Token_OpenBracket); Ast *elem = nullptr; Ast *underlying = nullptr; bool prev_allow_range = f->allow_range; f->allow_range = true; elem = parse_expr(f, true); f->allow_range = prev_allow_range; if (allow_token(f, Token_Semicolon)) { underlying = parse_type(f); } else if (allow_token(f, Token_Comma)) { String p = token_to_string(f->prev_token); syntax_error(token_end_of_line(f, f->prev_token), "Expected a semicolon, got a %.*s", LIT(p)); underlying = parse_type(f); } expect_token(f, Token_CloseBracket); return ast_bit_set_type(f, token, elem, underlying); } case Token_asm: { Token token = expect_token(f, Token_asm); Array param_types = {}; Ast *return_type = nullptr; if (allow_token(f, Token_OpenParen)) { param_types = array_make(heap_allocator()); while (f->curr_token.kind != Token_CloseParen && f->curr_token.kind != Token_EOF) { Ast *t = parse_type(f); array_add(¶m_types, t); if (f->curr_token.kind != Token_Comma || f->curr_token.kind == Token_EOF) { break; } advance_token(f); } expect_token(f, Token_CloseParen); if (allow_token(f, Token_ArrowRight)) { return_type = parse_type(f); } } bool has_side_effects = false; bool is_align_stack = false; InlineAsmDialectKind dialect = InlineAsmDialect_Default; while (f->curr_token.kind == Token_Hash) { advance_token(f); if (f->curr_token.kind == Token_Ident) { Token token = advance_token(f); String name = token.string; if (name == "side_effects") { if (has_side_effects) { syntax_error(token, "Duplicate directive on inline asm expression: '#side_effects'"); } has_side_effects = true; } else if (name == "align_stack") { if (is_align_stack) { syntax_error(token, "Duplicate directive on inline asm expression: '#align_stack'"); } is_align_stack = true; } else if (name == "att") { if (dialect == InlineAsmDialect_ATT) { syntax_error(token, "Duplicate directive on inline asm expression: '#att'"); } else if (dialect != InlineAsmDialect_Default) { syntax_error(token, "Conflicting asm dialects"); } else { dialect = InlineAsmDialect_ATT; } } else if (name == "intel") { if (dialect == InlineAsmDialect_Intel) { syntax_error(token, "Duplicate directive on inline asm expression: '#intel'"); } else if (dialect != InlineAsmDialect_Default) { syntax_error(token, "Conflicting asm dialects"); } else { dialect = InlineAsmDialect_Intel; } } } else { syntax_error(f->curr_token, "Expected an identifier after hash"); } } skip_possible_newline_for_literal(f); Token open = expect_token(f, Token_OpenBrace); Ast *asm_string = parse_expr(f, false); expect_token(f, Token_Comma); Ast *constraints_string = parse_expr(f, false); allow_token(f, Token_Comma); Token close = expect_closing_brace_of_field_list(f); return ast_inline_asm_expr(f, token, open, close, param_types, return_type, asm_string, constraints_string, has_side_effects, is_align_stack, dialect); } } return nullptr; } bool is_literal_type(Ast *node) { node = unparen_expr(node); switch (node->kind) { case Ast_BadExpr: case Ast_Ident: case Ast_SelectorExpr: case Ast_ArrayType: case Ast_StructType: case Ast_UnionType: case Ast_EnumType: case Ast_DynamicArrayType: case Ast_MapType: case Ast_BitSetType: case Ast_MatrixType: case Ast_CallExpr: return true; case Ast_MultiPointerType: // For better error messages return true; } return false; } Ast *parse_call_expr(AstFile *f, Ast *operand) { auto args = array_make(heap_allocator()); Token open_paren, close_paren; Token ellipsis = {}; isize prev_expr_level = f->expr_level; bool prev_allow_newline = f->allow_newline; f->expr_level = 0; f->allow_newline = ALLOW_NEWLINE; open_paren = expect_token(f, Token_OpenParen); while (f->curr_token.kind != Token_CloseParen && f->curr_token.kind != Token_EOF && ellipsis.pos.line == 0) { if (f->curr_token.kind == Token_Comma) { syntax_error(f->curr_token, "Expected an expression not ,"); } else if (f->curr_token.kind == Token_Eq) { syntax_error(f->curr_token, "Expected an expression not ="); } bool prefix_ellipsis = false; if (f->curr_token.kind == Token_Ellipsis) { prefix_ellipsis = true; ellipsis = expect_token(f, Token_Ellipsis); } Ast *arg = parse_expr(f, false); if (f->curr_token.kind == Token_Eq) { Token eq = expect_token(f, Token_Eq); if (prefix_ellipsis) { syntax_error(ellipsis, "'..' must be applied to value rather than the field name"); } Ast *value = parse_value(f); arg = ast_field_value(f, arg, value, eq); } array_add(&args, arg); if (!allow_field_separator(f)) { break; } } f->allow_newline = prev_allow_newline; f->expr_level = prev_expr_level; close_paren = expect_closing(f, Token_CloseParen, str_lit("argument list")); Ast *call = ast_call_expr(f, operand, args, open_paren, close_paren, ellipsis); Ast *o = unparen_expr(operand); if (o->kind == Ast_SelectorExpr && o->SelectorExpr.token.kind == Token_ArrowRight) { return ast_selector_call_expr(f, o->SelectorExpr.token, o, call); } return call; } Ast *parse_atom_expr(AstFile *f, Ast *operand, bool lhs) { if (operand == nullptr) { if (f->allow_type) return nullptr; Token begin = f->curr_token; syntax_error(begin, "Expected an operand"); fix_advance_to_next_stmt(f); operand = ast_bad_expr(f, begin, f->curr_token); } bool loop = true; while (loop) { switch (f->curr_token.kind) { case Token_OpenParen: operand = parse_call_expr(f, operand); break; case Token_Period: { Token token = advance_token(f); switch (f->curr_token.kind) { case Token_Ident: operand = ast_selector_expr(f, token, operand, parse_ident(f)); break; // case Token_Integer: // operand = ast_selector_expr(f, token, operand, parse_expr(f, lhs)); // break; case Token_OpenParen: { Token open = expect_token(f, Token_OpenParen); Ast *type = parse_type(f); Token close = expect_token(f, Token_CloseParen); operand = ast_type_assertion(f, operand, token, type); } break; case Token_Question: { Token question = expect_token(f, Token_Question); Ast *type = ast_unary_expr(f, question, nullptr); operand = ast_type_assertion(f, operand, token, type); } break; default: syntax_error(f->curr_token, "Expected a selector"); advance_token(f); operand = ast_bad_expr(f, ast_token(operand), f->curr_token); // operand = ast_selector_expr(f, f->curr_token, operand, nullptr); break; } } break; case Token_ArrowRight: { Token token = advance_token(f); operand = ast_selector_expr(f, token, operand, parse_ident(f)); // Ast *call = parse_call_expr(f, sel); // operand = ast_selector_call_expr(f, token, sel, call); break; } case Token_OpenBracket: { bool prev_allow_range = f->allow_range; f->allow_range = false; Token open = {}, close = {}, interval = {}; Ast *indices[2] = {}; bool is_interval = false; f->expr_level++; open = expect_token(f, Token_OpenBracket); switch (f->curr_token.kind) { case Token_Ellipsis: case Token_RangeFull: case Token_RangeHalf: // NOTE(bill): Do not err yet case Token_Colon: break; default: indices[0] = parse_expr(f, false); break; } switch (f->curr_token.kind) { case Token_Ellipsis: case Token_RangeFull: case Token_RangeHalf: syntax_error(f->curr_token, "Expected a colon, not a range"); /* fallthrough */ case Token_Comma: // matrix index case Token_Colon: interval = advance_token(f); is_interval = true; if (f->curr_token.kind != Token_CloseBracket && f->curr_token.kind != Token_EOF) { indices[1] = parse_expr(f, false); } break; } f->expr_level--; close = expect_token(f, Token_CloseBracket); if (is_interval) { if (interval.kind == Token_Comma) { if (indices[0] == nullptr || indices[1] == nullptr) { syntax_error(open, "Matrix index expressions require both row and column indices"); } operand = ast_matrix_index_expr(f, operand, open, close, interval, indices[0], indices[1]); } else { operand = ast_slice_expr(f, operand, open, close, interval, indices[0], indices[1]); } } else { operand = ast_index_expr(f, operand, indices[0], open, close); } f->allow_range = prev_allow_range; } break; case Token_Pointer: // Deference operand = ast_deref_expr(f, operand, expect_token(f, Token_Pointer)); break; case Token_or_return: operand = ast_or_return_expr(f, operand, expect_token(f, Token_or_return)); break; case Token_OpenBrace: if (!lhs && is_literal_type(operand) && f->expr_level >= 0) { operand = parse_literal_value(f, operand); } else { loop = false; } break; case Token_Increment: case Token_Decrement: if (!lhs) { Token token = advance_token(f); syntax_error(token, "Postfix '%.*s' operator is not supported", LIT(token.string)); } else { loop = false; } break; default: loop = false; break; } lhs = false; // NOTE(bill): 'tis not lhs anymore } return operand; } Ast *parse_unary_expr(AstFile *f, bool lhs) { switch (f->curr_token.kind) { case Token_transmute: case Token_cast: { Token token = advance_token(f); expect_token(f, Token_OpenParen); Ast *type = parse_type(f); expect_token(f, Token_CloseParen); Ast *expr = parse_unary_expr(f, lhs); return ast_type_cast(f, token, type, expr); } case Token_auto_cast: { Token token = advance_token(f); Ast *expr = parse_unary_expr(f, lhs); return ast_auto_cast(f, token, expr); } case Token_Add: case Token_Sub: case Token_Xor: case Token_And: case Token_Not: { Token token = advance_token(f); Ast *expr = parse_unary_expr(f, lhs); return ast_unary_expr(f, token, expr); } case Token_Increment: case Token_Decrement: { Token token = advance_token(f); syntax_error(token, "Unary '%.*s' operator is not supported", LIT(token.string)); Ast *expr = parse_unary_expr(f, lhs); return ast_unary_expr(f, token, expr); } case Token_Period: { Token token = expect_token(f, Token_Period); Ast *ident = parse_ident(f); return ast_implicit_selector_expr(f, token, ident); } } return parse_atom_expr(f, parse_operand(f, lhs), lhs); } bool is_ast_range(Ast *expr) { if (expr == nullptr) { return false; } if (expr->kind != Ast_BinaryExpr) { return false; } return is_token_range(expr->BinaryExpr.op.kind); } // NOTE(bill): result == priority i32 token_precedence(AstFile *f, TokenKind t) { switch (t) { case Token_Question: case Token_if: case Token_when: case Token_or_else: return 1; case Token_Ellipsis: case Token_RangeFull: case Token_RangeHalf: if (!f->allow_range) { return 0; } return 2; case Token_CmpOr: return 3; case Token_CmpAnd: return 4; case Token_CmpEq: case Token_NotEq: case Token_Lt: case Token_Gt: case Token_LtEq: case Token_GtEq: return 5; case Token_in: case Token_not_in: if (f->expr_level < 0 && !f->allow_in_expr) { return 0; } /*fallthrough*/ case Token_Add: case Token_Sub: case Token_Or: case Token_Xor: return 6; case Token_Mul: case Token_Quo: case Token_Mod: case Token_ModMod: case Token_And: case Token_AndNot: case Token_Shl: case Token_Shr: return 7; } return 0; } Ast *parse_binary_expr(AstFile *f, bool lhs, i32 prec_in) { Ast *expr = parse_unary_expr(f, lhs); for (;;) { Token op = f->curr_token; i32 op_prec = token_precedence(f, op.kind); if (op_prec < prec_in) { // NOTE(bill): This will also catch operators that are not valid "binary" operators break; } Token prev = f->prev_token; switch (op.kind) { case Token_if: case Token_when: if (prev.pos.line < op.pos.line) { // NOTE(bill): Check to see if the `if` or `when` is on the same line of the `lhs` condition goto loop_end; } break; } expect_operator(f); // NOTE(bill): error checks too if (op.kind == Token_Question) { Ast *cond = expr; // Token_Question Ast *x = parse_expr(f, lhs); Token token_c = expect_token(f, Token_Colon); Ast *y = parse_expr(f, lhs); expr = ast_ternary_if_expr(f, x, cond, y); } else if (op.kind == Token_if || op.kind == Token_when) { Ast *x = expr; Ast *cond = parse_expr(f, lhs); Token tok_else = expect_token(f, Token_else); Ast *y = parse_expr(f, lhs); switch (op.kind) { case Token_if: expr = ast_ternary_if_expr(f, x, cond, y); break; case Token_when: expr = ast_ternary_when_expr(f, x, cond, y); break; } } else { Ast *right = parse_binary_expr(f, false, op_prec+1); if (right == nullptr) { syntax_error(op, "Expected expression on the right-hand side of the binary operator '%.*s'", LIT(op.string)); } if (op.kind == Token_or_else) { // NOTE(bill): easier to handle its logic different with its own AST kind expr = ast_or_else_expr(f, expr, op, right); } else { expr = ast_binary_expr(f, op, expr, right); } } lhs = false; } loop_end:; return expr; } Ast *parse_expr(AstFile *f, bool lhs) { return parse_binary_expr(f, lhs, 0+1); } Array parse_expr_list(AstFile *f, bool lhs) { bool allow_newline = f->allow_newline; f->allow_newline = ALLOW_NEWLINE; auto list = array_make(heap_allocator()); for (;;) { Ast *e = parse_expr(f, lhs); array_add(&list, e); if (f->curr_token.kind != Token_Comma || f->curr_token.kind == Token_EOF) { break; } advance_token(f); } f->allow_newline = allow_newline; return list; } Array parse_lhs_expr_list(AstFile *f) { return parse_expr_list(f, true); } Array parse_rhs_expr_list(AstFile *f) { return parse_expr_list(f, false); } Array parse_ident_list(AstFile *f, bool allow_poly_names) { auto list = array_make(heap_allocator()); for (;;) { array_add(&list, parse_ident(f, allow_poly_names)); if (f->curr_token.kind != Token_Comma || f->curr_token.kind == Token_EOF) { break; } advance_token(f); } return list; } Ast *parse_type(AstFile *f) { Ast *type = parse_type_or_ident(f); if (type == nullptr) { Token token = advance_token(f); syntax_error(token, "Expected a type"); return ast_bad_expr(f, token, f->curr_token); } return type; } void parse_foreign_block_decl(AstFile *f, Array *decls) { Ast *decl = parse_stmt(f); switch (decl->kind) { case Ast_EmptyStmt: case Ast_BadStmt: case Ast_BadDecl: return; case Ast_WhenStmt: case Ast_ValueDecl: array_add(decls, decl); return; default: syntax_error(decl, "Foreign blocks only allow procedure and variable declarations"); return; } } Ast *parse_foreign_block(AstFile *f, Token token) { CommentGroup *docs = f->lead_comment; Ast *foreign_library = nullptr; if (f->curr_token.kind == Token_OpenBrace) { foreign_library = ast_ident(f, blank_token); } else { foreign_library = parse_ident(f); } Token open = {}; Token close = {}; auto decls = array_make(heap_allocator()); bool prev_in_foreign_block = f->in_foreign_block; defer (f->in_foreign_block = prev_in_foreign_block); f->in_foreign_block = true; skip_possible_newline_for_literal(f); open = expect_token(f, Token_OpenBrace); while (f->curr_token.kind != Token_CloseBrace && f->curr_token.kind != Token_EOF) { parse_foreign_block_decl(f, &decls); } close = expect_token(f, Token_CloseBrace); Ast *body = ast_block_stmt(f, decls, open, close); Ast *decl = ast_foreign_block_decl(f, token, foreign_library, body, docs); expect_semicolon(f); return decl; } Ast *parse_value_decl(AstFile *f, Array names, CommentGroup *docs) { bool is_mutable = true; Array values = {}; Ast *type = parse_type_or_ident(f); if (f->curr_token.kind == Token_Eq || f->curr_token.kind == Token_Colon) { Token sep = {}; if (!is_mutable) { sep = expect_token_after(f, Token_Colon, "type"); } else { sep = advance_token(f); is_mutable = sep.kind != Token_Colon; } values = parse_rhs_expr_list(f); if (values.count > names.count) { syntax_error(f->curr_token, "Too many values on the right hand side of the declaration"); } else if (values.count < names.count && !is_mutable) { syntax_error(f->curr_token, "All constant declarations must be defined"); } else if (values.count == 0) { syntax_error(f->curr_token, "Expected an expression for this declaration"); } } if (is_mutable) { if (type == nullptr && values.count == 0) { syntax_error(f->curr_token, "Missing variable type or initialization"); return ast_bad_decl(f, f->curr_token, f->curr_token); } } else { if (type == nullptr && values.count == 0 && names.count > 0) { syntax_error(f->curr_token, "Missing constant value"); return ast_bad_decl(f, f->curr_token, f->curr_token); } } if (values.data == nullptr) { values.allocator = heap_allocator(); } if (f->expr_level >= 0) { if (f->curr_token.kind == Token_CloseBrace && f->curr_token.pos.line == f->prev_token.pos.line) { } else { expect_semicolon(f); } } if (f->curr_proc == nullptr) { if (values.count > 0 && names.count != values.count) { syntax_error( values[0], "Expected %td expressions on the right hand side, got %td\n" "\tNote: Global declarations do not allow for multi-valued expressions", names.count, values.count ); } } return ast_value_decl(f, names, type, values, is_mutable, docs, f->line_comment); } Ast *parse_simple_stmt(AstFile *f, u32 flags) { Token token = f->curr_token; CommentGroup *docs = f->lead_comment; Array lhs = parse_lhs_expr_list(f); token = f->curr_token; switch (token.kind) { case Token_Eq: case Token_AddEq: case Token_SubEq: case Token_MulEq: case Token_QuoEq: case Token_ModEq: case Token_ModModEq: case Token_AndEq: case Token_OrEq: case Token_XorEq: case Token_ShlEq: case Token_ShrEq: case Token_AndNotEq: case Token_CmpAndEq: case Token_CmpOrEq: { if (f->curr_proc == nullptr) { syntax_error(f->curr_token, "You cannot use a simple statement in the file scope"); return ast_bad_stmt(f, f->curr_token, f->curr_token); } advance_token(f); Array rhs = parse_rhs_expr_list(f); if (rhs.count == 0) { syntax_error(token, "No right-hand side in assignment statement."); return ast_bad_stmt(f, token, f->curr_token); } return ast_assign_stmt(f, token, lhs, rhs); } break; case Token_in: if (flags&StmtAllowFlag_In) { allow_token(f, Token_in); bool prev_allow_range = f->allow_range; f->allow_range = true; Ast *expr = parse_expr(f, true); f->allow_range = prev_allow_range; auto rhs = array_make(heap_allocator(), 0, 1); array_add(&rhs, expr); return ast_assign_stmt(f, token, lhs, rhs); } break; case Token_Colon: expect_token_after(f, Token_Colon, "identifier list"); if ((flags&StmtAllowFlag_Label) && lhs.count == 1) { switch (f->curr_token.kind) { case Token_OpenBrace: // block statement case Token_if: case Token_for: case Token_switch: { Ast *name = lhs[0]; Ast *label = ast_label_decl(f, ast_token(name), name); Ast *stmt = parse_stmt(f); #define _SET_LABEL(Kind_, label_) case GB_JOIN2(Ast_, Kind_): (stmt->Kind_).label = label_; break switch (stmt->kind) { _SET_LABEL(BlockStmt, label); _SET_LABEL(IfStmt, label); _SET_LABEL(ForStmt, label); _SET_LABEL(RangeStmt, label); _SET_LABEL(SwitchStmt, label); _SET_LABEL(TypeSwitchStmt, label); default: syntax_error(token, "Labels can only be applied to a loop or switch statement"); break; } #undef _SET_LABEL return stmt; } break; } } return parse_value_decl(f, lhs, docs); } if (lhs.count > 1) { syntax_error(token, "Expected 1 expression"); return ast_bad_stmt(f, token, f->curr_token); } switch (token.kind) { case Token_Increment: case Token_Decrement: advance_token(f); syntax_error(token, "Postfix '%.*s' statement is not supported", LIT(token.string)); break; } #if 0 switch (token.kind) { case Token_Inc: case Token_Dec: advance_token(f); return ast_inc_dec_stmt(f, token, lhs[0]); } #endif return ast_expr_stmt(f, lhs[0]); } Ast *parse_block_stmt(AstFile *f, b32 is_when) { skip_possible_newline_for_literal(f); if (!is_when && f->curr_proc == nullptr) { syntax_error(f->curr_token, "You cannot use a block statement in the file scope"); return ast_bad_stmt(f, f->curr_token, f->curr_token); } return parse_body(f); } Ast *parse_results(AstFile *f, bool *diverging) { if (!allow_token(f, Token_ArrowRight)) { return nullptr; } if (allow_token(f, Token_Not)) { if (diverging) *diverging = true; return nullptr; } isize prev_level = f->expr_level; defer (f->expr_level = prev_level); // f->expr_level = -1; if (f->curr_token.kind != Token_OpenParen) { Token begin_token = f->curr_token; Array empty_names = {}; auto list = array_make(heap_allocator(), 0, 1); Ast *type = parse_type(f); Token tag = {}; array_add(&list, ast_field(f, empty_names, type, nullptr, 0, tag, nullptr, nullptr)); return ast_field_list(f, begin_token, list); } Ast *list = nullptr; expect_token(f, Token_OpenParen); list = parse_field_list(f, nullptr, FieldFlag_Results, Token_CloseParen, true, false); expect_token_after(f, Token_CloseParen, "parameter list"); return list; } ProcCallingConvention string_to_calling_convention(String const &s) { if (s == "odin") return ProcCC_Odin; if (s == "contextless") return ProcCC_Contextless; if (s == "cdecl") return ProcCC_CDecl; if (s == "c") return ProcCC_CDecl; if (s == "stdcall") return ProcCC_StdCall; if (s == "std") return ProcCC_StdCall; if (s == "fastcall") return ProcCC_FastCall; if (s == "fast") return ProcCC_FastCall; if (s == "none") return ProcCC_None; if (s == "naked") return ProcCC_Naked; if (s == "win64") return ProcCC_Win64; if (s == "sysv") return ProcCC_SysV; if (s == "system") { if (build_context.metrics.os == TargetOs_windows) { return ProcCC_StdCall; } return ProcCC_CDecl; } return ProcCC_Invalid; } Ast *parse_proc_type(AstFile *f, Token proc_token) { Ast *params = nullptr; Ast *results = nullptr; bool diverging = false; ProcCallingConvention cc = ProcCC_Invalid; if (f->curr_token.kind == Token_String) { Token token = expect_token(f, Token_String); auto c = string_to_calling_convention(string_value_from_token(f, token)); if (c == ProcCC_Invalid) { syntax_error(token, "Unknown procedure calling convention: '%.*s'", LIT(token.string)); } else { cc = c; } } if (cc == ProcCC_Invalid) { if (f->in_foreign_block) { cc = ProcCC_ForeignBlockDefault; } else { cc = default_calling_convention(); } } expect_token(f, Token_OpenParen); params = parse_field_list(f, nullptr, FieldFlag_Signature, Token_CloseParen, true, true); expect_token_after(f, Token_CloseParen, "parameter list"); results = parse_results(f, &diverging); u64 tags = 0; bool is_generic = false; for_array(i, params->FieldList.list) { Ast *param = params->FieldList.list[i]; ast_node(field, Field, param); if (field->type != nullptr) { if (field->type->kind == Ast_PolyType) { is_generic = true; goto end; } for_array(j, field->names) { Ast *name = field->names[j]; if (name->kind == Ast_PolyType) { is_generic = true; goto end; } } } } end: return ast_proc_type(f, proc_token, params, results, tags, cc, is_generic, diverging); } Ast *parse_var_type(AstFile *f, bool allow_ellipsis, bool allow_typeid_token) { if (allow_ellipsis && f->curr_token.kind == Token_Ellipsis) { Token tok = advance_token(f); Ast *type = parse_type_or_ident(f); if (type == nullptr) { syntax_error(tok, "variadic field missing type after '..'"); type = ast_bad_expr(f, tok, f->curr_token); } return ast_ellipsis(f, tok, type); } Ast *type = nullptr; if (allow_typeid_token && f->curr_token.kind == Token_typeid) { Token token = expect_token(f, Token_typeid); Ast *specialization = nullptr; if (allow_token(f, Token_Quo)) { specialization = parse_type(f); } type = ast_typeid_type(f, token, specialization); } else { type = parse_type(f); } return type; } struct ParseFieldPrefixMapping { String name; TokenKind token_kind; FieldFlag flag; }; gb_global ParseFieldPrefixMapping parse_field_prefix_mappings[] = { {str_lit("using"), Token_using, FieldFlag_using}, {str_lit("auto_cast"), Token_auto_cast, FieldFlag_auto_cast}, {str_lit("no_alias"), Token_Hash, FieldFlag_no_alias}, {str_lit("c_vararg"), Token_Hash, FieldFlag_c_vararg}, {str_lit("const"), Token_Hash, FieldFlag_const}, {str_lit("any_int"), Token_Hash, FieldFlag_any_int}, {str_lit("subtype"), Token_Hash, FieldFlag_subtype}, {str_lit("by_ptr"), Token_Hash, FieldFlag_by_ptr}, }; FieldFlag is_token_field_prefix(AstFile *f) { switch (f->curr_token.kind) { case Token_EOF: return FieldFlag_Invalid; case Token_using: return FieldFlag_using; case Token_auto_cast: return FieldFlag_auto_cast; case Token_Hash: advance_token(f); switch (f->curr_token.kind) { case Token_Ident: for (i32 i = 0; i < gb_count_of(parse_field_prefix_mappings); i++) { auto const &mapping = parse_field_prefix_mappings[i]; if (mapping.token_kind == Token_Hash) { if (f->curr_token.string == mapping.name) { return mapping.flag; } } } break; } return FieldFlag_Unknown; } return FieldFlag_Invalid; } u32 parse_field_prefixes(AstFile *f) { i32 counts[gb_count_of(parse_field_prefix_mappings)] = {}; for (;;) { FieldFlag flag = is_token_field_prefix(f); if (flag & FieldFlag_Invalid) { break; } if (flag & FieldFlag_Unknown) { syntax_error(f->curr_token, "Unknown prefix kind '#%.*s'", LIT(f->curr_token.string)); advance_token(f); continue; } for (i32 i = 0; i < gb_count_of(parse_field_prefix_mappings); i++) { if (parse_field_prefix_mappings[i].flag == flag) { counts[i] += 1; advance_token(f); break; } } } u32 field_flags = 0; for (i32 i = 0; i < gb_count_of(parse_field_prefix_mappings); i++) { if (counts[i] > 0) { field_flags |= parse_field_prefix_mappings[i].flag; if (counts[i] != 1) { auto const &mapping = parse_field_prefix_mappings[i]; String name = mapping.name; char const *prefix = ""; if (mapping.token_kind == Token_Hash) { prefix = "#"; } syntax_error(f->curr_token, "Multiple '%s%.*s' in this field list", prefix, LIT(name)); } } } return field_flags; } u32 check_field_prefixes(AstFile *f, isize name_count, u32 allowed_flags, u32 set_flags) { for (i32 i = 0; i < gb_count_of(parse_field_prefix_mappings); i++) { bool err = false; auto const &m = parse_field_prefix_mappings[i]; if ((set_flags & m.flag) != 0) { if (m.flag == FieldFlag_using && name_count > 1) { err = true; syntax_error(f->curr_token, "Cannot apply 'using' to more than one of the same type"); } if ((allowed_flags & m.flag) == 0) { err = true; char const *prefix = ""; if (m.token_kind == Token_Hash) { prefix = "#"; } syntax_error(f->curr_token, "'%s%.*s' in not allowed within this field list", prefix, LIT(m.name)); } } if (err) { set_flags &= ~m.flag; } } return set_flags; } struct AstAndFlags { Ast *node; u32 flags; }; Array convert_to_ident_list(AstFile *f, Array list, bool ignore_flags, bool allow_poly_names) { auto idents = array_make(heap_allocator(), 0, list.count); // Convert to ident list for_array(i, list) { Ast *ident = list[i].node; if (!ignore_flags) { if (i != 0) { syntax_error(ident, "Illegal use of prefixes in parameter list"); } } switch (ident->kind) { case Ast_Ident: case Ast_BadExpr: break; case Ast_PolyType: if (allow_poly_names) { if (ident->PolyType.specialization == nullptr) { break; } else { syntax_error(ident, "Expected a polymorphic identifier without any specialization"); } } else { syntax_error(ident, "Expected a non-polymorphic identifier"); } /*fallthrough*/ default: syntax_error(ident, "Expected an identifier"); ident = ast_ident(f, blank_token); break; } array_add(&idents, ident); } return idents; } bool allow_field_separator(AstFile *f) { Token token = f->curr_token; if (allow_token(f, Token_Comma)) { return true; } if (ALLOW_NEWLINE && token.kind == Token_Semicolon) { String p = token_to_string(token); syntax_error(token_end_of_line(f, f->prev_token), "Expected a comma, got a %.*s", LIT(p)); advance_token(f); return true; } return false; } Ast *parse_struct_field_list(AstFile *f, isize *name_count_) { Token start_token = f->curr_token; auto decls = array_make(heap_allocator()); isize total_name_count = 0; Ast *params = parse_field_list(f, &total_name_count, FieldFlag_Struct, Token_CloseBrace, false, false); if (name_count_) *name_count_ = total_name_count; return params; } // Returns true if any are polymorphic names bool check_procedure_name_list(Array const &names) { if (names.count == 0) { return false; } bool first_is_polymorphic = names[0]->kind == Ast_PolyType; bool any_polymorphic_names = first_is_polymorphic; for (isize i = 1; i < names.count; i++) { Ast *name = names[i]; if (first_is_polymorphic) { if (name->kind == Ast_PolyType) { any_polymorphic_names = true; } else { syntax_error(name, "Mixture of polymorphic and non-polymorphic identifiers"); return any_polymorphic_names; } } else { if (name->kind == Ast_PolyType) { any_polymorphic_names = true; syntax_error(name, "Mixture of polymorphic and non-polymorphic identifiers"); return any_polymorphic_names; } else { // Okay } } } return any_polymorphic_names; } Ast *parse_field_list(AstFile *f, isize *name_count_, u32 allowed_flags, TokenKind follow, bool allow_default_parameters, bool allow_typeid_token) { bool prev_allow_newline = f->allow_newline; defer (f->allow_newline = prev_allow_newline); f->allow_newline = ALLOW_NEWLINE; Token start_token = f->curr_token; CommentGroup *docs = f->lead_comment; auto params = array_make(heap_allocator()); auto list = array_make(heap_allocator()); defer (array_free(&list)); bool allow_poly_names = allow_typeid_token; isize total_name_count = 0; bool allow_ellipsis = allowed_flags&FieldFlag_ellipsis; bool seen_ellipsis = false; bool is_signature = (allowed_flags & FieldFlag_Signature) == FieldFlag_Signature; while (f->curr_token.kind != follow && f->curr_token.kind != Token_Colon && f->curr_token.kind != Token_EOF) { u32 flags = parse_field_prefixes(f); Ast *param = parse_var_type(f, allow_ellipsis, allow_typeid_token); if (param->kind == Ast_Ellipsis) { if (seen_ellipsis) syntax_error(param, "Extra variadic parameter after ellipsis"); seen_ellipsis = true; } else if (seen_ellipsis) { syntax_error(param, "Extra parameter after ellipsis"); } AstAndFlags naf = {param, flags}; array_add(&list, naf); if (!allow_field_separator(f)) { break; } } if (f->curr_token.kind == Token_Colon) { Array names = convert_to_ident_list(f, list, true, allow_poly_names); // Copy for semantic reasons if (names.count == 0) { syntax_error(f->curr_token, "Empty field declaration"); } bool any_polymorphic_names = check_procedure_name_list(names); u32 set_flags = 0; if (list.count > 0) { set_flags = list[0].flags; } set_flags = check_field_prefixes(f, names.count, allowed_flags, set_flags); total_name_count += names.count; Ast *type = nullptr; Ast *default_value = nullptr; Token tag = {}; expect_token_after(f, Token_Colon, "field list"); if (f->curr_token.kind != Token_Eq) { type = parse_var_type(f, allow_ellipsis, allow_typeid_token); Ast *tt = unparen_expr(type); if (tt == nullptr) { syntax_error(f->prev_token, "Invalid type expression in field list"); } else if (is_signature && !any_polymorphic_names && tt->kind == Ast_TypeidType && tt->TypeidType.specialization != nullptr) { syntax_error(type, "Specialization of typeid is not allowed without polymorphic names"); } } if (allow_token(f, Token_Eq)) { default_value = parse_expr(f, false); if (!allow_default_parameters) { syntax_error(f->curr_token, "Default parameters are only allowed for procedures"); default_value = nullptr; } } if (default_value != nullptr && names.count > 1) { syntax_error(f->curr_token, "Default parameters can only be applied to single values"); } if (allowed_flags == FieldFlag_Struct && default_value != nullptr) { syntax_error(default_value, "Default parameters are not allowed for structs"); default_value = nullptr; } if (type != nullptr && type->kind == Ast_Ellipsis) { if (seen_ellipsis) syntax_error(type, "Extra variadic parameter after ellipsis"); seen_ellipsis = true; if (names.count != 1) { syntax_error(type, "Variadic parameters can only have one field name"); } } else if (seen_ellipsis && default_value == nullptr) { syntax_error(f->curr_token, "Extra parameter after ellipsis without a default value"); } if (type != nullptr && default_value == nullptr) { if (f->curr_token.kind == Token_String) { tag = expect_token(f, Token_String); if ((allowed_flags & FieldFlag_Tags) == 0) { syntax_error(tag, "Field tags are only allowed within structures"); } } } allow_field_separator(f); Ast *param = ast_field(f, names, type, default_value, set_flags, tag, docs, f->line_comment); array_add(¶ms, param); while (f->curr_token.kind != follow && f->curr_token.kind != Token_EOF && f->curr_token.kind != Token_Semicolon) { CommentGroup *docs = f->lead_comment; u32 set_flags = parse_field_prefixes(f); Token tag = {}; Array names = parse_ident_list(f, allow_poly_names); if (names.count == 0) { syntax_error(f->curr_token, "Empty field declaration"); break; } bool any_polymorphic_names = check_procedure_name_list(names); set_flags = check_field_prefixes(f, names.count, allowed_flags, set_flags); total_name_count += names.count; Ast *type = nullptr; Ast *default_value = nullptr; expect_token_after(f, Token_Colon, "field list"); if (f->curr_token.kind != Token_Eq) { type = parse_var_type(f, allow_ellipsis, allow_typeid_token); Ast *tt = unparen_expr(type); if (is_signature && !any_polymorphic_names && tt->kind == Ast_TypeidType && tt->TypeidType.specialization != nullptr) { syntax_error(type, "Specialization of typeid is not allowed without polymorphic names"); } } if (allow_token(f, Token_Eq)) { default_value = parse_expr(f, false); if (!allow_default_parameters) { syntax_error(f->curr_token, "Default parameters are only allowed for procedures"); default_value = nullptr; } } if (default_value != nullptr && names.count > 1) { syntax_error(f->curr_token, "Default parameters can only be applied to single values"); } if (type != nullptr && type->kind == Ast_Ellipsis) { if (seen_ellipsis) syntax_error(type, "Extra variadic parameter after ellipsis"); seen_ellipsis = true; if (names.count != 1) { syntax_error(type, "Variadic parameters can only have one field name"); } } else if (seen_ellipsis && default_value == nullptr) { syntax_error(f->curr_token, "Extra parameter after ellipsis without a default value"); } if (type != nullptr && default_value == nullptr) { if (f->curr_token.kind == Token_String) { tag = expect_token(f, Token_String); if ((allowed_flags & FieldFlag_Tags) == 0) { syntax_error(tag, "Field tags are only allowed within structures"); } } } bool ok = allow_field_separator(f); Ast *param = ast_field(f, names, type, default_value, set_flags, tag, docs, f->line_comment); array_add(¶ms, param); if (!ok) { break; } } if (name_count_) *name_count_ = total_name_count; return ast_field_list(f, start_token, params); } for_array(i, list) { Ast *type = list[i].node; Token token = blank_token; if (allowed_flags&FieldFlag_Results) { // NOTE(bill): Make this nothing and not `_` token.string = str_lit(""); } auto names = array_make(heap_allocator(), 1); token.pos = ast_token(type).pos; names[0] = ast_ident(f, token); u32 flags = check_field_prefixes(f, list.count, allowed_flags, list[i].flags); Token tag = {}; Ast *param = ast_field(f, names, list[i].node, nullptr, flags, tag, docs, f->line_comment); array_add(¶ms, param); } if (name_count_) *name_count_ = total_name_count; return ast_field_list(f, start_token, params); } Ast *parse_type_or_ident(AstFile *f) { bool prev_allow_type = f->allow_type; isize prev_expr_level = f->expr_level; defer ({ f->allow_type = prev_allow_type; f->expr_level = prev_expr_level; }); f->allow_type = true; f->expr_level = -1; bool lhs = true; Ast *operand = parse_operand(f, lhs); Ast *type = parse_atom_expr(f, operand, lhs); return type; } Ast *parse_body(AstFile *f) { Array stmts = {}; Token open, close; isize prev_expr_level = f->expr_level; bool prev_allow_newline = f->allow_newline; // NOTE(bill): The body may be within an expression so reset to zero f->expr_level = 0; // f->allow_newline = false; open = expect_token(f, Token_OpenBrace); stmts = parse_stmt_list(f); close = expect_token(f, Token_CloseBrace); f->expr_level = prev_expr_level; f->allow_newline = prev_allow_newline; return ast_block_stmt(f, stmts, open, close); } Ast *parse_do_body(AstFile *f, Token const &token, char const *msg) { Token open, close; isize prev_expr_level = f->expr_level; bool prev_allow_newline = f->allow_newline; // NOTE(bill): The body may be within an expression so reset to zero f->expr_level = 0; f->allow_newline = false; Ast *body = convert_stmt_to_body(f, parse_stmt(f)); if (build_context.disallow_do) { syntax_error(body, "'do' has been disallowed"); } else if (token.pos.file_id != 0 && !ast_on_same_line(token, body)) { syntax_error(body, "The body of a 'do' must be on the same line as %s", msg); } f->expr_level = prev_expr_level; f->allow_newline = prev_allow_newline; return body; } bool parse_control_statement_semicolon_separator(AstFile *f) { Token tok = peek_token(f); if (tok.kind != Token_OpenBrace) { return allow_token(f, Token_Semicolon); } if (f->curr_token.string == ";") { return allow_token(f, Token_Semicolon); } return false; } Ast *parse_if_stmt(AstFile *f) { if (f->curr_proc == nullptr) { syntax_error(f->curr_token, "You cannot use an if statement in the file scope"); return ast_bad_stmt(f, f->curr_token, f->curr_token); } Token token = expect_token(f, Token_if); Ast *init = nullptr; Ast *cond = nullptr; Ast *body = nullptr; Ast *else_stmt = nullptr; isize prev_level = f->expr_level; f->expr_level = -1; bool prev_allow_in_expr = f->allow_in_expr; f->allow_in_expr = true; if (allow_token(f, Token_Semicolon)) { cond = parse_expr(f, false); } else { init = parse_simple_stmt(f, StmtAllowFlag_None); if (parse_control_statement_semicolon_separator(f)) { cond = parse_expr(f, false); } else { cond = convert_stmt_to_expr(f, init, str_lit("boolean expression")); init = nullptr; } } f->expr_level = prev_level; f->allow_in_expr = prev_allow_in_expr; if (cond == nullptr) { syntax_error(f->curr_token, "Expected condition for if statement"); } if (allow_token(f, Token_do)) { body = parse_do_body(f, cond ? ast_token(cond) : token, "the if statement"); } else { body = parse_block_stmt(f, false); } skip_possible_newline_for_literal(f); if (f->curr_token.kind == Token_else) { Token else_token = expect_token(f, Token_else); switch (f->curr_token.kind) { case Token_if: else_stmt = parse_if_stmt(f); break; case Token_OpenBrace: else_stmt = parse_block_stmt(f, false); break; case Token_do: expect_token(f, Token_do); else_stmt = parse_do_body(f, else_token, "'else'"); break; default: syntax_error(f->curr_token, "Expected if statement block statement"); else_stmt = ast_bad_stmt(f, f->curr_token, f->tokens[f->curr_token_index+1]); break; } } return ast_if_stmt(f, token, init, cond, body, else_stmt); } Ast *parse_when_stmt(AstFile *f) { Token token = expect_token(f, Token_when); Ast *cond = nullptr; Ast *body = nullptr; Ast *else_stmt = nullptr; isize prev_level = f->expr_level; f->expr_level = -1; cond = parse_expr(f, false); f->expr_level = prev_level; if (cond == nullptr) { syntax_error(f->curr_token, "Expected condition for when statement"); } if (allow_token(f, Token_do)) { body = parse_do_body(f, cond ? ast_token(cond) : token, "then when statement"); } else { body = parse_block_stmt(f, true); } skip_possible_newline_for_literal(f); if (f->curr_token.kind == Token_else) { Token else_token = expect_token(f, Token_else); switch (f->curr_token.kind) { case Token_when: else_stmt = parse_when_stmt(f); break; case Token_OpenBrace: else_stmt = parse_block_stmt(f, true); break; case Token_do: { expect_token(f, Token_do); else_stmt = parse_do_body(f, else_token, "'else'"); } break; default: syntax_error(f->curr_token, "Expected when statement block statement"); else_stmt = ast_bad_stmt(f, f->curr_token, f->tokens[f->curr_token_index+1]); break; } } return ast_when_stmt(f, token, cond, body, else_stmt); } Ast *parse_return_stmt(AstFile *f) { Token token = expect_token(f, Token_return); if (f->curr_proc == nullptr) { syntax_error(f->curr_token, "You cannot use a return statement in the file scope"); return ast_bad_stmt(f, token, f->curr_token); } if (f->expr_level > 0) { syntax_error(f->curr_token, "You cannot use a return statement within an expression"); return ast_bad_stmt(f, token, f->curr_token); } auto results = array_make(heap_allocator()); while (f->curr_token.kind != Token_Semicolon && f->curr_token.kind != Token_CloseBrace) { Ast *arg = parse_expr(f, false); array_add(&results, arg); if (f->curr_token.kind != Token_Comma || f->curr_token.kind == Token_EOF) { break; } advance_token(f); } expect_semicolon(f); return ast_return_stmt(f, token, results); } Ast *parse_for_stmt(AstFile *f) { if (f->curr_proc == nullptr) { syntax_error(f->curr_token, "You cannot use a for statement in the file scope"); return ast_bad_stmt(f, f->curr_token, f->curr_token); } Token token = expect_token(f, Token_for); Ast *init = nullptr; Ast *cond = nullptr; Ast *post = nullptr; Ast *body = nullptr; bool is_range = false; if (f->curr_token.kind != Token_OpenBrace && f->curr_token.kind != Token_do) { isize prev_level = f->expr_level; defer (f->expr_level = prev_level); f->expr_level = -1; if (f->curr_token.kind == Token_in) { Token in_token = expect_token(f, Token_in); Ast *rhs = nullptr; bool prev_allow_range = f->allow_range; f->allow_range = true; rhs = parse_expr(f, false); f->allow_range = prev_allow_range; if (allow_token(f, Token_do)) { body = parse_do_body(f, token, "the for statement"); } else { body = parse_block_stmt(f, false); } return ast_range_stmt(f, token, {}, in_token, rhs, body); } if (f->curr_token.kind != Token_Semicolon) { cond = parse_simple_stmt(f, StmtAllowFlag_In); if (cond->kind == Ast_AssignStmt && cond->AssignStmt.op.kind == Token_in) { is_range = true; } } if (!is_range && parse_control_statement_semicolon_separator(f)) { init = cond; cond = nullptr; if (f->curr_token.kind == Token_OpenBrace || f->curr_token.kind == Token_do) { syntax_error(f->curr_token, "Expected ';', followed by a condition expression and post statement, got %.*s", LIT(token_strings[f->curr_token.kind])); } else { if (f->curr_token.kind != Token_Semicolon) { cond = parse_simple_stmt(f, StmtAllowFlag_None); } if (f->curr_token.string != ";") { syntax_error(f->curr_token, "Expected ';', got %.*s", LIT(token_to_string(f->curr_token))); } else { expect_token(f, Token_Semicolon); } if (f->curr_token.kind != Token_OpenBrace && f->curr_token.kind != Token_do) { post = parse_simple_stmt(f, StmtAllowFlag_None); } } } } if (allow_token(f, Token_do)) { body = parse_do_body(f, token, "the for statement"); } else { body = parse_block_stmt(f, false); } if (is_range) { GB_ASSERT(cond->kind == Ast_AssignStmt); Token in_token = cond->AssignStmt.op; Slice vals = cond->AssignStmt.lhs; Ast *rhs = nullptr; if (cond->AssignStmt.rhs.count > 0) { rhs = cond->AssignStmt.rhs[0]; } return ast_range_stmt(f, token, vals, in_token, rhs, body); } cond = convert_stmt_to_expr(f, cond, str_lit("boolean expression")); return ast_for_stmt(f, token, init, cond, post, body); } Ast *parse_case_clause(AstFile *f, bool is_type) { Token token = f->curr_token; Array list = {}; expect_token(f, Token_case); bool prev_allow_range = f->allow_range; bool prev_allow_in_expr = f->allow_in_expr; f->allow_range = !is_type; f->allow_in_expr = !is_type; if (f->curr_token.kind != Token_Colon) { list = parse_rhs_expr_list(f); } f->allow_range = prev_allow_range; f->allow_in_expr = prev_allow_in_expr; expect_token(f, Token_Colon); Array stmts = parse_stmt_list(f); return ast_case_clause(f, token, list, stmts); } Ast *parse_switch_stmt(AstFile *f) { if (f->curr_proc == nullptr) { syntax_error(f->curr_token, "You cannot use a switch statement in the file scope"); return ast_bad_stmt(f, f->curr_token, f->curr_token); } Token token = expect_token(f, Token_switch); Ast *init = nullptr; Ast *tag = nullptr; Ast *body = nullptr; Token open, close; bool is_type_switch = false; auto list = array_make(heap_allocator()); if (f->curr_token.kind != Token_OpenBrace) { isize prev_level = f->expr_level; f->expr_level = -1; defer (f->expr_level = prev_level); if (allow_token(f, Token_in)) { auto lhs = array_make(heap_allocator(), 0, 1); auto rhs = array_make(heap_allocator(), 0, 1); Token blank_ident = token; blank_ident.kind = Token_Ident; blank_ident.string = str_lit("_"); Ast *blank = ast_ident(f, blank_ident); array_add(&lhs, blank); array_add(&rhs, parse_expr(f, true)); tag = ast_assign_stmt(f, token, lhs, rhs); is_type_switch = true; } else { tag = parse_simple_stmt(f, StmtAllowFlag_In); if (tag->kind == Ast_AssignStmt && tag->AssignStmt.op.kind == Token_in) { is_type_switch = true; } else if (parse_control_statement_semicolon_separator(f)) { init = tag; tag = nullptr; if (f->curr_token.kind != Token_OpenBrace) { tag = parse_simple_stmt(f, StmtAllowFlag_None); } } } } skip_possible_newline(f); open = expect_token(f, Token_OpenBrace); while (f->curr_token.kind == Token_case) { array_add(&list, parse_case_clause(f, is_type_switch)); } close = expect_token(f, Token_CloseBrace); body = ast_block_stmt(f, list, open, close); if (is_type_switch) { return ast_type_switch_stmt(f, token, tag, body); } tag = convert_stmt_to_expr(f, tag, str_lit("switch expression")); return ast_switch_stmt(f, token, init, tag, body); } Ast *parse_defer_stmt(AstFile *f) { if (f->curr_proc == nullptr) { syntax_error(f->curr_token, "You cannot use a defer statement in the file scope"); return ast_bad_stmt(f, f->curr_token, f->curr_token); } Token token = expect_token(f, Token_defer); Ast *stmt = parse_stmt(f); switch (stmt->kind) { case Ast_EmptyStmt: syntax_error(token, "Empty statement after defer (e.g. ';')"); break; case Ast_DeferStmt: syntax_error(token, "You cannot defer a defer statement"); stmt = stmt->DeferStmt.stmt; break; case Ast_ReturnStmt: syntax_error(token, "You cannot defer a return statement"); break; } return ast_defer_stmt(f, token, stmt); } enum ImportDeclKind { ImportDecl_Standard, ImportDecl_Using, }; Ast *parse_import_decl(AstFile *f, ImportDeclKind kind) { CommentGroup *docs = f->lead_comment; Token token = expect_token(f, Token_import); Token import_name = {}; switch (f->curr_token.kind) { case Token_Ident: import_name = advance_token(f); break; default: import_name.pos = f->curr_token.pos; break; } Token file_path = expect_token_after(f, Token_String, "import"); Ast *s = nullptr; if (f->curr_proc != nullptr) { syntax_error(import_name, "Cannot use 'import' within a procedure. This must be done at the file scope"); s = ast_bad_decl(f, import_name, file_path); } else { s = ast_import_decl(f, token, file_path, import_name, docs, f->line_comment); array_add(&f->imports, s); } if (kind != ImportDecl_Standard) { syntax_error(import_name, "'using import' is not allowed, please use the import name explicitly"); } expect_semicolon(f); return s; } Ast *parse_foreign_decl(AstFile *f) { CommentGroup *docs = f->lead_comment; Token token = expect_token(f, Token_foreign); switch (f->curr_token.kind) { case Token_Ident: case Token_OpenBrace: return parse_foreign_block(f, token); case Token_import: { Token import_token = expect_token(f, Token_import); Token lib_name = {}; switch (f->curr_token.kind) { case Token_Ident: lib_name = advance_token(f); break; default: lib_name.pos = token.pos; break; } if (is_blank_ident(lib_name)) { syntax_error(lib_name, "Illegal foreign import name: '_'"); } Array filepaths = {}; if (allow_token(f, Token_OpenBrace)) { array_init(&filepaths, heap_allocator()); while (f->curr_token.kind != Token_CloseBrace && f->curr_token.kind != Token_EOF) { Token path = expect_token(f, Token_String); array_add(&filepaths, path); if (!allow_field_separator(f)) { break; } } expect_closing_brace_of_field_list(f); } else { filepaths = array_make(heap_allocator(), 0, 1); Token path = expect_token(f, Token_String); array_add(&filepaths, path); } Ast *s = nullptr; if (filepaths.count == 0) { syntax_error(lib_name, "foreign import without any paths"); s = ast_bad_decl(f, lib_name, f->curr_token); } else if (f->curr_proc != nullptr) { syntax_error(lib_name, "You cannot use foreign import within a procedure. This must be done at the file scope"); s = ast_bad_decl(f, lib_name, filepaths[0]); } else { s = ast_foreign_import_decl(f, token, filepaths, lib_name, docs, f->line_comment); } expect_semicolon(f); return s; } } syntax_error(token, "Invalid foreign declaration"); return ast_bad_decl(f, token, f->curr_token); } Ast *parse_attribute(AstFile *f, Token token, TokenKind open_kind, TokenKind close_kind) { Array elems = {}; Token open = {}; Token close = {}; if (f->curr_token.kind == Token_Ident) { elems = array_make(heap_allocator(), 0, 1); Ast *elem = parse_ident(f); array_add(&elems, elem); } else { open = expect_token(f, open_kind); f->expr_level++; if (f->curr_token.kind != close_kind) { elems = array_make(heap_allocator()); while (f->curr_token.kind != close_kind && f->curr_token.kind != Token_EOF) { Ast *elem = nullptr; elem = parse_ident(f); if (f->curr_token.kind == Token_Eq) { Token eq = expect_token(f, Token_Eq); Ast *value = parse_value(f); elem = ast_field_value(f, elem, value, eq); } array_add(&elems, elem); if (!allow_field_separator(f)) { break; } } } f->expr_level--; close = expect_closing(f, close_kind, str_lit("attribute")); } Ast *attribute = ast_attribute(f, token, open, close, elems); skip_possible_newline(f); Ast *decl = parse_stmt(f); if (decl->kind == Ast_ValueDecl) { array_add(&decl->ValueDecl.attributes, attribute); } else if (decl->kind == Ast_ForeignBlockDecl) { array_add(&decl->ForeignBlockDecl.attributes, attribute); } else if (decl->kind == Ast_ForeignImportDecl) { array_add(&decl->ForeignImportDecl.attributes, attribute); }else { syntax_error(decl, "Expected a value or foreign declaration after an attribute, got %.*s", LIT(ast_strings[decl->kind])); return ast_bad_stmt(f, token, f->curr_token); } return decl; } Ast *parse_unrolled_for_loop(AstFile *f, Token unroll_token) { Token for_token = expect_token(f, Token_for); Ast *val0 = nullptr; Ast *val1 = nullptr; Token in_token = {}; Ast *expr = nullptr; Ast *body = nullptr; bool bad_stmt = false; if (f->curr_token.kind != Token_in) { Array idents = parse_ident_list(f, false); switch (idents.count) { case 1: val0 = idents[0]; break; case 2: val0 = idents[0]; val1 = idents[1]; break; default: syntax_error(for_token, "Expected either 1 or 2 identifiers"); bad_stmt = true; break; } } in_token = expect_token(f, Token_in); bool prev_allow_range = f->allow_range; isize prev_level = f->expr_level; f->allow_range = true; f->expr_level = -1; expr = parse_expr(f, false); f->expr_level = prev_level; f->allow_range = prev_allow_range; if (allow_token(f, Token_do)) { body = parse_do_body(f, for_token, "the for statement"); } else { body = parse_block_stmt(f, false); } if (bad_stmt) { return ast_bad_stmt(f, unroll_token, f->curr_token); } return ast_unroll_range_stmt(f, unroll_token, for_token, val0, val1, in_token, expr, body); } Ast *parse_stmt(AstFile *f) { Ast *s = nullptr; Token token = f->curr_token; switch (token.kind) { // Operands case Token_context: // Also allows for `context =` case Token_proc: case Token_Ident: case Token_Integer: case Token_Float: case Token_Imag: case Token_Rune: case Token_String: case Token_OpenParen: case Token_Pointer: case Token_asm: // Inline assembly // Unary Operators case Token_Add: case Token_Sub: case Token_Xor: case Token_Not: case Token_And: s = parse_simple_stmt(f, StmtAllowFlag_Label); expect_semicolon(f); return s; case Token_foreign: return parse_foreign_decl(f); case Token_import: return parse_import_decl(f, ImportDecl_Standard); case Token_if: return parse_if_stmt(f); case Token_when: return parse_when_stmt(f); case Token_for: return parse_for_stmt(f); case Token_switch: return parse_switch_stmt(f); case Token_defer: return parse_defer_stmt(f); case Token_return: return parse_return_stmt(f); case Token_break: case Token_continue: case Token_fallthrough: { Token token = advance_token(f); Ast *label = nullptr; if (token.kind != Token_fallthrough && f->curr_token.kind == Token_Ident) { label = parse_ident(f); } s = ast_branch_stmt(f, token, label); expect_semicolon(f); return s; } case Token_using: { CommentGroup *docs = f->lead_comment; Token token = expect_token(f, Token_using); if (f->curr_token.kind == Token_import) { return parse_import_decl(f, ImportDecl_Using); } Ast *decl = nullptr; Array list = parse_lhs_expr_list(f); if (list.count == 0) { syntax_error(token, "Illegal use of 'using' statement"); expect_semicolon(f); return ast_bad_stmt(f, token, f->curr_token); } if (f->curr_token.kind != Token_Colon) { expect_semicolon(f); return ast_using_stmt(f, token, list); } expect_token_after(f, Token_Colon, "identifier list"); decl = parse_value_decl(f, list, docs); if (decl != nullptr && decl->kind == Ast_ValueDecl) { decl->ValueDecl.is_using = true; return decl; } syntax_error(token, "Illegal use of 'using' statement"); return ast_bad_stmt(f, token, f->curr_token); } break; case Token_At: { Token token = expect_token(f, Token_At); return parse_attribute(f, token, Token_OpenParen, Token_CloseParen); } case Token_Hash: { Ast *s = nullptr; Token hash_token = expect_token(f, Token_Hash); Token name = expect_token(f, Token_Ident); String tag = name.string; if (tag == "bounds_check") { s = parse_stmt(f); return parse_check_directive_for_statement(s, name, StateFlag_bounds_check); } else if (tag == "no_bounds_check") { s = parse_stmt(f); return parse_check_directive_for_statement(s, name, StateFlag_no_bounds_check); } else if (tag == "type_assert") { s = parse_stmt(f); return parse_check_directive_for_statement(s, name, StateFlag_type_assert); } else if (tag == "no_type_assert") { s = parse_stmt(f); return parse_check_directive_for_statement(s, name, StateFlag_no_type_assert); } else if (tag == "partial") { s = parse_stmt(f); switch (s->kind) { case Ast_SwitchStmt: s->SwitchStmt.partial = true; break; case Ast_TypeSwitchStmt: s->TypeSwitchStmt.partial = true; break; case Ast_EmptyStmt: return parse_check_directive_for_statement(s, name, 0); default: syntax_error(token, "#partial can only be applied to a switch statement"); break; } return s; } else if (tag == "assert" || tag == "panic") { Ast *t = ast_basic_directive(f, hash_token, name); Ast *stmt = ast_expr_stmt(f, parse_call_expr(f, t)); expect_semicolon(f); return stmt; } else if (name.string == "force_inline" || name.string == "force_no_inline") { Ast *expr = parse_force_inlining_operand(f, name); Ast *stmt = ast_expr_stmt(f, expr); expect_semicolon(f); return stmt; } else if (tag == "unroll") { return parse_unrolled_for_loop(f, name); } else if (tag == "include") { syntax_error(token, "#include is not a valid import declaration kind. Did you mean 'import'?"); s = ast_bad_stmt(f, token, f->curr_token); } else { syntax_error(token, "Unknown tag directive used: '%.*s'", LIT(tag)); s = ast_bad_stmt(f, token, f->curr_token); } fix_advance_to_next_stmt(f); return s; } break; case Token_OpenBrace: return parse_block_stmt(f, false); case Token_Semicolon: s = ast_empty_stmt(f, token); expect_semicolon(f); return s; } // Error correction statements switch (token.kind) { case Token_else: expect_token(f, Token_else); syntax_error(token, "'else' unattached to an 'if' statement"); switch (f->curr_token.kind) { case Token_if: return parse_if_stmt(f); case Token_when: return parse_when_stmt(f); case Token_OpenBrace: return parse_block_stmt(f, true); case Token_do: { expect_token(f, Token_do); Ast *stmt = parse_do_body(f, {}, "the for statement"); if (build_context.disallow_do) { syntax_error(stmt, "'do' has been disallowed"); } return stmt; } break; default: fix_advance_to_next_stmt(f); return ast_bad_stmt(f, token, f->curr_token); } } syntax_error(token, "Expected a statement, got '%.*s'", LIT(token_strings[token.kind])); fix_advance_to_next_stmt(f); return ast_bad_stmt(f, token, f->curr_token); } Array parse_stmt_list(AstFile *f) { auto list = array_make(heap_allocator()); while (f->curr_token.kind != Token_case && f->curr_token.kind != Token_CloseBrace && f->curr_token.kind != Token_EOF) { Ast *stmt = parse_stmt(f); if (stmt && stmt->kind != Ast_EmptyStmt) { array_add(&list, stmt); if (stmt->kind == Ast_ExprStmt && stmt->ExprStmt.expr != nullptr && stmt->ExprStmt.expr->kind == Ast_ProcLit) { syntax_error(stmt, "Procedure literal evaluated but not used"); } } } return list; } ParseFileError init_ast_file(AstFile *f, String const &fullpath, TokenPos *err_pos) { GB_ASSERT(f != nullptr); f->fullpath = string_trim_whitespace(fullpath); // Just in case set_file_path_string(f->id, fullpath); thread_safe_set_ast_file_from_id(f->id, f); if (!string_ends_with(f->fullpath, str_lit(".odin"))) { return ParseFile_WrongExtension; } zero_item(&f->tokenizer); f->tokenizer.curr_file_id = f->id; TokenizerInitError err = init_tokenizer_from_fullpath(&f->tokenizer, f->fullpath, build_context.copy_file_contents); if (err != TokenizerInit_None) { switch (err) { case TokenizerInit_Empty: break; case TokenizerInit_NotExists: return ParseFile_NotFound; case TokenizerInit_Permission: return ParseFile_Permission; case TokenizerInit_FileTooLarge: return ParseFile_FileTooLarge; default: return ParseFile_InvalidFile; } } isize file_size = f->tokenizer.end - f->tokenizer.start; // NOTE(bill): Determine allocation size required for tokens isize token_cap = file_size/3ll; isize pow2_cap = gb_max(cast(isize)prev_pow2(cast(i64)token_cap)/2, 16); token_cap = ((token_cap + pow2_cap-1)/pow2_cap) * pow2_cap; isize init_token_cap = gb_max(token_cap, 16); array_init(&f->tokens, heap_allocator(), 0, gb_max(init_token_cap, 16)); if (err == TokenizerInit_Empty) { Token token = {Token_EOF}; token.pos.file_id = f->id; token.pos.line = 1; token.pos.column = 1; array_add(&f->tokens, token); return ParseFile_None; } u64 start = time_stamp_time_now(); for (;;) { Token *token = array_add_and_get(&f->tokens); tokenizer_get_token(&f->tokenizer, token); if (token->kind == Token_Invalid) { err_pos->line = token->pos.line; err_pos->column = token->pos.column; return ParseFile_InvalidToken; } if (token->kind == Token_EOF) { break; } } u64 end = time_stamp_time_now(); f->time_to_tokenize = cast(f64)(end-start)/cast(f64)time_stamp__freq(); f->prev_token_index = 0; f->curr_token_index = 0; f->prev_token = f->tokens[f->prev_token_index]; f->curr_token = f->tokens[f->curr_token_index]; array_init(&f->comments, heap_allocator(), 0, 0); array_init(&f->imports, heap_allocator(), 0, 0); f->curr_proc = nullptr; return ParseFile_None; } void destroy_ast_file(AstFile *f) { GB_ASSERT(f != nullptr); array_free(&f->tokens); array_free(&f->comments); array_free(&f->imports); } bool init_parser(Parser *p) { GB_ASSERT(p != nullptr); string_set_init(&p->imported_files, heap_allocator()); array_init(&p->packages, heap_allocator()); array_init(&p->package_imports, heap_allocator()); mutex_init(&p->wait_mutex); mutex_init(&p->import_mutex); mutex_init(&p->file_add_mutex); mutex_init(&p->file_decl_mutex); mutex_init(&p->packages_mutex); mpmc_init(&p->file_error_queue, heap_allocator(), 1024); return true; } void destroy_parser(Parser *p) { GB_ASSERT(p != nullptr); // TODO(bill): Fix memory leak for_array(i, p->packages) { AstPackage *pkg = p->packages[i]; for_array(j, pkg->files) { destroy_ast_file(pkg->files[j]); } array_free(&pkg->files); array_free(&pkg->foreign_files); } #if 0 for_array(i, p->package_imports) { // gb_free(heap_allocator(), p->package_imports[i].text); } #endif array_free(&p->packages); array_free(&p->package_imports); string_set_destroy(&p->imported_files); mutex_destroy(&p->wait_mutex); mutex_destroy(&p->import_mutex); mutex_destroy(&p->file_add_mutex); mutex_destroy(&p->file_decl_mutex); mutex_destroy(&p->packages_mutex); mpmc_destroy(&p->file_error_queue); } void parser_add_package(Parser *p, AstPackage *pkg) { mutex_lock(&p->packages_mutex); pkg->id = p->packages.count+1; array_add(&p->packages, pkg); mutex_unlock(&p->packages_mutex); } ParseFileError process_imported_file(Parser *p, ImportedFile imported_file); WORKER_TASK_PROC(parser_worker_proc) { ParserWorkerData *wd = cast(ParserWorkerData *)data; ParseFileError err = process_imported_file(wd->parser, wd->imported_file); if (err != ParseFile_None) { mpmc_enqueue(&wd->parser->file_error_queue, err); } return cast(isize)err; } void parser_add_file_to_process(Parser *p, AstPackage *pkg, FileInfo fi, TokenPos pos) { // TODO(bill): Use a better allocator ImportedFile f = {pkg, fi, pos, p->file_to_process_count++}; auto wd = gb_alloc_item(permanent_allocator(), ParserWorkerData); wd->parser = p; wd->imported_file = f; global_thread_pool_add_task(parser_worker_proc, wd); } WORKER_TASK_PROC(foreign_file_worker_proc) { ForeignFileWorkerData *wd = cast(ForeignFileWorkerData *)data; Parser *p = wd->parser; ImportedFile *imp = &wd->imported_file; AstPackage *pkg = imp->pkg; AstForeignFile foreign_file = {wd->foreign_kind}; String fullpath = string_trim_whitespace(imp->fi.fullpath); // Just in case char *c_str = alloc_cstring(heap_allocator(), fullpath); defer (gb_free(heap_allocator(), c_str)); gbFileContents fc = gb_file_read_contents(heap_allocator(), true, c_str); foreign_file.source.text = (u8 *)fc.data; foreign_file.source.len = fc.size; switch (wd->foreign_kind) { case AstForeignFile_S: // TODO(bill): Actually do something with it break; } mutex_lock(&p->file_add_mutex); array_add(&pkg->foreign_files, foreign_file); mutex_unlock(&p->file_add_mutex); return 0; } void parser_add_foreign_file_to_process(Parser *p, AstPackage *pkg, AstForeignFileKind kind, FileInfo fi, TokenPos pos) { // TODO(bill): Use a better allocator ImportedFile f = {pkg, fi, pos, p->file_to_process_count++}; auto wd = gb_alloc_item(permanent_allocator(), ForeignFileWorkerData); wd->parser = p; wd->imported_file = f; wd->foreign_kind = kind; global_thread_pool_add_task(foreign_file_worker_proc, wd); } // NOTE(bill): Returns true if it's added AstPackage *try_add_import_path(Parser *p, String const &path, String const &rel_path, TokenPos pos, PackageKind kind = Package_Normal) { String const FILE_EXT = str_lit(".odin"); mutex_lock(&p->import_mutex); defer (mutex_unlock(&p->import_mutex)); if (string_set_exists(&p->imported_files, path)) { return nullptr; } string_set_add(&p->imported_files, path); AstPackage *pkg = gb_alloc_item(permanent_allocator(), AstPackage); pkg->kind = kind; pkg->fullpath = path; array_init(&pkg->files, heap_allocator()); pkg->foreign_files.allocator = heap_allocator(); // NOTE(bill): Single file initial package if (kind == Package_Init && string_ends_with(path, FILE_EXT)) { FileInfo fi = {}; fi.name = filename_from_path(path); fi.fullpath = path; fi.size = get_file_size(path); fi.is_dir = false; pkg->is_single_file = true; parser_add_file_to_process(p, pkg, fi, pos); parser_add_package(p, pkg); return pkg; } Array list = {}; ReadDirectoryError rd_err = read_directory(path, &list); defer (array_free(&list)); if (list.count == 1) { GB_ASSERT(path != list[0].fullpath); } switch (rd_err) { case ReadDirectory_InvalidPath: syntax_error(pos, "Invalid path: %.*s", LIT(rel_path)); return nullptr; case ReadDirectory_NotExists: syntax_error(pos, "Path does not exist: %.*s", LIT(rel_path)); return nullptr; case ReadDirectory_Permission: syntax_error(pos, "Unknown error whilst reading path %.*s", LIT(rel_path)); return nullptr; case ReadDirectory_NotDir: syntax_error(pos, "Expected a directory for a package, got a file: %.*s", LIT(rel_path)); return nullptr; case ReadDirectory_Empty: syntax_error(pos, "Empty directory: %.*s", LIT(rel_path)); return nullptr; case ReadDirectory_Unknown: syntax_error(pos, "Unknown error whilst reading path %.*s", LIT(rel_path)); return nullptr; } for_array(list_index, list) { FileInfo fi = list[list_index]; String name = fi.name; String ext = path_extension(name); if (ext == FILE_EXT) { if (is_excluded_target_filename(name)) { continue; } parser_add_file_to_process(p, pkg, fi, pos); } else if (ext == ".S" || ext ==".s") { if (is_excluded_target_filename(name)) { continue; } parser_add_foreign_file_to_process(p, pkg, AstForeignFile_S, fi, pos); } } parser_add_package(p, pkg); return pkg; } gb_global Rune illegal_import_runes[] = { '"', '\'', '`', '\t', '\r', '\n', '\v', '\f', '\\', // NOTE(bill): Disallow windows style filepaths '!', '$', '%', '^', '&', '*', '(', ')', '=', '[', ']', '{', '}', ';', ':', // NOTE(bill): Disallow windows style absolute filepaths '#', '|', ',', '<', '>', '?', }; bool is_import_path_valid(String const &path) { if (path.len > 0) { u8 *start = path.text; u8 *end = path.text + path.len; u8 *curr = start; while (curr < end) { isize width = 1; Rune r = *curr; if (r >= 0x80) { width = utf8_decode(curr, end-curr, &r); if (r == GB_RUNE_INVALID && width == 1) { return false; } else if (r == GB_RUNE_BOM && curr-start > 0) { return false; } } for (isize i = 0; i < gb_count_of(illegal_import_runes); i++) { if (r == illegal_import_runes[i]) { return false; } } curr += width; } return true; } return false; } bool is_build_flag_path_valid(String const &path) { if (path.len > 0) { u8 *start = path.text; u8 *end = path.text + path.len; u8 *curr = start; isize index = 0; while (curr < end) { isize width = 1; Rune r = *curr; if (r >= 0x80) { width = utf8_decode(curr, end-curr, &r); if (r == GB_RUNE_INVALID && width == 1) { return false; } else if (r == GB_RUNE_BOM && curr-start > 0) { return false; } } for (isize i = 0; i < gb_count_of(illegal_import_runes); i++) { #if defined(GB_SYSTEM_WINDOWS) if (r == '\\') { break; } else if (r == ':') { break; } #endif if (r == illegal_import_runes[i]) { return false; } } curr += width; index += 1; } return true; } return false; } bool is_package_name_reserved(String const &name) { if (name == "builtin") { return true; } else if (name == "intrinsics") { return true; } return false; } bool determine_path_from_string(BlockingMutex *file_mutex, Ast *node, String base_dir, String const &original_string, String *path) { GB_ASSERT(path != nullptr); // NOTE(bill): if file_mutex == nullptr, this means that the code is used within the semantics stage gbAllocator a = heap_allocator(); String collection_name = {}; isize colon_pos = -1; for (isize j = 0; j < original_string.len; j++) { if (original_string[j] == ':') { colon_pos = j; break; } } bool has_windows_drive = false; #if defined(GB_SYSTEM_WINDOWS) if (file_mutex == nullptr) { if (colon_pos == 1 && original_string.len > 2) { if (original_string[2] == '/' || original_string[2] == '\\') { colon_pos = -1; has_windows_drive = true; } } } #endif String file_str = {}; if (colon_pos == 0) { syntax_error(node, "Expected a collection name"); return false; } if (original_string.len > 0 && colon_pos > 0) { collection_name = substring(original_string, 0, colon_pos); file_str = substring(original_string, colon_pos+1, original_string.len); } else { file_str = original_string; } if (has_windows_drive) { String sub_file_path = substring(file_str, 3, file_str.len); if (!is_import_path_valid(sub_file_path)) { syntax_error(node, "Invalid import path: '%.*s'", LIT(file_str)); return false; } } else if (!is_import_path_valid(file_str)) { syntax_error(node, "Invalid import path: '%.*s'", LIT(file_str)); return false; } if (collection_name.len > 0) { if (collection_name == "system") { if (node->kind != Ast_ForeignImportDecl) { syntax_error(node, "The library collection 'system' is restrict for 'foreign_library'"); return false; } else { *path = file_str; return true; } } else if (!find_library_collection_path(collection_name, &base_dir)) { // NOTE(bill): It's a naughty name syntax_error(node, "Unknown library collection: '%.*s'", LIT(collection_name)); return false; } } else { #if !defined(GB_SYSTEM_WINDOWS) // @NOTE(vassvik): foreign imports of shared libraries that are not in the system collection on // linux/mac have to be local to the executable for consistency with shared libraries. // Unix does not have a concept of "import library" for shared/dynamic libraries, // so we need to pass the relative path to the linker, and add the current // working directory of the exe to the library search paths. // Static libraries can be linked directly with the full pathname // if (node->kind == Ast_ForeignImportDecl && string_ends_with(file_str, str_lit(".so"))) { *path = file_str; return true; } #endif } if (is_package_name_reserved(file_str)) { *path = file_str; if (collection_name == "core") { return true; } else { syntax_error(node, "The package '%.*s' must be imported with the core library collection: 'core:%.*s'", LIT(file_str), LIT(file_str)); return false; } } if (file_mutex) mutex_lock(file_mutex); defer (if (file_mutex) mutex_unlock(file_mutex)); if (node->kind == Ast_ForeignImportDecl) { node->ForeignImportDecl.collection_name = collection_name; } if (has_windows_drive) { *path = file_str; } else { String fullpath = string_trim_whitespace(get_fullpath_relative(a, base_dir, file_str)); *path = fullpath; } return true; } void parse_setup_file_decls(Parser *p, AstFile *f, String const &base_dir, Slice &decls); void parse_setup_file_when_stmt(Parser *p, AstFile *f, String const &base_dir, AstWhenStmt *ws) { if (ws->body != nullptr) { auto stmts = ws->body->BlockStmt.stmts; parse_setup_file_decls(p, f, base_dir, stmts); } if (ws->else_stmt != nullptr) { switch (ws->else_stmt->kind) { case Ast_BlockStmt: { auto stmts = ws->else_stmt->BlockStmt.stmts; parse_setup_file_decls(p, f, base_dir, stmts); } break; case Ast_WhenStmt: parse_setup_file_when_stmt(p, f, base_dir, &ws->else_stmt->WhenStmt); break; } } } void parse_setup_file_decls(Parser *p, AstFile *f, String const &base_dir, Slice &decls) { for_array(i, decls) { Ast *node = decls[i]; if (!is_ast_decl(node) && node->kind != Ast_WhenStmt && node->kind != Ast_BadStmt && node->kind != Ast_EmptyStmt) { // NOTE(bill): Sanity check if (node->kind == Ast_ExprStmt) { Ast *expr = node->ExprStmt.expr; if (expr->kind == Ast_CallExpr && expr->CallExpr.proc->kind == Ast_BasicDirective) { f->directive_count += 1; continue; } } syntax_error(node, "Only declarations are allowed at file scope, got %.*s", LIT(ast_strings[node->kind])); } else if (node->kind == Ast_ImportDecl) { ast_node(id, ImportDecl, node); String original_string = string_trim_whitespace(string_value_from_token(f, id->relpath)); String import_path = {}; bool ok = determine_path_from_string(&p->file_decl_mutex, node, base_dir, original_string, &import_path); if (!ok) { decls[i] = ast_bad_decl(f, id->relpath, id->relpath); continue; } import_path = string_trim_whitespace(import_path); id->fullpath = import_path; if (is_package_name_reserved(import_path)) { continue; } try_add_import_path(p, import_path, original_string, ast_token(node).pos); } else if (node->kind == Ast_ForeignImportDecl) { ast_node(fl, ForeignImportDecl, node); auto fullpaths = array_make(permanent_allocator(), 0, fl->filepaths.count); for_array(fp_idx, fl->filepaths) { String file_str = string_trim_whitespace(string_value_from_token(f, fl->filepaths[fp_idx])); String fullpath = file_str; if (allow_check_foreign_filepath()) { String foreign_path = {}; bool ok = determine_path_from_string(&p->file_decl_mutex, node, base_dir, file_str, &foreign_path); if (!ok) { decls[i] = ast_bad_decl(f, fl->filepaths[fp_idx], fl->filepaths[fl->filepaths.count-1]); goto end; } fullpath = foreign_path; } array_add(&fullpaths, fullpath); } if (fullpaths.count == 0) { syntax_error(decls[i], "No foreign paths found"); decls[i] = ast_bad_decl(f, fl->filepaths[0], fl->filepaths[fl->filepaths.count-1]); goto end; } fl->fullpaths = slice_from_array(fullpaths); } else if (node->kind == Ast_WhenStmt) { ast_node(ws, WhenStmt, node); parse_setup_file_when_stmt(p, f, base_dir, ws); } end:; } } String build_tag_get_token(String s, String *out) { s = string_trim_whitespace(s); isize n = 0; while (n < s.len) { Rune rune = 0; isize width = utf8_decode(&s[n], s.len-n, &rune); if (n == 0 && rune == '!') { } else if (!rune_is_letter(rune) && !rune_is_digit(rune)) { isize k = gb_max(gb_max(n, width), 1); *out = substring(s, k, s.len); return substring(s, 0, k); } n += width; } out->len = 0; return s; } bool parse_build_tag(Token token_for_pos, String s) { String const prefix = str_lit("+build"); GB_ASSERT(string_starts_with(s, prefix)); s = string_trim_whitespace(substring(s, prefix.len, s.len)); if (s.len == 0) { return true; } bool any_correct = false; while (s.len > 0) { bool this_kind_correct = true; do { String p = string_trim_whitespace(build_tag_get_token(s, &s)); if (p.len == 0) break; if (p == ",") break; bool is_notted = false; if (p[0] == '!') { is_notted = true; p = substring(p, 1, p.len); if (p.len == 0) { syntax_error(token_for_pos, "Expected a build platform after '!'"); break; } } if (p.len == 0) { continue; } if (p == "ignore") { this_kind_correct = false; continue; } TargetOsKind os = get_target_os_from_string(p); TargetArchKind arch = get_target_arch_from_string(p); if (os != TargetOs_Invalid) { GB_ASSERT(arch == TargetArch_Invalid); if (is_notted) { this_kind_correct = this_kind_correct && (os != build_context.metrics.os); } else { this_kind_correct = this_kind_correct && (os == build_context.metrics.os); } } else if (arch != TargetArch_Invalid) { if (is_notted) { this_kind_correct = this_kind_correct && (arch != build_context.metrics.arch); } else { this_kind_correct = this_kind_correct && (arch == build_context.metrics.arch); } } if (os == TargetOs_Invalid && arch == TargetArch_Invalid) { syntax_error(token_for_pos, "Invalid build tag platform: %.*s", LIT(p)); break; } } while (s.len > 0); any_correct = any_correct || this_kind_correct; } return any_correct; } String dir_from_path(String path) { String base_dir = path; for (isize i = path.len-1; i >= 0; i--) { if (base_dir[i] == '\\' || base_dir[i] == '/') { break; } base_dir.len--; } return base_dir; } isize calc_decl_count(Ast *decl) { isize count = 0; switch (decl->kind) { case Ast_BlockStmt: for_array(i, decl->BlockStmt.stmts) { count += calc_decl_count(decl->BlockStmt.stmts.data[i]); } break; case Ast_WhenStmt: { isize inner_count = calc_decl_count(decl->WhenStmt.body); if (decl->WhenStmt.else_stmt) { inner_count = gb_max(inner_count, calc_decl_count(decl->WhenStmt.else_stmt)); } count += inner_count; } break; case Ast_ValueDecl: count = decl->ValueDecl.names.count; break; case Ast_ForeignBlockDecl: count = calc_decl_count(decl->ForeignBlockDecl.body); break; case Ast_ImportDecl: case Ast_ForeignImportDecl: count = 1; break; } return count; } bool parse_build_project_directory_tag(Token token_for_pos, String s) { String const prefix = str_lit("+build-project-name"); GB_ASSERT(string_starts_with(s, prefix)); s = string_trim_whitespace(substring(s, prefix.len, s.len)); if (s.len == 0) { return true; } bool any_correct = false; while (s.len > 0) { bool this_kind_correct = true; do { String p = string_trim_whitespace(build_tag_get_token(s, &s)); if (p.len == 0) break; if (p == ",") break; bool is_notted = false; if (p[0] == '!') { is_notted = true; p = substring(p, 1, p.len); if (p.len == 0) { syntax_error(token_for_pos, "Expected a build-project-name after '!'"); break; } } if (p.len == 0) { continue; } if (is_notted) { this_kind_correct = this_kind_correct && (p != build_context.ODIN_BUILD_PROJECT_NAME); } else { this_kind_correct = this_kind_correct && (p == build_context.ODIN_BUILD_PROJECT_NAME); } } while (s.len > 0); any_correct = any_correct || this_kind_correct; } return any_correct; } bool parse_file(Parser *p, AstFile *f) { if (f->tokens.count == 0) { return true; } if (f->tokens.count > 0 && f->tokens[0].kind == Token_EOF) { return true; } u64 start = time_stamp_time_now(); String filepath = f->tokenizer.fullpath; String base_dir = dir_from_path(filepath); if (f->curr_token.kind == Token_Comment) { consume_comment_groups(f, f->prev_token); } CommentGroup *docs = f->lead_comment; if (f->curr_token.kind != Token_package) { syntax_error(f->curr_token, "Expected a package declaration at the beginning of the file"); return false; } f->package_token = expect_token(f, Token_package); if (f->package_token.kind != Token_package) { return false; } if (docs != nullptr) { TokenPos end = token_pos_end(docs->list[docs->list.count-1]); if (end.line == f->package_token.pos.line || end.line+1 == f->package_token.pos.line) { // Okay } else { docs = nullptr; } } Token package_name = expect_token_after(f, Token_Ident, "package"); if (package_name.kind == Token_Ident) { if (package_name.string == "_") { syntax_error(package_name, "Invalid package name '_'"); } else if (f->pkg->kind != Package_Runtime && package_name.string == "runtime") { syntax_error(package_name, "Use of reserved package name '%.*s'", LIT(package_name.string)); } else if (is_package_name_reserved(package_name.string)) { syntax_error(package_name, "Use of reserved package name '%.*s'", LIT(package_name.string)); } } f->package_name = package_name.string; if (!f->pkg->is_single_file && docs != nullptr && docs->list.count > 0) { for_array(i, docs->list) { Token tok = docs->list[i]; GB_ASSERT(tok.kind == Token_Comment); String str = tok.string; if (string_starts_with(str, str_lit("//"))) { String lc = string_trim_whitespace(substring(str, 2, str.len)); if (lc.len > 0 && lc[0] == '+') { if (string_starts_with(lc, str_lit("+build-project-name"))) { if (!parse_build_project_directory_tag(tok, lc)) { return false; } } else if (string_starts_with(lc, str_lit("+build"))) { if (!parse_build_tag(tok, lc)) { return false; } } else if (string_starts_with(lc, str_lit("+private"))) { f->flags |= AstFile_IsPrivatePkg; String command = string_trim_starts_with(lc, str_lit("+private ")); command = string_trim_whitespace(command); if (lc == "+private") { f->flags |= AstFile_IsPrivatePkg; } else if (command == "package") { f->flags |= AstFile_IsPrivatePkg; } else if (command == "file") { f->flags |= AstFile_IsPrivateFile; } } else if (lc == "+lazy") { if (build_context.ignore_lazy) { // Ignore } else if (f->flags & AstFile_IsTest) { // Ignore } else if (f->pkg->kind == Package_Init && build_context.command_kind == Command_doc) { // Ignore } else { f->flags |= AstFile_IsLazy; } } } } } } Ast *pd = ast_package_decl(f, f->package_token, package_name, docs, f->line_comment); expect_semicolon(f); f->pkg_decl = pd; if (f->error_count == 0) { auto decls = array_make(heap_allocator()); while (f->curr_token.kind != Token_EOF) { Ast *stmt = parse_stmt(f); if (stmt && stmt->kind != Ast_EmptyStmt) { array_add(&decls, stmt); if (stmt->kind == Ast_ExprStmt && stmt->ExprStmt.expr != nullptr && stmt->ExprStmt.expr->kind == Ast_ProcLit) { syntax_error(stmt, "Procedure literal evaluated but not used"); } f->total_file_decl_count += calc_decl_count(stmt); if (stmt->kind == Ast_WhenStmt || stmt->kind == Ast_ExprStmt || stmt->kind == Ast_ImportDecl) { f->delayed_decl_count += 1; } } } f->decls = slice_from_array(decls); parse_setup_file_decls(p, f, base_dir, f->decls); } u64 end = time_stamp_time_now(); f->time_to_parse = cast(f64)(end-start)/cast(f64)time_stamp__freq(); for (int i = 0; i < AstDelayQueue_COUNT; i++) { mpmc_init(f->delayed_decls_queues+i, heap_allocator(), f->delayed_decl_count); } return f->error_count == 0; } ParseFileError process_imported_file(Parser *p, ImportedFile imported_file) { AstPackage *pkg = imported_file.pkg; FileInfo fi = imported_file.fi; TokenPos pos = imported_file.pos; AstFile *file = gb_alloc_item(permanent_allocator(), AstFile); file->pkg = pkg; file->id = cast(i32)(imported_file.index+1); TokenPos err_pos = {0}; ParseFileError err = init_ast_file(file, fi.fullpath, &err_pos); err_pos.file_id = file->id; file->last_error = err; if (err != ParseFile_None) { if (err == ParseFile_EmptyFile) { if (fi.fullpath == p->init_fullpath) { syntax_error(pos, "Initial file is empty - %.*s\n", LIT(p->init_fullpath)); gb_exit(1); } } else { switch (err) { case ParseFile_WrongExtension: syntax_error(pos, "Failed to parse file: %.*s; invalid file extension: File must have the extension '.odin'", LIT(fi.name)); break; case ParseFile_InvalidFile: syntax_error(pos, "Failed to parse file: %.*s; invalid file or cannot be found", LIT(fi.name)); break; case ParseFile_Permission: syntax_error(pos, "Failed to parse file: %.*s; file permissions problem", LIT(fi.name)); break; case ParseFile_NotFound: syntax_error(pos, "Failed to parse file: %.*s; file cannot be found ('%.*s')", LIT(fi.name), LIT(fi.fullpath)); break; case ParseFile_InvalidToken: syntax_error(err_pos, "Failed to parse file: %.*s; invalid token found in file", LIT(fi.name)); break; case ParseFile_EmptyFile: syntax_error(pos, "Failed to parse file: %.*s; file contains no tokens", LIT(fi.name)); break; case ParseFile_FileTooLarge: syntax_error(pos, "Failed to parse file: %.*s; file is too large, exceeds maximum file size of 2 GiB", LIT(fi.name)); break; } return err; } } if (build_context.command_kind == Command_test) { String name = file->fullpath; name = remove_extension_from_path(name); String test_suffix = str_lit("_test"); if (string_ends_with(name, test_suffix) && name != test_suffix) { file->flags |= AstFile_IsTest; } } if (parse_file(p, file)) { mutex_lock(&p->file_add_mutex); defer (mutex_unlock(&p->file_add_mutex)); array_add(&pkg->files, file); if (pkg->name.len == 0) { pkg->name = file->package_name; } else if (pkg->name != file->package_name) { if (file->tokens.count > 0 && file->tokens[0].kind != Token_EOF) { Token tok = file->package_token; tok.pos.file_id = file->id; tok.pos.line = gb_max(tok.pos.line, 1); tok.pos.column = gb_max(tok.pos.column, 1); syntax_error(tok, "Different package name, expected '%.*s', got '%.*s'", LIT(pkg->name), LIT(file->package_name)); } } p->total_line_count += file->tokenizer.line_count; p->total_token_count += file->tokens.count; } return ParseFile_None; } ParseFileError parse_packages(Parser *p, String init_filename) { GB_ASSERT(init_filename.text[init_filename.len] == 0); String init_fullpath = path_to_full_path(heap_allocator(), init_filename); if (!path_is_directory(init_fullpath)) { String const ext = str_lit(".odin"); if (!string_ends_with(init_fullpath, ext)) { error_line("Expected either a directory or a .odin file, got '%.*s'\n", LIT(init_filename)); return ParseFile_WrongExtension; } } else if (init_fullpath.len != 0) { String path = init_fullpath; if (path[path.len-1] == '/') { path.len -= 1; } if ((build_context.command_kind & Command__does_build) && build_context.build_mode == BuildMode_Executable) { String short_path = filename_from_path(path); char *cpath = alloc_cstring(heap_allocator(), short_path); defer (gb_free(heap_allocator(), cpath)); if (gb_file_exists(cpath)) { error_line("Please specify the executable name with -out: as a directory exists with the same name in the current working directory"); return ParseFile_DirectoryAlreadyExists; } } } { // Add these packages serially and then process them parallel mutex_lock(&p->wait_mutex); defer (mutex_unlock(&p->wait_mutex)); TokenPos init_pos = {}; { String s = get_fullpath_core(heap_allocator(), str_lit("runtime")); try_add_import_path(p, s, s, init_pos, Package_Runtime); } try_add_import_path(p, init_fullpath, init_fullpath, init_pos, Package_Init); p->init_fullpath = init_fullpath; if (build_context.command_kind == Command_test) { String s = get_fullpath_core(heap_allocator(), str_lit("testing")); try_add_import_path(p, s, s, init_pos, Package_Normal); } for_array(i, build_context.extra_packages) { String path = build_context.extra_packages[i]; String fullpath = path_to_full_path(heap_allocator(), path); // LEAK? if (!path_is_directory(fullpath)) { String const ext = str_lit(".odin"); if (!string_ends_with(fullpath, ext)) { error_line("Expected either a directory or a .odin file, got '%.*s'\n", LIT(fullpath)); return ParseFile_WrongExtension; } } AstPackage *pkg = try_add_import_path(p, fullpath, fullpath, init_pos, Package_Normal); if (pkg) { pkg->is_extra = true; } } } global_thread_pool_wait(); for (ParseFileError err = ParseFile_None; mpmc_dequeue(&p->file_error_queue, &err); /**/) { if (err != ParseFile_None) { return err; } } for (isize i = p->packages.count-1; i >= 0; i--) { AstPackage *pkg = p->packages[i]; for (isize j = pkg->files.count-1; j >= 0; j--) { AstFile *file = pkg->files[j]; if (file->error_count != 0) { if (file->last_error != ParseFile_None) { return file->last_error; } return ParseFile_GeneralError; } } } return ParseFile_None; } #undef ALLOW_NEWLINE