Compare commits
4 commits
b0e123f83a
...
4f7478d7b2
Author | SHA1 | Date | |
---|---|---|---|
4f7478d7b2 | |||
e2438f143b | |||
b591ecdf9b | |||
ac23c0d2d7 |
3 changed files with 67 additions and 48 deletions
|
@ -10,7 +10,7 @@ pub const NodeType = enum {
|
||||||
};
|
};
|
||||||
|
|
||||||
pub const FnDecl = struct {
|
pub const FnDecl = struct {
|
||||||
func_name: []const u8,
|
func_name: Token,
|
||||||
};
|
};
|
||||||
|
|
||||||
pub const Node = union(NodeType) {
|
pub const Node = union(NodeType) {
|
||||||
|
|
102
src/parser.zig
102
src/parser.zig
|
@ -17,7 +17,6 @@ pub const Parser = struct {
|
||||||
scanner: *Scanner,
|
scanner: *Scanner,
|
||||||
|
|
||||||
tokens: []Token = undefined,
|
tokens: []Token = undefined,
|
||||||
current: usize = 0,
|
|
||||||
|
|
||||||
pub fn init(allocator: *Allocator, scanner: *Scanner) Parser {
|
pub fn init(allocator: *Allocator, scanner: *Scanner) Parser {
|
||||||
return Parser{ .allocator = allocator, .scanner = scanner };
|
return Parser{ .allocator = allocator, .scanner = scanner };
|
||||||
|
@ -32,11 +31,11 @@ pub const Parser = struct {
|
||||||
}
|
}
|
||||||
|
|
||||||
fn peek(self: *Parser) Token {
|
fn peek(self: *Parser) Token {
|
||||||
return self.tokens[self.current];
|
return self.tokens[self.tokens.len - 1];
|
||||||
}
|
}
|
||||||
|
|
||||||
fn previous(self: *Parser) Token {
|
fn previous(self: *Parser) Token {
|
||||||
return self.tokens[self.current - 1];
|
return self.tokens[self.tokens.len - 2];
|
||||||
}
|
}
|
||||||
|
|
||||||
fn tokenError(self: *Parser, token: Token, msg: []const u8) Result!void {
|
fn tokenError(self: *Parser, token: Token, msg: []const u8) Result!void {
|
||||||
|
@ -53,85 +52,98 @@ pub const Parser = struct {
|
||||||
return self.peek().ttype == .EOF;
|
return self.peek().ttype == .EOF;
|
||||||
}
|
}
|
||||||
|
|
||||||
fn advance(self: *Parser) Token {
|
|
||||||
if (!self.isAtEnd()) self.current += 1;
|
|
||||||
return self.previous();
|
|
||||||
}
|
|
||||||
|
|
||||||
fn check(self: *Parser, ttype: TokenType) bool {
|
fn check(self: *Parser, ttype: TokenType) bool {
|
||||||
if (self.isAtEnd()) return false;
|
if (self.isAtEnd()) return false;
|
||||||
return self.peek().ttype == ttype;
|
return self.peek().ttype == ttype;
|
||||||
}
|
}
|
||||||
|
|
||||||
fn match(self: *Parser, ttypes: []TokenType) bool {
|
fn nextToken(self: *Parser) !Token {
|
||||||
for (ttypes) |ttype| {
|
var token: Token = undefined;
|
||||||
if (self.check(ttype)) {
|
|
||||||
_ = self.advance();
|
while (true) {
|
||||||
return true;
|
var next_token_opt = try self.scanner.nextToken();
|
||||||
|
if (next_token_opt) |token_nice| {
|
||||||
|
token = token_nice;
|
||||||
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return false;
|
self.tokens = try self.allocator.realloc(
|
||||||
|
self.tokens,
|
||||||
|
self.tokens.len + 1,
|
||||||
|
);
|
||||||
|
|
||||||
|
self.tokens[self.tokens.len - 1] = token;
|
||||||
|
return token;
|
||||||
}
|
}
|
||||||
|
|
||||||
fn matchSingle(self: *Parser, ttype: TokenType) bool {
|
fn consume(self: *Parser, ttype: TokenType, comptime msg: []const u8) !Token {
|
||||||
if (self.check(ttype)) {
|
if (self.check(ttype)) return try self.nextToken();
|
||||||
_ = self.advance();
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
fn consume(self: *Parser, ttype: TokenType, comptime msg: []const u8) Result!Token {
|
|
||||||
if (self.check(ttype)) return self.advance();
|
|
||||||
|
|
||||||
try self.tokenError(self.peek(), msg);
|
try self.tokenError(self.peek(), msg);
|
||||||
return Result.CompileError;
|
return Result.CompileError;
|
||||||
}
|
}
|
||||||
|
|
||||||
fn mkFnDecl(self: *Parser, name: []const u8) !*ast.Node {
|
fn consumeSingle(self: *Parser, ttype: TokenType) !Token {
|
||||||
var node = try self.allocator.create(Node.FnDecl);
|
std.debug.warn("consume {}, has {}\n", ttype, self.peek().ttype);
|
||||||
node.* = Node.FnDecl{ .name = name };
|
|
||||||
|
if (self.check(ttype)) {
|
||||||
|
var cur = self.peek();
|
||||||
|
_ = try self.nextToken();
|
||||||
|
std.debug.warn("now has {}\n", self.peek());
|
||||||
|
return cur;
|
||||||
|
}
|
||||||
|
|
||||||
|
var buf_main: [1000]u8 = undefined;
|
||||||
|
var buf = try std.fmt.bufPrint(
|
||||||
|
buf_main[0..],
|
||||||
|
"expected {}, got {}",
|
||||||
|
ttype,
|
||||||
|
self.peek().ttype,
|
||||||
|
);
|
||||||
|
try self.tokenError(self.peek(), buf);
|
||||||
|
|
||||||
|
return Result.CompileError;
|
||||||
|
}
|
||||||
|
|
||||||
|
fn mkFnDecl(self: *Parser, name: Token) !*ast.Node {
|
||||||
|
var node = try self.allocator.create(Node);
|
||||||
|
node.* = Node{ .FnDecl = ast.FnDecl{ .func_name = name } };
|
||||||
return node;
|
return node;
|
||||||
}
|
}
|
||||||
|
|
||||||
fn functionDecl(self: *Parser) !*ast.Node {
|
fn functionDecl(self: *Parser) !*ast.Node {
|
||||||
// get the name
|
_ = try self.consumeSingle(.Fn);
|
||||||
var name = try self.consume(.Identifier, "expected function name");
|
var name = try self.consumeSingle(.Identifier);
|
||||||
|
_ = try self.consumeSingle(.LeftParen);
|
||||||
|
|
||||||
return try self.mkFnDecl(name);
|
return try self.mkFnDecl(name);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn processToken(self: *Parser, token: Token) Result!ast.Node {
|
fn processToken(self: *Parser, token: Token) !*ast.Node {
|
||||||
switch (token.ttype) {
|
var node = switch (token.ttype) {
|
||||||
//.Fn => try self.functionDecl(),
|
.Fn => try self.functionDecl(),
|
||||||
else => blk: {
|
else => blk: {
|
||||||
try self.doError("TODO handle {}\n", token.ttype);
|
try self.doError("TODO handle {}\n", token.ttype);
|
||||||
return Result.CompileError;
|
return Result.CompileError;
|
||||||
},
|
},
|
||||||
}
|
};
|
||||||
|
|
||||||
|
return node;
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn parse(self: *Parser) !*ast.Node {
|
pub fn parse(self: *Parser) !*ast.Node {
|
||||||
self.tokens = try self.allocator.alloc(Token, 0);
|
self.tokens = try self.allocator.alloc(Token, 0);
|
||||||
var i: usize = 0;
|
|
||||||
var root = try ast.mkRoot(self.allocator);
|
var root = try ast.mkRoot(self.allocator);
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
var tok_opt = try self.scanner.nextToken();
|
var token = try self.nextToken();
|
||||||
|
|
||||||
if (tok_opt) |token| {
|
|
||||||
self.tokens = try self.allocator.realloc(self.tokens, i + 1);
|
|
||||||
self.tokens[i] = token;
|
|
||||||
i += 1;
|
|
||||||
|
|
||||||
if (token.ttype == .EOF) break;
|
if (token.ttype == .EOF) break;
|
||||||
|
|
||||||
var node = try self.processToken(token);
|
var node = try self.processToken(token);
|
||||||
try root.Root.append(&node);
|
std.debug.warn("{}\n", node.*);
|
||||||
} else {
|
try root.Root.append(node);
|
||||||
continue;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return root;
|
return root;
|
||||||
|
|
|
@ -110,6 +110,10 @@ pub const Scanner = struct {
|
||||||
return self.source[self.current - 1];
|
return self.source[self.current - 1];
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn rollback(self: *Scanner) void {
|
||||||
|
self.current -= 1;
|
||||||
|
}
|
||||||
|
|
||||||
pub fn currentLexeme(self: *Scanner) []const u8 {
|
pub fn currentLexeme(self: *Scanner) []const u8 {
|
||||||
return self.source[self.start..self.current];
|
return self.source[self.start..self.current];
|
||||||
}
|
}
|
||||||
|
@ -245,6 +249,9 @@ pub const Scanner = struct {
|
||||||
_ = self.advance();
|
_ = self.advance();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ugly hack.
|
||||||
|
self.rollback();
|
||||||
|
|
||||||
// after reading the identifier, we check
|
// after reading the identifier, we check
|
||||||
// if it is any of our keywords, if it is, then we add
|
// if it is any of our keywords, if it is, then we add
|
||||||
// the specificed keyword type. if not, just .IDENTIFIER
|
// the specificed keyword type. if not, just .IDENTIFIER
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue