reading the chumsky docs helps actually, wild /s

This commit is contained in:
LunarAkai 2025-08-07 02:29:37 +02:00
commit 2810913aae
3 changed files with 39 additions and 72 deletions

View file

@ -1,3 +1,4 @@
use chumsky::input::{Input, Stream};
use chumsky::Parser;
use logos::Logos;
@ -18,20 +19,27 @@ Simple Compiler -> 4 Stages:
fn main() {
let sourcecode = std::fs::read_to_string("sample.akai").unwrap();
let lexer = Token::lexer(&sourcecode);
let mut tokens = vec![];
for (token, span) in lexer.spanned() {
match token {
Ok(token) => tokens.push(token),
Err(e) => {
println!("lexer error at {:?}: {:?}", span, e);
return;
}
}
}
// Create a logos lexer over the source code
let token_iter = Token::lexer(&sourcecode)
.spanned()
// Convert logos errors into tokens. We want parsing to be recoverable and not fail at the lexing stage, so
// we have a dedicated `Token::Error` variant that represents a token error that was previously encountered
.map(|(tok, span)| match tok {
// Turn the `Range<usize>` spans logos gives us into chumsky's `SimpleSpan` via `Into`, because it's easier
// to work with
Ok(tok) => (tok, span.into()),
Err(()) => (Token::Error, span.into()),
});
match parser().parse(&tokens).into_result() {
// Turn the token iterator into a stream that chumsky can use for things like backtracking
let token_stream = Stream::from_iter(token_iter)
// Tell chumsky to split the (Token, SimpleSpan) stream into its parts so that it can handle the spans for us
// This involves giving chumsky an 'end of input' span: we just use a zero-width span at the end of the string
.map((0..sourcecode.len()).into(), |(t, s): (_, _)| (t, s));
match parser().parse(token_stream).into_result() {
Ok(ast) => match eval(&ast, &mut Vec::new(), &mut Vec::new()) {
Ok(output) => println!("{output}"),
Err(eval_err) => println!("Evaluation error: {eval_err}"),