formatted

This commit is contained in:
LunarAkai 2025-08-09 17:39:37 +02:00
commit 50526960ba
6 changed files with 101 additions and 80 deletions

View file

@ -2,4 +2,10 @@ run:
cargo run cargo run
test: test:
cargo test --verbose cargo test --verbose
doc:
cargo doc --no-deps
fmt:
cargo fmt

View file

@ -28,4 +28,4 @@ pub enum Expr {
FunctionExpr(Function), FunctionExpr(Function),
Error, Error,
} }

View file

@ -7,10 +7,9 @@ pub type BlockStatement = Vec<Statement>;
pub type Span = Range<usize>; pub type Span = Range<usize>;
#[derive(Clone, Debug, PartialEq)] #[derive(Clone, Debug, PartialEq)]
pub enum Statement { pub enum Statement {
Var(Ident, Option<Type>) Var(Ident, Option<Type>),
} }
#[derive(Clone, Debug, PartialEq)] #[derive(Clone, Debug, PartialEq)]
@ -79,7 +78,6 @@ impl Value {
} }
} }
//--------------------------------------- //---------------------------------------
// Structs // Structs
//--------------------------------------- //---------------------------------------
@ -99,8 +97,7 @@ pub struct Condition {
pub else_body: Option<BlockStatement>, pub else_body: Option<BlockStatement>,
} }
/// Example: `x = 5`
/// Example: `x = 5`
#[derive(Clone, Debug, PartialEq)] #[derive(Clone, Debug, PartialEq)]
pub struct Assignment { pub struct Assignment {
pub target: Box<Expr>, pub target: Box<Expr>,
@ -112,7 +109,7 @@ pub struct Assignment {
pub struct Var { pub struct Var {
pub ty: Option<Type>, pub ty: Option<Type>,
pub ident: String, pub ident: String,
pub value: Box<Expr> pub value: Box<Expr>,
} }
/// Example: `x++` /// Example: `x++`
@ -131,30 +128,29 @@ pub struct Binary {
} }
/// Represents the Structure of a `Function` in AkaiLang /// Represents the Structure of a `Function` in AkaiLang
/// ///
/// Examples: /// Examples:
///```AkaiLang ///```AkaiLang
///fun helloWorld() { ///fun helloWorld() {
/// print("Hello World") /// print("Hello World")
///} ///}
///``` ///```
/// <br> /// <br>
/// ///
///```AkaiLang ///```AkaiLang
///fun returnsIntPlusOne(i: int): int { ///fun returnsIntPlusOne(i: int): int {
/// -> i + 1 /// -> i + 1
///} ///}
///``` ///```
#[derive(Clone, Debug, PartialEq)] #[derive(Clone, Debug, PartialEq)]
pub struct Function { pub struct Function {
pub name: Rc<str>, pub name: Rc<str>,
pub params: Vec<(Ident, Type)>, pub params: Vec<(Ident, Type)>,
pub return_type: Option<Type>, pub return_type: Option<Type>,
pub body: Vec<Statement>, pub body: Vec<Statement>,
pub body_expr: Option<Type> pub body_expr: Option<Type>,
} }
#[derive(Clone, Debug, PartialEq)] #[derive(Clone, Debug, PartialEq)]
pub struct Call { pub struct Call {
pub callee: Box<Expr>, pub callee: Box<Expr>,

View file

@ -1,31 +1,50 @@
use chumsky::{ use chumsky::{
combinator::Or, error::Rich, extra, input::{Input, Stream, ValueInput}, prelude::{choice, end, just, nested_delimiters, recursive, skip_then_retry_until, via_parser}, primitive::select, recursive, select, select_ref, span::{self, SimpleSpan}, text::{self, ascii::{ident, keyword}, newline, whitespace}, Boxed, ConfigIterParser, IterParser, ParseResult, Parser Boxed, ConfigIterParser, IterParser, ParseResult, Parser,
combinator::Or,
error::Rich,
extra,
input::{Input, Stream, ValueInput},
prelude::{choice, end, just, nested_delimiters, recursive, skip_then_retry_until, via_parser},
primitive::select,
recursive, select, select_ref,
span::{self, SimpleSpan},
text::{
self,
ascii::{ident, keyword},
newline, whitespace,
},
}; };
use logos::{source, Logos}; use logos::{Logos, source};
use crate::language_frontend::{abstract_syntax_tree::{ast::Expr, definitions::*}, lexer::tokens::Token}; use crate::language_frontend::{
abstract_syntax_tree::{ast::Expr, definitions::*},
lexer::tokens::Token,
};
// goal of parsing is to construct an abstract syntax tree // goal of parsing is to construct an abstract syntax tree
pub fn parse(source: &str) ->Result<Vec<Expr>, Vec<Rich<'_, Token>>> { pub fn parse(source: &str) -> Result<Vec<Expr>, Vec<Rich<'_, Token>>> {
let token_iter = Token::lexer(source).spanned().map(|(token, span)| (token.unwrap_or(Token::Error), span.into())); let token_iter = Token::lexer(source)
.spanned()
.map(|(token, span)| (token.unwrap_or(Token::Error), span.into()));
let end_of_input: SimpleSpan = (0..source.len()).into(); let end_of_input: SimpleSpan = (0..source.len()).into();
let token_stream = Stream::from_iter(token_iter) let token_stream = Stream::from_iter(token_iter)
// Tell chumsky to split the (Token, SimpleSpan) stream into its parts so that it can handle the spans for us // Tell chumsky to split the (Token, SimpleSpan) stream into its parts so that it can handle the spans for us
// This involves giving chumsky an 'end of input' span: we just use a zero-width span at the end of the string // This involves giving chumsky an 'end of input' span: we just use a zero-width span at the end of the string
.map((0..end_of_input.into_iter().len()).into(), |(t, s): (_, _)| (t, s)); .map(
(0..end_of_input.into_iter().len()).into(),
|(t, s): (_, _)| (t, s),
);
parser().parse(token_stream).into_result() parser().parse(token_stream).into_result()
} }
fn parser<'src, I>() -> impl Parser<'src, I, Vec<Expr>, extra::Err<Rich<'src, Token>>>
fn parser<'src, I>() where
-> impl Parser<'src, I, Vec<Expr>, extra::Err<Rich<'src, Token>>>
where
I: ValueInput<'src, Token = Token, Span = SimpleSpan>, I: ValueInput<'src, Token = Token, Span = SimpleSpan>,
{ {
let ident = select! { Token::Identifier(s) => s, }; let ident = select! { Token::Identifier(s) => s, };
/* /*
let block = recursive(|block| { let block = recursive(|block| {
let indent = just(Token::NewLine) let indent = just(Token::NewLine)
.ignore_then(just(Token::Indent)) .ignore_then(just(Token::Indent))
@ -41,8 +60,9 @@ where
Token::FloatLiteral(x) => Expr::FloatLiteral(x), Token::FloatLiteral(x) => Expr::FloatLiteral(x),
Token::IntLiteral(x) => Expr::IntLiteral(x), Token::IntLiteral(x) => Expr::IntLiteral(x),
} }
.or(expr.clone().delimited_by(just(Token::LParen), just(Token::RParen))); .or(expr
.clone()
.delimited_by(just(Token::LParen), just(Token::RParen)));
let mul_div = atom.clone().foldl( let mul_div = atom.clone().foldl(
choice(( choice((
@ -52,12 +72,14 @@ where
.then(atom) .then(atom)
.then_ignore(just(Token::NewLine).or_not()) .then_ignore(just(Token::NewLine).or_not())
.repeated(), .repeated(),
|lhs, (op, rhs)| Expr::BinaryExpr ( Binary { |lhs, (op, rhs)| {
lhs: Box::new(lhs), Expr::BinaryExpr(Binary {
operator: op, lhs: Box::new(lhs),
rhs: Box::new(rhs), operator: op,
rhs: Box::new(rhs),
})
}, },
)); );
let add_sub = mul_div.clone().foldl( let add_sub = mul_div.clone().foldl(
choice(( choice((
@ -67,12 +89,14 @@ where
.then(mul_div) .then(mul_div)
.then_ignore(just(Token::NewLine).or_not()) .then_ignore(just(Token::NewLine).or_not())
.repeated(), .repeated(),
|lhs, (op, rhs)| Expr::BinaryExpr ( Binary { |lhs, (op, rhs)| {
lhs: Box::new(lhs), Expr::BinaryExpr(Binary {
operator: op, lhs: Box::new(lhs),
rhs: Box::new(rhs), operator: op,
rhs: Box::new(rhs),
})
}, },
)); );
add_sub add_sub
}); });
@ -82,35 +106,35 @@ where
.then_ignore(just(Token::Assign)) .then_ignore(just(Token::Assign))
.then(expr.clone()) .then(expr.clone())
.then_ignore(just(Token::NewLine).or_not()) .then_ignore(just(Token::NewLine).or_not())
.map(|(name, rhs)| Expr::VarExpr ( Var { .map(|(name, rhs)| {
ty: None, Expr::VarExpr(Var {
ident: name, ty: None,
value: Box::new(rhs), ident: name,
} value: Box::new(rhs),
)); })
});
/* /*
let fun = just(Token::Fun) let fun = just(Token::Fun)
.ignore_then(ident.clone()) .ignore_then(ident.clone())
.then_ignore(just(Token::LParen)) .then_ignore(just(Token::LParen))
.then_ignore(just(Token::RParen)) .then_ignore(just(Token::RParen))
.map(|(((name, args), ret), (body, body_expr))| Expr::Function(Function { .map(|(((name, args), ret), (body, body_expr))| Expr::Function(Function {
name, name,
params: args, params: args,
return_type: ret, return_type: ret,
body, body,
body_expr body_expr
})); */ })); */
var.or(expr) var.or(expr)
}); });
decl.repeated().collect() decl.repeated().collect()
} }
#[cfg(test)] #[cfg(test)]
mod tests { mod tests {
use super::*; use super::*;
#[test] #[test]
fn test_unary_expr() { fn test_unary_expr() {
let negate_two = parse("-2"); let negate_two = parse("-2");
@ -118,7 +142,7 @@ mod tests {
assert_eq!( assert_eq!(
negate_two.clone().unwrap(), negate_two.clone().unwrap(),
vec![Expr::UnaryExpr(Unary { vec![Expr::UnaryExpr(Unary {
operator: UnaryOp::Minus, operator: UnaryOp::Minus,
operand: Box::new(Expr::IntLiteral(2)), operand: Box::new(Expr::IntLiteral(2)),
})] })]
) )
@ -130,9 +154,9 @@ mod tests {
assert!(sum.is_ok()); assert!(sum.is_ok());
assert_eq!( assert_eq!(
sum.clone().unwrap(), sum.clone().unwrap(),
vec![Expr::BinaryExpr(Binary { vec![Expr::BinaryExpr(Binary {
lhs: Box::new(Expr::IntLiteral(1)), lhs: Box::new(Expr::IntLiteral(1)),
operator: BinaryOp::Add, operator: BinaryOp::Add,
rhs: Box::new(Expr::IntLiteral(2)) rhs: Box::new(Expr::IntLiteral(2))
})] })]
) )
@ -144,11 +168,11 @@ mod tests {
assert!(var_without_expl_type.is_ok()); assert!(var_without_expl_type.is_ok());
assert_eq!( assert_eq!(
var_without_expl_type.clone().unwrap(), var_without_expl_type.clone().unwrap(),
vec![Expr::VarExpr(Var { vec![Expr::VarExpr(Var {
ty: None, ty: None,
ident: String::from("x"), ident: String::from("x"),
value: Box::new(Expr::IntLiteral(12)) value: Box::new(Expr::IntLiteral(12))
})] })]
) )
} }
} }

View file

@ -1,6 +1,6 @@
use std::fmt; use std::fmt;
use logos::{Logos}; use logos::Logos;
#[derive(Logos, Debug, Clone, PartialEq)] #[derive(Logos, Debug, Clone, PartialEq)]
#[regex(r"[\t\f]+", logos::skip)] #[regex(r"[\t\f]+", logos::skip)]
@ -12,7 +12,7 @@ pub enum Token {
// Keywords // Keywords
#[token("fun")] #[token("fun")]
Fun, Fun,
#[token("class")] #[token("class")]
Class, Class,
@ -40,19 +40,18 @@ pub enum Token {
#[token("enum")] #[token("enum")]
Enum, Enum,
// Types // Types
#[token("int")] #[token("int")]
IntType, IntType,
#[token("float")] #[token("float")]
FloatType, FloatType,
#[token("bool")] #[token("bool")]
BoolType, BoolType,
#[token("String")] #[token("String")]
StringType, StringType,
// Literals // Literals
#[regex(r#""([^"\\]|\\.)*""#, |lex| lex.slice().to_owned())] #[regex(r#""([^"\\]|\\.)*""#, |lex| lex.slice().to_owned())]
@ -71,7 +70,7 @@ pub enum Token {
// Operators // Operators
#[token("=")] #[token("=")]
Assign, Assign,
#[token("==")] #[token("==")]
Equals, Equals,
@ -108,7 +107,6 @@ pub enum Token {
#[token("||")] #[token("||")]
Or, Or,
// Punctiuation // Punctiuation
#[token("(")] #[token("(")]
LParen, LParen,
@ -138,18 +136,15 @@ pub enum Token {
Dot, Dot,
// Special // Special
#[regex(r"\n")] #[regex(r"\n")]
NewLine, NewLine,
#[regex(r"//[^\r]*", logos::skip)] #[regex(r"//[^\r]*", logos::skip)]
#[regex(r"/\*([^*]|\*[^/])*\*/", logos::skip)] #[regex(r"/\*([^*]|\*[^/])*\*/", logos::skip)]
Comment, Comment,
#[regex(r"[ \t\f]+", logos::skip)] #[regex(r"[ \t\f]+", logos::skip)]
Whitespace, Whitespace,
Eof, Eof,
@ -207,4 +202,4 @@ impl fmt::Display for Token {
Token::Error => write!(f, "<error>"), Token::Error => write!(f, "<error>"),
} }
} }
} }

View file

@ -1,14 +1,14 @@
use chumsky::input::{Input, Stream};
use chumsky::Parser; use chumsky::Parser;
use chumsky::input::{Input, Stream};
use logos::Logos; use logos::Logos;
mod language_frontend; mod language_frontend;
use crate::{ use crate::{
language_frontend::lexer::tokens::Token, language_frontend::abstract_syntax_tree::parser::parse}; language_frontend::abstract_syntax_tree::parser::parse, language_frontend::lexer::tokens::Token,
};
use crate::language_frontend::abstract_syntax_tree::ast::{Expr}; use crate::language_frontend::abstract_syntax_tree::ast::Expr;
/* /*
Simple Compiler -> 4 Stages: Simple Compiler -> 4 Stages:
@ -27,10 +27,10 @@ fn main() {
println!("{:?}", sourcecode); println!("{:?}", sourcecode);
match parse(&sourcecode) { match parse(&sourcecode) {
Ok(res) => println!("{:#?}", res), Ok(res) => println!("{:#?}", res),
Err(e) => { Err(e) => {
panic!("{:#?}", e) panic!("{:#?}", e)
} }
}; };
} }