dyn trait for tokenizer buffer

This commit is contained in:
2024-11-21 19:28:20 -07:00
parent abaf58374f
commit ae45a3c848
3 changed files with 30 additions and 44 deletions

View File

@@ -35,7 +35,7 @@ fn run_logic() -> Result<(), StationlangError> {
let args = Args::parse();
let input_file = args.input_file;
let tokenizer: Tokenizer<_> = match input_file {
let tokenizer: Tokenizer = match input_file {
Some(input_file) => Tokenizer::from_path(&input_file)?,
None => {
let mut buf = String::new();

View File

@@ -90,16 +90,13 @@ macro_rules! token_matches {
};
}
pub struct Parser<R: Read + Seek> {
tokenizer: TokenizerBuffer<R>,
pub struct Parser {
tokenizer: TokenizerBuffer,
current_token: Option<Token>,
}
impl<R> Parser<R>
where
R: Read + Seek,
{
pub fn new(tokenizer: Tokenizer<R>) -> Self {
impl Parser {
pub fn new(tokenizer: Tokenizer) -> Self {
Parser {
tokenizer: TokenizerBuffer::new(tokenizer),
current_token: None,

View File

@@ -23,36 +23,20 @@ pub enum TokenizerError {
pub trait Tokenize: Read + Seek {}
pub(crate) struct Tokenizer<T>
where
T: Read + Seek,
{
reader: BufReader<T>,
impl<T> Tokenize for T where T: Read + Seek {}
pub(crate) struct Tokenizer {
reader: BufReader<Box<dyn Tokenize>>,
char_buffer: [u8; 1],
line: usize,
column: usize,
returned_eof: bool,
}
impl From<String> for Tokenizer<Cursor<Vec<u8>>> {
fn from(input: String) -> Self {
let cursor = Cursor::new(input.into_bytes());
let reader = BufReader::new(cursor);
Self {
reader,
line: 1,
column: 1,
char_buffer: [0],
returned_eof: false,
}
}
}
impl Tokenizer<File> {
impl Tokenizer {
pub fn from_path(input_file: impl Into<PathBuf>) -> Result<Self, TokenizerError> {
let file = std::fs::File::open(input_file.into())?;
let reader = BufReader::new(file);
let reader = BufReader::new(Box::new(file) as Box<dyn Tokenize>);
Ok(Self {
reader,
@@ -64,10 +48,21 @@ impl Tokenizer<File> {
}
}
impl<T> Tokenizer<T>
where
T: Read + Seek,
{
impl From<String> for Tokenizer {
fn from(input: String) -> Self {
let reader = BufReader::new(Box::new(Cursor::new(input)) as Box<dyn Tokenize>);
Self {
reader,
line: 1,
column: 1,
char_buffer: [0],
returned_eof: false,
}
}
}
impl Tokenizer {
/// Consumes the tokenizer and returns the next token in the stream
/// If there are no more tokens in the stream, this function returns None
/// If there is an error reading the stream, this function returns an error
@@ -415,20 +410,14 @@ where
}
}
pub struct TokenizerBuffer<T>
where
T: Read + Seek,
{
tokenizer: Tokenizer<T>,
pub struct TokenizerBuffer {
tokenizer: Tokenizer,
buffer: VecDeque<Token>,
history: VecDeque<Token>,
}
impl<T> TokenizerBuffer<T>
where
T: Seek + Read,
{
pub fn new(tokenizer: Tokenizer<T>) -> Self {
impl TokenizerBuffer {
pub fn new(tokenizer: Tokenizer) -> Self {
Self {
tokenizer,
buffer: VecDeque::new(),