dyn trait for tokenizer buffer
This commit is contained in:
@@ -35,7 +35,7 @@ fn run_logic() -> Result<(), StationlangError> {
|
||||
let args = Args::parse();
|
||||
let input_file = args.input_file;
|
||||
|
||||
let tokenizer: Tokenizer<_> = match input_file {
|
||||
let tokenizer: Tokenizer = match input_file {
|
||||
Some(input_file) => Tokenizer::from_path(&input_file)?,
|
||||
None => {
|
||||
let mut buf = String::new();
|
||||
|
||||
@@ -90,16 +90,13 @@ macro_rules! token_matches {
|
||||
};
|
||||
}
|
||||
|
||||
pub struct Parser<R: Read + Seek> {
|
||||
tokenizer: TokenizerBuffer<R>,
|
||||
pub struct Parser {
|
||||
tokenizer: TokenizerBuffer,
|
||||
current_token: Option<Token>,
|
||||
}
|
||||
|
||||
impl<R> Parser<R>
|
||||
where
|
||||
R: Read + Seek,
|
||||
{
|
||||
pub fn new(tokenizer: Tokenizer<R>) -> Self {
|
||||
impl Parser {
|
||||
pub fn new(tokenizer: Tokenizer) -> Self {
|
||||
Parser {
|
||||
tokenizer: TokenizerBuffer::new(tokenizer),
|
||||
current_token: None,
|
||||
|
||||
@@ -23,36 +23,20 @@ pub enum TokenizerError {
|
||||
|
||||
pub trait Tokenize: Read + Seek {}
|
||||
|
||||
pub(crate) struct Tokenizer<T>
|
||||
where
|
||||
T: Read + Seek,
|
||||
{
|
||||
reader: BufReader<T>,
|
||||
impl<T> Tokenize for T where T: Read + Seek {}
|
||||
|
||||
pub(crate) struct Tokenizer {
|
||||
reader: BufReader<Box<dyn Tokenize>>,
|
||||
char_buffer: [u8; 1],
|
||||
line: usize,
|
||||
column: usize,
|
||||
returned_eof: bool,
|
||||
}
|
||||
|
||||
impl From<String> for Tokenizer<Cursor<Vec<u8>>> {
|
||||
fn from(input: String) -> Self {
|
||||
let cursor = Cursor::new(input.into_bytes());
|
||||
let reader = BufReader::new(cursor);
|
||||
|
||||
Self {
|
||||
reader,
|
||||
line: 1,
|
||||
column: 1,
|
||||
char_buffer: [0],
|
||||
returned_eof: false,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl Tokenizer<File> {
|
||||
impl Tokenizer {
|
||||
pub fn from_path(input_file: impl Into<PathBuf>) -> Result<Self, TokenizerError> {
|
||||
let file = std::fs::File::open(input_file.into())?;
|
||||
let reader = BufReader::new(file);
|
||||
let reader = BufReader::new(Box::new(file) as Box<dyn Tokenize>);
|
||||
|
||||
Ok(Self {
|
||||
reader,
|
||||
@@ -64,10 +48,21 @@ impl Tokenizer<File> {
|
||||
}
|
||||
}
|
||||
|
||||
impl<T> Tokenizer<T>
|
||||
where
|
||||
T: Read + Seek,
|
||||
{
|
||||
impl From<String> for Tokenizer {
|
||||
fn from(input: String) -> Self {
|
||||
let reader = BufReader::new(Box::new(Cursor::new(input)) as Box<dyn Tokenize>);
|
||||
|
||||
Self {
|
||||
reader,
|
||||
line: 1,
|
||||
column: 1,
|
||||
char_buffer: [0],
|
||||
returned_eof: false,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl Tokenizer {
|
||||
/// Consumes the tokenizer and returns the next token in the stream
|
||||
/// If there are no more tokens in the stream, this function returns None
|
||||
/// If there is an error reading the stream, this function returns an error
|
||||
@@ -415,20 +410,14 @@ where
|
||||
}
|
||||
}
|
||||
|
||||
pub struct TokenizerBuffer<T>
|
||||
where
|
||||
T: Read + Seek,
|
||||
{
|
||||
tokenizer: Tokenizer<T>,
|
||||
pub struct TokenizerBuffer {
|
||||
tokenizer: Tokenizer,
|
||||
buffer: VecDeque<Token>,
|
||||
history: VecDeque<Token>,
|
||||
}
|
||||
|
||||
impl<T> TokenizerBuffer<T>
|
||||
where
|
||||
T: Seek + Read,
|
||||
{
|
||||
pub fn new(tokenizer: Tokenizer<T>) -> Self {
|
||||
impl TokenizerBuffer {
|
||||
pub fn new(tokenizer: Tokenizer) -> Self {
|
||||
Self {
|
||||
tokenizer,
|
||||
buffer: VecDeque::new(),
|
||||
|
||||
Reference in New Issue
Block a user