aboutsummaryrefslogtreecommitdiffstats
path: root/src/tokenizer.rs
diff options
context:
space:
mode:
authorLibravatar Titus Wormer <tituswormer@gmail.com>2022-07-22 18:10:55 +0200
committerLibravatar Titus Wormer <tituswormer@gmail.com>2022-07-22 18:10:55 +0200
commit002b6bfb54eebc70a9b5e9fb62175777044c9338 (patch)
tree1735d3e33408425d2b5d23450cbe4db2e4c88e88 /src/tokenizer.rs
parentb945e43103544fc31a0755841b380358b2c161e6 (diff)
downloadmarkdown-rs-002b6bfb54eebc70a9b5e9fb62175777044c9338.tar.gz
markdown-rs-002b6bfb54eebc70a9b5e9fb62175777044c9338.tar.bz2
markdown-rs-002b6bfb54eebc70a9b5e9fb62175777044c9338.zip
Refactor to pass more references around
Diffstat (limited to 'src/tokenizer.rs')
-rw-r--r--src/tokenizer.rs65
1 files changed, 39 insertions, 26 deletions
diff --git a/src/tokenizer.rs b/src/tokenizer.rs
index 85a2155..bb73430 100644
--- a/src/tokenizer.rs
+++ b/src/tokenizer.rs
@@ -475,9 +475,9 @@ impl<'a> Tokenizer<'a> {
state_fn,
|_code| false,
vec![],
- |result: (Vec<Code>, Vec<Code>), tokenizer: &mut Tokenizer, state| {
+ |result: (&[Code], &[Code]), tokenizer: &mut Tokenizer, state| {
if matches!(state, State::Ok(_)) {
- feed_impl(tokenizer, &result.1, after)
+ feed_impl(tokenizer, result.1, after)
} else {
State::Nok
}
@@ -498,9 +498,9 @@ impl<'a> Tokenizer<'a> {
state_fn,
until,
vec![],
- |result: (Vec<Code>, Vec<Code>), tokenizer: &mut Tokenizer, state| {
+ |result: (&[Code], &[Code]), tokenizer: &mut Tokenizer, state| {
tokenizer.consumed = true;
- feed_impl(tokenizer, &result.1, done(state))
+ feed_impl(tokenizer, result.1, done(state))
},
)
}
@@ -525,10 +525,14 @@ impl<'a> Tokenizer<'a> {
state_fn,
|_code| false,
vec![],
- |mut result: (Vec<Code>, Vec<Code>), tokenizer: &mut Tokenizer, state| {
+ |result: (&[Code], &[Code]), tokenizer: &mut Tokenizer, state| {
tokenizer.free(previous);
- result.0.append(&mut result.1);
- feed_impl(tokenizer, &result.0, done(matches!(state, State::Ok(_))))
+ feed_twice_impl(
+ tokenizer,
+ result.0,
+ result.1,
+ done(matches!(state, State::Ok(_))),
+ )
},
)
}
@@ -555,28 +559,21 @@ impl<'a> Tokenizer<'a> {
state_fn,
|_code| false,
vec![],
- |mut result: (Vec<Code>, Vec<Code>), tokenizer: &mut Tokenizer, state| {
+ |result: (&[Code], &[Code]), tokenizer: &mut Tokenizer, state| {
let ok = matches!(state, State::Ok(_));
if !ok {
tokenizer.free(previous);
}
- let codes = if ok {
- result.1
- } else {
- result.0.append(&mut result.1);
- result.0
- };
-
- log::debug!(
- "attempt: {:?}, codes: {:?}, at {:?}",
- ok,
- codes,
- tokenizer.point
- );
+ log::debug!("attempt: {:?}, at {:?}", ok, tokenizer.point);
- feed_impl(tokenizer, &codes, done(ok))
+ feed_twice_impl(
+ tokenizer,
+ if ok { &[] } else { result.0 },
+ result.1,
+ done(ok),
+ )
},
)
}
@@ -672,7 +669,7 @@ fn attempt_impl(
state: impl FnOnce(&mut Tokenizer, Code) -> State + 'static,
mut pause: impl FnMut(Code) -> bool + 'static,
mut codes: Vec<Code>,
- done: impl FnOnce((Vec<Code>, Vec<Code>), &mut Tokenizer, State) -> State + 'static,
+ done: impl FnOnce((&[Code], &[Code]), &mut Tokenizer, State) -> State + 'static,
) -> Box<StateFn> {
Box::new(|tokenizer, code| {
if !codes.is_empty() && pause(tokenizer.previous) {
@@ -682,7 +679,7 @@ fn attempt_impl(
vec![code]
};
- return done((codes, after), tokenizer, State::Fn(Box::new(state)));
+ return done((&codes, &after), tokenizer, State::Fn(Box::new(state)));
}
let state = state(tokenizer, code);
@@ -701,9 +698,9 @@ fn attempt_impl(
"`back` must be smaller than or equal to `codes.len()`"
);
let remaining = codes.split_off(codes.len() - back);
- done((codes, remaining), tokenizer, state)
+ done((&codes, &remaining), tokenizer, state)
}
- State::Nok => done((vec![], codes), tokenizer, state),
+ State::Nok => done((&[], &codes), tokenizer, state),
State::Fn(func) => State::Fn(attempt_impl(func, pause, codes, done)),
}
})
@@ -741,6 +738,22 @@ fn feed_impl(
state
}
+/// Feed a list of `codes` into `start`.
+fn feed_twice_impl(
+ tokenizer: &mut Tokenizer,
+ left: &[Code],
+ right: &[Code],
+ start: impl FnOnce(&mut Tokenizer, Code) -> State + 'static,
+) -> State {
+ let res = feed_impl(tokenizer, left, start);
+
+ match res {
+ State::Fn(func) => feed_impl(tokenizer, right, func),
+ State::Ok(back) => State::Ok(back + right.len()),
+ State::Nok => res,
+ }
+}
+
/// Flush `start`: pass `eof`s to it until done.
fn flush_impl(
tokenizer: &mut Tokenizer,