//! To do. use crate::constant::{LIST_ITEM_VALUE_SIZE_MAX, TAB_SIZE}; use crate::construct::{ blank_line::start as blank_line, partial_space_or_tab::space_or_tab_min_max, thematic_break::start as thematic_break, }; use crate::token::Token; use crate::tokenizer::{Code, Event, EventType, State, StateFnResult, Tokenizer}; use crate::util::{ edit_map::EditMap, skip, span::{codes as codes_from_span, from_exit_event}, }; /// Type of list. #[derive(Debug, PartialEq)] enum Kind { /// In a dot (`.`) list item. /// /// ## Example /// /// ```markdown /// 1. a /// ``` Dot, /// In a paren (`)`) list item. /// /// ## Example /// /// ```markdown /// 1) a /// ``` Paren, /// In an asterisk (`*`) list item. /// /// ## Example /// /// ```markdown /// * a /// ``` Asterisk, /// In a plus (`+`) list item. /// /// ## Example /// /// ```markdown /// + a /// ``` Plus, /// In a dash (`-`) list item. /// /// ## Example /// /// ```markdown /// - a /// ``` Dash, } impl Kind { // /// Turn the kind into a [char]. // fn as_char(&self) -> char { // match self { // Kind::Dot => '.', // Kind::Paren => ')', // Kind::Asterisk => '*', // Kind::Plus => '+', // Kind::Dash => '-', // } // } /// Turn a [char] into a kind. /// /// ## Panics /// /// Panics if `char` is not `.`, `)`, `*`, `+`, or `-`. fn from_char(char: char) -> Kind { match char { '.' => Kind::Dot, ')' => Kind::Paren, '*' => Kind::Asterisk, '+' => Kind::Plus, '-' => Kind::Dash, _ => unreachable!("invalid char"), } } /// Turn [Code] into a kind. /// /// ## Panics /// /// Panics if `code` is not `Code::Char('.' | ')' | '*' | '+' | '-')`. fn from_code(code: Code) -> Kind { match code { Code::Char(char) => Kind::from_char(char), _ => unreachable!("invalid code"), } } } /// To do. pub fn start(tokenizer: &mut Tokenizer, code: Code) -> StateFnResult { // To do: allow arbitrary when code (indented) is turned off. tokenizer.go(space_or_tab_min_max(0, TAB_SIZE - 1), before)(tokenizer, code) } /// To do. fn before(tokenizer: &mut Tokenizer, code: Code) -> StateFnResult { match code { // Unordered. Code::Char('*' | '+' | '-') => tokenizer.check(thematic_break, |ok| { let func = if ok { nok } else { before_unordered }; Box::new(func) })(tokenizer, code), // Ordered. Code::Char(char) if char.is_ascii_digit() => { tokenizer.enter(Token::ListItem); tokenizer.enter(Token::ListItemPrefix); tokenizer.enter(Token::ListItemValue); // To do: `interrupt || !1`? inside(tokenizer, code, 0) } _ => (State::Nok, None), } } /// To do. fn before_unordered(tokenizer: &mut Tokenizer, code: Code) -> StateFnResult { // To do: check if this is a thematic break? tokenizer.enter(Token::ListItem); tokenizer.enter(Token::ListItemPrefix); marker(tokenizer, code) } /// To do. fn inside(tokenizer: &mut Tokenizer, code: Code, mut size: usize) -> StateFnResult { size += 1; match code { Code::Char(char) if char.is_ascii_digit() && size < LIST_ITEM_VALUE_SIZE_MAX => { tokenizer.consume(code); (State::Fn(Box::new(move |t, c| inside(t, c, size))), None) } // To do: `(!self.interrupt || size < 2)` Code::Char('.' | ')') => { tokenizer.exit(Token::ListItemValue); marker(tokenizer, code) } _ => (State::Nok, None), } } /// To do. fn marker(tokenizer: &mut Tokenizer, code: Code) -> StateFnResult { let kind = Kind::from_code(code); println!("list item kind: {:?}", kind); tokenizer.enter(Token::ListItemMarker); tokenizer.consume(code); tokenizer.exit(Token::ListItemMarker); (State::Fn(Box::new(marker_after)), None) } /// To do. fn marker_after(tokenizer: &mut Tokenizer, code: Code) -> StateFnResult { let interrupt = tokenizer.interrupt; tokenizer.check(blank_line, move |ok| { println!("check:blank_line:after {:?} {:?}", ok, interrupt); let func = if ok { if interrupt { nok } else { on_blank } } else { marker_after_after }; Box::new(func) })(tokenizer, code) } /// To do. fn on_blank(tokenizer: &mut Tokenizer, code: Code) -> StateFnResult { // self.containerState.initialBlankLine = true // initialSize++ prefix_end(tokenizer, code) } /// To do. fn marker_after_after(tokenizer: &mut Tokenizer, code: Code) -> StateFnResult { let interrupt = tokenizer.interrupt; tokenizer.attempt(list_item_prefix_whitespace, move |ok| { println!("marker:after:after: {:?} {:?}", ok, interrupt); let func = if ok { prefix_end } else { prefix_other }; Box::new(func) })(tokenizer, code) } // To do: `on_blank`. /// To do. fn prefix_other(tokenizer: &mut Tokenizer, code: Code) -> StateFnResult { match code { Code::VirtualSpace | Code::Char('\t' | ' ') => { tokenizer.enter(Token::SpaceOrTab); tokenizer.consume(code); tokenizer.exit(Token::SpaceOrTab); (State::Fn(Box::new(prefix_end)), None) } _ => (State::Nok, None), } } /// To do. fn prefix_end(tokenizer: &mut Tokenizer, code: Code) -> StateFnResult { // To do: calculate size. tokenizer.exit(Token::ListItemPrefix); tokenizer.register_resolver_before("list_item".to_string(), Box::new(resolve)); (State::Ok, Some(vec![code])) } /// To do. fn list_item_prefix_whitespace(tokenizer: &mut Tokenizer, code: Code) -> StateFnResult { // To do: check how big this should be? tokenizer.go( space_or_tab_min_max(1, TAB_SIZE - 1), list_item_prefix_whitespace_after, )(tokenizer, code) } fn list_item_prefix_whitespace_after(_tokenizer: &mut Tokenizer, code: Code) -> StateFnResult { // To do: check some stuff? (State::Ok, Some(vec![code])) } /// To do. fn nok(_tokenizer: &mut Tokenizer, _code: Code) -> StateFnResult { (State::Nok, None) } /// To do. pub fn cont(tokenizer: &mut Tokenizer, code: Code) -> StateFnResult { tokenizer.check(blank_line, |ok| { println!("cont:check:blank:after: {:?}", ok); let func = if ok { blank_cont } else { not_blank_cont }; Box::new(func) })(tokenizer, code) } pub fn blank_cont(tokenizer: &mut Tokenizer, code: Code) -> StateFnResult { // self.containerState.furtherBlankLines = // self.containerState.furtherBlankLines || // self.containerState.initialBlankLine // We have a blank line. // Still, try to consume at most the items size. // To do: eat at most `size` whitespace. tokenizer.go(space_or_tab_min_max(0, TAB_SIZE), blank_cont_after)(tokenizer, code) } pub fn blank_cont_after(_tokenizer: &mut Tokenizer, code: Code) -> StateFnResult { println!("cont: blank: after"); (State::Ok, Some(vec![code])) } pub fn not_blank_cont(tokenizer: &mut Tokenizer, code: Code) -> StateFnResult { let index = tokenizer.events.len(); let currently_blank = index > 0 && tokenizer.events[index - 1].token_type == Token::BlankLineEnding; let mut further_blank = false; if currently_blank && index > 5 { let before = skip::opt_back(&tokenizer.events, index - 3, &[Token::SpaceOrTab]); further_blank = tokenizer.events[before].token_type == Token::BlankLineEnding; } if further_blank || !matches!(code, Code::VirtualSpace | Code::Char('\t' | ' ')) { println!("cont: not blank after further blank, or not blank w/o whitespace"); (State::Nok, None) } else { println!("cont: not blank"); // To do: eat exactly `size` whitespace. tokenizer.go(space_or_tab_min_max(TAB_SIZE, TAB_SIZE), blank_cont_after)(tokenizer, code) } } /// To do. pub fn end() -> Vec { vec![Token::ListItem] } /// To do. pub fn resolve(tokenizer: &mut Tokenizer) -> Vec { let mut edit_map = EditMap::new(); let mut index = 0; println!("list item:before: {:?}", tokenizer.events.len()); while index < tokenizer.events.len() { let event = &tokenizer.events[index]; println!( "ev: {:?} {:?} {:?} {:?} {:?} {:?}", index, event.event_type, event.token_type, event.content_type, event.previous, event.next ); index += 1; } let mut index = 0; let mut balance = 0; let mut lists_wip: Vec<(Kind, usize, usize, usize)> = vec![]; let mut lists: Vec<(Kind, usize, usize, usize)> = vec![]; // To do: track balance? Or, check what’s between them? while index < tokenizer.events.len() { let event = &tokenizer.events[index]; if event.token_type == Token::ListItem { if event.event_type == EventType::Enter { let end = skip::opt(&tokenizer.events, index, &[Token::ListItem]) - 1; let marker = skip::to(&tokenizer.events, index, &[Token::ListItemMarker]) + 1; let codes = codes_from_span( &tokenizer.parse_state.codes, &from_exit_event(&tokenizer.events, marker), ); let kind = Kind::from_code(codes[0]); let current = (kind, balance, index, end); let mut list_index = lists_wip.len(); let mut matched = false; while list_index > 0 { list_index -= 1; let previous = &lists_wip[list_index]; if previous.0 == current.0 && previous.1 == current.1 && skip::opt( &tokenizer.events, previous.3 + 1, &[Token::SpaceOrTab, Token::LineEnding, Token::BlankLineEnding], ) == current.2 { println!("prev:match {:?} {:?}", previous, current); let previous_mut = &mut lists_wip[list_index]; previous_mut.3 = current.3; let mut remainder = lists_wip.drain((list_index + 1)..).collect::>(); lists.append(&mut remainder); matched = true; break; } println!( "todo: move them over to `lists` at some point? {:?}", previous ); } if !matched { println!("prev:!match {:?} {:?}", lists_wip, current); lists_wip.push(current); } println!("enter: {:?}", event.token_type); balance += 1; } else { println!("exit: {:?}", event.token_type); balance -= 1; } } index += 1; } lists.append(&mut lists_wip); let mut index = 0; while index < lists.len() { let list_item = &lists[index]; let mut list_start = tokenizer.events[list_item.2].clone(); let token_type = if matches!(list_item.0, Kind::Paren | Kind::Dot) { Token::ListOrdered } else { Token::ListUnordered }; list_start.token_type = token_type.clone(); let mut list_end = tokenizer.events[list_item.3].clone(); list_end.token_type = token_type; println!("inject:list: {:?} {:?}", list_start, list_end); edit_map.add(list_item.2, 0, vec![list_start]); edit_map.add(list_item.3 + 1, 0, vec![list_end]); index += 1; } println!("list items: {:#?}", lists); let events = edit_map.consume(&mut tokenizer.events); let mut index = 0; println!("list item:after: {:?}", events.len()); while index < events.len() { let event = &events[index]; println!( "ev: {:?} {:?} {:?} {:?} {:?} {:?}", index, event.event_type, event.token_type, event.content_type, event.previous, event.next ); index += 1; } events }