diff options
author | Titus Wormer <tituswormer@gmail.com> | 2022-06-13 18:42:36 +0200 |
---|---|---|
committer | Titus Wormer <tituswormer@gmail.com> | 2022-06-13 18:42:36 +0200 |
commit | ef644f4def7d5cad3fb5307ec5e00fc7b0b025ff (patch) | |
tree | 1d284b657d2cade8e3d4e60db09750c768bbc76f /src/content/text.rs | |
parent | 06b4ff3531874c95ec07b8440de526795408ef86 (diff) | |
download | markdown-rs-ef644f4def7d5cad3fb5307ec5e00fc7b0b025ff.tar.gz markdown-rs-ef644f4def7d5cad3fb5307ec5e00fc7b0b025ff.tar.bz2 markdown-rs-ef644f4def7d5cad3fb5307ec5e00fc7b0b025ff.zip |
Add basic html (text)
* Add all states for html (text)
* Fix to link paragraph tokens together
* Add note about uncovered bug where linking paragraph tokens together
doesn’t work 😅
Diffstat (limited to 'src/content/text.rs')
-rw-r--r-- | src/content/text.rs | 14 |
1 files changed, 9 insertions, 5 deletions
diff --git a/src/content/text.rs b/src/content/text.rs index a7b40e7..3db82f5 100644 --- a/src/content/text.rs +++ b/src/content/text.rs @@ -7,7 +7,7 @@ //! //! * [Autolink][crate::construct::autolink] //! * Attention -//! * HTML (text) +//! * [HTML (text)][crate::construct::html-text] //! * Hard break escape //! * Code (text) //! * Line ending @@ -18,7 +18,7 @@ use crate::construct::{ autolink::start as autolink, character_escape::start as character_escape, - character_reference::start as character_reference, + character_reference::start as character_reference, html_text::start as html_text, }; use crate::tokenizer::{Code, State, StateFnResult, TokenType, Tokenizer}; @@ -34,9 +34,13 @@ use crate::tokenizer::{Code, State, StateFnResult, TokenType, Tokenizer}; pub fn start(tokenizer: &mut Tokenizer, code: Code) -> StateFnResult { match code { Code::None => (State::Ok, None), - _ => tokenizer.attempt_3(character_reference, character_escape, autolink, |ok| { - Box::new(if ok { start } else { before_data }) - })(tokenizer, code), + _ => tokenizer.attempt_4( + character_reference, + character_escape, + autolink, + html_text, + |ok| Box::new(if ok { start } else { before_data }), + )(tokenizer, code), } } |