//! HTML (text) is a construct that occurs in the [text][] content type.
//!
//! It forms with the following BNF:
//!
//! ```bnf
//! html_text ::= comment | instruction | declaration | cdata | tag_close | tag_open
//!
//! ; Restriction: the text is not allowed to start with `>`, `->`, or to contain `--`.
//! comment ::= ''
//! instruction ::= '' *code '?>'
//! declaration ::= ''
//! ; Restriction: the text is not allowed to contain `]]`.
//! cdata ::= ''
//! tag_close ::= '' tag_name whitespace_optional '>'
//! opening_tag ::= '<' tag_name *( whitespace attribute ) [ whitespace_optional '/' ] whitespace_optional '>'
//!
//! tag_name ::= ascii_alphabetic *( '-' | ascii_alphanumeric )
//! attribute ::= attribute_name [ whitespace_optional '=' whitespace_optional attribute_value ]
//! attribute_name ::= ( ':' | '_' | ascii_alphabetic ) *( '-' | '.' | ':' | '_' | ascii_alphanumeric )
//! attribute_value ::= '"' *( code - '"' ) '"' | "'" *( code - "'" ) "'" | 1*( code - space_or_tab - eol - '"' - "'" - '/' - '<' - '=' - '>' - '`')
//!
//! ; Note: blank lines can never occur in `text`.
//! whitespace ::= 1*space_or_tab | [ *space_or_tab eol *space_or_tab ]
//! whitespace_optional ::= [ whitespace ]
//! eol ::= '\r' | '\r\n' | '\n'
//! space_or_tab ::= ' ' | '\t'
//! ```
//!
//! The grammar for HTML in markdown does not resemble the rules of parsing
//! HTML according to the [*ยง 13.2 Parsing HTML documents* in the HTML
//! spec][html-parsing].
//! See the related flow construct [HTML (flow)][html_flow] for more info.
//!
//! Because the **tag open** and **tag close** productions in the grammar form
//! with just tags instead of complete elements, it is possible to interleave
//! (a word for switching between languages) markdown and HTML together.
//! For example:
//!
//! ```markdown
//! This is equivalent to *emphasised* code
.
//! ```
//!
//! ## Tokens
//!
//! * [`HtmlText`][Token::HtmlText]
//! * [`HtmlTextData`][Token::HtmlTextData]
//!
//! ## References
//!
//! * [`html-text.js` in `micromark`](https://github.com/micromark/micromark/blob/main/packages/micromark-core-commonmark/dev/lib/html-text.js)
//! * [*ยง 6.6 Raw HTML* in `CommonMark`](https://spec.commonmark.org/0.30/#raw-html)
//!
//! [text]: crate::content::text
//! [html_flow]: crate::construct::html_flow
//! [html-parsing]: https://html.spec.whatwg.org/multipage/parsing.html#parsing
use crate::constant::HTML_CDATA_PREFIX;
use crate::construct::partial_space_or_tab::space_or_tab;
use crate::token::Token;
use crate::tokenizer::{State, StateFn, Tokenizer};
/// Start of HTML (text)
///
/// ```markdown
/// > | a c
/// ^
/// ```
pub fn start(tokenizer: &mut Tokenizer) -> State {
if Some(b'<') == tokenizer.current && tokenizer.parse_state.constructs.html_text {
tokenizer.enter(Token::HtmlText);
tokenizer.enter(Token::HtmlTextData);
tokenizer.consume();
State::Fn(Box::new(open))
} else {
State::Nok
}
}
/// After `<`, before a tag name or other stuff.
///
/// ```markdown
/// > | a c
/// ^
/// > | a c
/// ^
/// > | a c
/// ^
/// ```
fn open(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
Some(b'!') => {
tokenizer.consume();
State::Fn(Box::new(declaration_open))
}
Some(b'/') => {
tokenizer.consume();
State::Fn(Box::new(tag_close_start))
}
Some(b'?') => {
tokenizer.consume();
State::Fn(Box::new(instruction))
}
// ASCII alphabetical.
Some(b'A'..=b'Z' | b'a'..=b'z') => {
tokenizer.consume();
State::Fn(Box::new(tag_open))
}
_ => State::Nok,
}
}
/// After ` | a c
/// ^
/// > | a c
/// ^
/// > | a &<]]> c
/// ^
/// ```
fn declaration_open(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
Some(b'-') => {
tokenizer.consume();
State::Fn(Box::new(comment_open_inside))
}
// ASCII alphabetical.
Some(b'A'..=b'Z' | b'a'..=b'z') => {
tokenizer.consume();
State::Fn(Box::new(declaration))
}
Some(b'[') => {
tokenizer.consume();
State::Fn(Box::new(|t| cdata_open_inside(t, 0)))
}
_ => State::Nok,
}
}
/// After ` | a c
/// ^
/// ```
fn comment_open_inside(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
Some(b'-') => {
tokenizer.consume();
State::Fn(Box::new(comment_start))
}
_ => State::Nok,
}
}
/// After `` or `` as
/// > empty comments.
/// > This is prohibited in html (text).
/// > See: .
///
/// ```markdown
/// > | a c
/// ^
/// ```
///
/// [html_flow]: crate::construct::html_flow
fn comment_start(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
None | Some(b'>') => State::Nok,
Some(b'-') => {
tokenizer.consume();
State::Fn(Box::new(comment_start_dash))
}
_ => comment(tokenizer),
}
}
/// After `` or `` as
/// > empty comments.
/// > This is prohibited in html (text).
/// > See: .
///
/// ```markdown
/// > | a c
/// ^
/// ```
///
/// [html_flow]: crate::construct::html_flow
fn comment_start_dash(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
None | Some(b'>') => State::Nok,
_ => comment(tokenizer),
}
}
/// In a comment.
///
/// ```markdown
/// > | a c
/// ^
/// ```
fn comment(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
None => State::Nok,
Some(b'\n') => at_line_ending(tokenizer, Box::new(comment)),
Some(b'-') => {
tokenizer.consume();
State::Fn(Box::new(comment_close))
}
_ => {
tokenizer.consume();
State::Fn(Box::new(comment))
}
}
}
/// In a comment, after `-`.
///
/// ```markdown
/// > | a c
/// ^
/// ```
fn comment_close(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
Some(b'-') => {
tokenizer.consume();
State::Fn(Box::new(end))
}
_ => comment(tokenizer),
}
}
/// After ` | a &<]]> b
/// ^^^^^^
/// ```
fn cdata_open_inside(tokenizer: &mut Tokenizer, size: usize) -> State {
if tokenizer.current == Some(HTML_CDATA_PREFIX[size]) {
tokenizer.consume();
if size + 1 == HTML_CDATA_PREFIX.len() {
State::Fn(Box::new(cdata))
} else {
State::Fn(Box::new(move |t| cdata_open_inside(t, size + 1)))
}
} else {
State::Nok
}
}
/// In CDATA.
///
/// ```markdown
/// > | a &<]]> b
/// ^^^
/// ```
fn cdata(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
None => State::Nok,
Some(b'\n') => at_line_ending(tokenizer, Box::new(cdata)),
Some(b']') => {
tokenizer.consume();
State::Fn(Box::new(cdata_close))
}
_ => {
tokenizer.consume();
State::Fn(Box::new(cdata))
}
}
}
/// In CDATA, after `]`.
///
/// ```markdown
/// > | a &<]]> b
/// ^
/// ```
fn cdata_close(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
Some(b']') => {
tokenizer.consume();
State::Fn(Box::new(cdata_end))
}
_ => cdata(tokenizer),
}
}
/// In CDATA, after `]]`.
///
/// ```markdown
/// > | a &<]]> b
/// ^
/// ```
fn cdata_end(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
Some(b'>') => end(tokenizer),
Some(b']') => cdata_close(tokenizer),
_ => cdata(tokenizer),
}
}
/// In a declaration.
///
/// ```markdown
/// > | a c
/// ^
/// ```
fn declaration(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
None | Some(b'>') => end(tokenizer),
Some(b'\n') => at_line_ending(tokenizer, Box::new(declaration)),
_ => {
tokenizer.consume();
State::Fn(Box::new(declaration))
}
}
}
/// In an instruction.
///
/// ```markdown
/// > | a c
/// ^
/// ```
fn instruction(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
None => State::Nok,
Some(b'\n') => at_line_ending(tokenizer, Box::new(instruction)),
Some(b'?') => {
tokenizer.consume();
State::Fn(Box::new(instruction_close))
}
_ => {
tokenizer.consume();
State::Fn(Box::new(instruction))
}
}
}
/// In an instruction, after `?`.
///
/// ```markdown
/// > | a c
/// ^
/// ```
fn instruction_close(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
Some(b'>') => end(tokenizer),
_ => instruction(tokenizer),
}
}
/// After ``, in a closing tag, before a tag name.
///
/// ```markdown
/// > | a c
/// ^
/// ```
fn tag_close_start(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
// ASCII alphabetical.
Some(b'A'..=b'Z' | b'a'..=b'z') => {
tokenizer.consume();
State::Fn(Box::new(tag_close))
}
_ => State::Nok,
}
}
/// After ` | a c
/// ^
/// ```
fn tag_close(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
// ASCII alphanumerical and `-`.
Some(b'-' | b'0'..=b'9' | b'A'..=b'Z' | b'a'..=b'z') => {
tokenizer.consume();
State::Fn(Box::new(tag_close))
}
_ => tag_close_between(tokenizer),
}
}
/// In a closing tag, after the tag name.
///
/// ```markdown
/// > | a c
/// ^
/// ```
fn tag_close_between(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
Some(b'\n') => at_line_ending(tokenizer, Box::new(tag_close_between)),
Some(b'\t' | b' ') => {
tokenizer.consume();
State::Fn(Box::new(tag_close_between))
}
_ => end(tokenizer),
}
}
/// After ` | a c
/// ^
/// ```
fn tag_open(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
// ASCII alphanumerical and `-`.
Some(b'-' | b'0'..=b'9' | b'A'..=b'Z' | b'a'..=b'z') => {
tokenizer.consume();
State::Fn(Box::new(tag_open))
}
Some(b'\t' | b'\n' | b' ' | b'/' | b'>') => tag_open_between(tokenizer),
_ => State::Nok,
}
}
/// In an opening tag, after the tag name.
///
/// ```markdown
/// > | a c
/// ^
/// ```
fn tag_open_between(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
Some(b'\n') => at_line_ending(tokenizer, Box::new(tag_open_between)),
Some(b'\t' | b' ') => {
tokenizer.consume();
State::Fn(Box::new(tag_open_between))
}
Some(b'/') => {
tokenizer.consume();
State::Fn(Box::new(end))
}
// ASCII alphabetical and `:` and `_`.
Some(b':' | b'A'..=b'Z' | b'_' | b'a'..=b'z') => {
tokenizer.consume();
State::Fn(Box::new(tag_open_attribute_name))
}
_ => end(tokenizer),
}
}
/// In an attribute name.
///
/// ```markdown
/// > | a d
/// ^
/// ```
fn tag_open_attribute_name(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
// ASCII alphabetical and `-`, `.`, `:`, and `_`.
Some(b'-' | b'.' | b'0'..=b'9' | b':' | b'A'..=b'Z' | b'_' | b'a'..=b'z') => {
tokenizer.consume();
State::Fn(Box::new(tag_open_attribute_name))
}
_ => tag_open_attribute_name_after(tokenizer),
}
}
/// After an attribute name, before an attribute initializer, the end of the
/// tag, or whitespace.
///
/// ```markdown
/// > | a d
/// ^
/// ```
fn tag_open_attribute_name_after(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
Some(b'\n') => at_line_ending(tokenizer, Box::new(tag_open_attribute_name_after)),
Some(b'\t' | b' ') => {
tokenizer.consume();
State::Fn(Box::new(tag_open_attribute_name_after))
}
Some(b'=') => {
tokenizer.consume();
State::Fn(Box::new(tag_open_attribute_value_before))
}
_ => tag_open_between(tokenizer),
}
}
/// Before an unquoted, double quoted, or single quoted attribute value,
/// allowing whitespace.
///
/// ```markdown
/// > | a e
/// ^
/// ```
fn tag_open_attribute_value_before(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
None | Some(b'<' | b'=' | b'>' | b'`') => State::Nok,
Some(b'\n') => at_line_ending(tokenizer, Box::new(tag_open_attribute_value_before)),
Some(b'\t' | b' ') => {
tokenizer.consume();
State::Fn(Box::new(tag_open_attribute_value_before))
}
Some(b'"' | b'\'') => {
let marker = tokenizer.current.unwrap();
tokenizer.consume();
State::Fn(Box::new(move |t| {
tag_open_attribute_value_quoted(t, marker)
}))
}
Some(_) => {
tokenizer.consume();
State::Fn(Box::new(tag_open_attribute_value_unquoted))
}
}
}
/// In a double or single quoted attribute value.
///
/// ```markdown
/// > | a e
/// ^
/// ```
fn tag_open_attribute_value_quoted(tokenizer: &mut Tokenizer, marker: u8) -> State {
match tokenizer.current {
None => State::Nok,
Some(b'\n') => at_line_ending(
tokenizer,
Box::new(move |t| tag_open_attribute_value_quoted(t, marker)),
),
Some(b'"' | b'\'') if tokenizer.current.unwrap() == marker => {
tokenizer.consume();
State::Fn(Box::new(tag_open_attribute_value_quoted_after))
}
_ => {
tokenizer.consume();
State::Fn(Box::new(move |t| {
tag_open_attribute_value_quoted(t, marker)
}))
}
}
}
/// In an unquoted attribute value.
///
/// ```markdown
/// > | a e
/// ^
/// ```
fn tag_open_attribute_value_unquoted(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
None | Some(b'"' | b'\'' | b'<' | b'=' | b'`') => State::Nok,
Some(b'\t' | b'\n' | b' ' | b'/' | b'>') => tag_open_between(tokenizer),
Some(_) => {
tokenizer.consume();
State::Fn(Box::new(tag_open_attribute_value_unquoted))
}
}
}
/// After a double or single quoted attribute value, before whitespace or the
/// end of the tag.
///
/// ```markdown
/// > | a e
/// ^
/// ```
fn tag_open_attribute_value_quoted_after(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
Some(b'\t' | b'\n' | b' ' | b'>' | b'/') => tag_open_between(tokenizer),
_ => State::Nok,
}
}
/// In certain circumstances of a complete tag where only an `>` is allowed.
///
/// ```markdown
/// > | a e
/// ^
/// ```
fn end(tokenizer: &mut Tokenizer) -> State {
match tokenizer.current {
Some(b'>') => {
tokenizer.consume();
tokenizer.exit(Token::HtmlTextData);
tokenizer.exit(Token::HtmlText);
State::Ok
}
_ => State::Nok,
}
}
/// At an allowed line ending.
///
/// > ๐ **Note**: we canโt have blank lines in text, so no need to worry about
/// > empty tokens.
///
/// ```markdown
/// > | a
/// ```
fn at_line_ending(tokenizer: &mut Tokenizer, return_state: Box) -> State {
match tokenizer.current {
Some(b'\n') => {
tokenizer.exit(Token::HtmlTextData);
tokenizer.enter(Token::LineEnding);
tokenizer.consume();
tokenizer.exit(Token::LineEnding);
State::Fn(Box::new(|t| after_line_ending(t, return_state)))
}
_ => unreachable!("expected eol"),
}
}
/// After a line ending.
///
/// > ๐ **Note**: we canโt have blank lines in text, so no need to worry about
/// > empty tokens.
///
/// ```markdown
/// | a
/// ^
/// ```
fn after_line_ending(tokenizer: &mut Tokenizer, return_state: Box) -> State {
tokenizer.attempt_opt(space_or_tab(), |t| {
after_line_ending_prefix(t, return_state)
})(tokenizer)
}
/// After a line ending, after indent.
///
/// > ๐ **Note**: we canโt have blank lines in text, so no need to worry about
/// > empty tokens.
///
/// ```markdown
/// | a
/// ^
/// ```
fn after_line_ending_prefix(tokenizer: &mut Tokenizer, return_state: Box) -> State {
tokenizer.enter(Token::HtmlTextData);
return_state(tokenizer)
}