aboutsummaryrefslogblamecommitdiffstats
path: root/src/construct/html_flow.rs
blob: 064da3501da8c8bb5d7b38e8ca21ea0d6d69f22e (plain) (tree)
1
                                                                        
























                                                                                                                                               
                                          













                                                                            




                                                                              


















                                                                               
                                                                          

                                                                          













                                                                            

             


                                             
   




                                                                                                                                                 
                                
                                            
                                            


                                                                               
 


                                                                                     
                       

                                                                          
                                                                                    
  
                        

                                         

                        
                           
           















                                                
                                      
                




                                                              

                                                  
                                                                         
              

 


                                                     

             
       
                                                  





                                                         




                                                                        



                                   
                    
            
                  
     




                                          

             
       
                                               
                                        
                                             
                            
                                 
            
                  





                                                





                  
       
                                             


                          
                                          
                         

                 

      
                             
                       

                                                              
         
                       
                                
                                               
                                                             
         
                       
                                          
                                

                                      

                                                                                         
                                                                             
         
                              
                                            
                                  
                                               
                                     
         
                        





                                                           





                       
       

                                                                         
                       
                                
                                      
                                                                 
         
                                            
                                
                                          

                                      
                                                                             
         





                                                               
                        





                                                      

                  
       

                                                                        
                       
                                

                                      
                                                                             
         
                        





                                                  

                       
       
                                                                          
                             
                                                                                        
                                
 

                                                                              





                                                                   
         
                        





                                                    

            
       
                                                                    
                             
                              
                                            
                                
                                                      
         
                        





                  



             
       

                                                                 

                                                                

                                                  
                                            








                                                                                             

                                                                                    
                                      

                                          
                                             
                                                                 
                                                           
                          

                                                                        
                        

                                              
                                                 

                    
                                           
 
                                                                  
                                                           
                              
                                          
                                                                   
                        
                                                               


                 
                                        
                                                                 
                                
                                                      
         
                              





                                              

              
       

                                                                       
                       
                                

                                      
                                                          
         
                        





                                                 

            
       

                                                                               
                               

                                                                        
         
                                           










                                                                             









              
       

                                                                                   



                                                                            
                       

                                                          
         
                                                
                                                                        

                                                                     
         
                                           





                         





              
       

                                                                            
                                                           
                                                                                      

                                                                     
         
                                                            






                                                                            



               
       

                                                                                  
                               

                                                                           
         



                                                                             
                                                             






                                                                        



                 
       

                                                                                        
                                                             
                               

                                                                             
         




                                                                             
                                                                





                                                 



                 
       

                                                                                    
                                         
                                                                           

                                                                                   

              

                                                                             






                                   

               
       

                                                                                      
                                                                                                
                                                          
         
                    

                                                                               







                                                                             

                 
       

                                                                                          
                                                                                            
                        





                                                                            

                 
       

                                                                 
                       

                                                            
         
                        





                                

            
       

                                                                   
                               

                                      
                                         
         
                               

                                                            
         
                              





                                         

                  
       

                                                                 














                                                                                   
                                                     

                                                                         
         
                                                 

                                                                       
         
                                                         

                                                                
         
                                                         

                                                                             
         
                                                   

                                                                                
         
              

                                                          



         
                               

               


            
       
                                                                       
                                                         
               
                                                                   
                
                                        
         
                 

 


                                                        


            
       

                                                                                
                        
                                               
                                
                                              
                                                                 
         



                                          


                                                           


           
       

                                                                        
                                                                  
              
                                                 
                                         






                                                              

                  
       

                                                                                
                       

                                                                             
         
                                           





                                                    

                                       
       
                                                                                  
                             
                       
                                
                                               
                                                                      
         
                                           





                                                                        

                                       
       

                                                                                 
                       

                                                  
                                            





                                                           

                                                        

                                                                    
                    
                                             

             


                                                                        
                                
                                                                      
         
              
                           
                                         
         





                                                     

                       
       

                                                                                       
                       

                                                                             
         
                                           





                                                                            









                       
       

                                                                                    
                       

                                                                
         
                                                     

                                                                             
         
                                           





                                                                              

                   
       

                                                                       
                               
                                                
                                         

              

                                                                



         


               

                   
       
                                                           




                                    
             

 


                                                 


              
       
                                                          
                                       
                        
                                      
                                   
 
//! HTML (flow) is a construct that occurs in the [flow][] content type.
//!
//! It forms with the following BNF:
//!
//! ```bnf
//! html_flow ::= raw | comment | instruction | declaration | cdata | basic | complete
//!
//! ; Note: closing tag name need to match opening tag name.
//! raw ::= '<' raw_tag_name [ [ ( whitespace | '>' ) *line ] *( eol *line ) ] [ '</' raw_tag_name *line ]
//! comment ::= '<!--' [ *'-' '>' *line | *line *( eol *line ) [ '-->' *line ] ]
//! instruction ::= '<?' [ '>' *line | *line *( eol *line ) [ '?>' *line ] ]
//! declaration ::= '<!' ascii_alphabetic *line *( eol *line ) [ '>' *line ]
//! cdata ::= '<![CDATA[' *line *( eol *line ) [ ']]>' *line ]
//! basic ::= '< [ '/' ] basic_tag_name [ [ '/' ] '>' *line *( eol 1*line ) ]
//! complete ::= ( opening_tag | closing_tag ) ( whitespace_optional *( eol 1*line ) | whitespace_optional )
//!
//! raw_tag_name ::= 'pre' | 'script' | 'style' | 'textarea' ; Note: case-insensitive.
//! basic_tag_name ::= 'address' | 'article' | 'aside' | ... ; See `constants.rs`, and note: case-insensitive.
//! opening_tag ::= '<' tag_name *( whitespace attribute ) [ whitespace_optional '/' ] whitespace_optional '>'
//! closing_tag ::= '</' tag_name whitespace_optional '>'
//! tag_name ::= ascii_alphabetic *( '-' | ascii_alphanumeric )
//! attribute ::= attribute_name [ whitespace_optional '=' whitespace_optional attribute_value ]
//! attribute_name ::= ( ':' | '_' | ascii_alphabetic ) *( '-' | '.' | ':' | '_' | ascii_alphanumeric )
//! attribute_value ::= '"' *( line - '"' ) '"' | "'" *( line - "'" )  "'" | 1*( line - space_or_tab - '"' - "'" - '/' - '<' - '=' - '>' - '`')
//!
//! whitespace ::= 1*space_or_tab
//! whitespace_optional ::= [ whitespace ]
//! line ::= code - eol
//! eol ::= '\r' | '\r\n' | '\n'
//! space_or_tab ::= ' ' | '\t'
//! ```
//!
//! The grammar for HTML in markdown does not resemble the rules of parsing
//! HTML according to the [*§ 13.2 Parsing HTML documents* in the HTML
//! spec][html-parsing].
//! As such, HTML in markdown *resembles* HTML, but is instead a (naïve?)
//! attempt to parse an XML-like language.
//! By extension, another notable property of the grammar is that it can
//! result in invalid HTML, in that it allows things that wouldn’t work or
//! wouldn’t work well in HTML, such as mismatched tags.
//!
//! Interestingly, most of the productions above have a clear opening and
//! closing condition (raw, comment, insutrction, declaration, cdata), but the
//! closing condition does not need to be satisfied.
//! In this case, the parser never has to backtrack.
//!
//! Because the **basic** and **complete** productions in the grammar form with
//! a tag, followed by more stuff, and stop at a blank line, it is possible to
//! interleave (a word for switching between languages) markdown and HTML
//! together, by placing the opening and closing tags on their own lines,
//! with blank lines between them and markdown.
//! For example:
//!
//! ```markdown
//! <div>This is a <code>div</code> but *this* is not emphasis.</div>
//!
//! <div>
//!
//! This is a paragraph in a `div` and *this* is emphasis.
//!
//! </div>
//! ```
//!
//! The **complete** production of HTML (flow) is not allowed to interrupt
//! content.
//! That means that a blank line is needed between a [paragraph][] and it.
//! However, [HTML (text)][html_text] has a similar production, which will
//! typically kick-in instead.
//!
//! The list of tag names allowed in the **raw** production are defined in
//! [`HTML_RAW_NAMES`][html_raw_names].
//! This production exists because there are a few cases where markdown
//! *inside* some elements, and hence interleaving, does not make sense.
//!
//! The list of tag names allowed in the **basic** production are defined in
//! [`HTML_BLOCK_NAMES`][html_block_names].
//! This production exists because there are a few cases where we can decide
//! early that something is going to be a flow (block) element instead of a
//! phrasing (inline) element.
//! We *can* interrupt and don’t have to care too much about it being
//! well-formed.
//!
//! ## Tokens
//!
//! *   [`HtmlFlow`][Token::HtmlFlow]
//! *   [`HtmlFlowData`][Token::HtmlFlowData]
//! *   [`LineEnding`][Token::LineEnding]
//!
//! ## References
//!
//! *   [`html-flow.js` in `micromark`](https://github.com/micromark/micromark/blob/main/packages/micromark-core-commonmark/dev/lib/html-flow.js)
//! *   [*§ 4.6 HTML blocks* in `CommonMark`](https://spec.commonmark.org/0.30/#html-blocks)
//!
//! [flow]: crate::content::flow
//! [html_text]: crate::construct::html_text
//! [paragraph]: crate::construct::paragraph
//! [html_raw_names]: crate::constant::HTML_RAW_NAMES
//! [html_block_names]: crate::constant::HTML_BLOCK_NAMES
//! [html-parsing]: https://html.spec.whatwg.org/multipage/parsing.html#parsing

use crate::constant::{
    HTML_BLOCK_NAMES, HTML_CDATA_PREFIX, HTML_RAW_NAMES, HTML_RAW_SIZE_MAX, TAB_SIZE,
};
use crate::construct::{
    blank_line::start as blank_line,
    partial_non_lazy_continuation::start as partial_non_lazy_continuation,
    partial_space_or_tab::{space_or_tab_with_options, Options as SpaceOrTabOptions},
};
use crate::token::Token;
use crate::tokenizer::{State, Tokenizer};
use crate::util::slice::Slice;

/// Kind of HTML (flow).
#[derive(Debug, PartialEq)]
enum Kind {
    /// Symbol for `<script>` (condition 1).
    Raw,
    /// Symbol for `<!---->` (condition 2).
    Comment,
    /// Symbol for `<?php?>` (condition 3).
    Instruction,
    /// Symbol for `<!doctype>` (condition 4).
    Declaration,
    /// Symbol for `<![CDATA[]]>` (condition 5).
    Cdata,
    /// Symbol for `<div` (condition 6).
    Basic,
    /// Symbol for `<x>` (condition 7).
    Complete,
}

/// State needed to parse HTML (flow).
#[derive(Debug)]
struct Info {
    /// Kind of HTML (flow).
    kind: Kind,
    /// Whether this is a start tag (`<` not followed by `/`).
    start_tag: bool,
    /// Start index of a tag name or cdata prefix.
    start: usize,
    /// Current quote, when in a double or single quoted attribute value.
    quote: u8,
}

/// Start of HTML (flow), before optional whitespace.
///
/// ```markdown
/// > | <x />
///     ^
/// ```
pub fn start(tokenizer: &mut Tokenizer) -> State {
    if tokenizer.parse_state.constructs.html_flow {
        tokenizer.enter(Token::HtmlFlow);
        tokenizer.go(
            space_or_tab_with_options(SpaceOrTabOptions {
                kind: Token::HtmlFlowData,
                min: 0,
                max: if tokenizer.parse_state.constructs.code_indented {
                    TAB_SIZE - 1
                } else {
                    usize::MAX
                },
                connect: false,
                content_type: None,
            }),
            before,
        )(tokenizer)
    } else {
        State::Nok
    }
}

/// After optional whitespace, before `<`.
///
/// ```markdown
/// > | <x />
///     ^
/// ```
fn before(tokenizer: &mut Tokenizer) -> State {
    if Some(b'<') == tokenizer.current {
        tokenizer.enter(Token::HtmlFlowData);
        tokenizer.consume();
        State::Fn(Box::new(open))
    } else {
        State::Nok
    }
}

/// After `<`, before a tag name or other stuff.
///
/// ```markdown
/// > | <x />
///      ^
/// > | <!doctype>
///      ^
/// > | <!--xxx-->
///      ^
/// ```
fn open(tokenizer: &mut Tokenizer) -> State {
    let mut info = Info {
        // Assume basic.
        kind: Kind::Basic,
        // Assume closing tag (or no tag).
        start_tag: false,
        start: 0,
        quote: 0,
    };

    match tokenizer.current {
        Some(b'!') => {
            tokenizer.consume();
            State::Fn(Box::new(|t| declaration_open(t, info)))
        }
        Some(b'/') => {
            tokenizer.consume();
            info.start = tokenizer.point.index;
            State::Fn(Box::new(|t| tag_close_start(t, info)))
        }
        Some(b'?') => {
            info.kind = Kind::Instruction;
            tokenizer.consume();
            // Do not form containers.
            tokenizer.concrete = true;
            // While we’re in an instruction instead of a declaration, we’re on a `?`
            // right now, so we do need to search for `>`, similar to declarations.
            State::Fn(Box::new(|t| continuation_declaration_inside(t, info)))
        }
        // ASCII alphabetical.
        Some(b'A'..=b'Z' | b'a'..=b'z') => {
            info.start_tag = true;
            info.start = tokenizer.point.index;
            tag_name(tokenizer, info)
        }
        _ => State::Nok,
    }
}

/// After `<!`, so inside a declaration, comment, or CDATA.
///
/// ```markdown
/// > | <!doctype>
///       ^
/// > | <!--xxx-->
///       ^
/// > | <![CDATA[>&<]]>
///       ^
/// ```
fn declaration_open(tokenizer: &mut Tokenizer, mut info: Info) -> State {
    match tokenizer.current {
        Some(b'-') => {
            tokenizer.consume();
            info.kind = Kind::Comment;
            State::Fn(Box::new(|t| comment_open_inside(t, info)))
        }
        Some(b'A'..=b'Z' | b'a'..=b'z') => {
            tokenizer.consume();
            info.kind = Kind::Declaration;
            // Do not form containers.
            tokenizer.concrete = true;
            State::Fn(Box::new(|t| continuation_declaration_inside(t, info)))
        }
        Some(b'[') => {
            tokenizer.consume();
            info.kind = Kind::Cdata;
            info.start = tokenizer.point.index;
            State::Fn(Box::new(|t| cdata_open_inside(t, info)))
        }
        _ => State::Nok,
    }
}

/// After `<!-`, inside a comment, before another `-`.
///
/// ```markdown
/// > | <!--xxx-->
///        ^
/// ```
fn comment_open_inside(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        Some(b'-') => {
            tokenizer.consume();
            // Do not form containers.
            tokenizer.concrete = true;
            State::Fn(Box::new(|t| continuation_declaration_inside(t, info)))
        }
        _ => State::Nok,
    }
}

/// After `<![`, inside CDATA, expecting `CDATA[`.
///
/// ```markdown
/// > | <![CDATA[>&<]]>
///        ^^^^^^
/// ```
fn cdata_open_inside(tokenizer: &mut Tokenizer, mut info: Info) -> State {
    match tokenizer.current {
        Some(byte) if byte == HTML_CDATA_PREFIX[tokenizer.point.index - info.start] => {
            tokenizer.consume();

            if tokenizer.point.index - info.start == HTML_CDATA_PREFIX.len() {
                info.start = 0;
                // Do not form containers.
                tokenizer.concrete = true;
                State::Fn(Box::new(|t| continuation(t, info)))
            } else {
                State::Fn(Box::new(|t| cdata_open_inside(t, info)))
            }
        }
        _ => State::Nok,
    }
}

/// After `</`, in a closing tag, before a tag name.
///
/// ```markdown
/// > | </x>
///       ^
/// ```
fn tag_close_start(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        // ASCII alphabetical.
        Some(b'A'..=b'Z' | b'a'..=b'z') => {
            tokenizer.consume();
            State::Fn(Box::new(|t| tag_name(t, info)))
        }
        _ => State::Nok,
    }
}

/// In a tag name.
///
/// ```markdown
/// > | <ab>
///      ^^
/// > | </ab>
///       ^^
/// ```
fn tag_name(tokenizer: &mut Tokenizer, mut info: Info) -> State {
    match tokenizer.current {
        None | Some(b'\t' | b'\n' | b' ' | b'/' | b'>') => {
            let slash = matches!(tokenizer.current, Some(b'/'));
            // Guaranteed to be valid ASCII bytes.
            let slice = Slice::from_indices(
                tokenizer.parse_state.bytes,
                info.start,
                tokenizer.point.index,
            );
            let name = slice
                .as_str()
                // The line ending case might result in a `\r` that is already accounted for.
                .trim()
                .to_ascii_lowercase();
            info.start = 0;

            if !slash && info.start_tag && HTML_RAW_NAMES.contains(&name.as_str()) {
                info.kind = Kind::Raw;
                // Do not form containers.
                tokenizer.concrete = true;
                continuation(tokenizer, info)
            } else if HTML_BLOCK_NAMES.contains(&name.as_str()) {
                // Basic is assumed, no need to set `kind`.
                if slash {
                    tokenizer.consume();
                    State::Fn(Box::new(|t| basic_self_closing(t, info)))
                } else {
                    // Do not form containers.
                    tokenizer.concrete = true;
                    continuation(tokenizer, info)
                }
            } else {
                info.kind = Kind::Complete;

                // Do not support complete HTML when interrupting.
                if tokenizer.interrupt && !tokenizer.lazy {
                    State::Nok
                } else if info.start_tag {
                    complete_attribute_name_before(tokenizer, info)
                } else {
                    complete_closing_tag_after(tokenizer, info)
                }
            }
        }
        // ASCII alphanumerical and `-`.
        Some(b'-' | b'0'..=b'9' | b'A'..=b'Z' | b'a'..=b'z') => {
            tokenizer.consume();
            State::Fn(Box::new(|t| tag_name(t, info)))
        }
        Some(_) => State::Nok,
    }
}

/// After a closing slash of a basic tag name.
///
/// ```markdown
/// > | <div/>
///          ^
/// ```
fn basic_self_closing(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        Some(b'>') => {
            tokenizer.consume();
            // Do not form containers.
            tokenizer.concrete = true;
            State::Fn(Box::new(|t| continuation(t, info)))
        }
        _ => State::Nok,
    }
}

/// After a closing slash of a complete tag name.
///
/// ```markdown
/// > | <x/>
///        ^
/// ```
fn complete_closing_tag_after(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        Some(b'\t' | b' ') => {
            tokenizer.consume();
            State::Fn(Box::new(|t| complete_closing_tag_after(t, info)))
        }
        _ => complete_end(tokenizer, info),
    }
}

/// At a place where an attribute name would be valid.
///
/// At first, this state is used after a complete tag name, after whitespace,
/// where it expects optional attributes or the end of the tag.
/// It is also reused after attributes, when expecting more optional
/// attributes.
///
/// ```markdown
/// > | <a />
///        ^
/// > | <a :b>
///        ^
/// > | <a _b>
///        ^
/// > | <a b>
///        ^
/// > | <a >
///        ^
/// ```
fn complete_attribute_name_before(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        Some(b'\t' | b' ') => {
            tokenizer.consume();
            State::Fn(Box::new(|t| complete_attribute_name_before(t, info)))
        }
        Some(b'/') => {
            tokenizer.consume();
            State::Fn(Box::new(|t| complete_end(t, info)))
        }
        // ASCII alphanumerical and `:` and `_`.
        Some(b'0'..=b'9' | b':' | b'A'..=b'Z' | b'_' | b'a'..=b'z') => {
            tokenizer.consume();
            State::Fn(Box::new(|t| complete_attribute_name(t, info)))
        }
        _ => complete_end(tokenizer, info),
    }
}

/// In an attribute name.
///
/// ```markdown
/// > | <a :b>
///         ^
/// > | <a _b>
///         ^
/// > | <a b>
///         ^
/// ```
fn complete_attribute_name(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        // ASCII alphanumerical and `-`, `.`, `:`, and `_`.
        Some(b'-' | b'.' | b'0'..=b'9' | b':' | b'A'..=b'Z' | b'_' | b'a'..=b'z') => {
            tokenizer.consume();
            State::Fn(Box::new(|t| complete_attribute_name(t, info)))
        }
        _ => complete_attribute_name_after(tokenizer, info),
    }
}

/// After an attribute name, before an attribute initializer, the end of the
/// tag, or whitespace.
///
/// ```markdown
/// > | <a b>
///         ^
/// > | <a b=c>
///         ^
/// ```
fn complete_attribute_name_after(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        Some(b'\t' | b' ') => {
            tokenizer.consume();
            State::Fn(Box::new(|t| complete_attribute_name_after(t, info)))
        }
        Some(b'=') => {
            tokenizer.consume();
            State::Fn(Box::new(|t| complete_attribute_value_before(t, info)))
        }
        _ => complete_attribute_name_before(tokenizer, info),
    }
}

/// Before an unquoted, double quoted, or single quoted attribute value,
/// allowing whitespace.
///
/// ```markdown
/// > | <a b=c>
///          ^
/// > | <a b="c">
///          ^
/// ```
fn complete_attribute_value_before(tokenizer: &mut Tokenizer, mut info: Info) -> State {
    match tokenizer.current {
        None | Some(b'<' | b'=' | b'>' | b'`') => State::Nok,
        Some(b'\t' | b' ') => {
            tokenizer.consume();
            State::Fn(Box::new(|t| complete_attribute_value_before(t, info)))
        }
        Some(b'"' | b'\'') => {
            info.quote = tokenizer.current.unwrap();
            tokenizer.consume();
            State::Fn(Box::new(|t| complete_attribute_value_quoted(t, info)))
        }
        _ => complete_attribute_value_unquoted(tokenizer, info),
    }
}

/// In a double or single quoted attribute value.
///
/// ```markdown
/// > | <a b="c">
///           ^
/// > | <a b='c'>
///           ^
/// ```
fn complete_attribute_value_quoted(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        None | Some(b'\n') => State::Nok,
        Some(b'"' | b'\'') if tokenizer.current.unwrap() == info.quote => {
            tokenizer.consume();
            State::Fn(Box::new(|t| complete_attribute_value_quoted_after(t, info)))
        }
        _ => {
            tokenizer.consume();
            State::Fn(Box::new(|t| complete_attribute_value_quoted(t, info)))
        }
    }
}

/// In an unquoted attribute value.
///
/// ```markdown
/// > | <a b=c>
///          ^
/// ```
fn complete_attribute_value_unquoted(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        None | Some(b'\t' | b'\n' | b' ' | b'"' | b'\'' | b'/' | b'<' | b'=' | b'>' | b'`') => {
            complete_attribute_name_after(tokenizer, info)
        }
        Some(_) => {
            tokenizer.consume();
            State::Fn(Box::new(|t| complete_attribute_value_unquoted(t, info)))
        }
    }
}

/// After a double or single quoted attribute value, before whitespace or the
/// end of the tag.
///
/// ```markdown
/// > | <a b="c">
///            ^
/// ```
fn complete_attribute_value_quoted_after(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        Some(b'\t' | b' ' | b'/' | b'>') => complete_attribute_name_before(tokenizer, info),
        _ => State::Nok,
    }
}

/// In certain circumstances of a complete tag where only an `>` is allowed.
///
/// ```markdown
/// > | <a b="c">
///             ^
/// ```
fn complete_end(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        Some(b'>') => {
            tokenizer.consume();
            State::Fn(Box::new(|t| complete_after(t, info)))
        }
        _ => State::Nok,
    }
}

/// After `>` in a complete tag.
///
/// ```markdown
/// > | <x>
///        ^
/// ```
fn complete_after(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        None | Some(b'\n') => {
            // Do not form containers.
            tokenizer.concrete = true;
            continuation(tokenizer, info)
        }
        Some(b'\t' | b' ') => {
            tokenizer.consume();
            State::Fn(Box::new(|t| complete_after(t, info)))
        }
        Some(_) => State::Nok,
    }
}

/// Inside continuation of any HTML kind.
///
/// ```markdown
/// > | <!--xxx-->
///          ^
/// ```
fn continuation(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        Some(b'\n') if info.kind == Kind::Basic || info.kind == Kind::Complete => {
            tokenizer.exit(Token::HtmlFlowData);
            tokenizer.check(blank_line_before, |ok| {
                if ok {
                    Box::new(continuation_after)
                } else {
                    Box::new(move |t| continuation_start(t, info))
                }
            })(tokenizer)
        }
        // Note: important that this is after the basic/complete case.
        None | Some(b'\n') => {
            tokenizer.exit(Token::HtmlFlowData);
            continuation_start(tokenizer, info)
        }
        Some(b'-') if info.kind == Kind::Comment => {
            tokenizer.consume();
            State::Fn(Box::new(|t| continuation_comment_inside(t, info)))
        }
        Some(b'<') if info.kind == Kind::Raw => {
            tokenizer.consume();
            State::Fn(Box::new(|t| continuation_raw_tag_open(t, info)))
        }
        Some(b'>') if info.kind == Kind::Declaration => {
            tokenizer.consume();
            State::Fn(Box::new(|t| continuation_close(t, info)))
        }
        Some(b'?') if info.kind == Kind::Instruction => {
            tokenizer.consume();
            State::Fn(Box::new(|t| continuation_declaration_inside(t, info)))
        }
        Some(b']') if info.kind == Kind::Cdata => {
            tokenizer.consume();
            State::Fn(Box::new(|t| continuation_character_data_inside(t, info)))
        }
        _ => {
            tokenizer.consume();
            State::Fn(Box::new(|t| continuation(t, info)))
        }
    }
}

/// In continuation, at an eol.
///
/// ```markdown
/// > | <x>
///        ^
///   | asd
/// ```
fn continuation_start(tokenizer: &mut Tokenizer, info: Info) -> State {
    tokenizer.check(partial_non_lazy_continuation, |ok| {
        if ok {
            Box::new(move |t| continuation_start_non_lazy(t, info))
        } else {
            Box::new(continuation_after)
        }
    })(tokenizer)
}

/// In continuation, at an eol, before non-lazy content.
///
/// ```markdown
/// > | <x>
///        ^
///   | asd
/// ```
fn continuation_start_non_lazy(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        Some(b'\n') => {
            tokenizer.enter(Token::LineEnding);
            tokenizer.consume();
            tokenizer.exit(Token::LineEnding);
            State::Fn(Box::new(|t| continuation_before(t, info)))
        }
        _ => unreachable!("expected eol"),
    }
}

/// In continuation, after an eol, before non-lazy content.
///
/// ```markdown
///   | <x>
/// > | asd
///     ^
/// ```
fn continuation_before(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        None | Some(b'\n') => continuation_start(tokenizer, info),
        _ => {
            tokenizer.enter(Token::HtmlFlowData);
            continuation(tokenizer, info)
        }
    }
}

/// In comment continuation, after one `-`, expecting another.
///
/// ```markdown
/// > | <!--xxx-->
///             ^
/// ```
fn continuation_comment_inside(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        Some(b'-') => {
            tokenizer.consume();
            State::Fn(Box::new(|t| continuation_declaration_inside(t, info)))
        }
        _ => continuation(tokenizer, info),
    }
}

/// In raw continuation, after `<`, expecting a `/`.
///
/// ```markdown
/// > | <script>console.log(1)</script>
///                            ^
/// ```
fn continuation_raw_tag_open(tokenizer: &mut Tokenizer, mut info: Info) -> State {
    match tokenizer.current {
        Some(b'/') => {
            tokenizer.consume();
            info.start = tokenizer.point.index;
            State::Fn(Box::new(|t| continuation_raw_end_tag(t, info)))
        }
        _ => continuation(tokenizer, info),
    }
}

/// In raw continuation, after `</`, expecting or inside a raw tag name.
///
/// ```markdown
/// > | <script>console.log(1)</script>
///                             ^^^^^^
/// ```
fn continuation_raw_end_tag(tokenizer: &mut Tokenizer, mut info: Info) -> State {
    match tokenizer.current {
        Some(b'>') => {
            // Guaranteed to be valid ASCII bytes.
            let slice = Slice::from_indices(
                tokenizer.parse_state.bytes,
                info.start,
                tokenizer.point.index,
            );
            let name = slice.as_str().to_ascii_lowercase();

            info.start = 0;

            if HTML_RAW_NAMES.contains(&name.as_str()) {
                tokenizer.consume();
                State::Fn(Box::new(|t| continuation_close(t, info)))
            } else {
                continuation(tokenizer, info)
            }
        }
        Some(b'A'..=b'Z' | b'a'..=b'z')
            if tokenizer.point.index - info.start < HTML_RAW_SIZE_MAX =>
        {
            tokenizer.consume();
            State::Fn(Box::new(|t| continuation_raw_end_tag(t, info)))
        }
        _ => {
            info.start = 0;
            continuation(tokenizer, info)
        }
    }
}

/// In cdata continuation, after `]`, expecting `]>`.
///
/// ```markdown
/// > | <![CDATA[>&<]]>
///                  ^
/// ```
fn continuation_character_data_inside(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        Some(b']') => {
            tokenizer.consume();
            State::Fn(Box::new(|t| continuation_declaration_inside(t, info)))
        }
        _ => continuation(tokenizer, info),
    }
}

/// In declaration or instruction continuation, waiting for `>` to close it.
///
/// ```markdown
/// > | <!-->
///         ^
/// > | <?>
///       ^
/// > | <!q>
///        ^
/// > | <!--ab-->
///             ^
/// > | <![CDATA[>&<]]>
///                   ^
/// ```
fn continuation_declaration_inside(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        Some(b'>') => {
            tokenizer.consume();
            State::Fn(Box::new(|t| continuation_close(t, info)))
        }
        Some(b'-') if info.kind == Kind::Comment => {
            tokenizer.consume();
            State::Fn(Box::new(|t| continuation_declaration_inside(t, info)))
        }
        _ => continuation(tokenizer, info),
    }
}

/// In closed continuation: everything we get until the eol/eof is part of it.
///
/// ```markdown
/// > | <!doctype>
///               ^
/// ```
fn continuation_close(tokenizer: &mut Tokenizer, info: Info) -> State {
    match tokenizer.current {
        None | Some(b'\n') => {
            tokenizer.exit(Token::HtmlFlowData);
            continuation_after(tokenizer)
        }
        _ => {
            tokenizer.consume();
            State::Fn(Box::new(|t| continuation_close(t, info)))
        }
    }
}

/// Done.
///
/// ```markdown
/// > | <!doctype>
///               ^
/// ```
fn continuation_after(tokenizer: &mut Tokenizer) -> State {
    tokenizer.exit(Token::HtmlFlow);
    // Feel free to interrupt.
    tokenizer.interrupt = false;
    // No longer concrete.
    tokenizer.concrete = false;
    State::Ok
}

/// Before a line ending, expecting a blank line.
///
/// ```markdown
/// > | <div>
///          ^
///   |
/// ```
fn blank_line_before(tokenizer: &mut Tokenizer) -> State {
    tokenizer.enter(Token::LineEnding);
    tokenizer.consume();
    tokenizer.exit(Token::LineEnding);
    State::Fn(Box::new(blank_line))
}