Server IP : 85.214.239.14 / Your IP : 3.148.105.127 Web Server : Apache/2.4.62 (Debian) System : Linux h2886529.stratoserver.net 4.9.0 #1 SMP Tue Jan 9 19:45:01 MSK 2024 x86_64 User : www-data ( 33) PHP Version : 7.4.18 Disable Function : pcntl_alarm,pcntl_fork,pcntl_waitpid,pcntl_wait,pcntl_wifexited,pcntl_wifstopped,pcntl_wifsignaled,pcntl_wifcontinued,pcntl_wexitstatus,pcntl_wtermsig,pcntl_wstopsig,pcntl_signal,pcntl_signal_get_handler,pcntl_signal_dispatch,pcntl_get_last_error,pcntl_strerror,pcntl_sigprocmask,pcntl_sigwaitinfo,pcntl_sigtimedwait,pcntl_exec,pcntl_getpriority,pcntl_setpriority,pcntl_async_signals,pcntl_unshare, MySQL : OFF | cURL : OFF | WGET : ON | Perl : ON | Python : ON | Sudo : ON | Pkexec : OFF Directory : /proc/2/root/usr/lib/python3/dist-packages/markdown_it/ |
Upload File : |
"""Block-level tokenizer.""" from __future__ import annotations import logging from . import rules_block from .ruler import Ruler from .rules_block.state_block import StateBlock from .token import Token LOGGER = logging.getLogger(__name__) _rules: list[tuple] = [ # First 2 params - rule name & source. Secondary array - list of rules, # which can be terminated by this one. ("table", rules_block.table, ["paragraph", "reference"]), ("code", rules_block.code), ("fence", rules_block.fence, ["paragraph", "reference", "blockquote", "list"]), ( "blockquote", rules_block.blockquote, ["paragraph", "reference", "blockquote", "list"], ), ("hr", rules_block.hr, ["paragraph", "reference", "blockquote", "list"]), ("list", rules_block.list_block, ["paragraph", "reference", "blockquote"]), ("reference", rules_block.reference), ("html_block", rules_block.html_block, ["paragraph", "reference", "blockquote"]), ("heading", rules_block.heading, ["paragraph", "reference", "blockquote"]), ("lheading", rules_block.lheading), ("paragraph", rules_block.paragraph), ] class ParserBlock: """ ParserBlock#ruler -> Ruler [[Ruler]] instance. Keep configuration of block rules. """ def __init__(self): self.ruler = Ruler() for data in _rules: name = data[0] rule = data[1] self.ruler.push(name, rule, {"alt": data[2] if len(data) > 2 else []}) def tokenize( self, state: StateBlock, startLine: int, endLine: int, silent: bool = False ) -> None: """Generate tokens for input range.""" rules = self.ruler.getRules("") line = startLine maxNesting = state.md.options.maxNesting hasEmptyLines = False while line < endLine: state.line = line = state.skipEmptyLines(line) if line >= endLine: break if state.sCount[line] < state.blkIndent: # Termination condition for nested calls. # Nested calls currently used for blockquotes & lists break if state.level >= maxNesting: # If nesting level exceeded - skip tail to the end. # That's not ordinary situation and we should not care about content. state.line = endLine break # Try all possible rules. # On success, rule should: # - update `state.line` # - update `state.tokens` # - return True for rule in rules: if rule(state, line, endLine, False): break # set state.tight if we had an empty line before current tag # i.e. latest empty line should not count state.tight = not hasEmptyLines line = state.line # paragraph might "eat" one newline after it in nested lists if (line - 1) < endLine and state.isEmpty(line - 1): hasEmptyLines = True if line < endLine and state.isEmpty(line): hasEmptyLines = True line += 1 state.line = line def parse( self, src: str, md, env, outTokens: list[Token], ords: tuple[int, ...] | None = None, ) -> list[Token] | None: """Process input string and push block tokens into `outTokens`.""" if not src: return None state = StateBlock(src, md, env, outTokens, ords) self.tokenize(state, state.line, state.lineMax) return state.tokens