#!/usr/bin/env python """ Python Markdown =============== Python Markdown converts Markdown to HTML and can be used as a library or called from the command line. ## Basic usage as a module: import markdown md = Markdown() html = md.convert(your_text_string) ## Basic use from the command line: python markdown.py source.txt > destination.html Run "python markdown.py --help" to see more options. ## Extensions See for more information and instructions on how to extend the functionality of Python Markdown. Read that before you try modifying this file. ## Authors and License Started by [Manfred Stienstra](http://www.dwerg.net/). Continued and maintained by [Yuri Takhteyev](http://www.freewisdom.org), [Waylan Limberg](http://achinghead.com/) and [Artem Yunusov](http://blog.splyer.com). Contact: markdown@freewisdom.org Copyright 2007, 2008 The Python Markdown Project (v. 1.7 and later) Copyright 2004, 2005, 2006 Yuri Takhteyev (v. 0.2-1.6b) Copyright 2004 Manfred Stienstra (the original version) License: BSD (see docs/LICENSE for details). """ version = "2.0-alpha" version_info = (2,0,0, "beta") import re import sys import codecs import htmlentitydefs import logging from logging import DEBUG, INFO, WARN, ERROR, CRITICAL from urlparse import urlparse, urlunparse """ CONSTANTS ============================================================================= """ """ Constants you might want to modify ----------------------------------------------------------------------------- """ # default logging level for command-line use COMMAND_LINE_LOGGING_LEVEL = CRITICAL TAB_LENGTH = 4 # expand tabs to this many spaces ENABLE_ATTRIBUTES = True # @id = xyz -> <... id="xyz"> SMART_EMPHASIS = True # this_or_that does not become thisorthat HTML_REMOVED_TEXT = "[HTML_REMOVED]" # text used instead of HTML in safe mode BLOCK_LEVEL_ELEMENTS = re.compile("p|div|h[1-6]|blockquote|pre|table|dl|ol|ul" +"|script|noscript|form|fieldset|iframe|math" +"|ins|del|hr|hr/|style|li|tr") """ Constants you probably do not need to change ----------------------------------------------------------------------------- """ RTL_BIDI_RANGES = ( (u'\u0590', u'\u07FF'), # Hebrew (0590-05FF), Arabic (0600-06FF), # Syriac (0700-074F), Arabic supplement (0750-077F), # Thaana (0780-07BF), Nko (07C0-07FF). (u'\u2D30', u'\u2D7F'), # Tifinagh ) EXECUTABLE_NAME_FOR_USAGE = "python markdown.py" """ The name used in the usage statement displayed for python versions < 2.3. (With python 2.3 and higher the usage statement is generated by optparse and uses the actual name of the executable called.) """ # Placeholders STX = u'\u0002' # Use STX ("Start of text") for start-of-placeholder ETX = u'\u0003' # Use ETX ("End of text") for end-of-placeholder HTML_PLACEHOLDER_PREFIX = STX+"wzxhzdk:" HTML_PLACEHOLDER = HTML_PLACEHOLDER_PREFIX + "%d" + ETX INLINE_PLACEHOLDER_PREFIX = STX+"klzzwxh:" INLINE_PLACEHOLDER = INLINE_PLACEHOLDER_PREFIX + "%s" + ETX AMP_SUBSTITUTE = STX+"amp"+ETX def wrapRe(raw_re) : return re.compile("^%s$" % raw_re, re.DOTALL) CORE_RE = { 'header': wrapRe(r'(#{1,6})[ \t]*(.*?)[ \t]*(#*)'), # # A title 'reference-def': wrapRe(r'(\ ?\ ?\ ?)\[([^\]]*)\]:\s*([^ ]*)(.*)'), # [Google]: http://www.google.com/ 'containsline': wrapRe(r'([-]*)$|^([=]*)'), # -----, =====, etc. 'ol': wrapRe(r'[ ]{0,3}[\d]*\.\s+(.*)'), # 1. text 'ul': wrapRe(r'[ ]{0,3}[*+-]\s+(.*)'), # "* text" 'isline1': wrapRe(r'(\**)'), # *** 'isline2': wrapRe(r'(\-*)'), # --- 'isline3': wrapRe(r'(\_*)'), # ___ 'tabbed': wrapRe(r'((\t)|( ))(.*)'), # an indented line 'quoted': wrapRe(r'[ ]{0,2}> ?(.*)'), # a quoted block ("> ...") 'containsline': re.compile(r'^([-]*)$|^([=]*)$', re.M), 'attr': re.compile("\{@([^\}]*)=([^\}]*)}") # {@id=123} } """Basic and reusable regular expressions.""" """ AUXILIARY GLOBAL FUNCTIONS ============================================================================= """ def message(level, text): """ A wrapper method for logging debug messages. """ logging.getLogger('MARKDOWN').log(level, text) def isString(s): """ Check if it's string """ return isinstance(s, unicode) or isinstance(s, str) ## Import def importETree(): """Import the best implementation of ElementTree, return a module object.""" etree_in_c = None try: # Is it Python 2.5+ with C implemenation of ElementTree installed? import xml.etree.cElementTree as etree_in_c except ImportError: try: # Is it Python 2.5+ with Python implementation of ElementTree? import xml.etree.ElementTree as etree except ImportError: try: # An earlier version of Python with cElementTree installed? import cElementTree as etree_in_c except ImportError: try: # An earlier version of Python with Python ElementTree? import elementtree.ElementTree as etree except ImportError: message(CRITICAL, "Failed to import ElementTree") sys.exit(1) if etree_in_c and etree_in_c.VERSION < "1.0": message(CRITICAL, "For cElementTree version 1.0 or higher is required.") sys.exit(1) elif etree_in_c : return etree_in_c elif etree.VERSION < "1.1": message(CRITICAL, "For ElementTree version 1.1 or higher is required") sys.exit(1) else : return etree def isBlockLevel(tag): """Check if the tag is a block level HTML tag.""" return BLOCK_LEVEL_ELEMENTS.match(tag) def handleAttributes(text, parent): """Set values of an element based on attribute definitions ({@id=123}).""" def attributeCallback(match): parent.set(match.group(1), match.group(2)) return CORE_RE['attr'].sub(attributeCallback, text) def dequote(string): """Remove quotes from around a string.""" if ( ( string.startswith('"') and string.endswith('"')) or (string.startswith("'") and string.endswith("'")) ): return string[1:-1] else: return string """ OVERALL DESIGN ============================================================================= Markdown processing takes place in three steps: 1. A bunch of "preprocessors" munge the input text. 2. MarkdownParser() parses the high-level structural elements of the pre-processed text into an ElementTree. 3. A bunch of Patterns are run against the ElementTree, detecting inline markup. 4. Some extra use-defined post-processors are run. 5. The output is written to a string. Those steps are put together by the Markdown() class. The code below is organized as follows: 1. MarkdownParser class - does basic parsing. 2. All the post-processors, patterns, etc. 3. Markdown class - does the high-level wrapping. """ """ CORE MARKDOWN PARSER ============================================================================= This class handles basic Markdown parsing. It doesn't concern itself with inline elements such as **bold** or *italics*, but rather just catches blocks, lists, quotes, etc. """ class MarkdownParser: """Parser Markdown into a ElementTree.""" def __init__(self): pass def parseDocument(self, lines): """Parse a markdown string into an ElementTree.""" # Create a ElementTree from the lines root = etree.Element("div") buffer = [] for line in lines: if line.startswith("#"): self.parseChunk(root, buffer) buffer = [line] else: buffer.append(line) self.parseChunk(root, buffer) return etree.ElementTree(root) def parseChunk(self, parent_elem, lines, inList=0, looseList=0): """Process a chunk of markdown-formatted text and attach the parse to an ElementTree node. Process a section of a source document, looking for high level structural elements like lists, block quotes, code segments, html blocks, etc. Some those then get stripped of their high level markup (e.g. get unindented) and the lower-level markup is processed recursively. Keyword arguments: * parent_elem: A ElementTree element to which the content will be added. * lines: a list of lines * inList: a level Returns: None """ # Loop through lines until none left. while lines: # Skipping empty line if not lines[0]: lines = lines[1:] continue # Check if this section starts with a list, a blockquote or # a code block processFn = { 'ul': self._processUList, 'ol': self._processOList, 'quoted': self._processQuote, 'tabbed': self._processCodeBlock} for regexp in ['ul', 'ol', 'quoted', 'tabbed']: m = CORE_RE[regexp].match(lines[0]) if m: processFn[regexp](parent_elem, lines, inList) return # We are NOT looking at one of the high-level structures like # lists or blockquotes. So, it's just a regular paragraph # (though perhaps nested inside a list or something else). If # we are NOT inside a list, we just need to look for a blank # line to find the end of the block. If we ARE inside a # list, however, we need to consider that a sublist does not # need to be separated by a blank line. Rather, the following # markup is legal: # # * The top level list item # # Another paragraph of the list. This is where we are now. # * Underneath we might have a sublist. # if inList: start, lines = self._linesUntil(lines, (lambda line: CORE_RE['ul'].match(line) or CORE_RE['ol'].match(line) or not line.strip())) self.parseChunk(parent_elem, start, inList-1, looseList=looseList) inList = inList-1 else: # Ok, so it's just a simple block paragraph, lines = self._linesUntil(lines, lambda line: not line.strip() or line[0] == '>') if len(paragraph) and paragraph[0].startswith('#'): self._processHeader(parent_elem, paragraph) elif len(paragraph) and \ CORE_RE["isline3"].match(paragraph[0]): self._processHR(parent_elem) lines = paragraph[1:] + lines elif paragraph: self._processParagraph(parent_elem, paragraph, inList, looseList) if lines and not lines[0].strip(): lines = lines[1:] # skip the first (blank) line def _processHR(self, parentElem): hr = etree.SubElement(parentElem, "hr") def _processHeader(self, parentElem, paragraph): m = CORE_RE['header'].match(paragraph[0]) if m: level = len(m.group(1)) h = etree.SubElement(parentElem, "h%d" % level) h.text = m.group(2).strip() else: message(CRITICAL, "We've got a problem header!") def _processParagraph(self, parentElem, paragraph, inList, looseList): if ( parentElem.tag == 'li' and not (looseList or parentElem.getchildren())): # If this is the first paragraph inside "li", don't # put

around it - append the paragraph bits directly # onto parentElem el = parentElem else: # Otherwise make a "p" element el = etree.SubElement(parentElem, "p") dump = [] # Searching for hr or header for line in paragraph: # it's hr if CORE_RE["isline3"].match(line): el.text = "\n".join(dump) self._processHR(el) dump = [] # it's header elif line.startswith("#"): el.text = "\n".join(dump) self._processHeader(parentElem, [line]) dump = [] else: dump.append(line) if dump: text = "\n".join(dump) el.text = text def _processUList(self, parentElem, lines, inList): self._processList(parentElem, lines, inList, listexpr='ul', tag = 'ul') def _processOList(self, parentElem, lines, inList): self._processList(parentElem, lines, inList, listexpr='ol', tag = 'ol') def _processList(self, parentElem, lines, inList, listexpr, tag): """ Given a list of document lines starting with a list item, finds the end of the list, breaks it up, and recursively processes each list item and the remainder of the text file. Keyword arguments: * parentElem: A ElementTree element to which the content will be added * lines: a list of lines * inList: a level Returns: None """ ul = etree.SubElement(parentElem, tag) # ul might actually be '

    ' looseList = 0 # Make a list of list items items = [] item = -1 i = 0 # a counter to keep track of where we are for line in lines: loose = 0 if not line.strip(): # If we see a blank line, this _might_ be the end of the list i += 1 loose = 1 # Find the next non-blank line for j in range(i, len(lines)): if lines[j].strip(): next = lines[j] break else: # There is no more text => end of the list break # Check if the next non-blank line is still a part of the list if ( CORE_RE[listexpr].match(next) or CORE_RE['tabbed'].match(next) ): # get rid of any white space in the line items[item].append(line.strip()) looseList = loose or looseList continue else: break # found end of the list # Now we need to detect list items (at the current level) # while also detabing child elements if necessary for expr in ['ul', 'ol', 'tabbed']: m = CORE_RE[expr].match(line) if m: if expr in ['ul', 'ol']: # We are looking at a new item #if m.group(1) : # Removed the check to allow for a blank line # at the beginning of the list item items.append([m.group(1)]) item += 1 elif expr == 'tabbed': # This line needs to be detabbed items[item].append(m.group(4)) #after the 'tab' i += 1 break else: items[item].append(line) # Just regular continuation i += 1 # added on 2006.02.25 else: i += 1 # Add the ElementTree elements for item in items: li = etree.SubElement(ul, "li") self.parseChunk(li, item, inList + 1, looseList = looseList) # Process the remaining part of the section self.parseChunk(parentElem, lines[i:], inList) def _linesUntil(self, lines, condition): """ A utility function to break a list of lines upon the first line that satisfied a condition. The condition argument should be a predicate function. """ i = -1 for line in lines: i += 1 if condition(line): break else: i += 1 return lines[:i], lines[i:] def _processQuote(self, parentElem, lines, inList): """ Given a list of document lines starting with a quote finds the end of the quote, unindents it and recursively processes the body of the quote and the remainder of the text file. Keyword arguments: * parentElem: ElementTree element to which the content will be added * lines: a list of lines * inList: a level Returns: None """ dequoted = [] i = 0 blank_line = False # allow one blank line between paragraphs for line in lines: m = CORE_RE['quoted'].match(line) if m: dequoted.append(m.group(1)) i += 1 blank_line = False elif not blank_line and line.strip() != '': dequoted.append(line) i += 1 elif not blank_line and line.strip() == '': dequoted.append(line) i += 1 blank_line = True else: break blockquote = etree.SubElement(parentElem, "blockquote") self.parseChunk(blockquote, dequoted, inList) self.parseChunk(parentElem, lines[i:], inList) def _processCodeBlock(self, parentElem, lines, inList): """ Given a list of document lines starting with a code block finds the end of the block, puts it into the ElementTree verbatim wrapped in ("
    ") and recursively processes the
            the remainder of the text file.
    
            Keyword arguments:
            
            * parentElem: ElementTree element to which the content will be added
            * lines: a list of lines
            * inList: a level
            
            Returns: None
            
            """
            detabbed, theRest = self.detectTabbed(lines)
    
            pre = etree.SubElement(parentElem, "pre")
            code = etree.SubElement(pre, "code")
            
            text = "\n".join(detabbed).rstrip()+"\n"
            code.text = AtomicString(text)
            self.parseChunk(parentElem, theRest, inList)        
    
        def detectTabbed(self, lines):
            """ Find indented text and remove indent before further proccesing.
    
            Keyword arguments:
            
            * lines: an array of strings
            * fn: a function that returns a substring of a string
               if the string matches the necessary criteria
            
            Returns: a list of post processes items and the unused
            remainder of the original list
            
            """
            items = []
            item = -1
            i = 0 # to keep track of where we are
    
            def detab(line):
                match = CORE_RE['tabbed'].match(line)
                if match:
                   return match.group(4)
    
            for line in lines:
                if line.strip(): # Non-blank line
                    line = detab(line)
                    if line:
                        items.append(line)
                        i += 1
                        continue
                    else:
                        return items, lines[i:]
    
                else: # Blank line: _maybe_ we are done.
                    i += 1 # advance
    
                    # Find the next non-blank line
                    for j in range(i, len(lines)):  
                        if lines[j].strip():
                            next_line = lines[j]; break
                    else:
                        break # There is no more text; we are done.
    
                    # Check if the next non-blank line is tabbed
                    if detab(next_line): # Yes, more work to do.
                        items.append("")
                        continue
                    else:
                        break # No, we are done.
            else:
                i += 1
    
            return items, lines[i:]
    
    
    
    
    """
    PRE-PROCESSORS
    =============================================================================
    
    Preprocessors work on source text before we start doing anything too
    complicated.  There are two types of preprocessors: TextPreprocessor and
    Preprocessor.
    """
    
    class TextPreprocessor:
        """
        TextPreprocessors are run before the text is broken into lines.
        
        Each TextPreprocessor implements a "run" method that takes a pointer to a
        text string of the document, modifies it as necessary and returns
        either the same pointer or a pointer to a new string.  
        
        TextPreprocessors must extend markdown.TextPreprocessor.
    
        """
    
        def run(self, text):
            """ 
            Each subclass of TextPreprocessor should override the `run` method, 
            which takes the document text as a single string and returns the 
            (possibly modified) document as a single string.
            
            """
            pass
    
    
    class Preprocessor:
        """
        Preprocessors are run after the text is broken into lines.
    
        Each preprocessor implements a "run" method that takes a pointer to a
        list of lines of the document, modifies it as necessary and returns
        either the same pointer or a pointer to a new list.  
        
        Preprocessors must extend markdown.Preprocessor.
        
        """
    
        def run(self, lines):
            """
            Each subclass of Preprocessor should override the `run` method, which
            takes the document as a list of strings split by newlines and returns
            the (possibly modified) list of lines.
    
            """
            pass
     
    
    class HtmlBlockPreprocessor(TextPreprocessor):
        """Remove html blocks from the text and store them for later retrieval."""
    
        right_tag_patterns = ["", "%s>"]
        
        def _get_left_tag(self, block):
            return block[1:].replace(">", " ", 1).split()[0].lower()
    
        def _get_right_tag(self, left_tag, block):        
            for p in self.right_tag_patterns:
                tag = p % left_tag
                i = block.rfind(tag)
                if i > 2:
                    return tag.lstrip("<").rstrip(">"), i + len(p)-2 + len(left_tag)
            return block.rstrip()[-len(left_tag)-2:-1].lower(), len(block)
    
        def _equal_tags(self, left_tag, right_tag):
            if left_tag == 'div' or left_tag[0] in ['?', '@', '%']: # handle PHP, etc.
                return True
            if ("/" + left_tag) == right_tag:
                return True
            if (right_tag == "--" and left_tag == "--"):
                return True
            elif left_tag == right_tag[1:] \
                and right_tag[0] != "<":
                return True
            else:
                return False
    
        def _is_oneliner(self, tag):
            return (tag in ['hr', 'hr/'])
    
        def run(self, text):
            new_blocks = []
            text = text.split("\n\n")        
            items = []
            left_tag = ''
            right_tag = ''
            in_tag = False # flag
    
            while text:
                block = text[0]
                if block.startswith("\n"):
                    block = block[1:]
                text = text[1:]
                
                if block.startswith("\n"):
                    block = block[1:]
    
                if not in_tag:
                    if block.startswith("<"):
                        left_tag = self._get_left_tag(block)
                        right_tag, data_index = self._get_right_tag(left_tag, block)
                        
                        if data_index < len(block):
                            text.insert(0, block[data_index:])
                            block = block[:data_index]
    
                        if not (isBlockLevel(left_tag) \
                            or block[1] in ["!", "?", "@", "%"]):
                            new_blocks.append(block)
                            continue
    
                        if self._is_oneliner(left_tag):
                            new_blocks.append(block.strip())
                            continue
                            
                        if block[1] == "!":
                            # is a comment block
                            left_tag = "--"
                            right_tag, data_index = self._get_right_tag(left_tag, block)
                            # keep checking conditions below and maybe just append
                            
                        if block.rstrip().endswith(">") \
                            and self._equal_tags(left_tag, right_tag):
                            new_blocks.append(
                                self.stash.store(block.strip()))
                            continue
                        else: #if not block[1] == "!":
                            # if is block level tag and is not complete
                            
                            if isBlockLevel(left_tag) or left_tag == "--" \
                            and not block.rstrip().endswith(">"):
                                items.append(block.strip())
                                in_tag = True
                            else:
                                new_blocks.append(
                                self.stash.store(block.strip()))
                                
                            continue
    
                    new_blocks.append(block)
    
                else:
                    items.append(block.strip())
                    
                    right_tag, data_index = self._get_right_tag(left_tag, block)
                    
                    if self._equal_tags(left_tag, right_tag):
                        # if find closing tag
                        in_tag = False
                        new_blocks.append(
                            self.stash.store('\n\n'.join(items)))
                        items = []
    
            if items:
                new_blocks.append(self.stash.store('\n\n'.join(items)))
                new_blocks.append('\n')
                
            return "\n\n".join(new_blocks)
    
    HTML_BLOCK_PREPROCESSOR = HtmlBlockPreprocessor()
    
    
    class HeaderPreprocessor(Preprocessor):
    
        """Replace underlined headers with hashed headers.
    
        (To avoid the need for lookahead later.)
    
        """
    
        def run (self, lines):
            i = -1
            while i+1 < len(lines):
                i = i+1
                if not lines[i].strip():
                    continue
    
                if lines[i].startswith("#"):
                    lines.insert(i+1, "\n")
    
                if (i+1 <= len(lines)
                      and lines[i+1]
                      and lines[i+1][0] in ['-', '=']):
    
                    underline = lines[i+1].strip()
    
                    if underline == "="*len(underline):
                        lines[i] = "# " + lines[i].strip()
                        lines[i+1] = ""
                    elif underline == "-"*len(underline):
                        lines[i] = "## " + lines[i].strip()
                        lines[i+1] = ""
    
            return lines
    
    HEADER_PREPROCESSOR = HeaderPreprocessor()
    
    
    class LinePreprocessor(Preprocessor):
        """Convert HR lines to "___" format."""
        blockquote_re = re.compile(r'^(> )+')
    
        def run (self, lines):
            for i in range(len(lines)):
                prefix = ''
                m = self.blockquote_re.search(lines[i])
                if m: 
                    prefix = m.group(0)
                if self._isLine(lines[i][len(prefix):]):
                    lines[i] = prefix + "___"
            return lines
    
        def _isLine(self, block):
            """Determine if a block should be replaced with an 
    """ if block.startswith(" "): return False # a code block text = "".join([x for x in block if not x.isspace()]) if len(text) <= 2: return False for pattern in ['isline1', 'isline2', 'isline3']: m = CORE_RE[pattern].match(text) if (m and m.group(1)): return True else: return False LINE_PREPROCESSOR = LinePreprocessor() class ReferencePreprocessor(Preprocessor): """Remove reference definitions from the text and store them for later use.""" def run (self, lines): new_text = []; for line in lines: m = CORE_RE['reference-def'].match(line) if m: id = m.group(2).strip().lower() t = m.group(4).strip() # potential title if not t: self.references[id] = (m.group(3), t) elif (len(t) >= 2 and (t[0] == t[-1] == "\"" or t[0] == t[-1] == "\'" or (t[0] == "(" and t[-1] == ")") ) ): self.references[id] = (m.group(3), t[1:-1]) else: new_text.append(line) else: new_text.append(line) return new_text #+ "\n" REFERENCE_PREPROCESSOR = ReferencePreprocessor() """ INLINE PATTERNS ============================================================================= Inline patterns such as *emphasis* are handled by means of auxiliary objects, one per pattern. Pattern objects must be instances of classes that extend markdown.Pattern. Each pattern object uses a single regular expression and needs support the following methods: pattern.getCompiledRegExp() # returns a regular expression pattern.handleMatch(m) # takes a match object and returns # an ElementTree element or just plain text All of python markdown's built-in patterns subclass from Pattern, but you can add additional patterns that don't. Also note that all the regular expressions used by inline must capture the whole block. For this reason, they all start with '^(.*)' and end with '(.*)!'. In case with built-in expression Pattern takes care of adding the "^(.*)" and "(.*)!". Finally, the order in which regular expressions are applied is very important - e.g. if we first replace http://.../ links with tags and _then_ try to replace inline html, we would end up with a mess. So, we apply the expressions in the following order: * escape and backticks have to go before everything else, so that we can preempt any markdown patterns by escaping them. * then we handle auto-links (must be done before inline html) * then we handle inline HTML. At this point we will simply replace all inline HTML strings with a placeholder and add the actual HTML to a hash. * then inline images (must be done before links) * then bracketed links, first regular then reference-style * finally we apply strong and emphasis """ """ The actual regular expressions for patterns ----------------------------------------------------------------------------- """ NOBRACKET = r'[^\]\[]*' BRK = ( r'\[(' + (NOBRACKET + r'(\[')*6 + (NOBRACKET+ r'\])*')*6 + NOBRACKET + r')\]' ) NOIMG = r'(?|((?:(?:\(.*?\))|[^\(\)]))*?)\s*((['"])(.*)\12)?\)''' # [text](url) or [text]() IMAGE_LINK_RE = r'\!' + BRK + r'\s*\((<.*?>|([^\)]*))\)' # ![alttxt](http://x.com/) or ![alttxt]() REFERENCE_RE = NOIMG + BRK+ r'\s*\[([^\]]*)\]' # [Google][3] IMAGE_REFERENCE_RE = r'\!' + BRK + '\s*\[([^\]]*)\]' # ![alt text][2] NOT_STRONG_RE = r'( \* )' # stand-alone * or _ AUTOLINK_RE = r'<((?:f|ht)tps?://[^>]*)>' # AUTOMAIL_RE = r'<([^> \!]*@[^> ]*)>' # HTML_RE = r'(\<([a-zA-Z/][^\>]*?|\!--.*?--)\>)' # <...> ENTITY_RE = r'(&[\#a-zA-Z0-9]*;)' # & LINE_BREAK_RE = r' \n' # two spaces at end of line LINE_BREAK_2_RE = r' $' # two spaces at end of text """ The pattern classes ----------------------------------------------------------------------------- """ class Pattern: """Base class that inline patterns subclass. """ def __init__ (self, pattern): """ Create an instant of an inline pattern. Keyword arguments: * pattern: A regular expression that matches a pattern """ self.pattern = pattern self.compiled_re = re.compile("^(.*?)%s(.*?)$" % pattern, re.DOTALL) # Api for Markdown to pass safe_mode into instance self.safe_mode = False def getCompiledRegExp (self): """ Return a compiled regular expression. """ return self.compiled_re def handleMatch(self, m): """Return a ElementTree element from the given match. Subclasses should override this method. Keyword arguments: * m: A re match object containing a match of the pattern. """ pass def type(self): """ Return class name, to define pattern type """ return self.__class__.__name__ BasePattern = Pattern # for backward compatibility class SimpleTextPattern (Pattern): """ Return a simple text of group(2) of a Pattern. """ def handleMatch(self, m): text = m.group(2) if text == INLINE_PLACEHOLDER_PREFIX: return None return text class SimpleTagPattern (Pattern): """ Return element of type `tag` with a text attribute of group(3) of a Pattern. """ def __init__ (self, pattern, tag): Pattern.__init__(self, pattern) self.tag = tag def handleMatch(self, m): el = etree.Element(self.tag) el.text = m.group(3) return el class SubstituteTagPattern (SimpleTagPattern): """ Return a eLement of type `tag` with no children. """ def handleMatch (self, m): return etree.Element(self.tag) class BacktickPattern (Pattern): """ Return a `` element containing the matching text. """ def __init__ (self, pattern): Pattern.__init__(self, pattern) self.tag = "code" def handleMatch(self, m): el = etree.Element(self.tag) el.text = AtomicString(m.group(3).strip()) return el class DoubleTagPattern (SimpleTagPattern): """Return a ElementTree element nested in tag2 nested in tag1. Useful for strong emphasis etc. """ def handleMatch(self, m): tag1, tag2 = self.tag.split(",") el1 = etree.Element(tag1) el2 = etree.SubElement(el1, tag2) el2.text = m.group(3) return el1 class HtmlPattern (Pattern): """ Store raw inline html and return a placeholder. """ def handleMatch (self, m): rawhtml = m.group(2) inline = True place_holder = self.stash.store(rawhtml) return place_holder class LinkPattern (Pattern): """ Return a link element from the given match. """ def handleMatch(self, m): el = etree.Element("a") el.text = m.group(2) title = m.group(11) href = m.group(9) if href: if href[0] == "<": href = href[1:-1] el.set("href", self.sanitize_url(href.strip())) else: el.set("href", "") if title: title = dequote(title) #.replace('"', """) el.set("title", title) return el def sanitize_url(self, url): """ Sanitize a url against xss attacks in "safe_mode". Rather than specifically blacklisting `javascript:alert("XSS")` and all its aliases (see ), we whitelist known safe url formats. Most urls contain a network location, however some are known not to (i.e.: mailto links). Script urls do not contain a location. Additionally, for `javascript:...`, the scheme would be "javascript" but some aliases will appear to `urlparse()` to have no scheme. On top of that relative links (i.e.: "foo/bar.html") have no scheme. Therefore we must check "path", "parameters", "query" and "fragment" for any literal colons. We don't check "scheme" for colons because it *should* never have any and "netloc" must allow the form: `username:password@host:port`. """ locless_schemes = ['', 'mailto', 'news'] scheme, netloc, path, params, query, fragment = url = urlparse(url) safe_url = False if netloc != '' or scheme in locless_schemes: safe_url = True for part in url[2:]: if ":" in part: safe_url = False if self.safe_mode and not safe_url: return '' else: return urlunparse(url) class ImagePattern(LinkPattern): """ Return a img element from the given match. """ def handleMatch(self, m): el = etree.Element("img") src_parts = m.group(9).split() if src_parts: src = src_parts[0] if src[0] == "<" and src[-1] == ">": src = src[1:-1] el.set('src', self.sanitize_url(src)) else: el.set('src', "") if len(src_parts) > 1: el.set('title', dequote(" ".join(src_parts[1:]))) if ENABLE_ATTRIBUTES: truealt = handleAttributes(m.group(2), el) else: truealt = m.group(2) el.set('alt', truealt) return el class ReferencePattern(LinkPattern): """ Match to a stored reference and return link element. """ def handleMatch(self, m): if m.group(9): id = m.group(9).lower() else: # if we got something like "[Google][]" # we'll use "google" as the id id = m.group(2).lower() if not self.references.has_key(id): # ignore undefined refs return None href, title = self.references[id] text = m.group(2) return self.makeTag(href, title, text) def makeTag(self, href, title, text): el = etree.Element('a') el.set('href', self.sanitize_url(href)) if title: el.set('title', title) el.text = text return el class ImageReferencePattern (ReferencePattern): """ Match to a stored reference and return img element. """ def makeTag(self, href, title, text): el = etree.Element("img") el.set("src", self.sanitize_url(href)) if title: el.set("title", title) el.set("alt", text) return el class AutolinkPattern (Pattern): """ Return a link Element given an autolink (``). """ def handleMatch(self, m): el = etree.Element("a") el.set('href', m.group(2)) el.text = AtomicString(m.group(2)) return el class AutomailPattern (Pattern): """ Return a mailto link Element given an automail link (``). """ def handleMatch(self, m): el = etree.Element('a') email = m.group(2) if email.startswith("mailto:"): email = email[len("mailto:"):] def codepoint2name(code): """Return entity definition by code, or the code if not defined.""" entity = htmlentitydefs.codepoint2name.get(code) if entity: return "%s%s;" % (AMP_SUBSTITUTE, entity) else: return "%s#%d;" % (AMP_SUBSTITUTE, code) letters = [codepoint2name(ord(letter)) for letter in email] el.text = AtomicString(''.join(letters)) mailto = "mailto:" + email mailto = "".join([AMP_SUBSTITUTE + '#%d;' % ord(letter) for letter in mailto]) el.set('href', mailto) return el ESCAPE_PATTERN = SimpleTextPattern(ESCAPE_RE) NOT_STRONG_PATTERN = SimpleTextPattern(NOT_STRONG_RE) BACKTICK_PATTERN = BacktickPattern(BACKTICK_RE) STRONG_PATTERN = SimpleTagPattern(STRONG_RE, 'strong') EMPHASIS_PATTERN = SimpleTagPattern(EMPHASIS_RE, 'em') EMPHASIS_PATTERN_2 = SimpleTagPattern(EMPHASIS_2_RE, 'em') STRONG_EM_PATTERN = DoubleTagPattern(STRONG_EM_RE, 'strong,em') LINE_BREAK_PATTERN = SubstituteTagPattern(LINE_BREAK_RE, 'br') LINE_BREAK_PATTERN_2 = SubstituteTagPattern(LINE_BREAK_2_RE, 'br') LINK_PATTERN = LinkPattern(LINK_RE) IMAGE_LINK_PATTERN = ImagePattern(IMAGE_LINK_RE) IMAGE_REFERENCE_PATTERN = ImageReferencePattern(IMAGE_REFERENCE_RE) REFERENCE_PATTERN = ReferencePattern(REFERENCE_RE) HTML_PATTERN = HtmlPattern(HTML_RE) ENTITY_PATTERN = HtmlPattern(ENTITY_RE) AUTOLINK_PATTERN = AutolinkPattern(AUTOLINK_RE) AUTOMAIL_PATTERN = AutomailPattern(AUTOMAIL_RE) """ POST-PROCESSORS ============================================================================= Markdown also allows post-processors, which are similar to preprocessors in that they need to implement a "run" method. However, they are run after core processing. There are two types of post-processors: Postprocessor and TextPostprocessor """ class Postprocessor: """ Postprocessors are run before the ElementTree serialization. Each Postprocessor implements a "run" method that takes a pointer to a ElementTree, modifies it as necessary and returns a ElementTree document. Postprocessors must extend markdown.Postprocessor. """ def run(self, root): """ Subclasses of Postprocessor should implement a `run` method, which takes a root Element. Method can return another Element, and global root Element will be replaced, or just modify current and return None. """ pass class TextPostprocessor: """ TextPostprocessors are run after the ElementTree it converted back into text. Each TextPostprocessor implements a "run" method that takes a pointer to a text string, modifies it as necessary and returns a text string. TextPostprocessors must extend markdown.TextPostprocessor. """ def run(self, text): """ Subclasses of TextPostprocessor should implement a `run` method, which takes the html document as a single text string and returns a (possibly modified) string. """ pass class PrettifyPostprocessor(Postprocessor): """Add linebreaks to the html document.""" def _prettifyETree(self, elem): """Recursively add linebreaks to ElementTree children.""" i = "\n" if isBlockLevel(elem.tag) and elem.tag not in ['code', 'pre']: if (not elem.text or not elem.text.strip()) \ and len(elem) and isBlockLevel(elem[0].tag): elem.text = i for e in elem: if isBlockLevel(e.tag): self._prettifyETree(e) if not elem.tail or not elem.tail.strip(): elem.tail = i if not elem.tail or not elem.tail.strip(): elem.tail = i def run(self, root): """.Add linebreaks to ElementTree root object.""" self._prettifyETree(root) # Do
    's seperately as they are often in the middle of # inline content and missed by _prettifyETree. brs = root.getiterator('br') for br in brs: if not br.tail or not br.tail.strip(): br.tail = '\n' else: br.tail = '\n%s' % br.tail PRETTIFYPOSTPROCESSOR = PrettifyPostprocessor() class RawHtmlTextPostprocessor(TextPostprocessor): """ Restore raw html to the document. """ def __init__(self): pass def run(self, text): """ Iterate over html stash and restore "safe" html. """ for i in range(self.stash.html_counter): html, safe = self.stash.rawHtmlBlocks[i] if self.safeMode and not safe: if str(self.safeMode).lower() == 'escape': html = self.escape(html) elif str(self.safeMode).lower() == 'remove': html = '' else: html = HTML_REMOVED_TEXT if safe or not self.safeMode: text = text.replace("

    %s

    " % (HTML_PLACEHOLDER % i), html + "\n") text = text.replace(HTML_PLACEHOLDER % i, html) return text def escape(self, html): """ Basic html escaping """ html = html.replace('&', '&') html = html.replace('<', '<') html = html.replace('>', '>') return html.replace('"', '"') RAWHTMLTEXTPOSTPROCESSOR = RawHtmlTextPostprocessor() class AndSubstitutePostprocessor(TextPostprocessor): """ Restore valid entities """ def __init__(self): pass def run(self, text): text = text.replace(AMP_SUBSTITUTE, "&") return text AMPSUBSTITUTETEXTPOSTPROCESSOR = AndSubstitutePostprocessor() """ MISC AUXILIARY CLASSES ============================================================================= """ class AtomicString(unicode): """A string which should not be further processed.""" pass class HtmlStash: """ This class is used for stashing HTML objects that we extract in the beginning and replace with place-holders. """ def __init__ (self): """ Create a HtmlStash. """ self.html_counter = 0 # for counting inline html segments self.rawHtmlBlocks=[] def store(self, html, safe=False): """ Saves an HTML segment for later reinsertion. Returns a placeholder string that needs to be inserted into the document. Keyword arguments: * html: an html segment * safe: label an html segment as safe for safemode Returns : a placeholder string """ self.rawHtmlBlocks.append((html, safe)) placeholder = HTML_PLACEHOLDER % self.html_counter self.html_counter += 1 return placeholder def rest(self): self.html_counter = 0 self.rawHtmlBlocks = [] class InlineStash: def __init__(self): """ Create a InlineStash. """ self.prefix = INLINE_PLACEHOLDER_PREFIX self.suffix = ETX self._nodes = {} self.phLength = 4 + len(self.prefix) + len(self.suffix) self._placeholder_re = re.compile(INLINE_PLACEHOLDER % r'([0-9]{4})') def _genPlaceholder(self, type): """ Generate a placeholder """ id = "%04d" % len(self._nodes) hash = INLINE_PLACEHOLDER % id return hash, id def extractId(self, data, index): """ Extract id from data string, start from index Keyword arguments: * data: string * index: index, from which we start search Returns: placeholder id and string index, after found placeholder """ m = self._placeholder_re.search(data, index) if m: return m.group(1), m.end() else: return None, index + 1 def isin(self, id): """ Check if node with given id exists in stash """ return self._nodes.has_key(id) def get(self, id): """ Return node by id """ return self._nodes.get(id) def add(self, node, type): """ Add node to stash """ pholder, id = self._genPlaceholder(type) self._nodes[id] = node return pholder def rest(self): """ Reset instance """ self._nodes = {} class Markdown: """Convert Markdown to HTML.""" def __init__(self, extensions=[], extension_configs={}, safe_mode = False): """ Creates a new Markdown instance. Keyword arguments: * extensions: A list of extensions. If they are of type string, the module mdx_name.py will be loaded. If they are a subclass of markdown.Extension, they will be used as-is. * extension-configs: Configuration setting for extensions. * safe_mode: Disallow raw html. One of "remove", "replace" or "escape". """ self.parser = MarkdownParser() self.safeMode = safe_mode self.registeredExtensions = [] self.docType = "" self.stripTopLevelTags = True self.textPreprocessors = [HTML_BLOCK_PREPROCESSOR] self.preprocessors = [HEADER_PREPROCESSOR, LINE_PREPROCESSOR, # A footnote preprocessor will # get inserted here REFERENCE_PREPROCESSOR] self.postprocessors = [PRETTIFYPOSTPROCESSOR, # a footnote postprocessor will get # inserted later ] self.textPostprocessors = [# a footnote postprocessor will get # inserted here RAWHTMLTEXTPOSTPROCESSOR, AMPSUBSTITUTETEXTPOSTPROCESSOR] self.prePatterns = [] self.inlinePatterns = [ BACKTICK_PATTERN, ESCAPE_PATTERN, REFERENCE_PATTERN, LINK_PATTERN, IMAGE_LINK_PATTERN, IMAGE_REFERENCE_PATTERN, AUTOLINK_PATTERN, AUTOMAIL_PATTERN, LINE_BREAK_PATTERN_2, LINE_BREAK_PATTERN, HTML_PATTERN, ENTITY_PATTERN, NOT_STRONG_PATTERN, STRONG_EM_PATTERN, STRONG_PATTERN, EMPHASIS_PATTERN, EMPHASIS_PATTERN_2 # The order of the handlers matters!!! ] self.inlineStash = InlineStash() self.references = {} self.htmlStash = HtmlStash() self.registerExtensions(extensions = extensions, configs = extension_configs) self.reset() def registerExtensions(self, extensions, configs): """ Register extensions with this instance of Markdown. Keyword aurguments: * extensions: A list of extensions, which can either be strings or objects. See the docstring on Markdown. * configs: A dictionary mapping module names to config options. """ for ext in extensions: if isinstance(ext, basestring): ext = load_extension(ext, configs.get(ext, [])) elif hasattr(ext, 'extendMarkdown'): # Looks like an Extension. # Nothing to do here. pass else: message(ERROR, "Incorrect type! Extension '%s' is " "neither a string or an Extension." %(repr(ext))) continue ext.extendMarkdown(self, globals()) def registerExtension(self, extension): """ This gets called by the extension """ self.registeredExtensions.append(extension) def reset(self): """ Resets all state variables so that we can start with a new text. """ self.inlineStash.rest() self.htmlStash.rest() self.references.clear() HTML_BLOCK_PREPROCESSOR.stash = self.htmlStash LINE_PREPROCESSOR.stash = self.htmlStash REFERENCE_PREPROCESSOR.references = self.references HTML_PATTERN.stash = self.htmlStash ENTITY_PATTERN.stash = self.htmlStash REFERENCE_PATTERN.references = self.references IMAGE_REFERENCE_PATTERN.references = self.references RAWHTMLTEXTPOSTPROCESSOR.stash = self.htmlStash RAWHTMLTEXTPOSTPROCESSOR.safeMode = self.safeMode for extension in self.registeredExtensions: extension.reset() for pattern in self.inlinePatterns: pattern.safe_mode = self.safeMode def _handleInline(self, data, patternIndex=0): """ Process string with inline patterns and replace it with placeholders Keyword arguments: * data: A line of Markdown text * patternIndex: The index of the inlinePattern to start with Returns: String with placeholders. """ if isinstance(data, AtomicString): return data startIndex = 0 while patternIndex < len(self.inlinePatterns): data, matched, startIndex = self._applyInline( self.inlinePatterns[patternIndex], data, patternIndex, startIndex) if not matched: patternIndex += 1 return data def _processElementText(self, node, subnode, isText=True): """ Process placeholders in Element.text or Element.tail of Elements popped from InlineStash Keywords arguments: * node: parent node * subnode: processing node * isText: bool variable, True - it's text, False - it's tail Returns: None """ if isText: text = subnode.text subnode.text = None else: text = subnode.tail subnode.tail = None childResult = self._processPlaceholders(text, subnode) if not isText and node is not subnode: pos = node.getchildren().index(subnode) node.remove(subnode) else: pos = 0 childResult.reverse() for newChild in childResult: node.insert(pos, newChild) def _processPlaceholders(self, data, parent): """ Process string with placeholders and generate ElementTree tree. Keyword arguments: * data: string with placeholders instead of ElementTree elements. * parent: Element, which contains processing inline data Returns: list with ElementTree elements with applied inline patterns. """ def linkText(text): if text: if result: if result[-1].tail: result[-1].tail += text else: result[-1].tail = text else: if parent.text: parent.text += text else: parent.text = text result = [] prefix = self.inlineStash.prefix strartIndex = 0 while data: index = data.find(prefix, strartIndex) if index != -1: id, phEndIndex = self.inlineStash.extractId(data, index) if self.inlineStash.isin(id): node = self.inlineStash.get(id) if index > 0: text = data[strartIndex:index] linkText(text) if not isString(node): # it's Element for child in [node] + node.getchildren(): if child.tail: if child.tail.strip(): self._processElementText(node, child, False) if child.text: if child.text.strip(): self._processElementText(child, child) else: # it's just a string linkText(node) strartIndex = phEndIndex continue strartIndex = phEndIndex result.append(node) else: # wrong placeholder end = index + len(prefix) linkText(data[strartIndex:end]) strartIndex = end else: text = data[strartIndex:] linkText(text) data = "" return result def _applyInline(self, pattern, data, patternIndex, startIndex=0): """ Check if the line fits the pattern, create the necessary elements, add it to InlineStash Keyword arguments: * data: the text to be processed * pattern: the pattern to be checked * patternIndex: index of current pattern * startIndex: string index, from which we starting search Returns: String with placeholders instead of ElementTree elements. """ match = pattern.getCompiledRegExp().match(data[startIndex:]) leftData = data[:startIndex] if not match: return data, False, 0 node = pattern.handleMatch(match) if node is None: return data, True, len(leftData) + match.span(len(match.groups()))[0] if not isString(node): if not isinstance(node.text, AtomicString): # We need to process current node too for child in [node] + node.getchildren(): if not isString(node): if child.text: child.text = self._handleInline(child.text, patternIndex + 1) if child.tail: child.tail = self._handleInline(child.tail, patternIndex) pholder = self.inlineStash.add(node, pattern.type()) return "%s%s%s%s" % (leftData, match.group(1), pholder, match.groups()[-1]), True, 0 def applyInlinePatterns(self, markdownTree): """ Iterate over ElementTree, find elements with inline tag, apply inline patterns and append newly created Elements to tree. If you don't want process your data with inline paterns, instead of normal string, use subclass AtomicString: node.text = AtomicString("data won't be processed with inline patterns") Arguments: * markdownTree: ElementTree object, representing Markdown tree. Returns: ElementTree object with applied inline patterns. """ stack = [markdownTree.getroot()] while stack: currElement = stack.pop() insertQueue = [] for child in currElement.getchildren(): if child.text and not isinstance(child.text, AtomicString): text = child.text child.text = None lst = self._processPlaceholders(self._handleInline( text), child) stack += lst insertQueue.append((child, lst)) if child.getchildren(): stack.append(child) for element, lst in insertQueue: if element.text: element.text = handleAttributes(element.text, element) i = 0 for newChild in lst: # Processing attributes if newChild.tail: newChild.tail = handleAttributes(newChild.tail, element) if newChild.text: newChild.text = handleAttributes(newChild.text, newChild) element.insert(i, newChild) i += 1 return markdownTree def convert (self, source): """Convert markdown to serialized XHTML.""" # Fixup the source text if not source: return u"" # a blank unicode string try: source = unicode(source) except UnicodeDecodeError: message(CRITICAL, 'UnicodeDecodeError: Markdown only accepts unicode or ascii input.') return u"" source = source.replace(STX, "") source = source.replace(ETX, "") source = source.replace("\r\n", "\n").replace("\r", "\n") source += "\n\n" source = source.expandtabs(TAB_LENGTH) # Run the text preprocessors for pp in self.textPreprocessors: source = pp.run(source) # Split into lines and run the line preprocessors. self.lines = source.split("\n") for prep in self.preprocessors : self.lines = prep.run(self.lines) # Parse the high-level elements. tree = self.parser.parseDocument(self.lines) root = self.applyInlinePatterns(tree).getroot() # Run the post-processors for postprocessor in self.postprocessors: postprocessor.stash = self.htmlStash newRoot = postprocessor.run(root) if newRoot: root = newRoot # Serialize _properly_. Strip top-level tags. xml, length = codecs.utf_8_decode(etree.tostring(root, encoding="utf8")) if self.stripTopLevelTags: xml = xml.strip()[44:-7] + "\n" # Run the text post-processors for pp in self.textPostprocessors: xml = pp.run(xml) return xml.strip() def convertFile(input = None, output = None, encoding = None): """Converts a markdown file and returns the HTML as a unicode string. Decodes the file using the provided encoding (defaults to utf-8), passes the file content to markdown, and outputs the html to either the provided stream or the file with provided name, using the same encoding as the source file. **Note:** This is the only place that decoding and encoding of unicode takes place in Python-Markdown. (All other code is unicode-in / unicode-out.) Keyword arguments: * input: Name of source text file. * output: Name of output file. Writes to stdout if `None`. * extensions: A list of extension names (may contain config args). * encoding: Encoding of input and output files. Defaults to utf-8. * safe_mode: Disallow raw html. One of "remove", "replace" or "escape". """ encoding = encoding or "utf-8" # Read the source input_file = codecs.open(input, mode="r", encoding=encoding) text = input_file.read() input_file.close() text = text.lstrip(u'\ufeff') # remove the byte-order mark # Convert html = self.convert(text) # Write to file or stdout if type(output) == type("string"): output_file = codecs.open(output, "w", encoding=encoding) output_file.write(html) output_file.close() else: output.write(html.encode(encoding)) """ Extensions ----------------------------------------------------------------------------- """ class Extension: """ Base class for extensions to subclass. """ def __init__(self, configs = {}): """Create an instance of an Extention. Keyword arguments: * configs: A dict of configuration setting used by an Extension. """ self.config = configs def getConfig(self, key): """ Return a setting for the given key or an empty string. """ if self.config.has_key(key): return self.config[key][0] else: return "" def getConfigInfo(self): """ Return all config settings as a list of tuples. """ return [(key, self.config[key][1]) for key in self.config.keys()] def setConfig(self, key, value): """ Set a config setting for `key` with the given `value`. """ self.config[key][0] = value def extendMarkdown(self, md, md_globals): """ Add the various proccesors and patterns to the Markdown Instance. This method must be overriden by every extension. Keyword arguments: * md: The Markdown instance. * md_globals: Global variables in the markdown module namespace. """ pass def load_extension(ext_name, configs = []): """Load extension by name, then return the module. The extension name may contain arguments as part of the string in the following format: "extname(key1=value1,key2=value2)" """ # Parse extensions config params (ignore the order) configs = dict(configs) pos = ext_name.find("(") # find the first "(" if pos > 0: ext_args = ext_name[pos+1:-1] ext_name = ext_name[:pos] pairs = [x.split("=") for x in ext_args.split(",")] configs.update([(x.strip(), y.strip()) for (x, y) in pairs]) # Setup the module names ext_module = 'markdown_extensions' module_name_new_style = '.'.join([ext_module, ext_name]) module_name_old_style = '_'.join(['mdx', ext_name]) # Try loading the extention first from one place, then another try: # New style (markdown_extensons.) module = __import__(module_name_new_style, {}, {}, [ext_module]) except ImportError: try: # Old style (mdx.) module = __import__(module_name_old_style) except ImportError: pass if module : # If the module is loaded successfully, we expect it to define a # function called makeExtension() try: return module.makeExtension(configs.items()) except: message(WARN, "Failed to instantiate extension '%s'" % ext_name) else: message(WARN, "Failed loading extension '%s' from '%s' or '%s'" % (ext_name, module_name_new_style, module_name_old_style)) def load_extensions(ext_names): """Loads multiple extensions""" extensions = [] for ext_name in ext_names: extension = load_extension(ext_name) if extension: extensions.append(extension) # Extensions should use "markdown.etree" instead of "etree" (or do `from # markdown import etree`). Do not import it by yourself. etree = importETree() """ EXPORTED FUNCTIONS ============================================================================= Those are the two functions we really mean to export: markdown() and markdownFromFile(). """ def markdown(text, extensions = [], safe_mode = False): """Convert a markdown string to HTML and return HTML as a unicode string. This is a shortcut function for `Markdown` class to cover the most basic use case. It initializes an instance of Markdown, loads the necessary extensions and runs the parser on the given text. Keyword arguments: * text: Markdown formatted text as Unicode or ASCII string. * extensions: A list of extensions or extension names (may contain config args). * safe_mode: Disallow raw html. One of "remove", "replace" or "escape". Returns: An HTML document as a string. """ md = Markdown(extensions=load_extensions(extensions), safe_mode = safe_mode) return md.convert(text) def markdownFromFile(input = None, output = None, extensions = [], encoding = None, safe = False): md = Markdown(extensions=load_extensions(extensions), safe_mode = safe_mode) md.convertFile(input, output, encoding) """ COMMAND-LINE SPECIFIC STUFF ============================================================================= The rest of the code is specifically for handling the case where Python Markdown is called from the command line. """ OPTPARSE_WARNING = """ Python 2.3 or higher required for advanced command line options. For lower versions of Python use: %s INPUT_FILE > OUTPUT_FILE """ % EXECUTABLE_NAME_FOR_USAGE def parse_options(): """ Define and parse `optparse` options for command-line usage. """ try: optparse = __import__("optparse") except: if len(sys.argv) == 2: return {'input': sys.argv[1], 'output': None, 'safe': False, 'extensions': [], 'encoding': None }, CRITICAL else: print OPTPARSE_WARNING return None, None parser = optparse.OptionParser(usage="%prog INPUTFILE [options]") parser.add_option("-f", "--file", dest="filename", help="write output to OUTPUT_FILE", metavar="OUTPUT_FILE") parser.add_option("-e", "--encoding", dest="encoding", help="encoding for input and output files",) parser.add_option("-q", "--quiet", default = CRITICAL, action="store_const", const=CRITICAL+10, dest="verbose", help="suppress all messages") parser.add_option("-v", "--verbose", action="store_const", const=INFO, dest="verbose", help="print info messages") parser.add_option("-s", "--safe", dest="safe", default=False, metavar="SAFE_MODE", help="safe mode ('replace', 'remove' or 'escape' user's HTML tag)") parser.add_option("--noisy", action="store_const", const=DEBUG, dest="verbose", help="print debug messages") parser.add_option("-x", "--extension", action="append", dest="extensions", help = "load extension EXTENSION", metavar="EXTENSION") (options, args) = parser.parse_args() if not len(args) == 1: parser.print_help() return None, None else: input_file = args[0] if not options.extensions: options.extensions = [] return {'input': input_file, 'output': options.filename, 'safe': options.safe, 'extensions': options.extensions, 'encoding': options.encoding }, options.verbose def command_line_run(): """Run Markdown from the command line.""" # Setup a logger manually for compatibility with Python 2.3 logger = logging.getLogger('MARKDOWN') logger.setLevel(COMMAND_LINE_LOGGING_LEVEL) logger.addHandler(logging.StreamHandler()) # Parse options and adjust logging level if necessary options, logging_level = parse_options() if not options: sys.exit(0) if logging_level: logging.getLogger('MARKDOWN').setLevel(logging_level) # Run markdownFromFile(**options) if __name__ == '__main__': command_line_run()