]> oss.titaniummirror.com Git - webber.git/blobdiff - plugins/read_markdown.py
Merge branch 'upstream'
[webber.git] / plugins / read_markdown.py
index 17e9c84a8eec23b27465542e92c515d02e04b0bd..d8006a27474c42e9eff7902c354f47f9f656013c 100644 (file)
 # -*- coding: iso-8859-1 -*-
 from webber import *
-from markdown2 import *
 
-# This code uses markdown2.py: # http://code.google.com/p/python-markdown2/
+# based on code from http://code.google.com/p/python-markdown2/
 # Copyright (c) 2007-2008 ActiveState Corp.
 # License: MIT (http://www.opensource.org/licenses/mit-license.php)
-#      
+#
+# I used version 1.0.1.16, but deleted:
+#      * file-vars (emacs-style settings inside the file)
+#      * Standardize line endings
+#      * call to _do_links()
+#      * logging
+#      * allow "= Header =" in addition to "# Header #"
+#
+
+import os, sys, re, codecs
+try:
+       from hashlib import md5
+except ImportError:
+       from md5 import md5
+from random import random, randint
+
+
+
+#---- Python version compat
+
+if sys.version_info[:2] < (2,4):
+       from sets import Set as set
+       def reversed(sequence):
+               for i in sequence[::-1]:
+                       yield i
+       def _unicode_decode(s, encoding, errors='xmlcharrefreplace'):
+               return unicode(s, encoding, errors)
+else:
+       def _unicode_decode(s, encoding, errors='strict'):
+               return s.decode(encoding, errors)
+
+
+#---- globals
+
+DEBUG = False
+
+DEFAULT_TAB_WIDTH = 4
+
+
+try:
+       import uuid
+except ImportError:
+       SECRET_SALT = str(randint(0, 1000000))
+else:
+       SECRET_SALT = str(uuid.uuid4())
+def _hash_ascii(s):
+       #return md5(s).hexdigest()       # Markdown.pl effectively does this.
+       return 'md5-' + md5(SECRET_SALT + s).hexdigest()
+def _hash_text(s):
+       return 'md5-' + md5(SECRET_SALT + s.encode("utf-8")).hexdigest()
+
+# Table of hash values for escaped characters:
+g_escape_table = dict([(ch, _hash_ascii(ch))
+                                          for ch in '\\`*_{}[]()>#+-.!'])
+
+
+
+#---- exceptions
+
+class MarkdownError(Exception):
+       pass
+
+
+
+#---- public api
+
+def markdown(text, html4tags=False, tab_width=DEFAULT_TAB_WIDTH,
+                        safe_mode=None, extras=None, link_patterns=None):
+       return Markdown(html4tags=html4tags, tab_width=tab_width,
+                                       safe_mode=safe_mode, extras=extras,
+                                       link_patterns=link_patterns).convert(text)
+
+class Markdown(object):
+       # The dict of "extras" to enable in processing -- a mapping of
+       # extra name to argument for the extra. Most extras do not have an
+       # argument, in which case the value is None.
+       #
+       # This can be set via (a) subclassing and (b) the constructor
+       # "extras" argument.
+       extras = None
+
+       urls = None
+       titles = None
+       html_blocks = None
+       html_spans = None
+       html_removed_text = "[HTML_REMOVED]"  # for compat with markdown.py
+
+       # Used to track when we're inside an ordered or unordered list
+       # (see _ProcessListItems() for details):
+       list_level = 0
+
+       _ws_only_line_re = re.compile(r"^[ \t]+$", re.M)
+
+       def __init__(self, html4tags=False, tab_width=4, safe_mode=None,
+                                extras=None, link_patterns=None):
+               if html4tags:
+                       self.empty_element_suffix = ">"
+               else:
+                       self.empty_element_suffix = " />"
+               self.tab_width = tab_width
+
+               # For compatibility with earlier markdown2.py and with
+               # markdown.py's safe_mode being a boolean,
+               #       safe_mode == True -> "replace"
+               if safe_mode is True:
+                       self.safe_mode = "replace"
+               else:
+                       self.safe_mode = safe_mode
+
+               if self.extras is None:
+                       self.extras = {}
+               elif not isinstance(self.extras, dict):
+                       self.extras = dict([(e, None) for e in self.extras])
+               if extras:
+                       if not isinstance(extras, dict):
+                               extras = dict([(e, None) for e in extras])
+                       self.extras.update(extras)
+               assert isinstance(self.extras, dict)
+               self._instance_extras = self.extras.copy()
+               self.link_patterns = link_patterns
+               self._outdent_re = re.compile(r'^(\t|[ ]{1,%d})' % tab_width, re.M)
+
+       def reset(self):
+               self.urls = {}
+               self.titles = {}
+               self.html_blocks = {}
+               self.html_spans = {}
+               self.list_level = 0
+               self.extras = self._instance_extras.copy()
+               self.encoding = 'utf-8'
+               if "footnotes" in self.extras:
+                       self.footnotes = {}
+                       self.footnote_ids = []
+
+       def convert(self, text, encoding=None):
+               """Convert the given text."""
+               # Main function. The order in which other subs are called here is
+               # essential. Link and image substitutions need to happen before
+               # _EscapeSpecialChars(), so that any *'s or _'s in the <a>
+               # and <img> tags get encoded.
+
+               # Clear the global hashes. If we don't clear these, you get conflicts
+               # from other articles when generating a page which contains more than
+               # one article (e.g. an index page that shows the N most recent
+               # articles):
+               self.reset()
+               if encoding:
+                       self.encoding = encoding
+
+               if not isinstance(text, unicode):
+                       text = unicode(text, self.encoding)
+
+               # Standardize line endings:
+               #text = re.sub("\r\n|\r", "\n", text)
+
+               # Make sure $text ends with a couple of newlines:
+               text += "\n\n"
+
+               # Convert all tabs to spaces.
+               text = self._detab(text)
+
+               # Strip any lines consisting only of spaces and tabs.
+               # This makes subsequent regexen easier to write, because we can
+               # match consecutive blank lines with /\n+/ instead of something
+               # contorted like /[ \t]*\n+/ .
+               text = self._ws_only_line_re.sub("", text)
+
+               if self.safe_mode:
+                       text = self._hash_html_spans(text)
+
+               # Turn block-level HTML blocks into hash entries
+               text = self._hash_html_blocks(text, raw=True)
+
+               # Strip link definitions, store in hashes.
+               if "footnotes" in self.extras:
+                       # Must do footnotes first because an unlucky footnote defn
+                       # looks like a link defn:
+                       #       [^4]: this "looks like a link defn"
+                       text = self._strip_footnote_definitions(text)
+               text = self._strip_link_definitions(text)
+
+               text = self._run_block_gamut(text)
+
+               if "footnotes" in self.extras:
+                       text = self._add_footnotes(text)
+
+               text = self._unescape_special_chars(text)
+
+               if self.safe_mode:
+                       text = self._unhash_html_spans(text)
+
+               text += "\n"
+               return text
+
+       # Cribbed from a post by Bart Lateur:
+       # <http://www.nntp.perl.org/group/perl.macperl.anyperl/154>
+       _detab_re = re.compile(r'(.*?)\t', re.M)
+       def _detab_sub(self, match):
+               g1 = match.group(1)
+               return g1 + (' ' * (self.tab_width - len(g1) % self.tab_width))
+       def _detab(self, text):
+               r"""Remove (leading?) tabs from a file.
+
+                       >>> m = Markdown()
+                       >>> m._detab("\tfoo")
+                       '        foo'
+                       >>> m._detab("  \tfoo")
+                       '        foo'
+                       >>> m._detab("\t  foo")
+                       '          foo'
+                       >>> m._detab("  foo")
+                       '  foo'
+                       >>> m._detab("  foo\n\tbar\tblam")
+                       '  foo\n        bar blam'
+               """
+               if '\t' not in text:
+                       return text
+               return self._detab_re.subn(self._detab_sub, text)[0]
+
+       _block_tags_a = 'p|div|h[1-6]|blockquote|pre|table|dl|ol|ul|script|noscript|form|fieldset|iframe|math|ins|del'
+       _strict_tag_block_re = re.compile(r"""
+               (                                               # save in \1
+                       ^                                       # start of line  (with re.M)
+                       <(%s)                           # start tag = \2
+                       \b                                      # word break
+                       (.*\n)*?                        # any number of lines, minimally matching
+                       </\2>                           # the matching end tag
+                       [ \t]*                          # trailing spaces/tabs
+                       (?=\n+|\Z)                      # followed by a newline or end of document
+               )
+               """ % _block_tags_a,
+               re.X | re.M)
+
+       _block_tags_b = 'p|div|h[1-6]|blockquote|pre|table|dl|ol|ul|script|noscript|form|fieldset|iframe|math'
+       _liberal_tag_block_re = re.compile(r"""
+               (                                               # save in \1
+                       ^                                       # start of line  (with re.M)
+                       <(%s)                           # start tag = \2
+                       \b                                      # word break
+                       (.*\n)*?                        # any number of lines, minimally matching
+                       .*</\2>                         # the matching end tag
+                       [ \t]*                          # trailing spaces/tabs
+                       (?=\n+|\Z)                      # followed by a newline or end of document
+               )
+               """ % _block_tags_b,
+               re.X | re.M)
+
+       def _hash_html_block_sub(self, match, raw=False):
+               html = match.group(1)
+               if raw and self.safe_mode:
+                       html = self._sanitize_html(html)
+               key = _hash_text(html)
+               self.html_blocks[key] = html
+               return "\n\n" + key + "\n\n"
+
+       def _hash_html_blocks(self, text, raw=False):
+               """Hashify HTML blocks
+
+               We only want to do this for block-level HTML tags, such as headers,
+               lists, and tables. That's because we still want to wrap <p>s around
+               "paragraphs" that are wrapped in non-block-level tags, such as anchors,
+               phrase emphasis, and spans. The list of tags we're looking for is
+               hard-coded.
+
+               @param raw {boolean} indicates if these are raw HTML blocks in
+                       the original source. It makes a difference in "safe" mode.
+               """
+               if '<' not in text:
+                       return text
+
+               # Pass `raw` value into our calls to self._hash_html_block_sub.
+               hash_html_block_sub = _curry(self._hash_html_block_sub, raw=raw)
+
+               # First, look for nested blocks, e.g.:
+               #       <div>
+               #               <div>
+               #               tags for inner block must be indented.
+               #               </div>
+               #       </div>
+               #
+               # The outermost tags must start at the left margin for this to match, and
+               # the inner nested divs must be indented.
+               # We need to do this before the next, more liberal match, because the next
+               # match will start at the first `<div>` and stop at the first `</div>`.
+               text = self._strict_tag_block_re.sub(hash_html_block_sub, text)
+
+               # Now match more liberally, simply from `\n<tag>` to `</tag>\n`
+               text = self._liberal_tag_block_re.sub(hash_html_block_sub, text)
+
+               # Special case just for <hr />. It was easier to make a special
+               # case than to make the other regex more complicated.
+               if "<hr" in text:
+                       _hr_tag_re = _hr_tag_re_from_tab_width(self.tab_width)
+                       text = _hr_tag_re.sub(hash_html_block_sub, text)
+
+               # Special case for standalone HTML comments:
+               if "<!--" in text:
+                       start = 0
+                       while True:
+                               # Delimiters for next comment block.
+                               try:
+                                       start_idx = text.index("<!--", start)
+                               except ValueError, ex:
+                                       break
+                               try:
+                                       end_idx = text.index("-->", start_idx) + 3
+                               except ValueError, ex:
+                                       break
+
+                               # Start position for next comment block search.
+                               start = end_idx
+
+                               # Validate whitespace before comment.
+                               if start_idx:
+                                       # - Up to `tab_width - 1` spaces before start_idx.
+                                       for i in range(self.tab_width - 1):
+                                               if text[start_idx - 1] != ' ':
+                                                       break
+                                               start_idx -= 1
+                                               if start_idx == 0:
+                                                       break
+                                       # - Must be preceded by 2 newlines or hit the start of
+                                       #       the document.
+                                       if start_idx == 0:
+                                               pass
+                                       elif start_idx == 1 and text[0] == '\n':
+                                               start_idx = 0  # to match minute detail of Markdown.pl regex
+                                       elif text[start_idx-2:start_idx] == '\n\n':
+                                               pass
+                                       else:
+                                               break
+
+                               # Validate whitespace after comment.
+                               # - Any number of spaces and tabs.
+                               while end_idx < len(text):
+                                       if text[end_idx] not in ' \t':
+                                               break
+                                       end_idx += 1
+                               # - Must be following by 2 newlines or hit end of text.
+                               if text[end_idx:end_idx+2] not in ('', '\n', '\n\n'):
+                                       continue
+
+                               # Escape and hash (must match `_hash_html_block_sub`).
+                               html = text[start_idx:end_idx]
+                               if raw and self.safe_mode:
+                                       html = self._sanitize_html(html)
+                               key = _hash_text(html)
+                               self.html_blocks[key] = html
+                               text = text[:start_idx] + "\n\n" + key + "\n\n" + text[end_idx:]
+
+               if "xml" in self.extras:
+                       # Treat XML processing instructions and namespaced one-liner
+                       # tags as if they were block HTML tags. E.g., if standalone
+                       # (i.e. are their own paragraph), the following do not get
+                       # wrapped in a <p> tag:
+                       #        <?foo bar?>
+                       #
+                       #        <xi:include xmlns:xi="http://www.w3.org/2001/XInclude" href="chapter_1.md"/>
+                       _xml_oneliner_re = _xml_oneliner_re_from_tab_width(self.tab_width)
+                       text = _xml_oneliner_re.sub(hash_html_block_sub, text)
+
+               return text
+
+       def _strip_link_definitions(self, text):
+               # Strips link definitions from text, stores the URLs and titles in
+               # hash references.
+               less_than_tab = self.tab_width - 1
+
+               # Link defs are in the form:
+               #       [id]: url "optional title"
+               _link_def_re = re.compile(r"""
+                       ^[ ]{0,%d}\[(.+)\]: # id = \1
+                         [ \t]*
+                         \n?                           # maybe *one* newline
+                         [ \t]*
+                       <?(.+?)>?                       # url = \2
+                         [ \t]*
+                       (?:
+                               \n?                             # maybe one newline
+                               [ \t]*
+                               (?<=\s)                 # lookbehind for whitespace
+                               ['"(]
+                               ([^\n]*)                # title = \3
+                               ['")]
+                               [ \t]*
+                       )?      # title is optional
+                       (?:\n+|\Z)
+                       """ % less_than_tab, re.X | re.M | re.U)
+               return _link_def_re.sub(self._extract_link_def_sub, text)
+
+       def _extract_link_def_sub(self, match):
+               id, url, title = match.groups()
+               key = id.lower()        # Link IDs are case-insensitive
+               self.urls[key] = self._encode_amps_and_angles(url)
+               if title:
+                       self.titles[key] = title.replace('"', '&quot;')
+               return ""
+
+       def _extract_footnote_def_sub(self, match):
+               id, text = match.groups()
+               text = _dedent(text, skip_first_line=not text.startswith('\n')).strip()
+               normed_id = re.sub(r'\W', '-', id)
+               # Ensure footnote text ends with a couple newlines (for some
+               # block gamut matches).
+               self.footnotes[normed_id] = text + "\n\n"
+               return ""
+
+       def _strip_footnote_definitions(self, text):
+               """A footnote definition looks like this:
+
+                       [^note-id]: Text of the note.
+
+                               May include one or more indented paragraphs.
+
+               Where,
+               - The 'note-id' can be pretty much anything, though typically it
+                 is the number of the footnote.
+               - The first paragraph may start on the next line, like so:
+
+                       [^note-id]:
+                               Text of the note.
+               """
+               less_than_tab = self.tab_width - 1
+               footnote_def_re = re.compile(r'''
+                       ^[ ]{0,%d}\[\^(.+)\]:   # id = \1
+                       [ \t]*
+                       (                                               # footnote text = \2
+                         # First line need not start with the spaces.
+                         (?:\s*.*\n+)
+                         (?:
+                               (?:[ ]{%d} | \t)  # Subsequent lines must be indented.
+                               .*\n+
+                         )*
+                       )
+                       # Lookahead for non-space at line-start, or end of doc.
+                       (?:(?=^[ ]{0,%d}\S)|\Z)
+                       ''' % (less_than_tab, self.tab_width, self.tab_width),
+                       re.X | re.M)
+               return footnote_def_re.sub(self._extract_footnote_def_sub, text)
+
+
+       _hr_res = [
+               re.compile(r"^[ ]{0,2}([ ]?\*[ ]?){3,}[ \t]*$", re.M),
+               re.compile(r"^[ ]{0,2}([ ]?\-[ ]?){3,}[ \t]*$", re.M),
+               re.compile(r"^[ ]{0,2}([ ]?\_[ ]?){3,}[ \t]*$", re.M),
+       ]
+
+       def _run_block_gamut(self, text):
+               # These are all the transformations that form block-level
+               # tags like paragraphs, headers, and list items.
+
+               text = self._do_headers(text)
+
+               # Do Horizontal Rules:
+               hr = "\n<hr"+self.empty_element_suffix+"\n"
+               for hr_re in self._hr_res:
+                       text = hr_re.sub(hr, text)
+
+               text = self._do_lists(text)
+
+               if "pyshell" in self.extras:
+                       text = self._prepare_pyshell_blocks(text)
+
+               text = self._do_code_blocks(text)
+
+               text = self._do_block_quotes(text)
+
+               # We already ran _HashHTMLBlocks() before, in Markdown(), but that
+               # was to escape raw HTML in the original Markdown source. This time,
+               # we're escaping the markup we've just created, so that we don't wrap
+               # <p> tags around block-level tags.
+               text = self._hash_html_blocks(text)
+
+               text = self._form_paragraphs(text)
+
+               return text
+
+       def _pyshell_block_sub(self, match):
+               lines = match.group(0).splitlines(0)
+               _dedentlines(lines)
+               indent = ' ' * self.tab_width
+               s = ('\n' # separate from possible cuddled paragraph
+                        + indent + ('\n'+indent).join(lines)
+                        + '\n\n')
+               return s
+
+       def _prepare_pyshell_blocks(self, text):
+               """Ensure that Python interactive shell sessions are put in
+               code blocks -- even if not properly indented.
+               """
+               if ">>>" not in text:
+                       return text
+
+               less_than_tab = self.tab_width - 1
+               _pyshell_block_re = re.compile(r"""
+                       ^([ ]{0,%d})>>>[ ].*\n   # first line
+                       ^(\1.*\S+.*\n)*                 # any number of subsequent lines
+                       ^\n                                             # ends with a blank line
+                       """ % less_than_tab, re.M | re.X)
+
+               return _pyshell_block_re.sub(self._pyshell_block_sub, text)
+
+       def _run_span_gamut(self, text):
+               # These are all the transformations that occur *within* block-level
+               # tags like paragraphs, headers, and list items.
+
+               text = self._do_code_spans(text)
+
+               text = self._escape_special_chars(text)
+
+               # Process anchor and image tags.
+               #text = self._do_links(text)
+
+               # Make links out of things like `<http://example.com/>`
+               # Must come after _do_links(), because you can use < and >
+               # delimiters in inline links like [this](<url>).
+               text = self._do_auto_links(text)
+
+               if "link-patterns" in self.extras:
+                       text = self._do_link_patterns(text)
+
+               text = self._encode_amps_and_angles(text)
+
+               text = self._do_italics_and_bold(text)
+
+               # Do hard breaks:
+               text = re.sub(r" {2,}\n", " <br%s\n" % self.empty_element_suffix, text)
+
+               return text
+
+       # "Sorta" because auto-links are identified as "tag" tokens.
+       _sorta_html_tokenize_re = re.compile(r"""
+               (
+                       # tag
+                       </?
+                       (?:\w+)                                                                         # tag name
+                       (?:\s+(?:[\w-]+:)?[\w-]+=(?:".*?"|'.*?'))*      # attributes
+                       \s*/?>
+                       |
+                       # auto-link (e.g., <http://www.activestate.com/>)
+                       <\w+[^>]*>
+                       |
+                       <!--.*?-->              # comment
+                       |
+                       <\?.*?\?>               # processing instruction
+               )
+               """, re.X)
+
+       def _escape_special_chars(self, text):
+               # Python markdown note: the HTML tokenization here differs from
+               # that in Markdown.pl, hence the behaviour for subtle cases can
+               # differ (I believe the tokenizer here does a better job because
+               # it isn't susceptible to unmatched '<' and '>' in HTML tags).
+               # Note, however, that '>' is not allowed in an auto-link URL
+               # here.
+               escaped = []
+               is_html_markup = False
+               for token in self._sorta_html_tokenize_re.split(text):
+                       if is_html_markup:
+                               # Within tags/HTML-comments/auto-links, encode * and _
+                               # so they don't conflict with their use in Markdown for
+                               # italics and strong.  We're replacing each such
+                               # character with its corresponding MD5 checksum value;
+                               # this is likely overkill, but it should prevent us from
+                               # colliding with the escape values by accident.
+                               escaped.append(token.replace('*', g_escape_table['*'])
+                                                                       .replace('_', g_escape_table['_']))
+                       else:
+                               escaped.append(self._encode_backslash_escapes(token))
+                       is_html_markup = not is_html_markup
+               return ''.join(escaped)
+
+       def _hash_html_spans(self, text):
+               # Used for safe_mode.
+
+               def _is_auto_link(s):
+                       if ':' in s and self._auto_link_re.match(s):
+                               return True
+                       elif '@' in s and self._auto_email_link_re.match(s):
+                               return True
+                       return False
+
+               tokens = []
+               is_html_markup = False
+               for token in self._sorta_html_tokenize_re.split(text):
+                       if is_html_markup and not _is_auto_link(token):
+                               sanitized = self._sanitize_html(token)
+                               key = _hash_text(sanitized)
+                               self.html_spans[key] = sanitized
+                               tokens.append(key)
+                       else:
+                               tokens.append(token)
+                       is_html_markup = not is_html_markup
+               return ''.join(tokens)
+
+       def _unhash_html_spans(self, text):
+               for key, sanitized in self.html_spans.items():
+                       text = text.replace(key, sanitized)
+               return text
+
+       def _sanitize_html(self, s):
+               if self.safe_mode == "replace":
+                       return self.html_removed_text
+               elif self.safe_mode == "escape":
+                       replacements = [
+                               ('&', '&amp;'),
+                               ('<', '&lt;'),
+                               ('>', '&gt;'),
+                       ]
+                       for before, after in replacements:
+                               s = s.replace(before, after)
+                       return s
+               else:
+                       raise MarkdownError("invalid value for 'safe_mode': %r (must be "
+                                                               "'escape' or 'replace')" % self.safe_mode)
+
+       _tail_of_inline_link_re = re.compile(r'''
+                 # Match tail of: [text](/url/) or [text](/url/ "title")
+                 \(                    # literal paren
+                       [ \t]*
+                       (?P<url>                        # \1
+                               <.*?>
+                               |
+                               .*?
+                       )
+                       [ \t]*
+                       (                                       # \2
+                         (['"])                        # quote char = \3
+                         (?P<title>.*?)
+                         \3                            # matching quote
+                       )?                                      # title is optional
+                 \)
+               ''', re.X | re.S)
+       _tail_of_reference_link_re = re.compile(r'''
+                 # Match tail of: [text][id]
+                 [ ]?                  # one optional space
+                 (?:\n[ ]*)?   # one optional newline followed by spaces
+                 \[
+                       (?P<id>.*?)
+                 \]
+               ''', re.X | re.S)
+
+       def _do_links(self, text):
+               """Turn Markdown link shortcuts into XHTML <a> and <img> tags.
+
+               This is a combination of Markdown.pl's _DoAnchors() and
+               _DoImages(). They are done together because that simplified the
+               approach. It was necessary to use a different approach than
+               Markdown.pl because of the lack of atomic matching support in
+               Python's regex engine used in $g_nested_brackets.
+               """
+               MAX_LINK_TEXT_SENTINEL = 3000  # markdown2 issue 24
+
+               # `anchor_allowed_pos` is used to support img links inside
+               # anchors, but not anchors inside anchors. An anchor's start
+               # pos must be `>= anchor_allowed_pos`.
+               anchor_allowed_pos = 0
+
+               curr_pos = 0
+               while True: # Handle the next link.
+                       # The next '[' is the start of:
+                       # - an inline anchor:   [text](url "title")
+                       # - a reference anchor: [text][id]
+                       # - an inline img:              ![text](url "title")
+                       # - a reference img:    ![text][id]
+                       # - a footnote ref:             [^id]
+                       #       (Only if 'footnotes' extra enabled)
+                       # - a footnote defn:    [^id]: ...
+                       #       (Only if 'footnotes' extra enabled) These have already
+                       #       been stripped in _strip_footnote_definitions() so no
+                       #       need to watch for them.
+                       # - a link definition:  [id]: url "title"
+                       #       These have already been stripped in
+                       #       _strip_link_definitions() so no need to watch for them.
+                       # - not markup:                 [...anything else...
+                       try:
+                               start_idx = text.index('[', curr_pos)
+                       except ValueError:
+                               break
+                       text_length = len(text)
+
+                       # Find the matching closing ']'.
+                       # Markdown.pl allows *matching* brackets in link text so we
+                       # will here too. Markdown.pl *doesn't* currently allow
+                       # matching brackets in img alt text -- we'll differ in that
+                       # regard.
+                       bracket_depth = 0
+                       for p in range(start_idx+1, min(start_idx+MAX_LINK_TEXT_SENTINEL,
+                                                                                       text_length)):
+                               ch = text[p]
+                               if ch == ']':
+                                       bracket_depth -= 1
+                                       if bracket_depth < 0:
+                                               break
+                               elif ch == '[':
+                                       bracket_depth += 1
+                       else:
+                               # Closing bracket not found within sentinel length.
+                               # This isn't markup.
+                               curr_pos = start_idx + 1
+                               continue
+                       link_text = text[start_idx+1:p]
+
+                       # Possibly a footnote ref?
+                       if "footnotes" in self.extras and link_text.startswith("^"):
+                               normed_id = re.sub(r'\W', '-', link_text[1:])
+                               if normed_id in self.footnotes:
+                                       self.footnote_ids.append(normed_id)
+                                       result = '<sup class="footnote-ref" id="fnref-%s">' \
+                                                        '<a href="#fn-%s">%s</a></sup>' \
+                                                        % (normed_id, normed_id, len(self.footnote_ids))
+                                       text = text[:start_idx] + result + text[p+1:]
+                               else:
+                                       # This id isn't defined, leave the markup alone.
+                                       curr_pos = p+1
+                               continue
+
+                       # Now determine what this is by the remainder.
+                       p += 1
+                       if p == text_length:
+                               return text
+
+                       # Inline anchor or img?
+                       if text[p] == '(': # attempt at perf improvement
+                               match = self._tail_of_inline_link_re.match(text, p)
+                               if match:
+                                       # Handle an inline anchor or img.
+                                       is_img = start_idx > 0 and text[start_idx-1] == "!"
+                                       if is_img:
+                                               start_idx -= 1
+
+                                       url, title = match.group("url"), match.group("title")
+                                       if url and url[0] == '<':
+                                               url = url[1:-1]  # '<url>' -> 'url'
+                                       # We've got to encode these to avoid conflicting
+                                       # with italics/bold.
+                                       url = url.replace('*', g_escape_table['*']) \
+                                                        .replace('_', g_escape_table['_'])
+                                       if title:
+                                               title_str = ' title="%s"' \
+                                                       % title.replace('*', g_escape_table['*']) \
+                                                                  .replace('_', g_escape_table['_']) \
+                                                                  .replace('"', '&quot;')
+                                       else:
+                                               title_str = ''
+                                       if is_img:
+                                               result = '<img src="%s" alt="%s"%s%s' \
+                                                       % (url.replace('"', '&quot;'),
+                                                          link_text.replace('"', '&quot;'),
+                                                          title_str, self.empty_element_suffix)
+                                               curr_pos = start_idx + len(result)
+                                               text = text[:start_idx] + result + text[match.end():]
+                                       elif start_idx >= anchor_allowed_pos:
+                                               result_head = '<a href="%s"%s>' % (url, title_str)
+                                               result = '%s%s</a>' % (result_head, link_text)
+                                               # <img> allowed from curr_pos on, <a> from
+                                               # anchor_allowed_pos on.
+                                               curr_pos = start_idx + len(result_head)
+                                               anchor_allowed_pos = start_idx + len(result)
+                                               text = text[:start_idx] + result + text[match.end():]
+                                       else:
+                                               # Anchor not allowed here.
+                                               curr_pos = start_idx + 1
+                                       continue
+
+                       # Reference anchor or img?
+                       else:
+                               match = self._tail_of_reference_link_re.match(text, p)
+                               if match:
+                                       # Handle a reference-style anchor or img.
+                                       is_img = start_idx > 0 and text[start_idx-1] == "!"
+                                       if is_img:
+                                               start_idx -= 1
+                                       link_id = match.group("id").lower()
+                                       if not link_id:
+                                               link_id = link_text.lower()      # for links like [this][]
+                                       if link_id in self.urls:
+                                               url = self.urls[link_id]
+                                               # We've got to encode these to avoid conflicting
+                                               # with italics/bold.
+                                               url = url.replace('*', g_escape_table['*']) \
+                                                                .replace('_', g_escape_table['_'])
+                                               title = self.titles.get(link_id)
+                                               if title:
+                                                       title = title.replace('*', g_escape_table['*']) \
+                                                                                .replace('_', g_escape_table['_'])
+                                                       title_str = ' title="%s"' % title
+                                               else:
+                                                       title_str = ''
+                                               if is_img:
+                                                       result = '<img src="%s" alt="%s"%s%s' \
+                                                               % (url.replace('"', '&quot;'),
+                                                                  link_text.replace('"', '&quot;'),
+                                                                  title_str, self.empty_element_suffix)
+                                                       curr_pos = start_idx + len(result)
+                                                       text = text[:start_idx] + result + text[match.end():]
+                                               elif start_idx >= anchor_allowed_pos:
+                                                       result = '<a href="%s"%s>%s</a>' \
+                                                               % (url, title_str, link_text)
+                                                       result_head = '<a href="%s"%s>' % (url, title_str)
+                                                       result = '%s%s</a>' % (result_head, link_text)
+                                                       # <img> allowed from curr_pos on, <a> from
+                                                       # anchor_allowed_pos on.
+                                                       curr_pos = start_idx + len(result_head)
+                                                       anchor_allowed_pos = start_idx + len(result)
+                                                       text = text[:start_idx] + result + text[match.end():]
+                                               else:
+                                                       # Anchor not allowed here.
+                                                       curr_pos = start_idx + 1
+                                       else:
+                                               # This id isn't defined, leave the markup alone.
+                                               curr_pos = match.end()
+                                       continue
+
+                       # Otherwise, it isn't markup.
+                       curr_pos = start_idx + 1
+
+               return text
+
+
+       _setext_h_re = re.compile(r'^(.+)[ \t]*\n(=+|-+)[ \t]*\n+', re.M)
+       def _setext_h_sub(self, match):
+               n = {"=": 1, "-": 2}[match.group(2)[0]]
+               demote_headers = self.extras.get("demote-headers")
+               if demote_headers:
+                       n = min(n + demote_headers, 6)
+               return "<h%d>%s</h%d>\n\n" \
+                          % (n, self._run_span_gamut(match.group(1)), n)
+
+       _atx_h_re = re.compile(r'''
+               ^([\#=]{1,6})  # \1 = string of #'s
+               [ \t]*
+               (.+?)           # \2 = Header text
+               [ \t]*
+               (?<!\\)         # ensure not an escaped trailing '#'
+               [\#=]*          # optional closing #'s (not counted)
+               \n+
+               ''', re.X | re.M)
+       def _atx_h_sub(self, match):
+               n = len(match.group(1))
+               demote_headers = self.extras.get("demote-headers")
+               if demote_headers:
+                       n = min(n + demote_headers, 6)
+               return "<h%d>%s</h%d>\n\n" \
+                          % (n, self._run_span_gamut(match.group(2)), n)
+
+       def _do_headers(self, text):
+               # Setext-style headers:
+               #         Header 1
+               #         ========
+               #
+               #         Header 2
+               #         --------
+               text = self._setext_h_re.sub(self._setext_h_sub, text)
+
+               # atx-style headers:
+               #       # Header 1
+               #       ## Header 2
+               #       ## Header 2 with closing hashes ##
+               #       ...
+               #       ###### Header 6
+               text = self._atx_h_re.sub(self._atx_h_sub, text)
+
+               return text
+
+
+       _marker_ul_chars  = '*+-'
+       _marker_any = r'(?:[%s]|\d+\.)' % _marker_ul_chars
+       _marker_ul = '(?:[%s])' % _marker_ul_chars
+       _marker_ol = r'(?:\d+\.)'
+
+       def _list_sub(self, match):
+               lst = match.group(1)
+               lst_type = match.group(3) in self._marker_ul_chars and "ul" or "ol"
+               result = self._process_list_items(lst)
+               if self.list_level:
+                       return "<%s>\n%s</%s>\n" % (lst_type, result, lst_type)
+               else:
+                       return "<%s>\n%s</%s>\n\n" % (lst_type, result, lst_type)
+
+       def _do_lists(self, text):
+               # Form HTML ordered (numbered) and unordered (bulleted) lists.
+
+               for marker_pat in (self._marker_ul, self._marker_ol):
+                       # Re-usable pattern to match any entire ul or ol list:
+                       less_than_tab = self.tab_width - 1
+                       whole_list = r'''
+                               (                                       # \1 = whole list
+                                 (                                     # \2
+                                       [ ]{0,%d}
+                                       (%s)                    # \3 = first list item marker
+                                       [ \t]+
+                                 )
+                                 (?:.+?)
+                                 (                                     # \4
+                                         \Z
+                                       |
+                                         \n{2,}
+                                         (?=\S)
+                                         (?!                   # Negative lookahead for another list item marker
+                                               [ \t]*
+                                               %s[ \t]+
+                                         )
+                                 )
+                               )
+                       ''' % (less_than_tab, marker_pat, marker_pat)
+
+                       # We use a different prefix before nested lists than top-level lists.
+                       # See extended comment in _process_list_items().
+                       #
+                       # Note: There's a bit of duplication here. My original implementation
+                       # created a scalar regex pattern as the conditional result of the test on
+                       # $g_list_level, and then only ran the $text =~ s{...}{...}egmx
+                       # substitution once, using the scalar as the pattern. This worked,
+                       # everywhere except when running under MT on my hosting account at Pair
+                       # Networks. There, this caused all rebuilds to be killed by the reaper (or
+                       # perhaps they crashed, but that seems incredibly unlikely given that the
+                       # same script on the same server ran fine *except* under MT. I've spent
+                       # more time trying to figure out why this is happening than I'd like to
+                       # admit. My only guess, backed up by the fact that this workaround works,
+                       # is that Perl optimizes the substition when it can figure out that the
+                       # pattern will never change, and when this optimization isn't on, we run
+                       # afoul of the reaper. Thus, the slightly redundant code to that uses two
+                       # static s/// patterns rather than one conditional pattern.
+
+                       if self.list_level:
+                               sub_list_re = re.compile("^"+whole_list, re.X | re.M | re.S)
+                               text = sub_list_re.sub(self._list_sub, text)
+                       else:
+                               list_re = re.compile(r"(?:(?<=\n\n)|\A\n?)"+whole_list,
+                                                                        re.X | re.M | re.S)
+                               text = list_re.sub(self._list_sub, text)
+
+               return text
+
+       _list_item_re = re.compile(r'''
+               (\n)?                           # leading line = \1
+               (^[ \t]*)                       # leading whitespace = \2
+               (?P<marker>%s) [ \t]+   # list marker = \3
+               ((?:.+?)                        # list item text = \4
+                (\n{1,2}))                     # eols = \5
+               (?= \n* (\Z | \2 (?P<next_marker>%s) [ \t]+))
+               ''' % (_marker_any, _marker_any),
+               re.M | re.X | re.S)
+
+       _last_li_endswith_two_eols = False
+       def _list_item_sub(self, match):
+               item = match.group(4)
+               leading_line = match.group(1)
+               leading_space = match.group(2)
+               if leading_line or "\n\n" in item or self._last_li_endswith_two_eols:
+                       item = self._run_block_gamut(self._outdent(item))
+               else:
+                       # Recursion for sub-lists:
+                       item = self._do_lists(self._outdent(item))
+                       if item.endswith('\n'):
+                               item = item[:-1]
+                       item = self._run_span_gamut(item)
+               self._last_li_endswith_two_eols = (len(match.group(5)) == 2)
+               return "<li>%s</li>\n" % item
+
+       def _process_list_items(self, list_str):
+               # Process the contents of a single ordered or unordered list,
+               # splitting it into individual list items.
+
+               # The $g_list_level global keeps track of when we're inside a list.
+               # Each time we enter a list, we increment it; when we leave a list,
+               # we decrement. If it's zero, we're not in a list anymore.
+               #
+               # We do this because when we're not inside a list, we want to treat
+               # something like this:
+               #
+               #               I recommend upgrading to version
+               #               8. Oops, now this line is treated
+               #               as a sub-list.
+               #
+               # As a single paragraph, despite the fact that the second line starts
+               # with a digit-period-space sequence.
+               #
+               # Whereas when we're inside a list (or sub-list), that line will be
+               # treated as the start of a sub-list. What a kludge, huh? This is
+               # an aspect of Markdown's syntax that's hard to parse perfectly
+               # without resorting to mind-reading. Perhaps the solution is to
+               # change the syntax rules such that sub-lists must start with a
+               # starting cardinal number; e.g. "1." or "a.".
+               self.list_level += 1
+               self._last_li_endswith_two_eols = False
+               list_str = list_str.rstrip('\n') + '\n'
+               list_str = self._list_item_re.sub(self._list_item_sub, list_str)
+               self.list_level -= 1
+               return list_str
+
+       def _get_pygments_lexer(self, lexer_name):
+               try:
+                       from pygments import lexers, util
+               except ImportError:
+                       return None
+               try:
+                       return lexers.get_lexer_by_name(lexer_name)
+               except util.ClassNotFound:
+                       return None
+
+       def _color_with_pygments(self, codeblock, lexer, **formatter_opts):
+               import pygments
+               import pygments.formatters
+
+               class HtmlCodeFormatter(pygments.formatters.HtmlFormatter):
+                       def _wrap_code(self, inner):
+                               """A function for use in a Pygments Formatter which
+                               wraps in <code> tags.
+                               """
+                               yield 0, "<code>"
+                               for tup in inner:
+                                       yield tup
+                               yield 0, "</code>"
+
+                       def wrap(self, source, outfile):
+                               """Return the source with a code, pre, and div."""
+                               return self._wrap_div(self._wrap_pre(self._wrap_code(source)))
+
+               formatter = HtmlCodeFormatter(cssclass="codehilite", **formatter_opts)
+               return pygments.highlight(codeblock, lexer, formatter)
+
+       def _code_block_sub(self, match):
+               codeblock = match.group(1)
+               codeblock = self._outdent(codeblock)
+               codeblock = self._detab(codeblock)
+               codeblock = codeblock.lstrip('\n')      # trim leading newlines
+               codeblock = codeblock.rstrip()          # trim trailing whitespace
+
+               if "code-color" in self.extras and codeblock.startswith(":::"):
+                       lexer_name, rest = codeblock.split('\n', 1)
+                       lexer_name = lexer_name[3:].strip()
+                       lexer = self._get_pygments_lexer(lexer_name)
+                       codeblock = rest.lstrip("\n")   # Remove lexer declaration line.
+                       if lexer:
+                               formatter_opts = self.extras['code-color'] or {}
+                               colored = self._color_with_pygments(codeblock, lexer,
+                                                                                                       **formatter_opts)
+                               return "\n\n%s\n\n" % colored
+
+               codeblock = self._encode_code(codeblock)
+               return "\n\n<pre><code>%s\n</code></pre>\n\n" % codeblock
+
+       def _do_code_blocks(self, text):
+               """Process Markdown `<pre><code>` blocks."""
+               code_block_re = re.compile(r'''
+                       (?:\n\n|\A)
+                       (                               # $1 = the code block -- one or more lines, starting with a space/tab
+                         (?:
+                               (?:[ ]{%d} | \t)  # Lines must start with a tab or a tab-width of spaces
+                               .*\n+
+                         )+
+                       )
+                       ((?=^[ ]{0,%d}\S)|\Z)   # Lookahead for non-space at line-start, or end of doc
+                       ''' % (self.tab_width, self.tab_width),
+                       re.M | re.X)
+
+               return code_block_re.sub(self._code_block_sub, text)
+
+
+       # Rules for a code span:
+       # - backslash escapes are not interpreted in a code span
+       # - to include one or or a run of more backticks the delimiters must
+       #       be a longer run of backticks
+       # - cannot start or end a code span with a backtick; pad with a
+       #       space and that space will be removed in the emitted HTML
+       # See `test/tm-cases/escapes.text` for a number of edge-case
+       # examples.
+       _code_span_re = re.compile(r'''
+                       (?<!\\)
+                       (`+)            # \1 = Opening run of `
+                       (?!`)           # See Note A test/tm-cases/escapes.text
+                       (.+?)           # \2 = The code block
+                       (?<!`)
+                       \1                      # Matching closer
+                       (?!`)
+               ''', re.X | re.S)
+
+       def _code_span_sub(self, match):
+               c = match.group(2).strip(" \t")
+               c = self._encode_code(c)
+               return "<code>%s</code>" % c
+
+       def _do_code_spans(self, text):
+               #       *       Backtick quotes are used for <code></code> spans.
+               #
+               #       *       You can use multiple backticks as the delimiters if you want to
+               #               include literal backticks in the code span. So, this input:
+               #
+               #                 Just type ``foo `bar` baz`` at the prompt.
+               #
+               #               Will translate to:
+               #
+               #                 <p>Just type <code>foo `bar` baz</code> at the prompt.</p>
+               #
+               #               There's no arbitrary limit to the number of backticks you
+               #               can use as delimters. If you need three consecutive backticks
+               #               in your code, use four for delimiters, etc.
+               #
+               #       *       You can use spaces to get literal backticks at the edges:
+               #
+               #                 ... type `` `bar` `` ...
+               #
+               #               Turns to:
+               #
+               #                 ... type <code>`bar`</code> ...
+               return self._code_span_re.sub(self._code_span_sub, text)
+
+       def _encode_code(self, text):
+               """Encode/escape certain characters inside Markdown code runs.
+               The point is that in code, these characters are literals,
+               and lose their special Markdown meanings.
+               """
+               replacements = [
+                       # Encode all ampersands; HTML entities are not
+                       # entities within a Markdown code span.
+                       ('&', '&amp;'),
+                       # Do the angle bracket song and dance:
+                       ('<', '&lt;'),
+                       ('>', '&gt;'),
+                       # Now, escape characters that are magic in Markdown:
+                       ('*', g_escape_table['*']),
+                       ('_', g_escape_table['_']),
+                       ('{', g_escape_table['{']),
+                       ('}', g_escape_table['}']),
+                       ('[', g_escape_table['[']),
+                       (']', g_escape_table[']']),
+                       ('\\', g_escape_table['\\']),
+               ]
+               for before, after in replacements:
+                       text = text.replace(before, after)
+               return text
+
+       _strong_re = re.compile(r"(\*\*|__)(?=\S)(.+?[*_]*)(?<=\S)\1", re.S)
+       _em_re = re.compile(r"(\*|_)(?=\S)(.+?)(?<=\S)\1", re.S)
+       _code_friendly_strong_re = re.compile(r"\*\*(?=\S)(.+?[*_]*)(?<=\S)\*\*", re.S)
+       _code_friendly_em_re = re.compile(r"\*(?=\S)(.+?)(?<=\S)\*", re.S)
+       def _do_italics_and_bold(self, text):
+               # <strong> must go first:
+               if "code-friendly" in self.extras:
+                       text = self._code_friendly_strong_re.sub(r"<strong>\1</strong>", text)
+                       text = self._code_friendly_em_re.sub(r"<em>\1</em>", text)
+               else:
+                       text = self._strong_re.sub(r"<strong>\2</strong>", text)
+                       text = self._em_re.sub(r"<em>\2</em>", text)
+               return text
+
+
+       _block_quote_re = re.compile(r'''
+               (                                                       # Wrap whole match in \1
+                 (
+                       ^[ \t]*>[ \t]?                  # '>' at the start of a line
+                         .+\n                                  # rest of the first line
+                       (.+\n)*                                 # subsequent consecutive lines
+                       \n*                                             # blanks
+                 )+
+               )
+               ''', re.M | re.X)
+       _bq_one_level_re = re.compile('^[ \t]*>[ \t]?', re.M);
+
+       _html_pre_block_re = re.compile(r'(\s*<pre>.+?</pre>)', re.S)
+       def _dedent_two_spaces_sub(self, match):
+               return re.sub(r'(?m)^  ', '', match.group(1))
+
+       def _block_quote_sub(self, match):
+               bq = match.group(1)
+               bq = self._bq_one_level_re.sub('', bq)  # trim one level of quoting
+               bq = self._ws_only_line_re.sub('', bq)  # trim whitespace-only lines
+               bq = self._run_block_gamut(bq)                  # recurse
+
+               bq = re.sub('(?m)^', '  ', bq)
+               # These leading spaces screw with <pre> content, so we need to fix that:
+               bq = self._html_pre_block_re.sub(self._dedent_two_spaces_sub, bq)
+
+               return "<blockquote>\n%s\n</blockquote>\n\n" % bq
+
+       def _do_block_quotes(self, text):
+               if '>' not in text:
+                       return text
+               return self._block_quote_re.sub(self._block_quote_sub, text)
+
+       def _form_paragraphs(self, text):
+               # Strip leading and trailing lines:
+               text = text.strip('\n')
+
+               # Wrap <p> tags.
+               grafs = []
+               for i, graf in enumerate(re.split(r"\n{2,}", text)):
+                       if graf in self.html_blocks:
+                               # Unhashify HTML blocks
+                               grafs.append(self.html_blocks[graf])
+                       else:
+                               cuddled_list = None
+                               if "cuddled-lists" in self.extras:
+                                       # Need to put back trailing '\n' for `_list_item_re`
+                                       # match at the end of the paragraph.
+                                       li = self._list_item_re.search(graf + '\n')
+                                       # Two of the same list marker in this paragraph: a likely
+                                       # candidate for a list cuddled to preceding paragraph
+                                       # text (issue 33). Note the `[-1]` is a quick way to
+                                       # consider numeric bullets (e.g. "1." and "2.") to be
+                                       # equal.
+                                       if (li and li.group("next_marker")
+                                               and li.group("marker")[-1] == li.group("next_marker")[-1]):
+                                               start = li.start()
+                                               cuddled_list = self._do_lists(graf[start:]).rstrip("\n")
+                                               assert cuddled_list.startswith("<ul>") or cuddled_list.startswith("<ol>")
+                                               graf = graf[:start]
+
+                               # Wrap <p> tags.
+                               graf = self._run_span_gamut(graf)
+                               grafs.append("<p>" + graf.lstrip(" \t") + "</p>")
+
+                               if cuddled_list:
+                                       grafs.append(cuddled_list)
+
+               return "\n\n".join(grafs)
+
+       def _add_footnotes(self, text):
+               if self.footnotes:
+                       footer = [
+                               '<div class="footnotes">',
+                               '<hr' + self.empty_element_suffix,
+                               '<ol>',
+                       ]
+                       for i, id in enumerate(self.footnote_ids):
+                               if i != 0:
+                                       footer.append('')
+                               footer.append('<li id="fn-%s">' % id)
+                               footer.append(self._run_block_gamut(self.footnotes[id]))
+                               backlink = ('<a href="#fnref-%s" '
+                                       'class="footnoteBackLink" '
+                                       'title="Jump back to footnote %d in the text.">'
+                                       '&#8617;</a>' % (id, i+1))
+                               if footer[-1].endswith("</p>"):
+                                       footer[-1] = footer[-1][:-len("</p>")] \
+                                               + '&nbsp;' + backlink + "</p>"
+                               else:
+                                       footer.append("\n<p>%s</p>" % backlink)
+                               footer.append('</li>')
+                       footer.append('</ol>')
+                       footer.append('</div>')
+                       return text + '\n\n' + '\n'.join(footer)
+               else:
+                       return text
+
+       # Ampersand-encoding based entirely on Nat Irons's Amputator MT plugin:
+       #       http://bumppo.net/projects/amputator/
+       _ampersand_re = re.compile(r'&(?!#?[xX]?(?:[0-9a-fA-F]+|\w+);)')
+       _naked_lt_re = re.compile(r'<(?![a-z/?\$!])', re.I)
+       _naked_gt_re = re.compile(r'''(?<![a-z?!/'"-])>''', re.I)
+
+       def _encode_amps_and_angles(self, text):
+               # Smart processing for ampersands and angle brackets that need
+               # to be encoded.
+               text = self._ampersand_re.sub('&amp;', text)
+
+               # Encode naked <'s
+               text = self._naked_lt_re.sub('&lt;', text)
+
+               # Encode naked >'s
+               # Note: Other markdown implementations (e.g. Markdown.pl, PHP
+               # Markdown) don't do this.
+               text = self._naked_gt_re.sub('&gt;', text)
+               return text
+
+       def _encode_backslash_escapes(self, text):
+               for ch, escape in g_escape_table.items():
+                       text = text.replace("\\"+ch, escape)
+               return text
+
+       _auto_link_re = re.compile(r'<((https?|ftp):[^\'">\s]+)>', re.I)
+       def _auto_link_sub(self, match):
+               g1 = match.group(1)
+               return '<a href="%s">%s</a>' % (g1, g1)
+
+       _auto_email_link_re = re.compile(r"""
+                 <
+                  (?:mailto:)?
+                 (
+                         [-.\w]+
+                         \@
+                         [-\w]+(\.[-\w]+)*\.[a-z]+
+                 )
+                 >
+               """, re.I | re.X | re.U)
+       def _auto_email_link_sub(self, match):
+               return self._encode_email_address(
+                       self._unescape_special_chars(match.group(1)))
+
+       def _do_auto_links(self, text):
+               text = self._auto_link_re.sub(self._auto_link_sub, text)
+               text = self._auto_email_link_re.sub(self._auto_email_link_sub, text)
+               return text
+
+       def _encode_email_address(self, addr):
+               #  Input: an email address, e.g. "foo@example.com"
+               #
+               #  Output: the email address as a mailto link, with each character
+               #          of the address encoded as either a decimal or hex entity, in
+               #          the hopes of foiling most address harvesting spam bots. E.g.:
+               #
+               #        <a href="&#x6D;&#97;&#105;&#108;&#x74;&#111;:&#102;&#111;&#111;&#64;&#101;
+               #               x&#x61;&#109;&#x70;&#108;&#x65;&#x2E;&#99;&#111;&#109;">&#102;&#111;&#111;
+               #               &#64;&#101;x&#x61;&#109;&#x70;&#108;&#x65;&#x2E;&#99;&#111;&#109;</a>
+               #
+               #  Based on a filter by Matthew Wickline, posted to the BBEdit-Talk
+               #  mailing list: <http://tinyurl.com/yu7ue>
+               chars = [_xml_encode_email_char_at_random(ch)
+                                for ch in "mailto:" + addr]
+               # Strip the mailto: from the visible part.
+               addr = '<a href="%s">%s</a>' \
+                          % (''.join(chars), ''.join(chars[7:]))
+               return addr
+
+       def _do_link_patterns(self, text):
+               """Caveat emptor: there isn't much guarding against link
+               patterns being formed inside other standard Markdown links, e.g.
+               inside a [link def][like this].
+
+               Dev Notes: *Could* consider prefixing regexes with a negative
+               lookbehind assertion to attempt to guard against this.
+               """
+               link_from_hash = {}
+               for regex, repl in self.link_patterns:
+                       replacements = []
+                       for match in regex.finditer(text):
+                               if hasattr(repl, "__call__"):
+                                       href = repl(match)
+                               else:
+                                       href = match.expand(repl)
+                               replacements.append((match.span(), href))
+                       for (start, end), href in reversed(replacements):
+                               escaped_href = (
+                                       href.replace('"', '&quot;')      # b/c of attr quote
+                                               # To avoid markdown <em> and <strong>:
+                                               .replace('*', g_escape_table['*'])
+                                               .replace('_', g_escape_table['_']))
+                               link = '<a href="%s">%s</a>' % (escaped_href, text[start:end])
+                               hash = _hash_text(link)
+                               link_from_hash[hash] = link
+                               text = text[:start] + hash + text[end:]
+               for hash, link in link_from_hash.items():
+                       text = text.replace(hash, link)
+               return text
+
+       def _unescape_special_chars(self, text):
+               # Swap back in all the special characters we've hidden.
+               for ch, hash in g_escape_table.items():
+                       text = text.replace(hash, ch)
+               return text
+
+       def _outdent(self, text):
+               # Remove one level of line-leading tabs or spaces
+               return self._outdent_re.sub('', text)
+
+
+class MarkdownWithExtras(Markdown):
+       """A markdowner class that enables most extras:
+
+       - footnotes
+       - code-color (only has effect if 'pygments' Python module on path)
+
+       These are not included:
+       - pyshell (specific to Python-related documenting)
+       - code-friendly (because it *disables* part of the syntax)
+       - link-patterns (because you need to specify some actual
+         link-patterns anyway)
+       """
+       extras = ["footnotes", "code-color"]
+
+
+#---- internal support functions
+
+# From http://aspn.activestate.com/ASPN/Cookbook/Python/Recipe/52549
+def _curry(*args, **kwargs):
+       function, args = args[0], args[1:]
+       def result(*rest, **kwrest):
+               combined = kwargs.copy()
+               combined.update(kwrest)
+               return function(*args + rest, **combined)
+       return result
+
+# Recipe: regex_from_encoded_pattern (1.0)
+def _regex_from_encoded_pattern(s):
+       """'foo'        -> re.compile(re.escape('foo'))
+          '/foo/'      -> re.compile('foo')
+          '/foo/i' -> re.compile('foo', re.I)
+       """
+       if s.startswith('/') and s.rfind('/') != 0:
+               # Parse it: /PATTERN/FLAGS
+               idx = s.rfind('/')
+               pattern, flags_str = s[1:idx], s[idx+1:]
+               flag_from_char = {
+                       "i": re.IGNORECASE,
+                       "l": re.LOCALE,
+                       "s": re.DOTALL,
+                       "m": re.MULTILINE,
+                       "u": re.UNICODE,
+               }
+               flags = 0
+               for char in flags_str:
+                       try:
+                               flags |= flag_from_char[char]
+                       except KeyError:
+                               raise ValueError("unsupported regex flag: '%s' in '%s' "
+                                                                "(must be one of '%s')"
+                                                                % (char, s, ''.join(flag_from_char.keys())))
+               return re.compile(s[1:idx], flags)
+       else: # not an encoded regex
+               return re.compile(re.escape(s))
+
+# Recipe: dedent (0.1.2)
+def _dedentlines(lines, tabsize=8, skip_first_line=False):
+       """_dedentlines(lines, tabsize=8, skip_first_line=False) -> dedented lines
+
+               "lines" is a list of lines to dedent.
+               "tabsize" is the tab width to use for indent width calculations.
+               "skip_first_line" is a boolean indicating if the first line should
+                       be skipped for calculating the indent width and for dedenting.
+                       This is sometimes useful for docstrings and similar.
+
+       Same as dedent() except operates on a sequence of lines. Note: the
+       lines list is modified **in-place**.
+       """
+       DEBUG = False
+       if DEBUG:
+               print "dedent: dedent(..., tabsize=%d, skip_first_line=%r)"\
+                         % (tabsize, skip_first_line)
+       indents = []
+       margin = None
+       for i, line in enumerate(lines):
+               if i == 0 and skip_first_line: continue
+               indent = 0
+               for ch in line:
+                       if ch == ' ':
+                               indent += 1
+                       elif ch == '\t':
+                               indent += tabsize - (indent % tabsize)
+                       elif ch in '\r\n':
+                               continue # skip all-whitespace lines
+                       else:
+                               break
+               else:
+                       continue # skip all-whitespace lines
+               if DEBUG: print "dedent: indent=%d: %r" % (indent, line)
+               if margin is None:
+                       margin = indent
+               else:
+                       margin = min(margin, indent)
+       if DEBUG: print "dedent: margin=%r" % margin
+
+       if margin is not None and margin > 0:
+               for i, line in enumerate(lines):
+                       if i == 0 and skip_first_line: continue
+                       removed = 0
+                       for j, ch in enumerate(line):
+                               if ch == ' ':
+                                       removed += 1
+                               elif ch == '\t':
+                                       removed += tabsize - (removed % tabsize)
+                               elif ch in '\r\n':
+                                       if DEBUG: print "dedent: %r: EOL -> strip up to EOL" % line
+                                       lines[i] = lines[i][j:]
+                                       break
+                               else:
+                                       raise ValueError("unexpected non-whitespace char %r in "
+                                                                        "line %r while removing %d-space margin"
+                                                                        % (ch, line, margin))
+                               if DEBUG:
+                                       print "dedent: %r: %r -> removed %d/%d"\
+                                                 % (line, ch, removed, margin)
+                               if removed == margin:
+                                       lines[i] = lines[i][j+1:]
+                                       break
+                               elif removed > margin:
+                                       lines[i] = ' '*(removed-margin) + lines[i][j+1:]
+                                       break
+                       else:
+                               if removed:
+                                       lines[i] = lines[i][removed:]
+       return lines
+
+def _dedent(text, tabsize=8, skip_first_line=False):
+       """_dedent(text, tabsize=8, skip_first_line=False) -> dedented text
+
+               "text" is the text to dedent.
+               "tabsize" is the tab width to use for indent width calculations.
+               "skip_first_line" is a boolean indicating if the first line should
+                       be skipped for calculating the indent width and for dedenting.
+                       This is sometimes useful for docstrings and similar.
+
+       textwrap.dedent(s), but don't expand tabs to spaces
+       """
+       lines = text.splitlines(1)
+       _dedentlines(lines, tabsize=tabsize, skip_first_line=skip_first_line)
+       return ''.join(lines)
+
+
+class _memoized(object):
+   """Decorator that caches a function's return value each time it is called.
+   If called later with the same arguments, the cached value is returned, and
+   not re-evaluated.
+
+   http://wiki.python.org/moin/PythonDecoratorLibrary
+   """
+   def __init__(self, func):
+         self.func = func
+         self.cache = {}
+   def __call__(self, *args):
+         try:
+                return self.cache[args]
+         except KeyError:
+                self.cache[args] = value = self.func(*args)
+                return value
+         except TypeError:
+                # uncachable -- for instance, passing a list as an argument.
+                # Better to not cache than to blow up entirely.
+                return self.func(*args)
+   def __repr__(self):
+         """Return the function's docstring."""
+         return self.func.__doc__
+
+
+def _xml_oneliner_re_from_tab_width(tab_width):
+       """Standalone XML processing instruction regex."""
+       return re.compile(r"""
+               (?:
+                       (?<=\n\n)               # Starting after a blank line
+                       |                               # or
+                       \A\n?                   # the beginning of the doc
+               )
+               (                                                       # save in $1
+                       [ ]{0,%d}
+                       (?:
+                               <\?\w+\b\s+.*?\?>       # XML processing instruction
+                               |
+                               <\w+:\w+\b\s+.*?/>      # namespaced single tag
+                       )
+                       [ \t]*
+                       (?=\n{2,}|\Z)           # followed by a blank line or end of document
+               )
+               """ % (tab_width - 1), re.X)
+_xml_oneliner_re_from_tab_width = _memoized(_xml_oneliner_re_from_tab_width)
+
+def _hr_tag_re_from_tab_width(tab_width):
+        return re.compile(r"""
+               (?:
+                       (?<=\n\n)               # Starting after a blank line
+                       |                               # or
+                       \A\n?                   # the beginning of the doc
+               )
+               (                                               # save in \1
+                       [ ]{0,%d}
+                       <(hr)                           # start tag = \2
+                       \b                                      # word break
+                       ([^<>])*?                       #
+                       /?>                                     # the matching end tag
+                       [ \t]*
+                       (?=\n{2,}|\Z)           # followed by a blank line or end of document
+               )
+               """ % (tab_width - 1), re.X)
+_hr_tag_re_from_tab_width = _memoized(_hr_tag_re_from_tab_width)
+
+
+def _xml_encode_email_char_at_random(ch):
+       r = random()
+       # Roughly 10% raw, 45% hex, 45% dec.
+       # '@' *must* be encoded. I [John Gruber] insist.
+       # Issue 26: '_' must be encoded.
+       if r > 0.9 and ch not in "@_":
+               return ch
+       elif r < 0.45:
+               # The [1:] is to drop leading '0': 0x63 -> x63
+               return '&#%s;' % hex(ord(ch))[1:]
+       else:
+               return '&#%s;' % ord(ch)
+
+
+
+
+text = """\
+Dies ist ein Text.
+
+---
+
+* Test
+* Mu
+* Blah
+"""
+
+#markdown(text, html4tags=False, tab_width=DEFAULT_TAB_WIDTH,
+#                        safe_mode=None, extras=None, link_patterns=None):
+#html = markdown(text, html4tags=False)
+#
+#print html
+
+
 
 @set_hook("read")
 def read(params):
@@ -32,7 +1629,7 @@ def htmlize(params):
                        "xml":True,
                        "demote-headers":1,
                        "code-color":{}})
-       html = _markdown.convert(params.file.contents)
+       html = _markdown.convert(params.file.contents, params.file.input_encoding)
        #print type(html)
        #print html
        return html