Fix link to lifetime elision rules
[rust-101.git] / pycco-rs
index d250b20e1e36976028ec2e653f6df7749a3abc3a..c8843112582228dfe3edc52884bdddff16d8991d 100755 (executable)
--- a/pycco-rs
+++ b/pycco-rs
@@ -11,16 +11,16 @@ def patch_html(source, marker, new_text):
     return source.replace(marker, marker + new_text, 1)
 
 # now, monkey-patch pycco for Rust support
-pycco.main.languages[".rs"] = { "name": "rust", "symbol": "//"}
-for ext, l in pycco.main.languages.items():
+pycco.main.supported_languages[".rs"] = { "name": "rust", "comment_symbol": "//"}
+for ext, l in pycco.main.supported_languages.items():
     # Does the line begin with a comment?
-    l["comment_matcher"] = re.compile(r"^\s*" + l["symbol"] + "\s?")
+    l["comment_matcher"] = re.compile(r"^\s*" + l["comment_symbol"] + "\s?")
     # The dividing token we feed into Pygments, to delimit the boundaries between
     # sections.
-    l["divider_text"] = "\n" + l["symbol"] + "DIVIDER\n"
+    l["divider_text"] = "\n" + l["comment_symbol"] + "DIVIDER\n"
     # The mirror of `divider_text` that we expect Pygments to return. We can split
     # on this to recover the original sections.
-    l["divider_html"] = re.compile(r'\n*<span class="c[1]?">' + l["symbol"] + 'DIVIDER</span>\n*')
+    l["divider_html"] = re.compile(r'\n*<span class="c[1]?">' + l["comment_symbol"] + 'DIVIDER</span>\n*')
     # Get the Pygments Lexer for this language.
     l["lexer"] = lexers.get_lexer_by_name(l["name"])