1
0

Allow txt content and deny robots

This commit is contained in:
Tim Van Baak 2023-09-11 17:48:03 -07:00
parent 7755f2be31
commit c652657ea7
2 changed files with 26 additions and 20 deletions

View File

@ -40,7 +40,7 @@ def main():
count += 1
# Future-proofing
if not filename.endswith(".html") and not filename.endswith(".md"):
if filename.rsplit(".")[-1] not in ("html", "md", "txt"):
raise Exception("Support for this filetype is not yet supported:", filename)
path = src / dirpath / filename
@ -58,30 +58,34 @@ def main():
content = md.convert(content)
meta = md.Meta
# Inject content into the template
page_content = bs4.BeautifulSoup(content, features="html.parser")
page = copy.copy(template)
article = page.new_tag("article")
article.append(page_content)
page.article.replace_with(article)
if dest.name.endswith("html"):
# Inject content into the template
page_content = bs4.BeautifulSoup(content, features="html.parser")
page = copy.copy(template)
article = page.new_tag("article")
article.append(page_content)
page.article.replace_with(article)
# Inject path into the nav
for i in range(len(dirpath.parts)):
a = page.new_tag("a")
a["href"] = "/" + "/".join(dirpath.parts[:i+1]) + "/"
a.string = dirpath.parts[i]
page.nav.append(a)
page.nav.append(page.new_string("/"))
# Inject path into the nav
for i in range(len(dirpath.parts)):
a = page.new_tag("a")
a["href"] = "/" + "/".join(dirpath.parts[:i+1]) + "/"
a.string = dirpath.parts[i]
page.nav.append(a)
page.nav.append(page.new_string("/"))
# Apply metadata to the template
if meta_title := meta.get("title"):
title = "".join(meta_title)
page.title.string = title
page.header.h1.string = title
# Apply metadata to the template
if meta_title := meta.get("title"):
title = "".join(meta_title)
page.title.string = title
page.header.h1.string = title
# The fully templated page is the new content
content = str(page)
# Write the destination file
print("Writing ", dest)
dest.write_text(str(page))
dest.write_text(content)
print("Processed", count, "files")

2
src/robots.txt Normal file
View File

@ -0,0 +1,2 @@
User-agent: *
Disallow: /