parent
13fb5dac1c
commit
43d40f7fce
@ -0,0 +1,7 @@ |
||||
{ |
||||
"site_root":"https://example.com", |
||||
"title":"Test Metadata", |
||||
"author": "Test User", |
||||
"author_email": "test_user@example.com", |
||||
"uuid_oid_root": "pixywerk-demo" |
||||
} |
@ -0,0 +1,33 @@ |
||||
<?xml version="1.0" encoding="utf-8"?> |
||||
|
||||
<feed xmlns="http://www.w3.org/2005/Atom"> |
||||
|
||||
<title>{{ metadata.title }}</title> |
||||
<subtitle>{{ metadata.subtitle }}</subtitle> |
||||
<link href="{{ metadata.site_root }}/{{ metadata.file_name }}" rel="self" /> |
||||
<link href="{{ metadata.site_root }}" /> |
||||
<id>urn:uuid:{{ metadata.uuid }}</id> |
||||
<updated>{{ get_time_iso8601(metadata['build-time']) }}</updated> |
||||
|
||||
{% set posts = get_file_list('blog_posts/*.cont') %} |
||||
{% for post in posts %} |
||||
{% set post_meta = get_file_metadata(post['file_path']) %} |
||||
<entry> |
||||
<title>{{ post_meta.title }}</title> |
||||
<link href="{{ metadata.site_root }}/{{post_meta.file_path}}" /> |
||||
<id>urn:uuid:{{ post_meta.uuid }}</id> |
||||
<updated>{{ get_time_iso8601(post_meta.stat.mtime) }}</updated> |
||||
<summary>{{post_meta.summary }}</summary> |
||||
<!-- this would be the snippet, more than summary chunk --> |
||||
<!-- <content type="xhtml"> --> |
||||
<!-- <div xmlns="http://www.w3.org/1999/xhtml"> --> |
||||
<!-- <p>{{ post_meta.summary }}</p> --> |
||||
<!-- </div> --> |
||||
<!-- </content> --> |
||||
<author> |
||||
<name>{{ post_meta.author }}</name> |
||||
<email>{{ post_meta.author_email }}</email> |
||||
</author> |
||||
</entry> |
||||
{% endfor %} |
||||
</feed> |
@ -0,0 +1,5 @@ |
||||
{ |
||||
"type": "templatable", |
||||
"title": "Test RSS Feed", |
||||
"subtitle": "Some Subtitle" |
||||
} |
@ -0,0 +1,5 @@ |
||||
Some more post |
||||
|
||||
|
||||
la la la |
||||
|
@ -0,0 +1,4 @@ |
||||
{ |
||||
"title":"Another Post(tm)", |
||||
"summary":"Yet another post" |
||||
} |
@ -0,0 +1 @@ |
||||
Some content. |
@ -0,0 +1,4 @@ |
||||
{ |
||||
"title":"Test.cont", |
||||
"summary":"Some empty test content" |
||||
} |
@ -0,0 +1 @@ |
||||
yo fresh |
@ -0,0 +1,5 @@ |
||||
{ |
||||
"foo":"bar", |
||||
"title":"A title", |
||||
"summary":"Just a post." |
||||
} |
@ -0,0 +1,19 @@ |
||||
<h1>Index of all content</h1> |
||||
{% for f in get_file_list('*', sort_order='file_name') %} |
||||
<a href="{{ get_file_name(f['file_name']) }}">{{get_file_name(f['file_name'])}}</a> |
||||
{% endfor %} |
||||
|
||||
<p>Including foo.cont.meta: |
||||
<pre> |
||||
{{ get_file_content('foo.cont.meta') }} |
||||
</pre> |
||||
</p> |
||||
|
||||
<h1>Metadata</h1> |
||||
<table class="metadata"> |
||||
<tr><th>key</th><th>value</th></tr> |
||||
{% set metadata = get_file_metadata('foo.cont') %} |
||||
{% for k in metadata.keys() %} |
||||
<tr><td>{{k}}</td><td>{{metadata[k]}}</td></tr> |
||||
{% endfor %} |
||||
</table> |
@ -0,0 +1,9 @@ |
||||
# README # |
||||
|
||||
This is a test of the emergency compiled HTML system. This is only a *test*. |
||||
|
||||
[Foo!](foo.html) |
||||
|
||||
{% for i in range(100) %} |
||||
* {{ i }} |
||||
{% endfor %} |
@ -0,0 +1,3 @@ |
||||
{ |
||||
"pragma":["no-proc"] |
||||
} |
@ -0,0 +1,9 @@ |
||||
# README # |
||||
|
||||
This is a test of the emergency compiled HTML system. This is only a *test*. |
||||
|
||||
[Foo!](foo.html) |
||||
|
||||
{% for i in range(100) %} |
||||
* {{ i }} |
||||
{% endfor %} |
@ -0,0 +1,3 @@ |
||||
{ |
||||
"title":"Yo, markdown" |
||||
} |
@ -0,0 +1,32 @@ |
||||
<!DOCTYPE html> |
||||
<head> |
||||
<title>Debug for {{path}}</title> |
||||
<style type="text/css"> |
||||
table { border: 1px solid black; } |
||||
div { border: 1px solid black; } |
||||
td { border: 1px solid black; } |
||||
</style> |
||||
</head> |
||||
<body> |
||||
<p>{{path}}</p> |
||||
<h1>Content</h1> |
||||
<div class="content"> |
||||
{{content}} |
||||
</div> |
||||
|
||||
<h1>Environment</h1> |
||||
<table class="environment"> |
||||
<tr><th>key</th><th>value</th></tr> |
||||
{% for k in environ.keys() %} |
||||
<tr><td>{{k}}</td><td>{{environ[k]}}</td></tr> |
||||
{% endfor %} |
||||
</table> |
||||
|
||||
<h1>Metadata</h1> |
||||
<table class="metadata"> |
||||
<tr><th>key</th><th>value</th></tr> |
||||
{% for k in metadata.keys() %} |
||||
<tr><td>{{k}}</td><td>{{metadata[k]}}</td></tr> |
||||
{% endfor %} |
||||
</table> |
||||
</body> |
@ -0,0 +1,6 @@ |
||||
<table class="werk-file-list"> |
||||
<tr class="werk-file-list-head"><th>file</th><th>type</th><th>size</th><th>last change</th></tr> |
||||
{% for f in files.keys() %} |
||||
<tr class="werk-file-list-item"><td><a href="/{{files[f].relpath}}">{{f}}</a></td><td>{{files[f].type}}</td><td>{{files[f].size}}</td><td>{{files[f].ctime | date}}</td></tr> |
||||
{% endfor %} |
||||
</table> |
@ -0,0 +1,13 @@ |
||||
<!DOCTYPE html> |
||||
<head> |
||||
<title>{{metadata.title}}</title> |
||||
<style type="text/css"> |
||||
table { border: 1px solid black; } |
||||
div { border: 1px solid black; } |
||||
td { border: 1px solid black; } |
||||
</style> |
||||
</head> |
||||
<body> |
||||
{{content}} |
||||
</body> |
||||
</html> |
@ -0,0 +1,119 @@ |
||||
# iterate source tree |
||||
# create directors in target tree |
||||
# for each item: |
||||
# run processor(s) on item, each processor could be in a chain or a branch |
||||
# Processors also provide filename munging |
||||
# output target based on processor output |
||||
|
||||
import argparse |
||||
import logging |
||||
import os |
||||
import sys |
||||
import time |
||||
|
||||
from typing import Dict, List, cast |
||||
|
||||
from .processchain import ProcessorChains |
||||
from .metadata import MetaTree |
||||
from .template_tools import file_list, file_name, file_content, file_metadata, time_iso8601 |
||||
|
||||
|
||||
logger = logging.getLogger() |
||||
|
||||
|
||||
def setup_logging(verbose: bool = False) -> None: |
||||
pass |
||||
|
||||
|
||||
def get_args(args: List[str]) -> argparse.Namespace: |
||||
parser = argparse.ArgumentParser("Compile a Pixywerk directory into an output directory.") |
||||
|
||||
parser.add_argument("root", help="The root of the pixywerk directory to process.") |
||||
parser.add_argument("output", help="The output directory to export post-compiled files to.") |
||||
|
||||
parser.add_argument( |
||||
"-c", "--clean", help="Remove the target tree before proceeding (by renaming to .bak).", action="store_true" |
||||
) |
||||
parser.add_argument("-s", "--safe", help="Abort if the target directory already exists.", action="store_true") |
||||
parser.add_argument("-t", "--template", help="The template directory (default: root/templates)", default=None) |
||||
parser.add_argument("-d", "--dry-run", help="Perform a dry-run.", action="store_true") |
||||
parser.add_argument("-v", "--verbose", help="Output verbosely.", action="store_true") |
||||
parser.add_argument("--processors", help="Specify a path to a processor configuration file.", default=None) |
||||
|
||||
result = parser.parse_args(args) |
||||
|
||||
# validate arguments |
||||
if not os.path.isdir(result.root): |
||||
raise FileNotFoundError("can't find root folder {}".format(result.root)) |
||||
|
||||
if not result.template: |
||||
result.template = os.path.join(result.root, "templates") |
||||
result.excludes = [result.template] |
||||
|
||||
return result |
||||
|
||||
|
||||
def main() -> int: |
||||
try: |
||||
args = get_args(sys.argv[1:]) |
||||
except FileNotFoundError as ex: |
||||
print("error finding arguments: {}".format(ex)) |
||||
return 1 |
||||
setup_logging(args.verbose) |
||||
if os.path.exists(args.output) and args.clean: |
||||
bak = "{}.bak-{}".format(args.output, int(time.time())) |
||||
print("cleaning target {} -> {}".format(args.output, bak)) |
||||
os.rename(args.output, bak) |
||||
|
||||
process_chains = ProcessorChains(args.processors) |
||||
|
||||
default_metadata = { |
||||
"templates": args.template, |
||||
"template": "default.jinja2", |
||||
"dir-template": "default-dir.jinja2", |
||||
"filters": {}, |
||||
"build-time": time.time(), |
||||
"build-datetime": time.ctime(), |
||||
"uuid-oid-root": "pixywerk", |
||||
} |
||||
meta_tree = MetaTree(args.root, default_metadata) |
||||
file_list_cache = cast(Dict, {}) |
||||
file_cont_cache = cast(Dict, {}) |
||||
file_name_cache = cast(Dict, {}) |
||||
default_metadata["globals"] = { |
||||
"get_file_list": file_list(args.root, file_list_cache), |
||||
"get_file_name": file_name(args.root, meta_tree, process_chains, file_name_cache), |
||||
"get_file_content": file_content(args.root, meta_tree, process_chains, file_cont_cache), |
||||
"get_file_metadata": file_metadata(meta_tree), |
||||
"get_time_iso8601": time_iso8601("UTC"), |
||||
} |
||||
|
||||
for root, _, files in os.walk(args.root): |
||||
workroot = os.path.relpath(root, args.root) |
||||
if workroot == ".": |
||||
workroot = "" |
||||
target_dir = os.path.join(args.output, workroot) |
||||
print("mkdir -> {}".format(target_dir)) |
||||
if not args.dry_run: |
||||
try: |
||||
os.mkdir(target_dir) |
||||
except FileExistsError: |
||||
if args.safe: |
||||
print("error, target directory exists, aborting") |
||||
return 1 |
||||
for f in files: |
||||
# fixme global generic filters |
||||
if f.endswith(".meta") or f.endswith("~"): |
||||
continue |
||||
metadata = meta_tree.get_metadata(os.path.join(workroot, f)) |
||||
chain = process_chains.get_chain_for_filename(os.path.join(root, f), ctx=metadata) |
||||
print("process {} -> {}".format(os.path.join(root, f), os.path.join(target_dir, chain.output_filename))) |
||||
if not args.dry_run: |
||||
with open(os.path.join(target_dir, chain.output_filename), "w") as outfile: |
||||
for line in chain.output: |
||||
outfile.write(line) |
||||
return 0 |
||||
|
||||
|
||||
if __name__ == "__main__": |
||||
sys.exit(main()) |
@ -0,0 +1,104 @@ |
||||
# Default: output == input |
||||
default: |
||||
extension: default |
||||
chain: |
||||
- passthrough |
||||
|
||||
# Any object that needs jinja scripts but no other explicit processing |
||||
templatable: |
||||
extension: null |
||||
chain: |
||||
- jinja2 |
||||
|
||||
# Markdown, BBCode and RST are first run through the templater, and then |
||||
# they are processed into HTML, and finally embedded in a page template. |
||||
markdown: |
||||
extension: |
||||
- md |
||||
chain: |
||||
- jinja2 |
||||
- process_md |
||||
- jinja2_page_embed |
||||
bbcode: |
||||
extension: |
||||
- bb |
||||
- pp |
||||
chain: |
||||
- jinja2 |
||||
- process_pp |
||||
- jinja2_page_embed |
||||
# FIXME implement RST processor |
||||
# restructured: |
||||
# extension: |
||||
# - rst |
||||
# chain: |
||||
# - jinja2 |
||||
# - process_rst |
||||
# - jinja2_page_embed |
||||
|
||||
# # JSON and YAML are split, passed through a pretty printer, and then output |
||||
# FIXME implement split chain processor, implement processor arguments |
||||
# json: |
||||
# extension: |
||||
# - json |
||||
# chain: |
||||
# - split (passthrough) |
||||
# - pp_json |
||||
# yaml: |
||||
# extension: |
||||
# - yml |
||||
# - yaml |
||||
# chain: |
||||
# - split (passthrough) |
||||
# - pp_yaml |
||||
|
||||
# Template-html is first passed through the templater, and then embedded |
||||
# in a page template |
||||
template-html: |
||||
extension: |
||||
- thtml |
||||
- cont |
||||
chain: |
||||
- jinja2 |
||||
- jinja2_page_embed |
||||
|
||||
# Smart CSS are simply converted to CSS. |
||||
sass: |
||||
extension: |
||||
- sass |
||||
- scss |
||||
chain: |
||||
- process_sass |
||||
less: |
||||
extension: |
||||
- less |
||||
chain: |
||||
- process_less |
||||
|
||||
stylus: |
||||
extension: |
||||
- styl |
||||
chain: |
||||
- process_styl |
||||
|
||||
# # Images are processed into thumbnails and sized in addition to being retained as their original |
||||
# FIXME implement split chain processor, implement processor arguments, |
||||
# image: |
||||
# extension: |
||||
# - jpg |
||||
# - jpeg |
||||
# - png |
||||
# chain: |
||||
# - split (image_bigthumb) |
||||
# - split (image_smallthumb) |
||||
# - passthrough |
||||
|
||||
# image_bigthumb: |
||||
# extension: |
||||
# chain: |
||||
# - smart_resize (big) |
||||
|
||||
# image_smallthumb: |
||||
# extension: |
||||
# chain: |
||||
# - smart_resize (small) |
@ -0,0 +1,145 @@ |
||||
"""Constructs a tree-like object containing the metadata for a given path, and caches said metadata.""" |
||||
|
||||
import logging |
||||
import mimetypes |
||||
import os |
||||
import uuid |
||||
|
||||
from typing import Dict, Optional, Union, List, Tuple, Any, cast |
||||
|
||||
import jstyleson |
||||
|
||||
from .utils import guess_mime |
||||
|
||||
# setup mimetypes with some extra ones |
||||
mimetypes.init() |
||||
mimetypes.add_type("text/html", "thtml") |
||||
mimetypes.add_type("text/html", "cont") |
||||
|
||||
logger = logging.getLogger(__name__) |
||||
|
||||
|
||||
class MetaCacheMiss(Exception): |
||||
"""Raised on cache miss.""" |
||||
|
||||
|
||||
class MetaCache: |
||||
"""This class provides an in-memory cache for metadata tree.""" |
||||
|
||||
def __init__(self, max_age: float = 200.0): |
||||
"""Initialize the cache. |
||||
|
||||
Arguments: |
||||
max_age (int): the number of seconds to age-out cache items |
||||
|
||||
""" |
||||
self._max_age = max_age |
||||
self._cache: Dict[str, Tuple[float, Any]] = {} |
||||
|
||||
def get(self, key: str, new_time_stamp: float) -> Any: |
||||
"""Get an item from the cache. |
||||
|
||||
Arguments: |
||||
key (str): the cache key to retieve |
||||
new_time_stamp (int): The time to use to compare the stored time with |
||||
|
||||
Returns: |
||||
:obj:misc: The previously stored value. |
||||
|
||||
Raises: |
||||
MetaCacheMiss: on missing key, or on aged out |
||||
|
||||
""" |
||||
if key not in self._cache: |
||||
raise MetaCacheMiss("no item for key {}".format(key)) |
||||
|
||||
if self._cache[key][0] + self._max_age <= new_time_stamp: |
||||
return self._cache[key][1] |
||||
|
||||
raise MetaCacheMiss("cache expired for key {}".format(key)) |
||||
|
||||
def put(self, key: str, value: Union[Dict, List, int, str, object], time_stamp: float) -> None: |
||||
"""Put an item into the cache. |
||||
|
||||
Arguments: |
||||
key (str): the key to store the cache item under |
||||
value (:obj:misc): the value to store |
||||
time_stamp (float): the time stamp to store the item under |
||||
|
||||
""" |
||||
self._cache[key] = (time_stamp, value) |
||||
|
||||
|
||||
class MetaTree: |
||||
"""This provides an interface to loading and caching tree metadata for a given directory tree.""" |
||||
|
||||
def __init__(self, root: str, default_metadata: Optional[Dict] = None): |
||||
"""Initialize the metadata tree object. |
||||
|
||||
Arguments: |
||||
root (str): The path to the root of the file tree to operate on. |
||||
default_metadata (dict, optional): The default metadata to apply to the tree |
||||
|
||||
""" |
||||
self._cache = MetaCache() |
||||
if default_metadata is None: |
||||
default_metadata = {} |
||||
self._default_metadata = default_metadata |
||||
if root[-1] != "/": |
||||
root += "/" |
||||
self._root = root |
||||
|
||||
def get_metadata(self, rel_path: str) -> Dict: |
||||
"""Retrieve the metadata for a given path |
||||
|
||||
The general procedure is to iterate the tree, at each level |
||||
m load .meta (JSON formatted dictionary) for that level, and |
||||
then finally load the path.meta, and merge these dictionaries |
||||
in descendant order. |
||||
|
||||
Arguments: |
||||
rel_path (str): The path to retrieve the metadata for (relative to root) |
||||
|
||||
Returns: |
||||
dict: A dictionary of metadata for that path tree. |
||||
|
||||
""" |
||||
metablob = dict(self._default_metadata) |
||||
# iterate path components from root to target path |
||||
comps = [self._root] + rel_path.split("/") |
||||
fullpath = "" |
||||
for pth in comps: |
||||
fullpath = os.path.join(fullpath, pth) |
||||
st = os.stat(fullpath) |
||||
|
||||
cachekey = fullpath + ".meta" |
||||
meta = cast(Dict, {}) |
||||
try: |
||||
st_meta = os.stat(cachekey) |
||||
meta = self._cache.get(cachekey, st_meta.st_mtime) |
||||
except FileNotFoundError: |
||||
st_meta = None # type: ignore |
||||
except MetaCacheMiss: |
||||
meta = {} |
||||
|
||||
if not meta and st_meta: |
||||
meta = jstyleson.load(open(cachekey, "r")) |
||||
self._cache.put(cachekey, meta, st_meta.st_mtime) |
||||
|
||||
metablob.update(meta) |
||||
|
||||
# return final dict |
||||
metablob["dir"], metablob["file_name"] = os.path.split(rel_path) |
||||
metablob["file_path"] = rel_path |
||||
metablob["uuid"] = uuid.uuid3( |
||||
uuid.NAMESPACE_OID, metablob["uuid-oid-root"] + os.path.join(self._root, rel_path) |
||||
) |
||||
metablob["os-path"], _ = os.path.split(fullpath) |
||||
metablob["guessed-type"] = guess_mime(os.path.join(self._root, rel_path)) |
||||
if "mime-type" not in metablob: |
||||
metablob["mime-type"] = metablob["guessed-type"] |
||||
metablob["stat"] = {} |
||||
for stk in ("st_mtime", "st_ctime", "st_atime", "st_mode", "st_size", "st_ino"): |
||||
metablob["stat"][stk.replace("st_", "")] = getattr(st, stk) |
||||
|
||||
return metablob |
@ -0,0 +1,180 @@ |
||||
"""Interface for chains of processors""" |
||||
|
||||
import os |
||||
import os.path |
||||
import random |
||||
|
||||
from typing import List, Iterable, Optional, Any, Dict, Type, cast |
||||
|
||||
import yaml |
||||
|
||||
from .processors.processors import Processor |
||||
|
||||
|
||||
class ProcessorChain: |
||||
"""This implements a wrapper for an arbitrary set of processors and an associated file stream.""" |
||||
|
||||
def __init__( |
||||
self, |
||||
processors: List[Processor], |
||||
file_name: str, |
||||
file_data: Iterable[str], |
||||
file_type: str, |
||||
ctx: Optional[Dict] = None, |
||||
): |
||||
"""Initialize the processing stream. |
||||
|
||||
Arguments: |
||||
processors (list): A list of processor objects. |
||||
file_data (Iterable): An iterable from which to retrieve the input |
||||
file_type (str): the specified file type for consumer information. |
||||
|
||||
""" |
||||
self._processors = processors |
||||
self._file_data = file_data |
||||
self._file_type = file_type |
||||
self._file_name = file_name |
||||
self._ctx: Dict = {} |
||||
if ctx is not None: |
||||
self._ctx = cast(Dict, ctx) |
||||
|
||||
@property |
||||
def output(self) -> Iterable: |
||||
"""Return an iterable for the output of the process chain |
||||
|
||||
Returns: |
||||
:obj:'iterable': the iterable |
||||
|
||||
""" |
||||
prev = self._file_data |
||||
for processor in self._processors: |
||||
if processor: |
||||
prev = processor.process(prev, self._ctx) |
||||
|
||||
return prev |
||||
|
||||
@property |
||||
def output_mime(self) -> str: |
||||
"""Return the post-processed MIME value from the processing chain |
||||
|
||||
Returns: |
||||
str: the mime type |
||||
|
||||
""" |
||||
fname = self._file_name |
||||
for processor in self._processors: |
||||
fname = processor.mime_type(fname, self._ctx) |
||||
return fname |
||||
|
||||
@property |
||||
def output_ext(self) -> str: |
||||
"""Return the post-processed extension from the processing chain |
||||
|
||||
Returns: |
||||
str: the extension |
||||
""" |
||||
fname = self._file_name |
||||
for processor in self._processors: |
||||
fname = processor.extension(fname, self._ctx) |
||||
return fname |
||||
|
||||
@property |
||||
def output_filename(self) -> str: |
||||
"""Return the post-processed filename from the processing chain |
||||
|
||||
Returns: |
||||
str: the new filename |
||||
|
||||
""" |
||||
fname = os.path.basename(self._file_name) |
||||
for processor in self._processors: |
||||
fname = processor.filename(fname, self._ctx) |
||||
return fname |
||||
|
||||
|
||||
class ProcessorChains: |
||||
"""Load a configuration for processor chains, and provide ability to process the chains given a particular input |
||||
file. |
||||
""" |
||||
|
||||
def __init__(self, config: Optional[str] = None): |
||||
"""Initialize, with a specified configuration file |
||||
|
||||
Arguments: |
||||
config (str, optional): The path to a yaml formatted configuration file. |
||||
|
||||
""" |
||||
if config is None: # pragma: no coverage |
||||
config = os.path.join(os.path.dirname(__file__), "defaults", "chains.yaml") |
||||
|
||||
self.chainconfig = yaml.load(open(config, "r")) |
||||
self.extensionmap: Dict[str, Any] = {} |
||||
self.processors: Dict[str, Type[Processor]] = {} |
||||
for ch, conf in self.chainconfig.items(): |
||||
if conf["extension"] == "default": |
||||
self.default = ch |
||||
else: |
||||
if conf["extension"]: |
||||
for ex in conf["extension"]: |
||||
if ex in self.extensionmap or ex is None: |
||||
# log an error or except or something we'll just override for now. |
||||
pass |
||||
self.extensionmap[ex] = ch |
||||
for pr in conf["chain"]: |
||||
if pr in self.processors: |
||||
continue |
||||
processor_module = __import__("processors", globals(), locals(), [pr], 1) |
||||
self.processors[pr] = processor_module.__dict__[pr].processor |
||||
|
||||
def get_chain_for_filename(self, filename: str, ctx: Optional[Dict] = None) -> ProcessorChain: |
||||
"""Get the ProcessorChain, as configured for a given file by extension. |
||||
|
||||
Arguments: |
||||
filename (str): The name of the file to get a chain for. |
||||
|
||||
Returns: |
||||
ProcessorChain: the constructed processor chain. |
||||
""" |
||||
r = filename.rsplit(".", 1) |
||||
ftype = "default" |
||||
if r: |
||||
ftype = r[-1] |
||||
if ctx and "pragma" in ctx: |
||||
if "no-proc" in ctx["pragma"]: |
||||
ftype = "default" |
||||
|
||||
if ctx and "type" in ctx: |
||||
ftype = ctx["type"] |
||||
return self.get_chain_for_file(open(filename, "r"), ftype, filename, ctx) |
||||
|
||||
def get_chain_for_file( |
||||
self, file_obj: Iterable, file_ext: str, file_name: Optional[str] = None, ctx: Optional[Dict] = None |
||||
) -> ProcessorChain: |
||||
"""Get the ProcessorChain for a given iterable object based on the specified file type |
||||
|
||||
Arguments: |
||||
file_obj (:obj:`iterable`): The input file stream |
||||
file_ext (str): The type (extension) of the input stream |
||||
|
||||
Returns: |
||||
ProcessorChain: the constructed processor chain. |
||||
|
||||
""" |
||||
if file_ext not in self.extensionmap or not self.extensionmap[file_ext]: |
||||
if file_ext in self.chainconfig: |
||||
file_type = file_ext |
||||
else: |
||||
file_type = "default" |
||||
else: |
||||
file_type = self.extensionmap[file_ext] |
||||
|
||||
if not (bool(file_name)): |
||||
file_name = hex(random.randint(0, 65536)) |
||||
|
||||
return ProcessorChain( |
||||
[self.processors[x]() for x in self.chainconfig[file_type]["chain"]], |
||||
cast(str, file_name), |
||||
file_obj, |
||||
file_type, |
||||
ctx, |
||||
) |
@ -0,0 +1 @@ |
||||
# processors metadata here |
@ -0,0 +1,32 @@ |
||||
"""Define a Jinja2 Processor which applies programmable templating to the input stream.""" |
||||
|
||||
from typing import Iterable, Optional, Dict, cast |
||||
|
||||
from jinja2 import Environment, FileSystemLoader |
||||
|
||||
from .passthrough import PassThrough |
||||
|
||||
|
||||
class Jinja2(PassThrough): |
||||
"""Pass the input stream through Jinja2 for scritable templating.""" |
||||
|
||||
def process(self, input_file: Iterable, ctx: Optional[Dict] = None) -> Iterable: |
||||
"""Return an iterable object of the post-processed file. |
||||
|
||||
Arguments: |
||||
input_file (iterable): An input stream |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
|
||||
Returns: |
||||
iterable: The post-processed output stream |
||||
""" |
||||
ctx = cast(Dict, ctx) |
||||
template_env = Environment(loader=FileSystemLoader(ctx["templates"])) |
||||
template_env.globals.update(ctx["globals"]) |
||||
template_env.filters.update(ctx["filters"]) |
||||
tmpl = template_env.from_string("".join([x for x in input_file])) |
||||
return tmpl.render(metadata=ctx) |
||||
|
||||
|
||||
processor = Jinja2 |
@ -0,0 +1,76 @@ |
||||
"""Define a Jinja2 processor which embeds the (presumably HTML) input stream into a Page Template |
||||
as defined in the ctx metadata (the ``content`` variable is assigned to the input stream and |
||||
the target template is rendered).""" |
||||
|
||||
import os |
||||
|
||||
from typing import Iterable, Optional, Dict, cast |
||||
|
||||
from jinja2 import Environment, FileSystemLoader |
||||
|
||||
from .processors import Processor |
||||
|
||||
|
||||
class Jinja2PageEmbed(Processor): |
||||
"""Embed input stream as ``content`` variable in page template defined in context key ``template``.""" |
||||
|
||||
def filename(self, oldname: str, ctx: Optional[Dict] = None) -> str: |
||||
"""Return the filename of the post-processed file. |
||||
|
||||
Arguments: |
||||
oldname (str): the previous name for the file. |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
Returns: |
||||
str: the new name for the file |
||||
|
||||
""" |
||||
|
||||
return os.path.splitext(oldname)[0] + ".html" |
||||
|
||||
def mime_type(self, oldname: str, ctx: Optional[Dict] = None) -> str: |
||||
"""Return the mimetype of the post-processed file. |
||||
|
||||
Arguments: |
||||
oldname (str): the input filename |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
Returns: |
||||
str: the new mimetype of the file after processing |
||||
|
||||
""" |
||||
return "text/html" |
||||
|
||||
def process(self, input_file: Iterable, ctx: Optional[Dict] = None) -> Iterable: |
||||
"""Return an iterable object of the post-processed file. |
||||
|
||||
Arguments: |
||||
input_file (iterable): An input stream |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
Returns: |
||||
iterable: The post-processed output stream |
||||
""" |
||||
ctx = cast(Dict, ctx) |
||||
template_env = Environment(loader=FileSystemLoader(ctx["templates"])) |
||||
template_env.globals.update(ctx["globals"]) |
||||
template_env.filters.update(ctx["filters"]) |
||||
tmpl = template_env.get_template(ctx["template"]) |
||||
content = "".join([x for x in input_file]) |
||||
return tmpl.render(content=content, metadata=ctx) |
||||
|
||||
def extension(self, oldname: str, ctx: Optional[Dict] = None) -> str: |
||||
"""Return the mimetype of the post-processed file. |
||||
|
||||
Arguments: |
||||
oldname (str): the input filename |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
Returns: |
||||
str: the new extension of the file after processing |
||||
|
||||
""" |
||||
return "html" |
||||
|
||||
|
||||
processor = Jinja2PageEmbed |
@ -0,0 +1,68 @@ |
||||
"""Passthrough progcessor which takes input and returns it.""" |
||||
|
||||
import os |
||||
|
||||
from .processors import Processor |
||||
from ..utils import guess_mime |
||||
from typing import Iterable, Optional, Dict, cast |
||||
|
||||
|
||||
class PassThrough(Processor): |
||||
"""A simple passthrough processor that takes input and sends it to output.""" |
||||
|
||||
def filename(self, oldname: str, ctx: Optional[Dict] = None) -> str: |
||||
"""Return the filename of the post-processed file. |
||||
|
||||
Arguments: |
||||
oldname (str): the previous name for the file. |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
Returns: |
||||
str: the new name for the file |
||||
|
||||
""" |
||||
return oldname |
||||
|
||||
def mime_type(self, oldname: str, ctx: Optional[Dict] = None) -> str: |
||||
"""Return the mimetype of the post-processed file. |
||||
|
||||
Arguments: |
||||
oldname (str): the input filename |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
Returns: |
||||
str: the new mimetype of the file after processing |
||||
|
||||
""" |
||||
result = cast(str, guess_mime(oldname)) |
||||
if result == "directory": |
||||
result = "DIR" |
||||
return result |
||||
|
||||
def process(self, input_file: Iterable, ctx: Optional[Dict] = None) -> Iterable: |
||||
"""Return an iterable object of the post-processed file. |
||||
|
||||
Arguments: |
||||
input_file (iterable): An input stream |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
Returns: |
||||
iterable: The post-processed output stream |
||||
""" |
||||
return input_file |
||||
|
||||
def extension(self, oldname: str, ctx: Optional[Dict] = None) -> str: |
||||
"""Return the mimetype of the post-processed file. |
||||
|
||||
Arguments: |
||||
oldname (str): the input filename |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
Returns: |
||||
str: the new extension of the file after processing |
||||
|
||||
""" |
||||
return os.path.splitext(oldname)[-1] |
||||
|
||||
|
||||
processor = PassThrough |
@ -0,0 +1 @@ |
||||
processor = None |
@ -0,0 +1,69 @@ |
||||
"""Convert an MD stream into an HTML stream""" |
||||
|
||||
import io |
||||
import os |
||||
|
||||
from typing import Iterable, Optional, Dict |
||||
|
||||
import markdown |
||||
|
||||
from .processors import Processor |
||||
|
||||
|
||||
class MarkdownProcessor(Processor): |
||||
"""Convert an MD stream into an HTML stream""" |
||||
|
||||
def filename(self, oldname: str, ctx: Optional[Dict] = None) -> str: |
||||
"""Return the filename of the post-processed file. |
||||
|
||||
Arguments: |
||||
oldname (str): the previous name for the file. |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
Returns: |
||||
str: the new name for the file |
||||
|
||||
""" |
||||
return os.path.splitext(oldname)[0] + ".html" |
||||
|
||||
def mime_type(self, oldname: str, ctx: Optional[Dict] = None) -> str: |
||||
"""Return the mimetype of the post-processed file. |
||||
|
||||
Arguments: |
||||
oldname (str): the input filename |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
Returns: |
||||
str: the new mimetype of the file after processing |
||||
|
||||
""" |
||||
return "text/html" |
||||
|
||||
def extension(self, oldname: str, ctx: Optional[Dict] = None) -> str: |
||||
"""Return the mimetype of the post-processed file. |
||||
|
||||
Arguments: |
||||
oldname (str): the input filename |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
Returns: |
||||
str: the new extension of the file after processing |
||||
|
||||
""" |
||||
return "html" |
||||
|
||||
def process(self, input_file: Iterable, ctx: Optional[Dict] = None) -> Iterable: |
||||
"""Return an iterable object of the post-processed file. |
||||
|
||||
Arguments: |
||||
input_file (iterable): An input stream |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
Returns: |
||||
iterable: The post-processed output stream |
||||
""" |
||||
md = u"".join([x for x in input_file]) |
||||
return io.StringIO(markdown.markdown(md, extensions=["extra", "admonition", "wikilinks"])) |
||||
|
||||
|
||||
processor = MarkdownProcessor # pylint: disable=invalid-name |
@ -0,0 +1 @@ |
||||
processor = None |
@ -0,0 +1 @@ |
||||
processor = None |
@ -0,0 +1 @@ |
||||
processor = None |
@ -0,0 +1,63 @@ |
||||
import abc |
||||
|
||||
from typing import Iterable, Optional, Dict |
||||
|
||||
|
||||
class ProcessorException(Exception): # pragma: no cover |
||||
"""A base exception class to be used by processor objects.""" |
||||
|
||||
|
||||
class Processor(abc.ABC): # pragma: no cover |
||||
def __init__(self, *args, **kwargs): |
||||
"""Initialize the class.""" |
||||
|
||||
@abc.abstractmethod |
||||
def filename(self, oldname: str, ctx: Optional[Dict] = None) -> str: |
||||
"""Return the filename of the post-processed file. |
||||
|
||||
Arguments: |
||||
oldname (str): the previous name for the file. |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
Returns: |
||||
str: the new name for the file |
||||
|
||||
""" |
||||
|
||||
@abc.abstractmethod |
||||
def mime_type(self, oldname: str, ctx: Optional[Dict] = None) -> str: |
||||
"""Return the mimetype of the post-processed file. |
||||
|
||||
Arguments: |
||||
oldname (str): the input filename |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
Returns: |
||||
str: the new mimetype of the file after processing |
||||
|
||||
""" |
||||
|
||||
@abc.abstractmethod |
||||
def extension(self, oldname: str, ctx: Optional[Dict] = None) -> str: |
||||
"""Return the mimetype of the post-processed file. |
||||
|
||||
Arguments: |
||||
oldname (str): the input filename |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
Returns: |
||||
str: the new extension of the file after processing |
||||
|
||||
""" |
||||
|
||||
@abc.abstractmethod |
||||
def process(self, input_file: Iterable, ctx: Optional[Dict] = None) -> Iterable: |
||||
"""Return an iterable object of the post-processed file. |
||||
|
||||
Arguments: |
||||
input_file (iterable): An input stream |
||||
ctx (dict, optional): A context object generated from the processor configuration |
||||
|
||||
Returns: |
||||
iterable: The post-processed output stream |
||||
""" |
@ -0,0 +1,80 @@ |
||||
import datetime |
||||
import glob |
||||
import itertools |
||||
import os |
||||
import pytz |
||||
from typing import Callable, Dict, List, Iterable, Union, cast |
||||
|
||||
from .metadata import MetaTree |
||||
from .processchain import ProcessorChains |
||||
|
||||
|
||||
def file_list(root: str, listcache: Dict) -> Callable: |
||||
def get_file_list(path_glob: str, *, sort_order: str = "ctime", reverse: bool = False, limit: int = 0) -> Iterable: |
||||
stattable = cast(List, []) |
||||
if path_glob in listcache: |
||||
stattable = listcache[path_glob] |
||||
else: |
||||
for fil in glob.glob(os.path.join(root, path_glob)): |
||||
if os.path.isdir(fil): |
||||
continue |
||||
if fil.endswith(".meta") or fil.endswith("~"): |
||||
continue |
||||
st = os.stat(fil) |
||||
stattable.append( |
||||
{ |
||||
"file_path": os.path.relpath(fil, root), |
||||
"file_name": os.path.split(fil)[-1], |
||||
"mtime": st.st_mtime, |
||||
"ctime": st.st_ctime, |
||||
"size": st.st_size, |
||||
"ext": os.path.splitext(fil)[1], |
||||
} |
||||
) |
||||
listcache[path_glob] = stattable |
||||
ret = sorted(stattable, key=lambda x: x[sort_order], reverse=reverse) |
||||
if limit > 0: |
||||
return itertools.islice(ret, limit) |
||||
return ret |
||||
|
||||
return get_file_list |
||||
|
||||
|
||||
def file_name(root: str, metatree: MetaTree, processor_chains: ProcessorChains, namecache: Dict) -> Callable: |
||||
def get_file_name(file_name: str) -> Dict: |
||||
if file_name in namecache: |
||||
return namecache[file_name] |
||||
metadata = metatree.get_metadata(file_name) |
||||
chain = processor_chains.get_chain_for_filename(os.path.join(root, file_name), ctx=metadata) |
||||
namecache[file_name] = chain.output_filename |
||||
return namecache[file_name] |
||||
|
||||
return get_file_name |
||||
|
||||
|
||||
def file_content(root: str, metatree: MetaTree, processor_chains: ProcessorChains, contcache: Dict) -> Callable: |
||||
def get_file_content(file_name: str) -> Iterable: |
||||
if file_name in contcache: |
||||
return contcache[file_name] |
||||
metadata = metatree.get_metadata(file_name) |
||||
chain = processor_chains.get_chain_for_filename(os.path.join(root, file_name), ctx=metadata) |
||||
contcache[file_name] = chain.output |
||||
return chain.output |
||||
|
||||
return get_file_content |
||||
|
||||
|
||||
def file_metadata(metatree: MetaTree) -> Callable: |
||||
def get_file_metadata(file_name: str) -> Dict: |
||||
return metatree.get_metadata(file_name) |
||||
|
||||
return get_file_metadata |
||||
|
||||
|
||||
def time_iso8601(timezone: str) -> Callable: |
||||
tz = pytz.timezone(timezone) |
||||
|
||||
def get_time_iso8601(time_t: Union[int, float]) -> str: |
||||
return datetime.datetime.fromtimestamp(time_t, tz).isoformat("T") |
||||
|
||||
return get_time_iso8601 |
@ -0,0 +1,6 @@ |
||||
class TestProcessChain: |
||||
def test_process_chain(self): |
||||
pass |
||||
|
||||
def test_processor_chain(self): |
||||
pass |
@ -0,0 +1,42 @@ |
||||
import mimetypes |
||||
import os |
||||
|
||||
from typing import Dict, Optional |
||||
|
||||
|
||||
def merge_dicts(dict_a: Dict, dict_b: Dict) -> Dict: |
||||
"""Merge two dictionaries. |
||||
|
||||
Arguments: |
||||
dict_a (dict): The dictionary to use as the base. |
||||
dict_b (dict): The dictionary to update the values with. |
||||
|
||||
Returns: |
||||
dict: A new merged dictionary. |
||||
|
||||
""" |
||||
dict_z = dict_a.copy() |
||||
dict_z.update(dict_b) |
||||
return dict_z |
||||
|
||||
|
||||
def guess_mime(path: str) -> Optional[str]: |
||||
"""Guess the mime type for a given path. |
||||
|
||||
Arguments: |
||||
root (str): the root path of the file tree |
||||
path (str): the sub-path within the file tree |
||||
|
||||
Returns: |
||||
str: the guessed mime-type |
||||
|
||||
""" |
||||
mtypes = mimetypes.guess_type(path) |
||||
ftype = None |
||||
if os.path.isdir(path): |
||||
ftype = "directory" |
||||
elif os.access(path, os.F_OK) and mtypes[0]: |
||||
ftype = mtypes[0] |
||||
else: |
||||
ftype = "application/octet-stream" |
||||
return ftype |
@ -0,0 +1,58 @@ |
||||
"""Package configuration.""" |
||||
from setuptools import find_packages, setup |
||||
|
||||
LONG_DESCRIPTION = """Pixywerk 2 is a DWIM filesystem based static site generator.""" |
||||
|
||||
INSTALL_REQUIRES = ["yaml-1.3", "markdown", "jstyleson", "jinja2"] |
||||
|
||||
# Extra dependencies |
||||
EXTRAS_REQUIRE = { |
||||
# Test dependencies |
||||
"tests": [ |
||||
"black", |
||||
"bandit>=1.1.0", |
||||
"flake8>=3.2.1", |
||||
"mypy>=0.470", |
||||
"prospector[with_everything]>=0.12.4", |
||||
"pytest-cov>=1.8.0", |
||||
"pytest-xdist>=1.15.0", |
||||
"pytest>=3.0.3", |
||||
"sphinx_rtd_theme>=0.1.6", |
||||
"sphinx-argparse>=0.1.15", |
||||
"Sphinx>=1.4.9", |
||||
] |
||||
} |
||||
|
||||
SETUP_REQUIRES = ["pytest-runner>=2.7.1", "setuptools_scm>=1.15.0"] |
||||
setup( |
||||
author="Cassowary Rusnov", |
||||
author_email="rusnovn@gmail.com", |
||||
classifiers=[ |
||||
"Development Status :: 1 - Pre-alpha", |
||||
"Environment :: Console", |
||||
"License :: OSI Approved :: MIT", |
||||
"Operating System :: POSIX :: Linux", |
||||
"Programming Language :: Python :: 3.6", |
||||
"Programming Language :: Python :: 3.7", |
||||
"Programming Language :: Python :: 3 :: Only", |
||||
"Topic :: Software Development :: Libraries :: Python Modules", |
||||
], |
||||
description="A filesystem-based DWIM website generator / CMS", |
||||
# entry_points={ |
||||
# 'console_scripts': [ |
||||
# 'cookbook = spicerack.cookbook:main', |
||||
# ], |
||||
# }, |
||||
extras_require=EXTRAS_REQUIRE, |
||||
install_requires=INSTALL_REQUIRES, |
||||
keywords=["wmf", "automation", "orchestration"], |
||||
license="MIT", |
||||
long_description=LONG_DESCRIPTION, |
||||
name="pixywerk2", # Must be the same used for __version__ in __init__.py |
||||
packages=find_packages(exclude=["*.tests", "*.tests.*"]), |
||||
platforms=["GNU/Linux"], |
||||
setup_requires=SETUP_REQUIRES, |
||||
use_scm_version=True, |
||||
url="https://git.antpanethon.com/cas/pixywerk2", |
||||
zip_safe=False, |
||||
) |
@ -0,0 +1,22 @@ |
||||
[tox] |
||||
envlist=py{36,37}-{code-quality, unit} #, py37-sphinx |
||||
skipsdist = true |
||||
|
||||
[testenv] |
||||
setenv = |
||||
LANG = en_US.UTF-8 |
||||
deps = .[tests] |
||||
commands = |
||||
unit: py.test --strict --cov-report=term-missing --cov=pixywerk2 pixywerk2/tests/unit {posargs} |
||||
code-quality: flake8 pixywerk2 |
||||
code-quality: black -l 120 --check pixywerk2 |
||||
code-quality: - prospector -A |
||||
code-quality: - mypy --ignore-missing-imports pixywerk2 |
||||
# sphinx: python setup.py build_sphinx -b html |
||||
# sphinx: python setup.py build_sphinx -b man |
||||
basepython = |
||||
py36: python3.6 |
||||
py37: python3.7 |
||||
|
||||
[flake8] |
||||
max-line-length = 120 |
Reference in new issue