@gaditb but the shorter version here is that i’ve already resigned myself to requiring libxml2 and libxslt, which are widely‐available if not POSIX, but don’t want to have to choose between Info-Zip, libzip, and libarchive, and would rather if pax just had a zip format added instead
@gaditb there aren’t POSIX XML capabilities, but i’d draw a distinction between content creation and content packaging. i think it’s reasonable to allow OS‐specific tools for content creation while expecting content packaging to be portable across OS’s. (i also would love if XML were POSIX, namely the core functionality of xmllint and xsltproc, but that’s never going to happen.)
i would switch to a B·S·D if every B·S·D didn’t give vibes of being staffed entirely by missing stairs
what version of unicode is used for APFS, for example? the original specification says Unicode 9.0, but i can make files named “🥠” on my computer today
so many formats these days are just specially-engineered zip files and there is no Posix way of engaging with them
the use of <书社:link xlink:href="about:shushe?include=things/"> as a shorthand for doing a link with every file nested in things/ is definitely a little conceptually sus, especially once i get to supporting @xlink:show values other than "embed", but it is so convenient 😔
the approach is to define a
<x:wrapper>
<书社:link xlink:href="about:shushe?include=things/" xlink:show="embed"/>
</x:wrapper>
which then gets all the files in $(INCLUDEDIR)/things/ and transforms to a
<书社:archive 书社:expanded="">
<html:article 书社:archived-as="index.html"><!--…--></html:article>
<!--…-->
</书社:archive>
not hard to do at all
the thing with the approach of “treat everything as an include, create a conceptual archive containing the entire site, and then extract that into public” is that (1) being able to just xpath into any file at any time is actually extremely powerful, and (2) this does just mean loading your entire site into memory and doing every transformation on it all in one go
one time someone in fandom coders was complaining because their static site generator took like 10 seconds to rebuild their very simple site and i was just like “…what are they doing in there?”
unrelatedly, i AM working on building a unified RDF/OWL ontology for my projects using ⛩️📰 书社 and it IS definitely stress-testing its capacity to handle “actual workloads”
jury is still out whether something complexly interconnected like that or something with Just A Lot Of Data (e·g a dictionary) will bog it down more
but i’m pretty sure it’s still faster than Any Static Site Generator
Administrator / Public Relations for GlitchCat. Not actually glitchy, nor a cat. I wrote the rules for this instance.
“Constitutionally incapable of not going hard” — @aescling
“Fedi Cassandra” – @Satsuma
I HAVE EXPERIENCE IN THINGS. YOU CAN JUST @ ME.
I work for a library but I post about Zelda fanfiction.