Or perhaps MDN is wrong and #Safari supports Client API?
Proof-of-Concept of the new signalling system done without removing the old one.
Can anyone test on Safari please? Open a new tab, open the JS console, and navigate here:
Then, reload (so that the service worker kicks in); you should see "ServiceWorker: yes" in orange.
Make sure that you see this commit ID in the console and in both places at the page bottom: c223b08c
If all of this is true, check if in the console you have messages saying: "SamizdatInfo received!"
Done some serious work on #Samizdat. Fixed some bugs, almost finished implementing the new messaging system (based on client.postMessage() in the end), ripped the old Indexed DB-based system out completely. Introduced new bugs to fix next.
Merge request here:
Still work in progress though.
Merged! #Samizdat now uses message passing instead of Indexed DB for ServiceWorker to inform the window clients of things. I CAN HAZ nice things, liek:
- info that a resource was fetched from cache, but fetching it via Gun+IPFS is running in background;
- near-instant info on resources being fetched and status of that;
- info when all resources get initially fetched (in the future this is when "stuff fetched from cache, but newer versions available, reload please" message will be displayed).
The Merge Request of Doom:
You might need to reload the service worker (refer to browser docs). Automagic reloading of the service worker code will come... one day, inshallah!
Also, probably doesn't work on Safari, because crapple refuses to implement things. Graceful degradation will come... one day, inshallah!
So I guess the roadmap to #Samizdat 1.0-beta would be something along the lines of:
- fix the issues (like caching plugin use is double-counted; when reloading soon after a load there is no indication how/where the resources were loaded from);
- implement the "stuff loaded from cache but newer content available, reload to see" message;
- cleanup the browser window / UI side of things so that it's easy to include on any site.
A *lot* of work, but hey, now at least we kinda have a roadmap!
Ok, back to playing with #Samizdat after some traveling.
- caching plugin not double-counted anymore;
- finally there is a proper project website at https://samizdat.is/
Need to fix Gun+IPFS for the new domain, today is a good time.
Main project home still https://git.occrp.org/libre/samizdat/ for the time being, but hoping to move it to a public GitLab instance soon.
That means now when you load the site in Firefox you should get the favicon. Favicon does not exist on the server, but exists in IPFS, for the purpose of testing all works.
In Chrome/Chromium it should show up after a reload or two (take your time though, Chrome/Chromium caches things in weird ways).
Woo! That means our migration of Samizdat is complete. It's on it's own domain, and on an open GitLab instance. 🎉 🎈
tl;dr: there needs to be a way to measure how many times Samizdat made it possible to circumvent censorship.
That's something that will have to run on reader's browser, and so there are serious privacy considerations.
But without being able to show it works, it will be hard to convince people (and site admins) it does.
In the meantime, working on cache invalidation for #Samizdat. One of the Two Hard Problems in IT (cache invalidation, naming things, and off-by-one errors)!
Anyway, trying to keep some context in cache using "x-samizdat-*" headers. But the Cache API doesn't seem to cache all headers, just some:
Of course, there is no mention of it anywhere in the docs (or I have not found it after hours of looking).
I *think* I figured out how to do cache invalidation in #Samizdat in a more-or-less sane way, *assuming that* only a single live plugin is in use.
I might have an idea how to do it across plugins too.
Relevant branch here:
Boom! Cache (or, rather, locally stashed version) invalidation implemented in #samizdat https://0xacab.org/rysiek/samizdat/merge_requests/14
From now on if you visit the site once load the current Service Worker, stuff gets stashed, and then when you happen to visit the site on a blocked connection, it is *assumed* Gun+IPFS version is fresher.
If you visit again, and have the Gun+IPFS version stashed, IPFS addresses are compared to check freshness.
If a fresh version is available, a message is displayed to the reader.
What's the difference between a "cached" and "stashed" resource in #Samizdat, you ask? Excellent question!
There can be multiple Samizdat plugins that implement the basic idea of keeping a version of a resource locally. One plugin currently implementing this is called "cache" and uses the Cache API:
So, to avoid confusion, whenever I'm talking in general about keeping versions locally, I will call it "stashing".
This will be made clear here: https://0xacab.org/rysiek/samizdat/blob/master/ARCHITECTURE.md
Worked on the documenation for #Samizdat a bit. Also, started working on implementing the standalone interface. MR: https://0xacab.org/rysiek/samizdat/merge_requests/15
The idea is to have the basic interface defined in samizdat.js so that all an admin needs to do is include that file. Currently the interface is tightly tied to index.html.
And we now have a standalone user UI in #Samizdat:
Check it out here:
Or here, to see it on a page that does not use the regular Samizdat CSS:
The UI only shows up if there are resources that seem to be unavailable via HTTPS (on samizdat.is that's the case with the favicon).
The only thing that needs to be included by website admins is a single JS file (samizdat.js).
Next step: creating a standalone admin UI.
Like measuring usage:
It *seems* like it's complicated, until it becomes clear that 3rd party tracking is not going to be affected by most website blocking scenarios. So the only thing that needs to be handled is when a website is using log analytics or their own tracker.
And the relevant merge request:
Did some code cleanup, and the samizdat-cli now can get a user's pubkey (will be needed later), and *almost* register a new Gun user.
More fun soon!
Working on implementing some basic user management in #Samizdat's samizdat-cli, as a necessary foundation for more sane deployment procedure. Relevant ticket and merge request:
Almost works, but for *some* reason users created using it are unusable. Specifically, it seems impossible to auth() as them. Moar debugging tomorrow. *sigh*
I have no clue what's wrong with my #Samizdat CLI code. When I create a user using samizdat-cli, it's impossible to auth() as that user (neither using the CLI, nor in a browser window):
But if I create a user using the same functions in a browser window, all works fine. I can then auth() as that user both in the browser window *and* via the CLI.
Relevant (fugly!) code here:
I've reported one bug already:
More to come.
Oh, did I write a test harness just for that? Yes. Yes I did:
(GitHub because Gun is hosted there; personally I prefer unifficial Gitlab instances, obviously)
I have a few things I can focus on in #Samizdat once I report all the NodeJS-related bugs (and before they get fixed).
I am very tempted to finally write the IPFS/IPNS plugin (completely side-stepping Gun), or a dat:// plugin. But perhaps I should do some boring stuff from the Beta milestone?
So, a poll! What should I focus on in Samizdat?
And so, the People have spoken. I'll bump implementing dat:// up on the ToDo list for #Samizdat. However, for Beta I really need to have documentation and Admin UI I guess. Eh.
Yesterday I noticed #Samizdat is not working. Spent most of the day debugging. Turns out four things happened at the same time:
- major code changes on my side
- some code changes on Gun side
- Samizdat stopped using the test Gun instance run by @OCCRP
- the public Gun peer started deleting stuff
Ooof! This was pretty damn annoying to deal with, but all is well again. As an added bonus:
- there is a Gun peer running at samizdat.is
- got an idea how to simplify deployment significantly.
I am also more and more considering moving #Samizdat away from Gun. Gun is currently used to map from a well know address ( Gun user pubkey) to the content-adressed resources in IPFS. This can be done using #IPNS.
So far my experience with Gun has been bumpy. It seems a bit easier to use than IPNS, but with all the trouble I've had with it... not sure it's worth it.
I'll probably develop gun+ipfs plugin a tiny bit more, and then move focus to IPNS/IPFS. Added benefit: fewer dependencies.
Had a good chat with Sam from dat:// project about #Samizdat. Got a bunch of good input and great links (including the lunet thing).
Good news: dat:// protocol v2 has a bunch of improvements and is almost ready for being released.
Bad news: dat:// v2 is incompatible with v1, has no pure JS implementation, and it's unlikely it will get one soon.
Ugly news: this means it most likely doesn't make much sense to implement dat:// in #Samizdat at this moment.
Ok, so it might in fact make sense to implement dat:// in #Samizdat, since the API is not expected to change between v1 and v2.
Many thanks to @syntax for his contribution to #Samizdat:
This is a much-needed nudge for me to get back to hacking on this project.
I have taken a way-too-long sabbatical from working on #Samizdat, but finally getting back into it.
First step (making sure pipelines work again) was easier than expected: my #GunJS superpeer was down. All green:
And need to improve how the pipeline verifies stuff is available in IPFS, pretty sure the 504s there are because we get throttled by gateways:
That oversight has just been fixed:
Work on the #Samizdat overview document is proceeding nicely and I am starting to be pretty happy with it:
On friend's advice I shortened the Philosophy section substantially, and expanded on it in a separate document:
As always, comments, suggestions, and patches welcome!
#Samizdat overview now has flowcharts:
I have no idea if they're useful. Only one way to find out!
"I appreciate SDF but it's a general-purpose server and the name doesn't make it obvious that it's about art." - Eugen Rochko