spoiler

made you look

  • 0 Posts
  • 190 Comments
Joined 2 years ago
cake
Cake day: July 27th, 2024

help-circle
  • The Category 5e specification improves upon the Category 5 specification by further mitigating crosstalk. The bandwidth (100 MHz) and physical construction are the same between the two, and most Cat 5 cables actually happen to meet Cat 5e specifications even though they are not certified as such.

    I did kinda mean it as a joke originally, but yeah 5/5e are the same outside of crosstalk resistance.
















  • Compared to e.g. pushing a button in VS code and having your browser pop up with a pre-filled in github PR page? It’s clunky, but that doesn’t mean it’s not useful.

    For starters it’s entirely decentralised, a single email address is all you need to commit to anything, regardless of where and how it’s hosted. There was actually an article on lobsters recently that I thought was quite neat, how the combination of a patch-based workflow and email allows for entirely offline development, something that’s simply not possible with things like github or codeberg.

    https://ploum.net/2026-01-31-offline-git-send-email.html

    The fact that you can “send” an email without actually sending it means you can queue the patch submissions up offline and then send them whenever you’re ready, along with downloading the replies.


  • Sourcehut uses it, it’s actually the only way to interact with repos hosted on it.

    It definitely feels outdated, yet it’s also how git is designed to work well with. Like git makes it really easy to re-write commit history, while also warning you not to force push re-written history to a public repo (Like e.g. a PR), that’s because none of that is an issue with the email workflow, where each email is always an entirely isolated new commit.



  • I’ve got some numbers, took longer than I’d have liked because of ISP issues. Each period is about a day, give or take.

    With the default TTL, my unbound server saw 54,087 total requests, 17,022 got a cache hit, 37,065 a cache miss. So a 31.5% cache hit rate.

    With clamping it saw 56,258 requests, 30,761 were hits, 25,497 misses. A 54.7% cache hit rate.

    And the important thing, and the most “unscientific”, I didn’t encounter any issues with stale DNS results. In that everything still seemed to work and I didn’t get random error pages while browsing or such.

    I’m kinda surprised the total query counts were so close, I would have assumed a longer TTL would also cause clients to cache results for longer, making less requests (Though e.g. Firefox actually caps TTL to 600 seconds or so). My working idea is that for things like e.g. YouTube video, instead of using static hostnames and rotating out IPs, they’re doing the opposite and keeping the addresses fixed but changing the domain names, effectively cache-busting DNS.