• 0 Posts
  • 29 Comments
Joined 1 year ago
cake
Cake day: July 30th, 2023

help-circle





  • This is true of a even some public universities in the US. I can’t remember if it was a rule where I was, but definitely most freshman did just live in dorms.

    Lot of folks brought their own desktops to set up, and we were allowed Ethernet switches to hook up multiple devices - had to be wired. Wireless had two options, WPA# 802.1X or unencrypted captive portal guest. If your device didn’t support that, it had to be wired by policy.

    And they weren’t wrong, I did a radio scan and they had the full sized enterprise access points about as good as they could (with a few low signal exceptions, and the air waves were still overloaded with too many people. The building uplink was perfectly fine, it was just overcrowded wireless.



  • I’m an American android user and I’m confused too. At least in my area, contactless is pretty ubiquitous now. (I accept adoption is slower, but it’s getting there)

    Sure Apple Pay seemed to come to a lot of terminals first, but NFC Google wallet or whatever it is the phone does automatically I’ve only seen fail at certain terminals. In that rare case, usually someone behind me with Apple Pay often also fails, so I’d be more likely to attribute it to a system glitch rather than lack of support.





  • The first few screens look like a combination reverse job board and Coinbase. But your description says “post stuff and create tokens”, which doesn’t quite seem to line up

    Why?

    I don’t think Web3 contracts have tested case law yet, so who knows if it’s enforceable in court, at most it may only be as strong as a gentleman’s agreement. And the token part looks like an easy way to create rug pull coins, just on the ETH database instead of an independent database.


  • I’d like to politely disagree

    Finding alternatives to large software packages is great, don’t think I’m not saying that - but any time you have competitor X and competitor Y, be they both commercial, both F/OSS, or some combination thereof, the competitors must be cognizant of each other when setting up features.

    Burying your head in the sand and ignoring Microsoft, Apple, and Google is a very solidly Microsoft-Apple-Google-style play. It’s the play of someone who believes the other side offers no competition. That’s how you get unwieldy features these tech giants implement because they know they can make a 70% effort and people won’t be annoyed enough to leave.

    Every tool they make has a reason someone made it. Many tools are very important - for one example, the Microsoft Office document format is considered to be almost a universal format in presentations, spreadsheets, and plain documents for message passing between businesses.

    But as we as a society design alternatives to those various monopolies (as we should), we need users to want to use the new thing. We have to take what people like and keeps them on their old platform, and best preserve the intent of what they want on the new platform. Doing so requires discussing the features those big tech companies

    And as users, when we select the platforms we use, we need to weigh the cost of going with an alternative vs going with a giant. No solution is a perfect solution for everyone, and the chooser needs to weigh the maintenance cost (in hours or money) they will incur, how their users will like/dislike it, and maybe even look at a piece of software and decide “nah the vibes are off”.

    I’d love a world where those three tech giants had proper competition in all fields, and I think their business practices are scummy and need improvement. But the real alternatives to each need some polish before they’re ready to be used by [arbitrary tech illiterate grandmother].


  • Others have some good information here - all I’d like to add to the root is that Windows and Mac have a built-in DNS cache and it’s pretty straightforward to add a DNS cache to systemd distros (if it’s not already installed or in use) using systemd-resolved or dnsmasq if you really dislike systemd. Some distros enable this from install time.

    Systems that utilize a DNS cache will keep copies of DNS query results for a period of time, making the application-level name lookup speed essentially 0ms for a cached result. Cold results obviously incur the latency of the DNS server itself.




  • Far-UVC has a lot of potential once it’s scaled up. Right now, we’re still learning about best practices.

    Institutions should be adopting this tech at scale.

    If we’re still learning about best practices why are we talking about deploying this at scale? Self contradictory article…

    It should be the other way around. Figure out if it works academically, then test small scale, then scale up with proven and reproducible results. That’s how science works. Best practices can be formulated and adjusted at each stage as more knowledge is gained. That’s how we don’t make a massive health mistake and give an entire convention center indoor sunburns. Especially for people who might be more sensitive to sunburns.




  • TLDR: probably a lot of people continue using the thing that they know if it just works as long as it works well enough not to be a bother.

    Many many years ago when I learned, I think the only ones I found were Apache and IIS. I had a Mac at the time which came pre installed with Apache2, so I learned Apache2 and got okay at it. While by release dates Nginx and HAProxy most definitely existed, I don’t think I came across either in my research. I don’t have any notes from the time because I didn’t take any because I was in high school.

    When I started Linux things, I kept using Apache for a while because I knew it. Found Nginx, learned it in a snap because the config is more natural language and hierarchical than Apache’s XMLish monstrosity. Then for the next decade I kept using Nginx whenever I needed a webserver fast because I knew it would work with minimal tinkering.

    Now, as of a few years ago, I knew that haproxy, caddy, and traefik all existed. I even tried out Caddy on my homelab reverse proxy server (which has about a dozen applications routed through it), and the first few sites were easy - just let the auto-LetsEncrypt do its job - but once I got to the sites that needed manual TLS (I have both an internal CA and utilize Cloudflare’ origin HTTPS cert), and other special config, Caddy started becoming as cumbersome as my Nginx conf.d directory. At the time, I also didn’t have a way to get software updates easily on my then-CentOS 7 server, so Caddy was okay-enough, but it was back to Nginx with me because it was comparatively easier to manage.

    HAProxy is something I’ve added to my repertoire more recently. It took me quite a while and lots of trial and error to figure out the config syntax which is quite different from anything I’d used before (except maybe kinda like Squid, which I had learned not a year prior…), but once it clicked, it clicked. Now I have an internal high availability (+keepalived) load balancer than can handle so many backend servers and do wildcard TLS termination and validate backend TLS certs. I even got LDAP and LDAPS load balancing to AD working on that for services like Gitea that don’t behave well when there’s more than one LDAPS backend server.

    So, at some point I’ll get around to converting that everything reverse proxy to HAProxy. But I’ll probably need to deploy another VM or two because the existing one also has a static web server and I’ve been meaning to break up that server’s roles anyways (long ago, it was my everything server before I used VMs).