![](https://programming.dev/pictrs/image/b9f776a4-dee1-485e-91ec-ac0e4ed5dc80.jpeg)
![](https://fry.gs/pictrs/image/c6832070-8625-4688-b9e5-5d519541e092.png)
It’s clear that you don’t agree with my original opinions. And that’s ok. But it really doesn’t seem so simple and clear. Take a look at the ratio of up to down votes.
It’s clear that you don’t agree with my original opinions. And that’s ok. But it really doesn’t seem so simple and clear. Take a look at the ratio of up to down votes.
Thanks for pointing this out. I keep looking back at this thread as new people grow annoyed at my comments 🙂.
At the time I writing this, there are currently 15 upvotes and 28 downvotes on my original comment. That’s clearly negative and that’s ok. But that also makes it the third most voted on and the 4th most upvoted comment in the entire post. Seems there’s a very split opinion in the community here. This is now officially entertaining!
Thanks for the happy comment. But it’s all good. People are allowed to not like my comment. I’m not exactly swayed by the downvotes but maybe I could be just wrong here.
I am not
deleted by creator
Potentially. See my edit shove
Ok that is an impressive number but it feels a little disingenuous. You still need to something on your machine to interpret the js code, right? Is that included in the 13k? How much storage does that take?
EDIT: Well this is by far my most negative comment here. That’s almost entertaining. I’ll share a few more of my thoughts here rather than respond to individual comments. Maybe the context will make this more palatable.
First, I expect that the js language is doing most of the work here. Which makes sense. But having a browser installed as a prerequisite is an enormous dependency.
How would that stack up against other languages? Can I build a 13k binary using C? How about C#? I think Go is maybe the most interesting because the binary is entirely self contained by default. No external dependencies aside from the OS. I don’t think this or a similar game is viable with only 13k. Which is fine! I just that I find 13k is disingenuous.
That brings up the question of whether or not we should include the OS in the storage size. I would think not. But that’s only because the OS is (usually) the least common denominator when we talk about developing software. It’s generally assumed by default. But if someone wants to compare with a game that interfaces with hardware directly, then yes, we should absolutely include the OS as a dependency.
Now that I’m giving this more thought, I suspect that the devs wrote 13k of code + assets to make the game functional. Still impressive. But the more I think about this, the more meaningless that number gets. Does pre or post compiling matter more? What if we compress the thing as tarball? There’s just too many ways to manipulate this number.
Ya… being paid to perform isn’t immoral. Honestly, I hope he took a ton of cash from Amazon for the show.
Amazon is the crowd doing evil crap. Their immorality doesn’t automatically spread to everyone they interact with. Especially, people that aren’t actually aiding their efforts. This one is corporate waste
Open source software literally means that the source code is available to anyone. In GitHub, that just means that your repo is public rather than private. But your method technically doesn’t matter. You could publish to a forum if you wish. That’s still open source!
Free OSS just means that anyone is free to use and modify the source code for any purpose. The details are usually defined in a LICENSE file.
I feel like you’re really asking about the common practices and methods used in FOSS. Right? If so, that’s entirely up to you as the maintainer. As the project matures, you may attract other contributors which will in turn will motivate change to your tools and methods.
Start with what works for you. Model after similar projects if you wish. Adjust as change is needed.
lol got it. Definitely not email then
Uh email? It’s not exactly exciting but there are loads of tools available for automating emails. Definitely asynchronous. Does it fit your needs?
Anyone know why?
Ethically, it should apply. In practice, it doesn’t because the rich make the rules.
I’m generally not a big fan of big social media like e.g. Facebook where you might have many thousands of followers, purposefully grow the numbers, etc. I personally think these things are an everyday evil. Yes, it’s a bit melodramatic 🙂but that’s how I feel. Reddit, and now Lemmy are about as far as I like to go with it.
So the isolation of geo-local-only federation is a feature. The feature, actually. I want an entire social media platform that isn’t capable of focusing on single accounts. Where you are near guaranteed to interact with your local community only. Where it would take a dramatic effort for a single actor to influence global opinions. I want a social media platform that isn’t so easy to manipulate. I could go on and on.
After reading your responses, it seems like we’re describing two different methods of building this system.
Your ideas seems to depend on having many instances for various regions, where all instances are federated with each other. So my local instance somewhere in the US would still be federated with for example, an instance in Germany. But the content I receive would be heavily focused on “nearby” content. Interesting
My ideas are based on an important difference. An instance for my town would only federate with instances for the surrounding towns. Maybe one or two more “hops” away. So sharing content between my local instance and one in Germany would be impossible. Content on my local instance would only be accessible to users in nearby instances. Local content enforced by local federation.
I spent several weeks thinking about this exact idea.
Federation is cool. You could set up each instance to only federate with instances for nearby towns and cities. Maybe a “2 district” radius. Users would only see content for their local communities. Local news stays local. Local government could officially participate if they wish. People you talk to are actually neighbors you might see in person. Larger regions like counties, states, provinces, or even countries, could also have dedicated instances and federate similarly. I think this is the big appeal and it sounds awesome!
There are a few problems 🙂
First is a little bit of confusion with posting. Let’s say that I see a post about a cool new restaurant in my town. I share it with a friend who lives a few towns away and that’s outside the “federation radius”. I can’t share the post with that friend very easily. Maybe the tools could be enhanced to make this viable?
Second is a matter of privacy. How do you know that new accounts belong to people associated with the geographic location of each instance? If you don’t validate, the system will certainly be abused. If you do validate, then users need to supply some real info! Home address, ID, etc. that’s a big deal for users and instance admins.
Third. What happens if you move? Do you have to abandon your old account and start over? Again, the system itself can be developed further to solve this. But that’ll take time and money.
Next is the operating costs. You would need to build thousands of instances to build this system up. And each one would have to be tied to a geographic region. You need new features to handle signups this way. You have the simple cost of running these servers. You probably need a lot of staff to manage it all. This is an expensive platform for one party to run. Alternatively…
It doesn’t have to be one party running this entire system. That’s the point of the Fediverse, right? The operational costs go way down if anyone can run their own instance. But how do you enforce the rules of federating with instances for geographically nearby locations? I don’t see a reasonable way to solve this one.
I could probably keep listing issues. But these are the big ones IMO. If you solve these, the system is viable and could be amazing.
Unfortunately, I don’t remember the source so we may need to go digging. But I recall reading that something like 1/3 of all bugs are related to memory safety. And those bugs translate to things like buffer overflow and privilege escalation attacks.
The proclaimed advantage is that by making the entirety of Rust memory safe, that entire class of bugs simply won’t exist for projects written in Rust. When they do happen, the bugs will be addressed by the language rather than many thousands of downstream projects. It should be an enormous gain in development performance for the world.
I think the idea makes sense. Time will tell us how well that works.
I ditched chrome (chromium + google propriety spyware) some years ago in favor of Brave browser (chromium + Brave stuff). It was a decent user experience but Brave also does some shady stuff, which you can google easily if interested.
Last year, google poisoned chromium with DRM stuff. They rolled back the changes after a few months but the damage was already done. I, and many others, jumped ship to Firefox and other non-chromium based browsers. Firefox isn’t perfect, but it’s an excellent browser. I’m sticking with it for the foreseeable future. And absolutely use uBlock Origin. Between that and proton VPN features, I don’t see ads anymore. It’s fantastic.
Strange. I’m not exactly keeping track. But isn’t the current going in just the opposite direction? Seems like tons of utilities are being rewritten in Rust to avoid memory safety bugs
Thanks for asking the question. Apparently I need to check out opensuse!