• 0 Posts
  • 89 Comments
Joined 1 year ago
cake
Cake day: June 1st, 2023

help-circle




  • Kata1yst@kbin.socialtoTechnology@lemmy.worldThe decline of Intel..
    link
    fedilink
    arrow-up
    30
    arrow-down
    1
    ·
    edit-2
    3 months ago

    2009 era was also when Intel leveraged their position in the compiler market to cripple all non-Intel processors. Nearly every benchmarking tool used that complier and put an enormous handicap on AMD processors by locking them to either no SSE or, later, back to SSE2.

    My friends all thought I was crazy for buying AMD, but accusations had started circulating about the complier heavily favoring Intel at least as early as 2005, and they were finally ordered to stop in 2010 by the FTC… Though of course they have been caught cheating in several other ways since.

    Everyone has this picture in their heads of AMD being the scrappy underdog and Intel being the professional choice, but Intel hasn’t really worn the crown since the release of Athlon. Except during Bulldozer/Piledriver, but who can blame AMD for trying something crazy after 10 years of frustration?


  • They even literally have a section of the article that says they “see Fair Software as an alternative model to the free and open source software model”, and they think it’s superior because the “developers can profit”.

    Newsflash: the developers usually see fractions of those cents while most of the money goes to the management and shareholders of the company that employs them. Hmm, doesn’t seem fair to me.

    Also, developers can and do profit from FOSS in many ways, but the most popular models are with commercial support, SaaS offerings, and additional functionality (like providing a web interface, clustering manager or other external piece of the puzzle to solve the problem at scale in enterprise).

    Like you said so succinctly: propaganda website to make rug pullers like Elastic and Hashicorp look better.






  • Accurate, but not bad, yes. It turns out standardized base systems and ABIs are important to an ecosystem.

    Linux tried the disorganized free-for-all for two decades, and what we got was fragmented “Ubuntu admins”, “debian admins”, “redhat admins”, “suse admins”, and a whole shitload of duplicated effort in the packaging ecosystem, only for half the packages out there to be locked to Ubuntu or RHEL. So the corporate interests, and a fair number of the community efforts, centralized their problems and solutions into a small standardized suite in Mesa+Wayland+systemd+Pipewire+flatpak, etc

    The result is a ton more interoperability, a truly open ecosystem where switching your distro doesn’t mean hiring different people and using different software, and a lot more stability and maturity.

    And hey, if a user or distro wants to do their own thing, they can make and own their niche, same as before. Nothing lost.

    It’s been kind of wild to watch over the past 15 years or so, makes me very hopeful for the next 15.


  • No no you don’t understand. The evil corporate overlords abused their power to force a choice on a developer, even though that choice was objectively the right choice and the developer was throwing a tantrum.

    This is truly awful. We must not let evil corporations, no matter their credentials, expertise, and decades of beneficial partnership with open source, tell immature and short sighted developers how to develop.




  • Never ask a man his pay, a woman her weight, or a data horder the contents of their stash.

    Jk. Mostly.

    I have a similar-ish set up to @Davel23 , I have a couple of cool use cases.

    • I seed the last 5 arch and opensuse (a few different flavors) ISOs at all times

    • I run an ArchiveBot for archive.org

    • I scan nontrivial mail (the paper kind) and store it in docspell for later OCR searches, tax purposes etc.

    • I help keep Sci-Hub healthy

    • I host several services for de-googling, including Nextcloud, Blocky, Immich, and Searxng

    • I run Navidrome, that has mostly (and hopefully will soon completely) replace Spotify for my family.

    • I run Plex (hoping to move to Jellyfin sometime, but there’s inertial resistance to that) that has completely replaced Disney streaming, Netflix streaming, etc for me and my extended family.

    • I host backups for my family and close friends with an S3 and WebDAV backup target

    I run 4x14TB, 2x8TB, 2x4TB, all from serverpartsdeals, in a ZFS RAID10 with two 1TB cache dives, so half of the spinning rust usable at ~35TB, and right now I’m at 62% utilization. I usually expand at about 85%




  • Kata1yst@kbin.socialtolinuxmemes@lemmy.worldArch with XZ
    link
    fedilink
    arrow-up
    2
    arrow-down
    1
    ·
    5 months ago

    Amazingly, for someone so eager to give a lesson in linguistics, you managed to ignore literal definitions of the words in question and entirely skip relevant information in my (quite short) reply.

    Both are widely used in that context. Language is like that.

    Further, the textbook definition of Stability-

    the quality, state, or degree of being stable: such as

    a: the strength to stand or endure : firmness

    b: the property of a body that causes it when disturbed from a condition of equilibrium or steady motion to develop forces or moments that restore the original condition

    c: resistance to chemical change or to physical disintegration

    Pay particular attention to “b”.

    The state of my system is “running”. Something changes. If the system doesn’t continue to be state “running”, the system is unstable BY TEXTBOOK DEFINITION.



  • I think the confusion comes from the meaning of stable. In software there are two relevant meanings:

    1. Unchanging, or changing the least possible amount.

    2. Not crashing / requiring intervention to keep running.

    Debian, for example, focuses on #1, with the assumption that #2 will follow. And it generally does, until you have to update and the changes are truly massive and the upgrade is brittle, or you have to run software with newer requirements and your hacks to get it working are brittle.

    Arch, for example, instead focuses on the second definition, by attempting to ensure that every change, while frequent, is small, with a handful of notable exceptions.

    Honestly, both strategies work well. I’ve had debian systems running for 15 years and Arch systems running for 12+ years (and that limitation is really only due to the system I run Arch on, rather than their update strategy.

    It really depends on the user’s needs and maintenance frequency.