If they had said “locally hosted in our datacenter”
Then that would also be an oxymoron.
Local is the opposite of remote. This is a remote server. Remote servers are not local. This is not a matter of interpretation.
If they had said “locally hosted in our datacenter”
Then that would also be an oxymoron.
Local is the opposite of remote. This is a remote server. Remote servers are not local. This is not a matter of interpretation.
Why does local mean local? I’m not sure I understand your question.
“Locally hosted” means it’s running on the local host. In this case, that would mean on the same computer running Firefox.
Calling something that is only accessible over the internet “locally hosted” is outrageous doublespeak.
Orbit currently uses a version of Mistral LLM (Mistral 7B) that is locally hosted on Mozilla’s Google Cloud Platform instance.
Hmm.
>locally hosted
>Google Cloud
Hmmmmmmmmmmmmmmmm.
So probably there will be some systems other than Linux that do use Rust
There’s one called Redox that is entirely written in Rust. Still in fairly early stages, though. https://www.redox-os.org/
Have you used Facebook in the last 5 years?
The UX is godawful. More than half my feed is just random crap suggestions and ads.
Installing Linux after Windows should be fine without disconnecting drives.
The reverse is troublesome. Microsoft’s installer is all too happy to shit on your drives, even the ones you’re not using for installation. But Linux installers are much more friendly to dual-booting and all kinds of complex setups.
Or about a quarter of a single Super Smash Bros fanfic.
That’s like 10-20 adult novels. Respect.
Same on macOS. Apple has “case-sensitive HFS+” as an option for UNIX compatibility (or at least they used to) but actually running a system on it is a bad idea in general.
Haven’t heard of Hiren’s BootCD in like 15 years. Good to see it’s still around!
Yeah, I had to disconnect all my SATA HDs to stop the Windows installer from shitting all over them.
I’d be worried about Windows updates doing the same thing now, after the the recent glitch that broke bootloaders.
F-Droid link for the lazy: https://f-droid.org/packages/com.junkfood.seal/
Definitely going to check this out. I’ve been using yt-dlp via command line in Termux but that experience is less than ideal.
It was bought out and cleaned up a few years ago. It’s legit again now, though I don’t think it’ll ever really recover from that fiasco.
Totally agree. Their product line was an absolute mess back then. Their current lineup is getting a little bloated too. I don’t know why they bother having two laptop product lines anymore when they are so similar.
Apple tried to allow clones, but ran into the same problem because the clone makers could make cheaper machines by slapping together parts.
Yeah, this is exactly what happened, although some of the clone brands were perfectly high-quality (Power Computing in particular made great machines, usually the fastest on the market). In the Mac community at the time, a lot of people (myself included) wished Apple would just exit the hardware business and focus on what they were good at: software.
Then Steve Jobs came back and did exactly the opposite of that. First order of business was to kill cloning. Then came the iPod.
To be fair, the next generation of Power Macs after that were about half the price of the previous gen.
Most of Apple’s history, actually.
Macs have a reputation for being expensive because people compare the cheapest Mac to the cheapest PC, or to a custom-built PC. That’s reasonable if the cheapest PC meets your needs or if you’re into building your own PC, but if you compare a similarly-equipped name-brand PC, the numbers shift a LOT.
From the G3-G5 era ('97-2006) through most of the Intel era (2006-2020), if you went to Dell or HP and configured a machine to match Apple’s specs as closely as possible, you’d find the Macs were almost never much more expensive, and often cheaper. I say this as someone who routinely did such comparisons as part of their job. There were some notable exceptions, like most of the Intel MacBook Air models (they ranged from “okay” to “so bad it feels like a personal insult”), but that was never the rule. Even in the early-mid 90s, while Apple’s own hardware was grossly overpriced, you could by Mac clones for much cheaper (clones were licensed third-parties who made Macs, and they were far and away the best value in the pre-G3 PowerPC era).
Macs also historically have a lower total cost of ownership, factoring in lifespan (cheap PCs fail frequently), support costs, etc. One of the most recent and extensive analyses of this I know if comes from IBM. See https://www.computerworld.com/article/1666267/ibm-mac-users-are-happier-and-more-productive.html
Toward the tail end of the Intel era, let’s say around 2016-2020, Apple put out some real garbage. e.g. butterfly keyboards and the aforementioned craptastic Airs. But historically those are the exceptions, not the rule.
As for the “does more”, well, that’s debatable. Considering this is using Apple’s 90s logo, I think it’s pretty fair. Compare System 7 (released in '91) to Windows 3.1 (released in '92), and there is no contest. Windows was shit. This was generally true up until the 2000s, when the first few versions of OS X were half-baked and Apple was only just exiting its “beleaguered” period, and the mainstream press kept ringing the death knell. Windows lagged behind its competition by at least a few years up until Microsoft successfully killed or sufficiently hampered all that competition. I don’t think you can make an honest argument in favor of Windows compared to any of its contemporaries in the 90s (e.g. Macintosh, OS/2, BeOS) that doesn’t boil down to “we’re used to it” or “we’re locked in”.
Chromium itself will. Other Chromium-based browser vendors have confirmed that they will maintain v2 support for as long as they can. So perhaps try something like Vivaldi. I haven’t tried PWAs in Vivaldi myself, but it supports them according to the docs.
Debian still supports Pentium IIs. They axed support for the i586 architecture (original Pentium) a few years back, but Debian 12 (current stable, AKA Bookworm) still supports i686 chips like the P2.
Not sure how the rest of the hardware in that Compaq will work.
See: https://www.debian.org/releases/stable/i386/ch02s01.en.html
This is a FAQ for end users, about a feature in software running on end users’ computers.
It is absolutely doublespeak to call it “local”. Are we supposed to invent an entirely new term now to distinguish between remote and local? Please do not accept this usage. It will make meaningful communication much harder.