All they say is that there's a (niche) trend of a few people using feature phones with expected combined sales of $2.8 million. Versus the $200 billions of iPhones alone.
I did it for 3 months. I really enjoyed my time doing it and learned a lot about my usage. It was a cheap $50 experiment. After I went back to my smartphone, I uninstalled ALL social media apps. Turned off ALL notifications but left calls and messages as an exception. My smartphone is now essentially a feature phone. It's not 100% the same since the big screen does lure you in to use it but my usage is still way down and because I don't have any social media there's no reason for me to be on my phone around other people. I wholeheartedly recommend trying it for those curious.
That reminded me how a local wanna-be influencer did a smartphone detox for a week, immediately after the completion she posted an FB story: Part 1 - Reflection, how eyeopening the experience was, how much time she suddenly had for the things that truly matter etc. Subscribe to not miss the Part 2!
Just the other day I saw an article with the exact opposite headline about how Gen z is sticking with the iPhone. Now I don't know which one is full of shit; but it's obviously one of them.
On the one hand, this could be filed under "yeah, no shit, we all know stuff in the cloud is forever".
On the other hand, it's something that's easy to forget with the ubiquitous omnipresence of compute in our lives. We become numb to it, and everyone has moments of crisis or weakness where they may let their guard down.
The US needs better privacy and consumer protection laws. But we're always behind Europe, and way behind technology, when it comes to our crappy legal system.
That's not to say the two men don't think AI will be helpful in the future. Indeed, Torvalds noted one good side effect already: "NVIDIA has gotten better at talking to Linux kernel developers and working with Linux memory management," because of its need for Linux to run AI's large language models (LLMs) efficiently.
Hohndel agreed but added that the industry needs to support these smaller projects -- and not only with money. "Companies need to engage with these projects. Have your company adopt a couple of such projects and just participate. Read the code, review the patches, and provide moral support to the maintainers. It's as simple as that."
Really glad he said this, I keep seeing posts about how all these big companies could solve the problem by just throwing money at small projects and while that is better than nothing it would help way more to have their own developers helping to review and fix issues.
It means anyone including microsoft or apple can use the code contribution or take the entire softwarw and make some modifications and sell it proprietary. Any optimisations or features made by community can be proprietarised
Just because it has a CVE number doesn't mean it's exploitable. Of the 800 CVEs, which ones are in the KEV catalogue? What are the attack vectors? What mitigations are available?
The idea that it is somehow possible to determine that for each and every bug is a crazy fantasy by the people who don't like to update to the latest version.
Security is not a binary variable, but managed in terms of risk. Update your stuff, don't expose it to the open Internet if it doesn't need it, and so on. If it's a server, it should probably have unattended upgrades.
If it's a server, it should probably have unattended upgrades.
Interesting opinion, I've always heard that unattended upgrades were a terrible option for servers because it might randomly break your system or reboot when an important service is running.
There are two schools of thought here. The "never risk anything that could potentially break something" school and the "make stuff robust enough that it will deal with broken states". Usually the former doesn't work so well once something actually breaks.
Both my Debian 12 servers run with unattended upgrades. I've never had anything break from the changes in packages, I think. I tend to use docker and on one even lxc containers (proxmox), but the lxc containers also have unattended upgrades running.
Do you just update your stuff manually or do you not update at all? I'm subscribed to the Debian security mailing list, and they frequently find something that means people should upgrade, recently something with the glibc.
Debian especially is focused on being very stable, so updating should never break anything that wasn't broken before. Sometimes docker containers don't like to restart so they refuse, but then I did something stupid.
I used to check the cockpit web interface every once in a while, but I've tried to enable unattended updates today. It doesn't actually seem to work, but I planned on switching to Nix anyway.
I don't use Cockpit, I just followed the Debian wiki guide to enabling unattended upgrades. As fast as I remember you have to apt install something and change a few lines in the config file.
It's also good to have SMTP set up, so your server will notify you when something happens, you can configure what exactly.
Any "How To" that doesn't just use Rufus isn't worth the page its text is rendered on. Rufus can do Linux boot disks, but is indispensable for Windows boot disk utilities. It's one of the only ways I know of to make a Windows ToGo installation (equivalent of a Linux Live USB), which I used to install Windows on a friends SD card for their Steam Deck so they can dual-boot.
Man, Google really does suck now. It feels nearly impossible to get something like a how-to deep in the Debian FAQs to come up, as it mostly surfaces this auto-generated SEO crap for How To's.
Very cool, I'd assumed there was a simple command line set of commands, just was failing to find it. Thanks.
Man, Google really does suck now. It feels nearly impossible to get something like a how-to deep in the Debian FAQs to come up, as it mostly surfaces this auto-generated SEO crap
By design. The longer you're Googling, the more ads they can sell.
...Ben Gomes – a long-tenured googler who helped define the company during its best years – lost a fight with Prabhakar Raghavan, a computer scientist turned manager whose tactic for increasing the number of search queries (and thus the number of ads the company could show to searchers) was to decrease the quality of search. That way, searchers would have to spend more time on Google before they found what they were looking for.
I don't remember where, but i read that this method only works because linux distributors "abuse" the ISO format to allow this. If I remember right, it's not possible to use this ISOs on regular disks
Of course the command you provided is right and it's what I use, it's just a fun fact
Yes and no, it's the other way round. The ISOs often are hybrid images which you can burn onto a CD/DVD or dd onto a USB pen drive. Until approximately 10-15 years ago, if I remember correctly, the distributed Linux ISOs where standard not hybrid images, thus you always needed some other program to create bootable USB media.
qrmu-debootstrap is also super useful if you want to customize and image for a different architecture (for example building custom RPi images).
Super useful information, thanks!
EDIT: Is this anything like the isorespinner.sh? I've previously used that to get Linux on an RCA Cambio W101 because it needed a fancy ISO since it has a 32-bit bootloader and a 64-bit CPU.
I tried Windows ToGo on a few USB keys (including two high-speed ones), never managed to get something I could actually use that was not laggy AF, to the point it's not usable (dozens of minutes to boot, lags of entire minutes and so on). Did I do something wrong?
This works because block devices like /dev/sdX are just files. If you cp a file onto another file, it overwrites the data of the destination with the source. A block device represents the device itself, not the filesystem; if you wanted to put the ISO inside the filesystem, you'd have to mount it first.
Not technically. unetbootin and some similar tools like rufus take the USB, partition it, and copy the contents of the disk to it after manually setting up a bootloader on it. This is not required for most Linux ISOs though where you can just cp or dd the image directly to the USB as they are already setup with all that on the image. But other ISOs, like I believe Windows ones have a filesystem on them that is not vfat so cannot be directly copied. Although these days for windows you just need to format the USB as vfat and copy the contents of the windows ISO (aka the files inside it, not the iso filesystem) to the filesystem.
I tend to find unetbootin and rufus break more ISOs then they actually help with though. Personally I find ventoy is the better approach overall, just copy the ISO as a file to the USB filesystem (and you can copy multiple ones as well).
zdnet.com
Oldest