I assume the KDE implementation resizes to default when you stop shaking it.
I could totally see someone coding a function that increases the mouse pointer by x% every y mouse shakes, and then neglecting to put in a size cap.
I assume the KDE implementation resizes to default when you stop shaking it.
I could totally see someone coding a function that increases the mouse pointer by x% every y mouse shakes, and then neglecting to put in a size cap.
Yeah, and Windows and OS X both do it as well.
Though there being no upper limit to the size is amusing.
really effects performance that much
Depending on the exact flags, some workloads will be faster, some will be identical, and some will be slower. Compilier optimization is some dark magic that relies on a ton of factors, but you can’t just assume that going from like -O2 to -O3 will provide better performance, since the optimizations also rely on the underlying code as to what they’ll actually make happen… and is why, for the most part, everyone suggests you stop at -O2 since you can start getting unexpected behavior the further up the curve you go.
And we’re talking low single digit performance improvements at best, not anything that anyone who is doing anything that’s not running benchmarks 24/7 would ever even notice in real world performance.
Disclaimer: there are workloads that are going to show different performance uplifts, but we’re talking Firefox and KDE and games here, per the OP’s comments.
Also they do default to a different scheduler, which is almost certainly why anyone using it will notice it feels “faster”, but it’s mainlined in the kernel so it’s not like you can’t use that anywhere else.
Absolutely.
2.0 was 100% not the same game, but it was vastly improved and perfectly playable well before then.
I played at launch, but on PC, and it was… fine. In that, unlike Starfield, it was a game with characters and a story that was interesting enough to carry the buggy world and somewhat less than fleshed out side-quest mechanics.
But, like, there were enough buildings and set pieces and people and stories to actually sit down and spend 200 hours exploring the world without seeing the same stupid PoIs over and over and over again, while trying to care about the least interesting NPC companions I’ve probably ever dealt with.
And Phantom Liberty is fucking fantastic, so they took a bit of a turd at launch and turned it into an amazing game.
Perhaps it was just a little too simple?
And the amazing variety of points of interests and quests!
You can have literally dozens of combinations!
(Still salty about Starfield, don’t mind me.)
Does !12345:p do what you want?
Edit: that also makes hitting the up arrow result in whatever command that was, so if you wanted to edit the line or whatever, you could !12345:p, up, then edit and execute.
Uh, are you sure your shell you’re using is bash and not zsh or something else?
Bash is indeed just !12345.
Why not save time and do it the other way?
Install the minimal/netinstall image, and then add what you need.
You’ll probably spend less time adding than trying to figure out what’s installed that you do or don’t need and trying to remove random packages without breaking anything.
Kudos for the unique delivery method, I guess?
(Also I do not like this sudden appearance of QR codes in daily life like for menus and shit for exactly this reason: just give me a URL if you must, or print something out and stop being cheap.)
two commands: dd and resize2fs, assuming you’re using ext4 and not something more exotic.
one makes a block-level copy of one device to another like so: dd if=/dev/source-drive of=/dev/destination-drive
the other is used to resize the filesystem from whatever size it was, to whatever size you tell it (or the whole disk; I’d have to go read a manpage since it’s been a bit)
the dd is completely safe, but the resize2fs command can break things, but you’d still have the data on the original drive, so you could always start over if it does - i’d unplug the source drive before you start doing any expansion stuff.
dd then resize the fs?
Edit: one caveat here I forgot: if your fstab is using UUIDs, you’re going to have to update that, since the new drive won’t be the same UUID because, well, it’s not the same drive.
Also, if you like htop, youre going to love btop.
Yeah DNS is, in general, just goofy and weird and a lot of the interactions I wouldn’t expect someone who’s done it for years to necessarily know.
And besides, the round-robin thing is my favorite weird DNS fact so any excuse to share it is great.
Uh, don’t do that if you expect your mail to be delivered.
Multiple PTRs, depending on how the DNS service is set up, may be returned in round-robin fashion, and if you return a PTR that doesn’t match what your HELO claims you are, then congrats on your mail being likely tossed in the trash.
Pick the most accurate name (that is, match your HELO domain), and only set one PTR.
(Useless fact of the day: multiple A records behave the same way and you can use that as a poverty-spec version of a load balancer.)
sudo smartctl -a /dev/yourssd
You’re looking for the Media_Wearout_Indicator which is a percentage starting at 100% and going to 0%, with 0% being no more spare sectors available and thus “failed”. A very important note here, though, is that a 0% drive isn’t going to always result in data loss.
Unless you have the shittiest SSD I’ve ever heard of or seen, it’ll almost certainly just go read-only and all your data will be there, you just won’t be able to write more data to the drive.
Also you’ll probably be interested in the Total_LBAs_Written variable, which is (usually) going to be converted to gigabytes and will tell you how much data has been written to the drive.
As a FunFact™, you’re more likely to have the SSD controller die than the flash wear out at this point.
Even really cheap SSDs will do hundreds and hundreds of TB written these days, and on a normal consumer workload we’re talking years and years and years and years of expected lifespan.
Even the cheap SSDs in my home server have been fine: they’re pushing 5 years on this specific build, and about 200 TBW on the drives and they’re still claiming 90% life left.
At that rate, I’ll be dead well before those drives fail, lol.
It’s usable-ish, but still kinda crashy and prone to occasionally imploding.
I wouldn’t really use it as my sole daily driver, but for certain people doing certain things, it’s probably fine.
(It needs another year, honestly.)
Well, I know what I’m doing this weekend.
Honestly, I would have assumed 1080p was an acceptable default assumption.
Is this just a case of older hardware, or are there still laptops that don’t have 1080p panels at this point?
A quick review of stuff on BestBuy indicates that $150 laptops have 1080p displays now, and anything more than that does as well, so uh, what devices are still using these?